Teaching computers to understand the visual world
We want to change the way you interact with visual data. We want to make your photos magical, we want to deeply understand images and videos from cameras everywhere: in your phone, on your Xbox, in your fridge, on robots, in cars, anywhere. We want you to be able to find your stuff, answer questions, make fantastic new images. And we do that by inventing new algorithms and thinking of new mathematical models for how images come to be.
J. Margeta, A.Criminisi, D.C.Lee, and N.Ayache, Recognizing Cardiac Magnetic Resonance Acquisition Planes using Finetuned Convolutional Neural Networks, in To appear in Computer Methods in Biomechanics and Biomedical Engineering, December 2015.
Gerard Pons-Moll, Jonathan Taylor, Jamie Shotton, Aaron Hertzmann, and Andrew Fitzgibbon, Metric Regression Forests for Correspondence Estimation, in IJCV, Springer, August 2015.
Hao Fang, Saurabh Gupta, Forrest Iandola, Rupesh Srivastava, Li Deng, Piotr Dollar, Jianfeng Gao, Xiaodong He, Margaret Mitchell, John Platt, Lawrence Zitnick, and Geoffrey Zweig, From Captions to Visual Concepts and Back, in The proceedings of CVPR, IEEE – Institute of Electrical and Electronics Engineers, June 2015.
Hyon Lim and Sudipta N. Sinha, Monocular Localization of a moving person onboard a Quadrotor MAV, in 2014 IEEE International Conference on Robotics and Automation, IEEE – Institute of Electrical and Electronics Engineers, 27 May 2015.
Toby Sharp, Cem Keskin, Duncan Robertson, Jonathan Taylor, Jamie Shotton, David Kim, Christoph Rhemann, Ido Leichter, Alon Vinnikov, Yichen Wei, Daniel Freedman, Pushmeet Kohli, Eyal Krupka, Andrew Fitzgibbon, and Shahram Izadi, Accurate, Robust, and Flexible Real-time Hand Tracking, CHI, April 2015.
Join us! Do you love to turn mathematics into code? Do you want to build the future? Then Apply here.
|All hands, no keyboard: New technology can track detailed hand motion|
|Why isn't what's good For Microsoft good for the country?|
|Diversity in data science: Microsoft Research's summer school aims high|
|Lab of Things Enables Research and Teaching|
- From Captions to Visual Concepts and Back
- Eye Gaze Keyboard
- Human activity detection in RGBD videos
- Fully Articulated Hand Tracking
- ATL Cairo GPSP - Projects Ideas
- Learning to be a depth camera for close-range human capture and interaction
- Sparse Reflections Analysis
- User-Specific Hand Modeling from Monocular Depth Sequences
- Real-Time RGB-D Camera Relocalization
- Real-time 3D Reconstruction at Scale using Voxel Hashing
Our research page.