The mHealthHub is a virtual forum where technologists, researchers and clinicians connect, learn, share, and innovate on mHealth tools to transform healthcare.

Tools & Datasets

Reach out or find us on Social Media.

365 Innovation Drive, Suite 335, Memphis, TN 38152

mHealthHUB@MD2K.org

Join our Community.

Stay up-to-date on the latest mHealth news and training.

Invalid email address
We promise not to spam you. You can unsubscribe at any time.

8/6/15: Introduction to First Person Vision

August 6, 2015

 

James M. Rehg, PhD
Deputy Director, MD2K Center
Professor, School of Interactive Computing
Georgia Tech University

About the Webinar:

Recent progress in miniaturizing digital cameras and improving battery life has created a growing market for wearable cameras, exemplified by products such as GoPro and Google Glass. At the same time, the field of computer vision, which is concerned with the automatic extraction of information about the world from images and video, has also made rapid progress due to the increasing availability of image data, increases in computational power, and the emergence of machine learning methods such as deep learning.

The analysis of video captured from body-worn cameras is an emerging subfield of computer vision known as First Person Vision (FPV). FPV provides new opportunities to model and analyze human behavior, create personalized records of visual experiences, and improve the treatment of a broad range of mental and physical health conditions. In this talk I will provide an introduction to some of the concepts and methods from computer vision which underlie the analysis of first person videos. In particular, I will focus the automatic analysis of video to track the motion of the camera and recover the 3D geometry of the scene, recognize activities, and detect and recognize objects of interest. I will also briefly discuss the role of visual attention in FPV. The presentation won’t assume any prior knowledge of computer vision. This is the first talk in a series of two talks, and the second presentation, scheduled for Aug. 20, will focus on specific FPV technologies in the context of MD2K.

Learning Objectives:

Following the presentation, attendees will be able to:

  • Describe some basic analysis goals for first person video and identify some of the challenges posed by automatic video analysis.
  • Summarize the relationship between the movement of a body-worn camera in 3D, the motion induced in a video sequence, and methods for estimating video motion.
  • Outline a basic approach to activity recognition in first person video using either object or motion features, including the major system components and sources of error.

Suggested Reading

About James Rehg:

Dr. James M. Rehg  (pronounced “ray”) is a Professor in the School of Interactive Computing at the Georgia Institute of Technology, where he is co-Director of the Computational Perception Lab (CPL) and Director of the Center for Behavioral Imaging.He also serves as the Deputy Director of the NIH Center of Excellence on Mobile Sensor Data-to-Knowledge (MD2K). More about James Rehg.

Leave a comment