Jump to content

Egocentric vision: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
No edit summary
Line 11: Line 11:
== History ==
== History ==


The idea of using a wearable camera to gather visual data from a first-person perspective dates back to the 70s, when [[Steve Mann]] invented "Eye Glass", a device that, when worn, causes the human eye itself to effectively become both an electronic camera and a television display. But it was only after the introduction to the market of the [[Microsoft SenseCam]] in 2006 that wearable cameras were used for the first time in large scale experimental health research works. The interest of the computer vision community into the egocentric paradigm has been arising slowly entering the 2010s and it is rapidly growing in recent years, boosted by both the impressive advanced in the field of [[wearable technology]] and by the increasingly number of potential applications.
The idea of using a wearable camera to gather visual data from a first-person perspective dates back to the 70s, when [[Steve Mann]] invented "Eye Glass", a device that, when worn, causes the human eye itself to effectively become both an electronic camera and a television display. But it was only after the introduction to the market of the [[Microsoft SenseCam]] in 2006 that wearable cameras were used for the first time in large scale experimental health research works <ref>Doherty, A. R., Hodges, S. E., King, A. C., Smeaton, A. F., Berry, E., Moulin, C. J., ... & Foster, C. (2013). Wearable cameras in health. American journal of preventive medicine, 44(3), 320-323.</ref>. The interest of the computer vision community into the egocentric paradigm has been arising slowly entering the 2010s and it is rapidly growing in recent years <ref>Bolanos, M., Dimiccoli, M., & Radeva, P. (2017). Toward storytelling from visual lifelogging: An overview. IEEE Transactions on Human-Machine Systems, 47(1), 77-90.</ref>, boosted by both the impressive advanced in the field of [[wearable technology]] and by the increasingly number of potential applications.


The prototypical first-person vision system described by Kanade and Hebert,<ref>Kanade, T., & Hebert, M. (2012). First-person vision. Proceedings of the IEEE, 100(8), 2442-2453.</ref> in 2012 is composed by three basic components: a localization component able to estimate the surrounding, a recognition component able to identify object and people, and an activity recognition component, able to provide information about the current activity of the user. Together, these three components provide a complete situational awareness of the user, which in turn can be used to provide assistance to the itself or to the caregiver. Following this idea, the first computational techniques for egocentric analysis focused on hand-related activities and social interaction analysis. Also, given the unconstrained nature of the video and the huge amount of data generated, temporal segmentation and summarization where among the first problem addressed. After almost ten years of egocentric vision (2007 - 2017), the field is still undergoing diversification. Emerging research topics include:
The prototypical first-person vision system described by Kanade and Hebert,<ref>Kanade, T., & Hebert, M. (2012). First-person vision. Proceedings of the IEEE, 100(8), 2442-2453.</ref> in 2012 is composed by three basic components: a localization component able to estimate the surrounding, a recognition component able to identify object and people, and an activity recognition component, able to provide information about the current activity of the user. Together, these three components provide a complete situational awareness of the user, which in turn can be used to provide assistance to the itself or to the caregiver. Following this idea, the first computational techniques for egocentric analysis focused on hand-related activities <ref>Fathi, A., Farhadi, A., & Rehg, J. M. (2011, November). Understanding egocentric activities. In Computer Vision (ICCV), 2011 IEEE International Conference on (pp. 407-414). IEEE.</ref>and social interaction analysis <ref>Fathi, A., Hodgins, J. K., & Rehg, J. M. (2012, June). Social interactions: A first-person perspective. In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on (pp. 1226-1233). IEEE.</ref>. Also, given the unconstrained nature of the video and the huge amount of data generated, temporal segmentation <ref>Poleg, Y., Arora, C., & Peleg, S. (2014). Temporal segmentation of egocentric videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 2537-2544).</ref>and summarization <ref>Lee, Y. J., Ghosh, J., & Grauman, K. (2012, June). Discovering important people and objects for egocentric video summarization. In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on (pp. 1346-1353). IEEE.</ref>where among the first problem addressed. After almost ten years of egocentric vision (2007 - 2017), the field is still undergoing diversification. Emerging research topics include:


* Social saliency estimation <ref>Park, H. S., Jain, E., & Sheikh, Y. (2012). 3d social saliency from head-mounted cameras. In Advances in Neural Information Processing Systems (pp. 422-430).</ref>
* Social saliency estimation <ref>Park, H. S., Jain, E., & Sheikh, Y. (2012). 3d social saliency from head-mounted cameras. In Advances in Neural Information Processing Systems (pp. 422-430).</ref>

Revision as of 20:59, 8 January 2018

Egocentric vision or first-person vision is a sub-field of computer vision that entails analyzing images and videos captured by a wearable camera, which is typically worn on the head or on the chest and naturally approximates the visual field of the camera wearer. Consequently, visual data capture the part of the scene on which the user focuses to carry out the task at hand and offer a valuable perspective to understand the user's activities and their context in a naturalistic setting.

The wearable camera looking forwards is often supplemented with a camera looking inward at the user’s eye and able to measure a user’s eye gaze, which is useful to reveal attention and to better understand the user’s activity and intentions.

History

The idea of using a wearable camera to gather visual data from a first-person perspective dates back to the 70s, when Steve Mann invented "Eye Glass", a device that, when worn, causes the human eye itself to effectively become both an electronic camera and a television display. But it was only after the introduction to the market of the Microsoft SenseCam in 2006 that wearable cameras were used for the first time in large scale experimental health research works [1]. The interest of the computer vision community into the egocentric paradigm has been arising slowly entering the 2010s and it is rapidly growing in recent years [2], boosted by both the impressive advanced in the field of wearable technology and by the increasingly number of potential applications.

The prototypical first-person vision system described by Kanade and Hebert,[3] in 2012 is composed by three basic components: a localization component able to estimate the surrounding, a recognition component able to identify object and people, and an activity recognition component, able to provide information about the current activity of the user. Together, these three components provide a complete situational awareness of the user, which in turn can be used to provide assistance to the itself or to the caregiver. Following this idea, the first computational techniques for egocentric analysis focused on hand-related activities [4]and social interaction analysis [5]. Also, given the unconstrained nature of the video and the huge amount of data generated, temporal segmentation [6]and summarization [7]where among the first problem addressed. After almost ten years of egocentric vision (2007 - 2017), the field is still undergoing diversification. Emerging research topics include:

  • Social saliency estimation [8]
  • Multi-agent egocentric vision systems
  • Privacy preserving techniques and applications
  • Attention-based activity analysis [9]
  • Social interaction analysis [10]
  • Hand pose analysis [11]
  • Ego graphical User Interfaces (EUI)
  • Understanding social dynamics and attention [12]
  • Revisiting robotic vision as egocentric sensing
  • Activity forecasting [13]

Technical challenges

Today's wearable cameras are small and lightweight digital recording devices that can acquire images and videos automatically, without the user intervention, with different resolutions and frame rates, and from a first-person point of view. Therefore, wearable cameras are naturally primed to gather visual information from our everyday interactions since they offer an intimate perspective of the visual field of the camera wearer.

Depending on the frame rate, it is common to distinguish between photo-cameras (also called lifelogging cameras) and video-cameras.

  • The former (e.g., Narrative Clip and Microsoft SenseCam ), are commonly worn on the chest, and are characterized by a very low frame rate (up to 2fpm) that allows to capture images over a long period of time without the need of recharging the battery. Consequently, they offer considerable potential for inferring knowledge about e.g. behaviour patterns, habits or lifestyle of the user. However, due the low frame-rate and the free motion of the camera, temporally adjacent images typically present abrupt appearance changes so that motion features cannot be reliably estimated.
  • The latter (e.g., Google Glass, GoPro), are commonly mounted on the head, and capture conventional video (around 35fps) that allows to capture fine temporal details of interactions. Consequently, they offer potential for in-depth analysis of daily or special activities. However, since the camera is moving with the wearer head, it becomes more difficult to estimate the global motion of the wearer and in the case of abrupt movements, the images can result blurred.

In both cases, since the camera is worn in a naturalistic setting, visual data present an huge variability in terms of illumination conditions and object appearance. Moreover, the camera wearer is not visible in the image and what he/she is doing has to be inferred from the information in the visual field of the camera, implying that important information about the wearer, such for instance as pose or facial expression estimation, is not available.

Applications

A collection of studies published in a special theme issue of the American Journal of Preventive Medicine[14] has demonstrated the potential of lifelogs captured through wearable cameras from a number of viewpoints. In particular, it has been shown that used as a tool for understanding and tracking lifestyle behaviour, lifelogs would enable the prevention of noncommunicable diseases associated to unhealthy trends and risky profiles (such as obesity, depression, etc.). In addition, used as a tool of re-memory cognitive training, lifelogs would enable the prevention of cognitive and functional decline in elderly people.

More recently, egocentric cameras have been used to study human and animal cognition, human-human social interaction, human-robot interaction, human expertise in complex tasks. Other applications include navigation/assistive technologies for the blind,[15] monitoring and assistance of industrial workflows.[16][17]

See also

References

  1. ^ Doherty, A. R., Hodges, S. E., King, A. C., Smeaton, A. F., Berry, E., Moulin, C. J., ... & Foster, C. (2013). Wearable cameras in health. American journal of preventive medicine, 44(3), 320-323.
  2. ^ Bolanos, M., Dimiccoli, M., & Radeva, P. (2017). Toward storytelling from visual lifelogging: An overview. IEEE Transactions on Human-Machine Systems, 47(1), 77-90.
  3. ^ Kanade, T., & Hebert, M. (2012). First-person vision. Proceedings of the IEEE, 100(8), 2442-2453.
  4. ^ Fathi, A., Farhadi, A., & Rehg, J. M. (2011, November). Understanding egocentric activities. In Computer Vision (ICCV), 2011 IEEE International Conference on (pp. 407-414). IEEE.
  5. ^ Fathi, A., Hodgins, J. K., & Rehg, J. M. (2012, June). Social interactions: A first-person perspective. In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on (pp. 1226-1233). IEEE.
  6. ^ Poleg, Y., Arora, C., & Peleg, S. (2014). Temporal segmentation of egocentric videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 2537-2544).
  7. ^ Lee, Y. J., Ghosh, J., & Grauman, K. (2012, June). Discovering important people and objects for egocentric video summarization. In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on (pp. 1346-1353). IEEE.
  8. ^ Park, H. S., Jain, E., & Sheikh, Y. (2012). 3d social saliency from head-mounted cameras. In Advances in Neural Information Processing Systems (pp. 422-430).
  9. ^ Su, Y. C., & Grauman, K. (2016, October). Detecting engagement in egocentric video. In European Conference on Computer Vision (pp. 454-471). Springer International Publishing.
  10. ^ Fathi, A., Hodgins, J. K., & Rehg, J. M. (2012, June). Social interactions: A first-person perspective. In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on (pp. 1226-1233). IEEE.
  11. ^ Rogez, G., Supancic, J. S., & Ramanan, D. (2015). First-person pose recognition using egocentric workspaces. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 4325-4333).
  12. ^ Bettadapura, V., Essa, I., & Pantofaru, C. (2015, January). Egocentric field-of-view localization using first-person point-of-view devices. In Applications of Computer Vision (WACV), 2015 IEEE Winter Conference on (pp. 626-633). IEEE
  13. ^ Bokhari, S. Z., & Kitani, K. M. (2016, November). Long-Term Activity Forecasting Using First-Person Vision. In Asian Conference on Computer Vision (pp. 346-360). Springer, Cham
  14. ^ Doherty, A. R., Hodges, S. E., King, A. C., Smeaton, A. F., Berry, E., Moulin, C. J., ... & Foster, C. (2013). Wearable cameras in health. American journal of preventive medicine, 44(3), 320-323.
  15. ^ Yagi, T., Mangalam, K., Yonetani, R., & Sato, Y. (2017). Future Person Localization in First-Person Videos. arXiv preprint arXiv:1711.11217.
  16. ^ Leelasawassuk, T., Damen, D., & Mayol-Cuevas, W. (2017, March). Automated capture and delivery of assistive task guidance with an eyewear computer: the GlaciAR system
  17. ^ Edmunds, S. R., Rozga, A., Li, Y., Karp, E. A., Ibanez, L. V., Rehg, J. M., & Stone, W. L. (2017). Brief Report: Using a Point-of-View Camera to Measure Eye Gaze in Young Children with Autism Spectrum Disorder During Naturalistic Social Interactions: A Pilot Study. Journal of autism and developmental disorders, 47(3), 898-904.