skip to main content
Geometry and photometry in three-dimensional visual recognition
Publisher:
  • Massachusetts Institute of Technology
  • 201 Vassar Street, W59-200 Cambridge, MA
  • United States
Order Number:Not available from Univ. Microfilms Int.
Bibliometrics
Skip Abstract Section
Abstract

This thesis addresses the problem of visual recognition under two sources of variability: geometric and photometric. The geometric deals with the relation between 3D objects and their views under parallel, perspective, and central projection. The photometric deals with the relation between 3D matte objects and their images under changing illumination conditions. Taken together, an alignment-based method is presented for recognizing objects viewed from arbitrary viewing positions and illuminated by arbitrary settings of light sources.

In the first part of the thesis we show that a relative non-metric structure invariant that holds under both parallel and central projection models can be defined relative to four points in space and, moreover, can be uniquely recovered from two views regardless of whether one or the other was created by means of parallel or central projection. As a result, we propose a method that is useful for purposes of recognition (via alignment) and structure from motion, and that has the following properties: (i) the transition between projection models is natural and transparent, (ii) camera calibration is not required, and (iii) structure is defined relative to the object and does not involve the center of projection.

The second part of this thesis addresses the photometric aspect of recognition under changing illumination. First, we argue that image properties alone do not appear to be generally sufficient for dealing with the effects of changing illumination; we propose a model-based approach instead. Second, we observe that the process responsible for factoring out the illumination during the recognition process appears to require more than just contour information, but just slightly more. Taken together, we introduce a model-based alignment method that compensates for the effects of changing illumination by linearly combining model images of the object. The model images, each taken from a different illumination condition, can be converted onto novel images of the object regardless of whether the image is represented by grey-values, sign-bits, or other forms of reduced representations.

The third part of this thesis addresses the problem of achieving full correspondence between model views and puts together the geometric and photometric components into a single recognition system. The method for achieving correspondence is based on combining affine or projective geometry and optical flow techniques into a single working framework. (Copies available exclusively from MIT Libraries, Rm. 14-0551, Cambridge, MA 02139-4307. Ph. 617-253-5668; Fax 617-253-1690.)

Cited By

  1. Mazumdar A and Bora P (2019). Estimation of lighting environment for exposing image splicing forgeries, Multimedia Tools and Applications, 78:14, (19839-19860), Online publication date: 1-Jul-2019.
  2. ACM
    Kemelmacher-Shlizerman I, Shechtman E, Garg R and Seitz S (2014). Moving portraits, Communications of the ACM, 57:9, (93-99), Online publication date: 1-Sep-2014.
  3. Zhang X, Gao Y and Caelli T Parametric manifold of an object under different viewing directions Proceedings of the 12th European conference on Computer Vision - Volume Part V, (186-199)
  4. ACM
    Kemelmacher-Shlizerman I, Shechtman E, Garg R and Seitz S Exploring photobios ACM SIGGRAPH 2011 papers, (1-10)
  5. An G, Wu J and Ruan Q (2010). An illumination normalization model for face recognition under varied lighting conditions, Pattern Recognition Letters, 31:9, (1056-1067), Online publication date: 1-Jul-2010.
  6. Franco A and Nanni L (2009). Fusion of classifiers for illumination robust face recognition, Expert Systems with Applications: An International Journal, 36:5, (8946-8954), Online publication date: 1-Jul-2009.
  7. Han F and Zhu S (2007). A Two-Level Generative Model for Cloth Representation and Shape from Shading, IEEE Transactions on Pattern Analysis and Machine Intelligence, 29:7, (1230-1243), Online publication date: 1-Jul-2007.
  8. Olivares-Mercado J, Sanchez-Perez G, Nakano-Miyatake M and Perez-Meana H Feature extraction and face verification using Gabor and Gaussian mixture models Proceedings of the artificial intelligence 6th Mexican international conference on Advances in artificial intelligence, (769-778)
  9. Tan P, Lin S and Quan L Resolution-enhanced photometric stereo Proceedings of the 9th European conference on Computer Vision - Volume Part III, (58-71)
  10. Kong H, Wang L, Teoh E, Li X, Wang J and Venkateswarlu R (2005). 2005 Special Issue, Neural Networks, 18:5-6, (585-594), Online publication date: 1-Jun-2005.
  11. ACM
    Zhao W, Chellappa R, Phillips P and Rosenfeld A (2003). Face recognition, ACM Computing Surveys (CSUR), 35:4, (399-458), Online publication date: 1-Dec-2003.
  12. Seitz S and Kutulakos K (2019). Plenoptic Image Editing, International Journal of Computer Vision, 48:2, (115-129), Online publication date: 1-Jul-2002.
  13. Maki A, Watanabe M and Wiles C (2019). Geotensity, International Journal of Computer Vision, 48:2, (75-90), Online publication date: 1-Jul-2002.
  14. Carceroni R and Kutulakos K (2019). Multi-View Scene Capture by Surfel Sampling, International Journal of Computer Vision, 49:2-3, (175-214), Online publication date: 1-Sep-2002.
  15. Georghiades A, Belhumeur P and Kriegman D (2001). From Few to Many, IEEE Transactions on Pattern Analysis and Machine Intelligence, 23:6, (643-660), Online publication date: 1-Jun-2001.
  16. Nishino K, Sato Y and Ikeuchi K (2001). Eigen-Texture Method, IEEE Transactions on Pattern Analysis and Machine Intelligence, 23:11, (1257-1265), Online publication date: 1-Nov-2001.
  17. Wiles C, Maki A and Matsuda N (2001). Hyperpatches for 3D Model Acquisition and Tracking, IEEE Transactions on Pattern Analysis and Machine Intelligence, 23:12, (1391-1403), Online publication date: 1-Dec-2001.
  18. La Cascia M, Sclaroff S and Athitsos V (2000). Fast, Reliable Head Tracking under Varying Illumination, IEEE Transactions on Pattern Analysis and Machine Intelligence, 22:4, (322-336), Online publication date: 1-Apr-2000.
  19. Belhumeur P, Kriegman D and Yuille A (2019). The Bas-Relief Ambiguity, International Journal of Computer Vision, 35:1, (33-44), Online publication date: 1-Nov-1999.
  20. Yuille A, Snow D, Epstein R and Belhumeur P (1999). Determining Generative Models of Objects Under Varying Illumination, International Journal of Computer Vision, 35:3, (203-222), Online publication date: 1-Dec-1999.
  21. Donato G, Bartlett M, Hager J, Ekman P and Sejnowski T (1999). Classifying Facial Actions, IEEE Transactions on Pattern Analysis and Machine Intelligence, 21:10, (974-989), Online publication date: 1-Oct-1999.
  22. Sapiro G (1999). Color and Illuminant Voting, IEEE Transactions on Pattern Analysis and Machine Intelligence, 21:11, (1210-1215), Online publication date: 1-Nov-1999.
  23. Belhumeur P and Kriegman D (1998). What Is the Set of Images of an Object Under All Possible Illumination Conditions?, International Journal of Computer Vision, 28:3, (245-260), Online publication date: 1-Jul-1998.
  24. Jones M and Poggio T (1998). Multidimensional Morphable Models, International Journal of Computer Vision, 29:2, (107-131), Online publication date: 29-Aug-1998.
  25. Kutulakos K and Vallino J (1998). Calibration-Free Augmented Reality, IEEE Transactions on Visualization and Computer Graphics, 4:1, (1-20), Online publication date: 1-Jan-1998.
Contributors
  • Hebrew University of Jerusalem

Recommendations