MIT Artificial Intelligence Laboratory

The primary goal of the Artificial Intelligence Laboratory is to understand how computers can be made to exhibit intelligence. Current research in the area of vision at the Laboratory includes work on such diverse issues as object recognition, navigation, scene understanding, and active vision. We also have projects which address the topics of visual learning and biological vision. Some particularly exciting applications of our vision research include our image-guided surgery project and our intelligent room endeavor. The faculty members involved in vision related research include Professor Eric Grimson, Professor Berthold Horn, Professor Tomas Lozano-Perez, Professor Tomaso Poggio, and new faculty member Paul Viola.

Demonstration Abstracts

The MIT Cheap Vision Machine (CVM)
Chris Barnhart and Ian Horswill
We will demonstrate a very low cost vision system that packages a 60MFLOP digital signal processor with high-speed memory mapped a memory-mapped RGB/NTSC frame grabber and display. The system uses less than 10 Watts and is small enough for embedded applications. It forms the basis of several of the real-time vision demos. Click here for more information.

Fast Object Recognition in Noisy Images Using Simulated Annealing
Margrit Betke
We will present an automatic object recognition system based on fast simulated annealing. The object recognition problem is addressed as the problem of best describing a match between a hypothesized object and an image. The normalized correlation coefficient is used as a measure of the match. Templates are generated on-line during the search by transforming model images. Simulated annealing reduces the search time by orders of magnitude with respect to an exhaustive search. The algorithm is applied to the problem of how landmarks, e.g., traffic signs, can be recognized by a navigating robot. The performance of the system is illustrated with real-world images of complicated scenes with traffic signs.

Synthesizing Virtual Views of Faces
David Beymer
Given one view of a face, we demonstrate a technique for synthesizing new views of the face as seen from different viewpoints or expressions. To synthesize these "virtual" views, a 2D deformation is measured from 2D views of a prototype face and then mapped onto the given target face. Come see virtual views of your face that are synthesized using a completely automatic technique in near real time! An example.

Medical Image Registration
Gil Ettinger
A key problem in effective analysis of 3D medical imagery is the registration of the scans to different coordinate frames: across modalities, across time, or across patients. We are developing automated 3D medical image registration algorithms which employ a combination of energy-minimization surface alignment techniques to achieve accurate and robust alignment of 3D data sets. We have applied these techniques to the problems of: (1) image-guided surgery, in which we register MR imagery to the patient's coordinate frame for generating "enhanced reality visualizations" of the patient's internal anatomy (An example image-guided surgery), and (2) 3D change detection, in which structural anatomical changes are tracked over time (An example change detection).

Grounding language in visual routines
Ian Horswill
We present a simple natural language understanding program that uses a real-time implementation of Ullman's visual routine processor theory to find the referents of simple noun phrases without a fully-articulated world model.

Image Analysis and Synthesis
Mike Jones and Steve Lines
We present an image analysis and synthesis system which analyses line drawings of cartoon faces and then synthesizes a real image which has roughly the same facial expression and pose. Both the analysis and synthesis modules use 2D prototype images to build a model.

An Active Attentive Visual System for Object Recognition
Aparna Lakshmi Ratan
We present an active and attentive vision system which finds target objects in a scene by integrating color and stereo cues to fixate candidate regions in which to recognize the target objects using alignment-style recognition methods. Click here for more information

Model Guided Correspondence for Recognition
Pamela Lipson
We present a model-guided approach to correspondence that efficiently and robustly establishes a pointwise correspondence between a model and image picture. We have tested our approach within the framework of a linear combination recognition scheme.

Visually-Guided Navigation in Rough Outdoor Terrain
Liana Lorigo
The task is autonomous obstacle avoidance in unmapped rocky terrain, and the platform is a small mobile robot. Preliminary results will be demonstrated.

Enhanced Reality Visualization
J.P. Mellor
Enhanced reality visualization is the process of enhancing an image by adding to it information which is not present in the original image. A wide variety of information can be added to an image ranging from hidden lines or surfaces to textual or iconic data about a particular part of the image. We will demonstrate enhancements which require geometrically accurate positioning. Click here for a sample

Real-time Face Verification
Raquel Romano
We present a real-time face verification system which grabs a live image and automatically authenticates a given user by determining whether a frontal view of the subject is present in the image.

Vision and Touch Guided Manipulation
Salisbury and Slotine (PIs)
A high performance robot and vision system which is capable of autonomously grasping stationary and moving objects.

Haptics
K. Salisbury (PI)
Systems which permit touching and physical interaction with virtual objects.

Human Face Detection in Cluttered Scenes
Kah-Kay Sung
We present a distribution-based modeling cum example-based learning technique for finding human faces in cluttered scenes. During the demo, we will grab live images of subjects in a background of their choice and have the face detection algorithm locate faces in the images.

Real-time vision-based robot mapping
Robert Thau
We present a robot which maps its immediate environs based on camera data, in real-time. The robot runs in an unrestricted office environment; camera and odometry are the only sensors.

Gesture Recognition for Presentation Support
Mark Torrance
A demonstration using a motion-based visual tracking system developed by Sajit Rao, in combination with a continuous speech recognition system developed by the Spoken Language Systems group in the Laboratory for Computer Science, to support the use of audio visual tools during a presentation.

Reubens: A Modular Visual Tracking System
Mike Wessler
Reubens is an active vision system that runs at video rate on a single C-31 DSP chip. It integrates independent saccade and smooth pursuit modules to make the complete system more robust than either of the modules alone.

Alignment by Maximization of Mutual Information
Paul Viola and William Wells
Over the last 30 years the problems of image registration and recognition have proven more difficult than even the most pessimistic might have predicted. Progress has been hampered by the sheer complexity of the relationship between an image and an object, which involves the object's shape, surface properties, position, and illumination. We will present an alignment technique based on mutual information that can work both for real objects/images and for medical registration problems.

Poster Titles

Error Propagation in Full 3d-from-2d Object Recognition
Tao Alter

An Analysis of Shashua and Ullman's Saliency Network
Tao Alter

Extracting Salient Contours Using Shortest Paths
Tao Alter

Segmentation of Brain Tissue from Magnetic Resonance Images
Tina Kapur

Object Recognition via Image Invariances
Pawan Sinha

Accurate Internal Camera Calibration using Rotation, with Analysis of Sources of Error.
Gideon Stein


Back to ICCV '95 Home Page