Press Release 05-005
New Image Sensor will Show what the Eyes See, and a Camera Cannot
Software behind the technology already finding its way into photo editing
January 12, 2005
ARLINGTON, Va.— Researchers are developing new technologies that may give robots the visual-sensing edge they need to monitor dimly lit airports, pilot vehicles in extreme weather and direct unmanned combat vehicles.
The researchers intend to create an imaging chip that defeats the harmful effects of arbitrary illumination, allowing robotic vision to leave the controlled lighting of a laboratory and enter the erratic lighting of the natural world. In a first step, the researchers have now developed software that simulates the chip circuitry, a program that alone is capable of uncovering hidden detail in existing images.
Designed by robot-vision expert, Vladimir Brajovic, and his colleagues at Intrigue Technologies, Inc.—a spin-off of the team’s Carnegie Mellon University research—the new optical device will work more like a retina than a standard imaging sensor.
Just as neurons in the eye process information before sending signals to the brain, the pixels of the new device will "talk" to each other about what they see. The pixels will use the information to modify their behavior and adapt to lighting, ultimately gathering visual information even under adverse conditions.
Through an online demonstration, the simulator software plug-in, dubbed Shadow Illuminator, has processed more than 80,000 pictures from around the world. By balancing exposure across images, clearing away "noise" and improving contrast, the software revealed missing textures, exposed concealed individuals and even uncovered obscured features in medical x-ray film.
This new approach counters a persistent problem for computer-vision cameras – when capturing naturally lit scenes, a camera can be as much of an obstacle as it is a tool. Despite careful attention to shutter speeds and other settings, the brightly illuminated parts of the image are often washed out, and shadowy parts of the image are completely black.
The mathematical churning behind that process will allow pixels to "perceive" reflectance—a surface property that determines how much incoming light reflects off an object, light that a camera can capture.
Light illuminating an object helps reveal reflectance to a camera or an eye. However, illumination is a necessary evil, says Brajovic.
"Most of the problems in robotic imaging can be traced back to having too much light in some parts of the image and too little in others," he says, "and yet we need light to reveal the objects in a field of view."
To produce images that appear uniformly illuminated, the researchers created a system that widens the range of light intensities a sensor can accommodate.
According to Brajovic, limitations in standard imaging sensors have hindered many vision applications, such as security and surveillance, intelligent transportation systems, and defense systems – not to mention ruining a few cherished family photos.
The researchers hope the new technology will yield high-quality image data, despite natural lighting, and ultimately improve the reliability of machine-vision systems, such as those for biometric identification, enhanced X-ray diagnostics and space exploration imagers.
Additional comments from the researcher:
"The washed out and underexposed images captured by today’s digital cameras are simply too confusing for machines to interpret, ultimately leading to failure of their vision systems in many critical applications." – Vladimir Brajovic, Carnegie Mellon University and Intrigue Technologies, Inc.
"Often, when we take a picture with a digital or film camera, we are disappointed that many details we remember seeing appear in the image buried in deep shadows or washed out in overexposed regions. This is because our eyes have a built-in mechanism to adapt to local illumination conditions, while our cameras don't. Because of this camera deficiency, robot vision often fails." – Vladimir Brajovic
Vladimir Brajovic's image sensor research has been supported by two NSF awards:
SBIR Phase I: Reflectance Sensitive Image Sensor for Illumination-Invariant Visual Perception
ITR: Sensory Level Computation and Information Encoding for Robust Imaging
Joshua A. Chamot, NSF, (703) 292-8070, email@example.com
Anne Watzman, Carnegie Mellon University, (412) 268-3830, firstname.lastname@example.org
Junku Yuh, NSF, (703) 292-8930, email@example.com
Muralidharan S. Nair, NSF, (703) 292-7059, firstname.lastname@example.org
Winslow Sargeant, NSF, (703) 292-7313, email@example.com
Vladimir Brajovic, Carnegie Mellon University, Intrigue Technologies, (412) 855-8780, firstname.lastname@example.org
Additional information on Shadow Illuminator: http://www.intriguetek.com/
Carnegie Mellon press release: http://news.cs.cmu.edu/Releases/demo/132.html
Additional Carnegie Mellon press release: http://www-2.cs.cmu.edu/afs/cs/usr/brajovic/www/labweb/in_the_news.htm
The National Science Foundation (NSF) is an independent federal agency that supports fundamental research and education across all fields of science and engineering. In fiscal year (FY) 2014, its budget is $7.2 billion. NSF funds reach all 50 states through grants to nearly 2,000 colleges, universities and other institutions. Each year, NSF receives about 50,000 competitive requests for funding, and makes about 11,500 new funding awards. NSF also awards about $593 million in professional and service contracts yearly.
Get News Updates by Email
Useful NSF Web Sites:
NSF Home Page: http://www.nsf.gov
NSF News: http://www.nsf.gov/news/
For the News Media: http://www.nsf.gov/news/newsroom.jsp
Science and Engineering Statistics: http://www.nsf.gov/statistics/
Awards Searches: http://www.nsf.gov/awardsearch/