![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
. | ![]() |
. |
![]() by Staff Writers Seattle WA (SPX) Oct 16, 2015
Peering into a grocery store bin, it's hard to tell if a peach or tomato or avocado is starting to go bad underneath its skin. But an affordable camera technology being developed by the University of Washington and Microsoft Research might enable consumers of the future to tell which piece of fruit is perfectly ripe or what's rotting in the fridge. The team of computer science and electrical engineers developed HyperCam, a lower-cost hyperspectral camera that uses both visible and invisible near-infrared light to "see" beneath surfaces and capture unseen details. This type of camera is typically used in industrial applications and can cost between several thousand to tens of thousands of dollars. In a paper presented at the UbiComp 2015 conference, the team detailed a hardware solution that costs roughly $800, or potentially as little as $50 to add to a mobile phone camera. They also developed intelligent software that easily finds "hidden" differences between what the hyperspectral camera captures and what can be seen with the naked eye. When HyperCam captured images of a person's hand, for instance, they revealed detailed vein and skin texture patterns that are unique to that individual. That can aid in everything from gesture recognition to biometrics to distinguishing between two different people playing the same video game. As a preliminary investigation of HyperCam's utility as a biometric tool, in a test of 25 different users, the system was able differentiate between hand images of users with 99 percent accuracy. In another test, the team also took hyperspectral images of 10 different fruits, from strawberries to mangoes to avocados, over the course of a week. The HyperCam images predicted the relative ripeness of the fruits with 94 percent accuracy, compared with only 62 percent for a typical camera. "It's not there yet, but the way this hardware was built you can probably imagine putting it in a mobile phone," said Shwetak Patel, Washington Research Foundation Endowed Professor of Computer Science and Engineering and Electrical Engineering at the UW. "With this kind of camera, you could go to the grocery store and know what produce to pick by looking underneath the skin and seeing if there's anything wrong inside. It's like having a food safety app in your pocket," Patel said. Hyperspectral imaging is used today in everything from satellite imaging and energy monitoring to infrastructure and food safety inspections, but the technology's high cost has historically limited its use to industrial or commercial purposes. The UW and Microsoft Research team wanted to see if they could make a relatively simple and affordable hyperspectral camera for consumer uses. "Existing systems are costly and hard to use, so we decided to create an inexpensive hyperspectral camera and explore these uses ourselves," said Neel Joshi, a Microsoft researcher who collaborated on the project. "After building the camera we just started pointing it at everyday objects - really anything we could find in our homes and offices - and we were amazed at all the hidden information it revealed." A typical camera divides visible light into three bands - red, green and blue - and generates images using different combinations of those colors. But cameras that utilize other wavelengths in the electromagnetic spectrum can reveal invisible differences. Near-infrared cameras, for instance, can reveal whether crops are healthy or a work of art is genuine. Thermal infrared cameras can visualize where heat is escaping from leaky windows or an overloaded electrical circuit. "When you look at a scene with a naked eye or a normal camera, you're mostly seeing colors. You can say, 'Oh, that's a pair of blue pants,'" said lead author Mayank Goel, a UW computer science and engineering doctoral student and Microsoft Research graduate fellow. "With a hyperspectral camera, you're looking at the actual material that something is made of. You can see the difference between blue denim and blue cotton." HyperCam, which uses the visible and near-infrared parts of the electromagnetic spectrum, illuminates a scene with 17 different wavelengths and generates an image for each. One challenge in hyperspectral imaging is sorting through the sheer volume of frames produced. The UW software analyzes the images and finds ones that are most different from what the naked eye sees, essentially zeroing in on ones that the user is likely to find most revealing. "It mines all the different possible images and compares it to what a normal camera or the human eye will see and tries to figure out what scenes look most different," Goel said. One remaining challenge is that the technology doesn't work particularly well in bright light, Goel said. Next research steps will include addressing that problem and making the camera small enough to be incorporated into mobile phones and other devices, he said. Co-authors include Eric Whitmire, Alex Mariakakis and the late Gaetano Borriello of the UW's Department of Computer Science and Engineering and T. Scott Saponas, Neel Joshi, Dan Morris, Brian Guenter and Marcel Gavriliu at Microsoft Research.
Related Links University of Washington Satellite-based Internet technologies
|
|
The content herein, unless otherwise known to be public domain, are Copyright 1995-2024 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. General Data Protection Regulation (GDPR) Statement Our advertisers use various cookies and the like to deliver the best ad banner available at one time. All network advertising suppliers have GDPR policies (Legitimate Interest) that conform with EU regulations for data collection. By using our websites you consent to cookie based advertising. If you do not agree with this then you must stop using the websites from May 25, 2018. Privacy Statement. Additional information can be found here at About Us. |