
By combining the information from the Kinect depth frame in (a) with polarized photographs, MIT researchers reconstructed the 3-D surface shown in (c). Polarization cues can allow coarse depth sensors like Kinect to achieve laser scan quality (b).
Courtesy of the researchers
Algorithms exploiting light’s polarization boost resolution of commercial depth sensors 1,000-fold
MIT researchers have shown that by exploiting the polarization of light — the physical phenomenon behind polarized sunglasses and most 3-D movie systems — they can increase the resolution of conventional 3-D imaging devices as much as 1,000 times.
The technique could lead to high-quality 3-D cameras built into cellphones, and perhaps to the ability to snap a photo of an object and then use a 3-D printer to produce a replica.
Further out, the work could also abet the development of driverless cars.
“Today, they can miniaturize 3-D cameras to fit on cellphones,” says Achuta Kadambi, a PhD student in the MIT Media Lab and one of the system’s developers. “But they make compromises to the 3-D sensing, leading to very coarse recovery of geometry. That’s a natural application for polarization, because you can still use a low-quality sensor, and adding a polarizing filter gives you something that’s better than many machine-shop laser scanners.”
The researchers describe the new system, which they call Polarized 3D, in a paper they’re presenting at the International Conference on Computer Vision in December. Kadambi is the first author, and he’s joined by his thesis advisor, Ramesh Raskar, associate professor of media arts and sciences in the MIT Media Lab; Boxin Shi, who was a postdoc in Raskar’s group and is now a research fellow at the Rapid-Rich Object Search Lab; and Vage Taamazyan, a master’s student at the Skolkovo Institute of Science and Technology in Russia, which MIT helped found in 2011.
When polarized light gets the bounce
If an electromagnetic wave can be thought of as an undulating squiggle, polarization refers to the squiggle’s orientation. It could be undulating up and down, or side to side, or somewhere in-between.
Polarization also affects the way in which light bounces off of physical objects. If light strikes an object squarely, much of it will be absorbed, but whatever reflects back will have the same mix of polarizations that the incoming light did. At wider angles of reflection, however, light within a certain range of polarizations is more likely to be reflected.
This is why polarized sunglasses are good at cutting out glare: Light from the sun bouncing off asphalt or water at a low angle features an unusually heavy concentration of light with a particular polarization. So the polarization of reflected light carries information about the geometry of the objects it has struck.
This relationship has been known for centuries, but it’s been hard to do anything with it, because of a fundamental ambiguity about polarized light. Light with a particular polarization, reflecting off of a surface with a particular orientation and passing through a polarizing lens is indistinguishable from light with the opposite polarization, reflecting off of a surface with the opposite orientation.
This means that for any surface in a visual scene, measurements based on polarized light offer two equally plausible hypotheses about its orientation. Canvassing all the possible combinations of either of the two orientations of every surface, in order to identify the one that makes the most sense geometrically, is a prohibitively time-consuming computation.
Polarization plus depth sensing
To resolve this ambiguity, the Media Lab researchers use coarse depth estimates provided by some other method, such as the time a light signal takes to reflect off of an object and return to its source. Even with this added information, calculating surface orientation from measurements of polarized light is complicated, but it can be done in real-time by a graphics processing unit, the type of special-purpose graphics chip found in most video game consoles.
The researchers’ experimental setup consisted of a Microsoft Kinect — which gauges depth using reflection time — with an ordinary polarizing photographic lens placed in front of its camera. In each experiment, the researchers took three photos of an object, rotating the polarizing filter each time, and their algorithms compared the light intensities of the resulting images.
On its own, at a distance of several meters, the Kinect can resolve physical features as small as a centimeter or so across. But with the addition of the polarization information, the researchers’ system could resolve features in the range of tens of micrometers, or one-thousandth the size.
Read more: Making 3-D imaging 1,000 times better
The Latest on: 3-D imaging
via Google News
The Latest on: 3-D imaging
- Atlanta’s Top Rated OB-Gyn Medical Group Practice Now Offers 3D & 4D Ultrasound Imaging.on March 1, 2021 at 3:22 pm
Leading OB-Gyn medical group extends their services by providing the latest 3D/4D obstetrical ultrasound in Snellville and Atlanta areas.
- Elevated Demand For Mini C-Arms Devices Boosts Surgical Imaging Arms Market Growthon March 1, 2021 at 2:05 pm
High demand for mini-C-arms devices is expected to boost market growth during the forecast period. Globally, companies are engaged in the development and manufacturing of novel eq ...
- Pulmonary blood volume estimation in mice by magnetic particle imaging and magnetic resonance imagingon March 1, 2021 at 6:16 am
This methodical work describes the measurement and calculation of pulmonary blood volume in mice based on two imaging techniques namely by using magnetic particle imaging (MPI) and cardiac magnetic ...
- Fusion Robotics wins FDA clearance for 3D imaging robotic targeting systemon February 26, 2021 at 9:56 am
Fusion Robotics this week announced that it has received FDA 510(k) clearance for its robotics system. The Fusion Robotics System is a 3D imaging compatible navigation and robotic targeting system for ...
- Direct imaging and electronic structure modulation of moiré superlattices at the 2D/3D interfaceon February 26, 2021 at 2:40 am
Here, advanced scanning transmission electron microscopy techniques are used to image the atomic structure at the interface between 2D MoS2 and 3D Au nanoislands, revealing a moiré superlattice and ...
- Nanocrystals for high-resolution X-ray imaging of curved 3D objectson February 25, 2021 at 12:37 am
X-ray imaging is widely used in areas such as healthcare and forensic science, but existing X-ray machines are unable to capture curved three-dimensional (3D) objects at high resolution, and they are ...
- Global 3D Imaging in Smartphone Market 2020 Key Business Strategies, Technology Innovation and Regional Data Analysis to 2025on February 24, 2021 at 7:24 pm
Global 3D Imaging in Smartphone Market 2020 by Company, Regions, Type and Application, Forecast to 2025 is the latest addition by MarketsandResearch.biz that is based on the market and extends over ...
- 3D Imaging Market Analysis, Segments, Growth and Value Chain 2027 | By Kenneth Researchon February 22, 2021 at 5:28 am
Global 3D Imaging Market is valued approximately at USD 12.9 billion in 2019 and is anticipated to grow with a ...
- Researchers use 3D imaging to detect diseaseson February 19, 2021 at 3:06 pm
With continued advancements in medical technology, officials are discovering more and more new ways to detect life-threatening diseases. A research team organized by UMass Lowell is ...
- 3D Imaging Market Size Expected to Reach $21,341 Million Globally by the Upcoming Years 2025on February 17, 2021 at 9:46 pm
Allied Analytics –:The 3D imaging market is driven by factors such as technological advancements in electronics sector and increase in demand for high quality resolution and 3D displays. Moreover, ...
via Bing News