A schematic of the how the lens-less imaging process works, from light collection through encoding the signal to post-processing with computing algorithms.
CREDIT : Xiuxi Pan from Tokyo Tech
A camera usually requires a lens system to capture a focused image, and the lensed camera has been the dominant imaging solution for centuries. A lensed camera requires a complex lens system to achieve high-quality, bright, and aberration-free imaging. Recent decades have seen a surge in the demand for smaller, lighter, and cheaper cameras. There is a clear need for next-generation cameras with high functionality, which are compact enough to be installed anywhere. However, the miniaturization of the lensed camera is restricted by the lens system and the focusing distance required by refractive lenses.
Recent advances in computing technology can simplify the lens system by substituting some parts of the optical system with computing. The entire lens can be abandoned thanks to the use of image reconstruction computing, allowing for a lens-less camera, which is ultra-thin, lightweight, and low-cost. The lens-less camera is gaining traction recently. But thus far, the image reconstruction technique has not been established, resulting in inadequate imaging quality and tedious computation time for the lens-less camera.
Recently, researchers have developed a new image reconstruction method that improves computation time and provides high-quality images. Describing the initial motivation behind the research, a core member of the research team, Prof. Masahiro Yamaguchi of Tokyo Tech, says, “Without the limitations of a lens, the lens-less camera could be ultra-miniature, which could allow new applications that are beyond our imagination.” Their work has been published in Optics Letters.
The typical optical hardware of the lens-less camera simply consists of a thin mask and an image sensor. The image is then reconstructed using a mathematical algorithm, as shown in Fig. 1. The mask and the sensor can be fabricated together in established semiconductor manufacturing processes for future production. The mask optically encodes the incident light and casts patterns on the sensor. Though the casted patterns are completely non-interpretable to the human eye, they can be decoded with explicit knowledge of the optical system.
However, the decoding process—based on image reconstruction technology—remains challenging. Traditional model-based decoding methods approximate the physical process of the lens-less optics and reconstruct the image by solving a “convex” optimization problem. This means the reconstruction result is susceptible to the imperfect approximations of the physical model. Moreover, the computation needed for solving the optimization problem is time-consuming because it requires iterative calculation. Deep learning could help avoid the limitations of model-based decoding, since it can learn the model and decode the image by a non-iterative direct process instead. However, existing deep learning methods for lens-less imaging, which utilize a convolutional neural network (CNN), cannot produce good-quality images. They are inefficient because CNN processes the image based on the relationships of neighboring, “local”, pixels, whereas lens-less optics transform local information in the scene into overlapping “global” information on all the pixels of the image sensor, through a property called “multiplexing”.
The TokyoTech research team is studying this multiplexing property and have now proposed a novel, dedicated machine learning algorithm for image reconstruction. The proposed algorithm, shown in Fig. 2, is based on a leading-edge machine learning technique called Vision Transformer (ViT), which is better at global feature reasoning. The novelty of the algorithm lies in the structure of the multistage transformer blocks with overlapped ”patchify” modules. This allows it to efficiently learn image features in a hierarchical representation. Consequently, the proposed method can well address the multiplexing property and avoid the limitations of conventional CNN-based deep learning, allowing better image reconstruction.
While conventional model-based methods require long computation times for iterative processing, the proposed method is faster because the direct reconstruction is possible with an iterative-free processing algorithm designed by machine learning. The influence of model approximation errors is also dramatically reduced because the machine learning system learns the physical model. Furthermore, the proposed ViT-based method uses global features in the image and is suitable for processing casted patterns over a wide area on the image sensor, whereas conventional machine learning-based decoding methods mainly learn local relationships by CNN.
In summary, the proposed method solves the limitations of conventional methods such as iterative image reconstruction-based processing and CNN-based machine learning with the ViT architecture, enabling the acquisition of high-quality images in a short computing time. The research team further performed optical experiments—as reported in their latest publication in—which suggest that the lens-less camera with the proposed reconstruction method can produce high-quality and visually appealing images while the speed of post-processing computation is high enough for real-time capture. The assembled lens-less camera and the experimental results are shown in Fig. 3 and Fig. 4, respectively.
“We realize that miniaturization should not be the only advantage of the lens-less camera. The lens-less camera can be applied to invisible light imaging, in which the use of a lens is impractical or even impossible. In addition, the underlying dimensionality of captured optical information by the lens-less camera is greater than two, which makes one-shot 3D imaging and post-capture refocusing possible. We are exploring more features of the lens-less camera. The ultimate goal of a lens-less camera is being miniature-yet-mighty. We are excited to be leading in this new direction for next-generation imaging and sensing solutions,” says the lead author of the study, Mr. Xiuxi Pan of TokyoTech, while talking about their future work.
More from: Tokyo Institute of Technology
The Latest Updates from Bing News & Google News
Go deeper with Bing News on:
Image reconstruction computing
- New technology can reveal what’s hidden behind objects using algorithm
When light hits an object, it creates a shadow on the surface behind it. Murray-Bruce and Czajkowski's algorithm analyzes these shadows to create a 3D model of the hidden scene. The technology can ...
- Everything You Need To Know About OpenAI's Sora Video Generator
The newest innovation from OpenAI, Sora, can create videos from simple text prompts, just as DALL-E can do with images. Here is everything we know so far.
- Unravelling the threat of data poisoning to generative AI
Much like the cloud heralded in a new era, so will generative AI, bringing with it new cybersecurity challenges and a significantly changed attack surface. One of the most insidious threats as a ...
- Deciphering dynamics of electric charge
Spiral-tip motion combined with image reconstruction techniques is an approach that can help scientists better understand the behavior of an electric charge at the microscopic level, essential for ...
- New tomographic reconstruction algorithm sets world record
These images are then fed ... recently developed a new reconstruction algorithm, TomoCAM, that leverages advanced mathematical techniques and GPU-based computing. A paper detailing TomoCAM was ...
Go deeper with Google Headlines on:
Image reconstruction computing
[google_news title=”” keyword=”image reconstruction computing” num_posts=”5″ blurb_length=”0″ show_thumb=”left”]
Go deeper with Bing News on:
- Cadence (CDNS) Expands Its Tensilica Portfolio for Sensor Fusion
Cadence Design Systems CDNS expanded its Tensilica IP lineup to cater to the rising computational demands in automotive sensor fusion applications. The latest high-performance Cadence Tensilica Vision ...
- Cadence Expands Tensilica Vision Family with Radar Accelerator and New DSPs Optimized for Automotive Applications
Cadence Design Systems, Inc. (Nasdaq: CDNS) today expanded its Tensilica IP portfolio to address the increasing computational requirements associated with automotive sensor fusion applications. The ...
- Writer unveils Palmyra-Vision, a multimodal AI to reimagine enterprise workflows
Vision, an enterprise-focused multimodal AI model with advanced visual capabilities, high accuracy benchmarks, and seamless integration to reimagine business workflows.
- Here's why the $3,500 Apple Vision Pro headset is so expensive
Apple's Vision Pro headset starts at $3,499. The Vision Pro includes lots of pricey state-of-the-art parts. One estimate from research firm Omdia puts the "bill of materials" for the headset at $ ...
- Arrest made in connection with theft of transformer
Authorities in Ssezibwa region, in collaboration with Kayunga Central Police Station, have apprehended a suspect identified as Henry Kigozi, in connection with the theft of a transformer belonging ...
Go deeper with Google Headlines on:
[google_news title=”” keyword=”Vision Transformer” num_posts=”5″ blurb_length=”0″ show_thumb=”left”]