Wafer-level optics is considered to yield imaging lenses for cameras of the smallest possible form factor. The high accuracy of the applied microsystem technologies and the parallel fabrication of thousands of modules on the wafer level make it a hot topic for high-volume applications with respect to quality and costs. However, the adaption of existing materials and technologies from microoptics for the manufacturing of millimeter scale lens diameters led to yield problems due to material shrinkage and z-height accuracy. A multi-aperture approach to real-time vision systems is proposed that overcomes these issues because it relies on microlens arrays. The demonstrated prototype achieves VGA (Video Graphics Array, 640×480 pixels) resolution with a thickness of 1.4 mm, which is a thickness reduction of 50% compared to single-aperture equivalents. The partial images that are separately recorded in different channels are stitched together to form a final image of the whole field of view by means of image processing. Distortion is corrected within the processing chain. The microlens arrays are realized by state-of-the-art micro-optical fabrication techniques on wafer level that are suitable for a potential application in high volume, e.g., for consumer electronic products
We propose a microoptical approach to ultra-compact optics for real-time vision systems that are inspired by the compound eyes of insects. The demonstrated module achieves 720p resolution with a total track length of 2.0 mm which is about 1.5 times shorter than comparable conventional miniaturized optics. The partial images that are separately recorded in multiple optical channels are stitched together to form a final image of the whole FOV by means of image processing. The microlens arrays are realized by microoptical fabrication techniques on wafer-level which are suitable for a potential application in high volume e.g. for consumer electronic products
A digital camera samples the continuous real world. As with any sampling process, questions of aliasing for certain sampling frequencies and the prevention thereof arise. In this paper we will discuss the spatial domain sampling and prevention of aliasing in digital cameras. We focus on the widely used birefringent anti alias filters that are often called optical low pass filters (OLPF). We show 2D models for all contributions to spatial domain sampling and derive optimum filter parameters for minimum aliasing and best possible image sharpness. Compared to previously used selection rules, we can show that the optimum selection of filter parameters can easily deliver more sharpness and reduce aliasing by a factor of 2. The simulated results are finally confirmed in real world experiments
Multi-aperture imaging systems inspired by insect compound eyes promise advances in both miniaturization and cost reduction of digital camera systems. Instead of a single lens stack with size and sag in the order of a few millimeters, the optical system consists of an array of microlenses. At a given field of view of the complete system, the focal lengths of the microlenses is a fraction of the focal length of a single-aperture system, reducing track length and increasing depth of field significantly. As each microimage spans only a small field of view, the optical systems can be simple. Because the microlenses have a diameter of hundreds of microns and a sag of tens of microns, they can be manufactured cost-effectively on wafer scale and with high precision. However, reaching a sufficient resolution for applications such as camera phones has been a challenge so far. We demonstrate a multi-aperture color camera system with approximately VGA resolution (700x550 pixels) and a remarkably short track length of 1.4 mm. The algorithm for correcting optical distortion of the microlenses and combining the microimages into a single image is the focus of this presentation
Adding an array of microlenses in front of the sensor transforms the capabilities of a conventional camera to capture both spatial and angular information within a single shot. This plenoptic camera is capable of obtaining depth information and providing it for a multitude of applications, e.g. artificial re-focusing of photographs. Without the need of active illumination it represents a compact and fast optical 3D acquisition technique with reduced effort in system alignment. Since the extent of the aperture limits the range of detected angles, the observed parallax is reduced compared to common stereo imaging systems, which results in a decreased depth resolution. Besides, the gain of angular information implies a degraded spatial resolution. This trade-off requires a careful choice of the optical system parameters. We present a comprehensive assessment of possible degrees of freedom in the design of plenoptic systems. Utilizing a custom-built simulation tool, the optical performance is quantified with respect to particular starting conditions. Furthermore, a plenoptic camera prototype is demonstrated in order to verify the predicted optical characteristics
Typical image sensors in digital cameras have a fixed sensitivity, and the amount of captured light energy is often controlled by adjusting exposure time and lens aperture. For high end motion imaging these settings are not available as they are used to set motion blur and depth of field, respectively. In many cases a proper exposure is achieved with additional optical filtering, using so called "neutral density" (ND) filters. We propose a digital equivalent of a neutral density filter, which can replace the handling of optical filters for camera systems. It consists of an adjusted sensor readout and in-camera processing of images. Instead of a single long exposure we capture N short exposures. These images are then combined by averaging. The short exposures reduce the sensitivity by a factor of N, while averaging reconstructs motion blur. In addition we also achieve a reduction of both dynamic and fixed pattern noise which leads to an overall increase in dynamic range. The digital ND filter can be used with regular image sensors and does not require hardware modifications
In this contribution, a microoptical imaging system is demonstrated that is inspired by the insect compound eye. The array camera module achieves HD resolution with a z-height of 2.0 mm, which is about 50% compared to traditional cameras with comparable parameters. The FOV is segmented by multiple optical channels imaging in parallel. The partial images are stitched together to form a final image of the whole FOV by image processing software. The system is able to acquire depth maps along with the 2D video and it includes light field imaging features such as software refocusing. The microlens arrays are realized by microoptical technologies on wafer-level which are suitable for a potential fabrication in high volume
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.