User:Toweryang/sandbox

An oversampled binary image sensor is a new image sensor that is reminiscent of traditional photographic film. Each pixel in the sensor has a binary response, giving only a one-bit quantized measurement of the local light intensity. The response function of the image sensor is non-linear and similar to a logarithmic function, which makes the sensor suitable for high dynamic range imaging.

Introduction
Before the advent of digital image sensors, photography, for the most part of its history, used film to record light information. At the heart of every photographic film are a large number of light-sensitive grains of silver-halide crystals. During exposure, each micron-sized grain has a binary fate: Either it is struck by some incident photons and becomes "exposed", or it is missed by the photon bombardment and remains "unexposed". In the subsequent film development process, exposed grains, due to their altered chemical properties, are converted to silver metal, contributing to opaque spots on the film; unexposed grains are washed away in a chemical bath, leaving behind them transparent regions on the film. Thus, in essence, photographic film is a binary imaging medium, using local densities of opaque silver grains to encode the original light intensity information. Thanks to the small size and large number of these grains, one hardly notices this quantized nature of film when viewing it at a distance, observing only a continuous gray tone.

The oversampled binary image sensor is reminiscent of photographic film. Each pixel in the sensor has a binary response, giving only a one-bit quantized measurement of the local light intensity. At the start of the exposure period, all pixels are set to 0. A pixel is then set to 1 if the number of photons reaching it during the exposure is at least equal to a given threshold q. One way to build such binary sensors is to modify standard memory chip technology, where each memory bit cell is designed to be sensitive to visible light. With current CMOS technology, the level of integration of such systems can exceed 109~1010 (i.e., 1 giga to 10 giga) pixels per chip. In this case, the corresponding pixel sizes (around 50~nm ) are far below the diffraction limit of light, and thus the image sensor is oversampling the optical resolution of the light field. Intuitively, one can exploit this spatial redundancy to compensate for the information loss due to one-bit quantizations, as is classic in oversampling delta-sigma conversions.

Building a binary sensor that emulates the photographic film process was first envisioned by Fossum, who coined the name digital film sensor. The original motivation was mainly out of technical necessity. The miniaturization of camera systems calls for the continuous shrinking of pixel sizes. At a certain point, however, the limited full-well capacity (i.e., the maximum photon-electrons a pixel can hold) of small pixels becomes a bottleneck, yielding very low signal-to-noise ratios (SNRs) and poor dynamic ranges. In contrast, a binary sensor whose pixels only need to detect a few photon-electrons around a small threshold q has much less requirement for full-well capacities, allowing pixel sizes to shrink further.

The lens
Consider a simplified camera model shown in Fig.1. The $$\lambda_0(x)$$ is the incoming light intensity field. By assuming that light intensities remain constant within a short exposure period, the field can be modeled as only a function of the spatial variable $$x$$. After passing through the optical system, the original light field $$\lambda_0(x)$$ gets filtered by the lens, which acts like a linear system with a given impulse response. Due to imperfections (e.g., aberrations) in the lens, the impulse response, a.k.a. the point spread function (PSF) of the optical system, cannot be a Dirac delta, thus, imposing a limit on the resolution of the observable light field. However, a more fundamental physical limit is due to light diffraction. As a result, even if the lens is ideal, the PSF is still unavoidably a small blurry spot. In optics, such diffraction-limited spot is often called the Airy disk, whose radius $$R_a$$ can be computed as


 * $$R_a = 1.22 \, w f,$$

where $$w$$ is the wavelength of the light and $$f$$ is the F-number of the optical system. Due to the lowpass (smoothing) nature of the PSF, the resulting $$\lambda(x)$$ has a finite spatial-resolution, i.e., it has a finite number of degrees of freedom per unit space.

The sensor


Fig.2 illustrates the binary sensor model. The $$s_m$$ denote the exposure values accumulated by the sensor pixels. Depending on the local values of $$s_m$$, each pixel (depicted as "buckets" in the figure) collects a different number of photons hitting on its surface. $$y_m$$ is the number of photons impinging on the surface of the $$m$$th pixel during an exposure period. The relation between $$s_m$$ and the photon count $$y_m$$ is stochastic. More specifically, $$y_m$$ can be modeled as realizations of a Poisson random variable, whose intensity parameter is equal to $$s_m$$,

As a photosensitive device, each pixel in the image sensor converts photons to electrical signals, whose amplitude is proportional to the number of photons impinging on that pixel. In a conventional sensor design, the analog electrical signals are then quantized by an A/D converter into 8 to 14 bits (usually the more bits the better). But in the binary sensor, the quantizer is 1 bit. In Fig.2, $$b_m$$ is the quantized output of the $$m$$th pixel. Since the photon counts $$y_m$$ are drawn from random variables, so are the binary sensor output $$b_m$$.

Spatial and temporal oversampling
If it is allowed to have temporal oversampling, i.e.,taking multiple consecutive and independent frames without changing the total exposure time $$\tau$$, the performance of the binary sensor is equivalent to the sensor with same number of spatial oversampling under certain condition. It means that people can make trade off between spatial oversampling and temporal oversampling. This is quite important, since technology usually gives limitation on the size of the pixels and the exposure time.

Advantages over traditional sensors
Due to the limited full-well capacity of conventional image pixel, the pixel will saturate when the light intensity is too strong. This is the reason that the dynamic range of the pixel is low. For the oversampled binary image sensor, the dynamic range is not defined for a single pixel, but a group of pixels, which makes the dynamic range high.

Reconstruction


One of the most important problem in the oversampled binary image sensor is to reconstruct the light intensity $$\lambda(x)$$ from the binary measurement $$b_m$$. Maximum likelihood estimator can be used for solving this problem. Fig. 4 shows the results of reconstructing the light intensity from 4096 binary images taken by single photon avalanche diodes (SPADs) camera.