# See Light Move: Compressed Sensing and the World’s Fastest 2-D Camera

By Hans Kaper

It may take one, ten, or even a hundred years, but beautiful ideas always find their way to applications.

In an ordinary digital camera, a large amount of information is collected by the camera’s sensor. This information is immediately processed by compression software, and a much smaller amount is then stored. This process takes time, contributing largely to what is called “shutter lag.” Is it possible to collect only what is needed and thus reduce or even eliminate shutter lag? Thanks to some clever physics and compressed sensing, the answer is “yes.”

The mathematical concepts underlying *compressed* or *compressive sensing* date back to the 1970s, when the merits of the \(\ell^1\) norm were recognized in convex programming. The theory developed rapidly in the 1990s, and a breakthrough occurred in the early years of the 21st century with the work of Candès, Tao, Romberg, and Donoho [2, 3, 5]. Collectively, they realized that one can design an efficient sensing or sampling protocol by capturing useful information in a sparse signal and compressing it into a small amount of data.

These beautiful mathematical ideas have found an application in the world’s fastest 2-D camera. By adding a digital micromirror device for spatial encoding and applying reconstruction algorithms from compressed sensing, Lihong Wang and colleagues at Washington University in St. Louis transformed a traditional one-dimensional streak camera into a 2-D ultrafast imaging camera. Now, for the first time, humans can see the movement of light on the fly (see Figure 1).

Instead of collecting all of the information about a signal or image in pixels in a large \(n\)-dimensional vector \(x\), compressed sensing collects information in a much smaller \(k\)-dimensional vector \(y = Ax\). Here, \(A\) is a (rectangular) matrix that represents the measurement process. If the signal is sparse, \(k\) need not be very large to recover the original signal \(x\) from \(y\); in practice, it usually needs to be no more than four times the sparsity level of \(x\).

**Figure 1.** Still frames from movies captured with compressed ultrafast photography (CUP)shows a light pulse (a) bouncing off a mirror and (b) refracting at an air-resin interface.

## Compressed Sensing

Compressed sensing is based on the observation that most signals or images contain only a small amount of crucial information. By strategically discarding unnecessary data, one can reduce the image’s size while retaining sufficient information to faithfully reconstruct the original. The full image is then reconstructed from the encoded data using linear programming algorithms. To make this possible, compressed sensing relies on two principles: *sparsity* and *incoherence*. Sparsity implies that the image has a concise representation when expressed in the proper basis, while incoherence indicates that unlike the image of interest, the sampling/sensing waveforms have an extremely dense representation in the sparsifying basis. Thus, compressed sensing is a simple and efficient signal acquisition protocol that samples—in a signal-independent fashion—at a low rate and later uses computational power for reconstruction from what appears to be an incomplete set of measurements.

The underdetermined system \(Ax = y\) can be solved in a remarkably efficient way by minimizing the \(\ell^1\) norm \(\| x \|_{1} = \sum_{i=1}^n | x_i |\) of \(x\), subject to the constraint \(y = Ax\). This convex problem can be converted to a linear programming problem, for which we have fast algorithms. When the signal is sufficiently sparse, the recovery via \(\ell^1\) minimization is provably exact [1].

Signals and images are rarely exactly sparse; more often they are only approximately sparse. Additionally, measured data are invariably corrupted by at minimum a small amount of noise; at the very least, small perturbations in the data should cause small perturbations in the reconstruction. So for practical reasons, compressed sensing must be able to deal with both nearly sparse signals and with noise. As was shown by Candès, Romberg, and Tao [4], if \(x\) is not exactly sparse, then the quality of the recovered signal is as good as if one knew the location of the largest values of \(x\) ahead of time and decided to measure those directly. In other words, the reconstruction is nearly as good as that provided by an oracle which, with full and perfect knowledge about \(x\), extracts the most significant pieces of information.

**
**

**The 2-D Camera**

**
**
The 2-D camera, which can record non-repetitive, time-evolving events at one hundred billion frames per second, was developed by Wang and his colleagues using a dynamic 2-D imaging technique called *compressed ultrafast photography* (CUP). CUP combines*streak photography*, an ultrafast imaging technique, with compressed sensing [6].

While traditional cameras capture discrete images by repeated opening and closing of a shutter, streak cameras deflect incoming light in a way that maps time onto a spatial dimension on the camera’s two-dimensional pixel array. The light is deflected by an amount proportional to its time of arrival, a process known as* shearing*.

Traditional streak cameras can film in only one spatial dimension. When a one-dimensional scene like the one shown in Figure 2—where the object moves from left to right—is sheared vertically, the scene’s time domain is translated to the detector’s \(y\)-domain. A movie can then be reconstructed by taking each row of pixels as a frame.

**Figure 2.** A one-dimensional scene is sheared vertically and the scene’s time domain maps to the detector’s y-domain. A movie can be reconstructed by taking each row of pixels as a frame.

By masking the 2-D scene with a known pixel pattern, Wang’s group recorded a 2-D scene in a single shot. The Washington University team was able to achieve this by taking their cues from signal processing. CUP facilitates detection of information about all three dimensions—spatial coordinates \(x\) and \(y\), and time \(t\)—on a single pixel array by encoding the input scene in the spatial domain. The resultant image is then sheared in the streak camera. A 2-D detector array with a single snapshot measures this encoded and sheared three-dimensional scene \((x, y, t)\). The encoded data can retrieve time information from the input scene in the subsequent image reconstruction.

**
**

**Applications**

**
**
The single-shot ultrafast camera is like a microscope for time, one that has many potential applications in fundamental and applied science. One remarkable medical application for the camera is in magnetic resonance imaging (MRI). MRI recovers images from a human body by taking a large number of measurements before reprocessing the data. The scan can take extended periods of time, much to the disadvantage of patients who have to be kept completely still. In the worst cases, the individual’s breathing is stopped, depriving the patient of oxygen for the entire scan. Compressed sensing techniques alleviate this issue by reconstructing the image using many fewer samples. These techniques significantly reduce the number of measurements required for the MRI from minutes to mere seconds.

CUP can potentially be coupled with microscopes to enable the observation of transient events in cell structures. For instance, it could be used to observe energy metabolism in cellular mitochondria. It could also enable understanding of light’s passage through tissues, which could yield important insights into therapies that use lasers to destroy abnormal or diseased tissue, with the objective of keeping healthy tissue unharmed.

As Wang puts it, the generic nature of the camera means it can be coupled with a wide variety of tools—from microscopes to telescopes—and thus be used to film anything from cellular functions to collapsing supernovae [7].

**Acknowledgments**

This article is based on a presentation by Dr. Fleming Crim, NSF Assistant Director for Mathematical and Physical Sciences, for the MPS Advisory Committee on November 5, 2015. The author acknowledges input from Dr. Tie Luo, Acting Deputy Division Director, NSF/MPS/Division of Mathematical Sciences.

**References**

[1] Candès, E., & Romberg, J. (2007). Sparsity and incoherence in compressive sampling. *Inverse Prob., 23*(3), 969985.

[2] Candès, E., & Tao, T. (2006, December). Near optimal signal recovery from random projections: Universal encoding strategies?*IEEE Trans. Inform. Theory, 52*(12), 5406–5425.

[3] Candès, E., Romberg, J., & Tao, T. (2006, February). Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. *IEEE Trans. Inform. Theory, 52*(2), 489–509.

[4] Candès, E, Romberg, J., & Tao, T. (2006, August). Stable signal recovery from incomplete and inaccurate measurements. *Comm. Pure Appl. Math., 59*(8), 1207–1223.

[5] Donoho, D. (2006, April). Compressed sensing. *IEEE Trans. Inform. Theory, 52*(4), 1289–1306.

[6] Gao, L., Liang, J., Li, C., & Wang, L.V. (2014, December 4). Single-shot compressed ultrafast photography at one hundred billion frames per second. *Nature, 516*, 74–77. doi:10.1038/nature14005.

[7] Smart, A.G. (2015, February). An ultrafast camera films light at light speed. *Physics Today, 68*(2), 12. doi: http://dx.doi.org/10.1063/PT.3.2673.

Hans Kaper, founding chair of SIAG/MPE and editor-in-chief of SIAM News, is an adjunct professor of mathematics at Georgetown University.