ADI ToF Depth Sensing Technology: New and Emerging Applications in Industrial, Automotive Markets, and More

Time-of-flight (ToF) cameras have gained attention as the depth sensing method of choice for its smaller form factor, wide dynamic range of sensing, and its ability to operate in a variety of environments. Though ToF technology has been used for years in the scientific and military fields, it has become more prevalent starting in the early 2000s with advances in image sensing technology. This evolution in performance means that technologies such as ADI’s ToF technology will become more ubiquitously deployed beyond the consumer market, where it is currently being designed into smartphones, consumer devices, and gaming devices. As the technology matures, there will be opportunities to further leverage mainstream manufacturing processes to increase system efficiencies in the design, manufacturing, and transport of goods.

Figure 1. Time of flight (ToF): a technology to detect the distance to an object.

Logistics, quality inspection, navigation, robotics, facial recognition, security, surveillance, safety, healthcare, and driver monitoring are all application use cases that can leverage 3D depth sensing ToF technology to solve many problems that traditional 2D technology struggles with today. Combining high resolution depth data with powerful classification algorithms and AI will uncover new applications.

This article will examine the basics of, and the two dominant methods for, ToF depth sensing, as well as compare it to other prevalent depth sensing technologies. It will then provide a detailed introduction to ADI’s 3D depth sensing ToF technology—leveraging the ADDI9036 analog front end, a complete ToF signal processing device that integrates a depth processor that processes the raw image data from a VGA CCD sensor into depth/pixel data. We will also discuss how ADI can scale this technology to our broad market customer base through an ecosystem of hardware partners.

Basic Theory of Operation

Figure 2. Simple diagram of time-of-flight measurement.

A ToF camera measures distance by actively illuminating an object with a modulated light source (such as a laser or an LED), and a sensor that is sensitive to the laser’s wavelength captures the reflected light (Figure 2). The sensor measures the time delay ∆T between when the light is emitted and when the reflected light is received by the camera. The time delay is proportional to twice the distance between the camera and the object (round-trip); therefore the depth can be estimated as:

Equation 1

where c is the speed of light. The goal of a ToF camera is to estimate the delay between the emitted signal and the return signal.

There are different methods for measuring ∆T, of which two have become the most prevalent: the continuous-wave (CW) method and the pulse-based method.

Continuous-Wave Method

Figure 3. Illustration of a continuous-wave ToF system.

In the CW method, a periodic modulated signal is used for the active illumination (Figure 3), and the phase shift of the reflected light is measured by homodyne demodulation of the received signal.

For example, a sinusoidal modulation can be used, where the emitted signal is:

Equation 2


  • As is the amplitude of the signal
  • Bs is the offset of the signal
  • fmod is the modulation frequency, fmod – 1/Tmod where Tmod is the modulation period.

The received signal r(t) is a delayed and attenuated version of the emitted signal:

Equation 3

0 ≤ α < 1 is an attenuation coefficient that depends on the distance and the surface reflectivity, and ∆T is the delay of the return signal.

Continuous-wave time-of-flight sensors measure the distance at each pixel by sampling the correlation function between the received signal r(t) and a demodulating signal g(t) with the same frequency as s(t). In the ideal case, the demodulating signal is also a sine wave:

Equation 4

The operation performed by the pixel is a correlation operation:

Equation 5

When both the emitted signal and the demodulating signal are sine waves, the correlation value as a function of the delay τ applied to the demodulating signal is:

Equation 6

where A = AgAsα and B = αBgBs.

The correlation function c(τ) is then sampled at four equal steps over one period (by changing the illumination phase shift in steps of 90°), as shown in Equation 3. The phase offset Φ=2πfmodΔT between the emitted signal and the demodulating signal can be estimated using Equation 7:

Equation 7

and the depth, which is proportional to the phase shift is:

Equation 8
Figure 4. Illustration of the correlation function sampling process.

Pulse-Based Method

In the pulse-based method, an illumination source emits a series of N short light pulses, which are reflected back to a sensor equipped with an electronic shutter that enables the capture of light in a series of short temporal windows. In Figure 5, three shutter windows, or pulses, are used to capture the reflected light pulse. The BG window captures the ambient light, which is then subtracted from the other measurements.

Figure 5. Illustration of the shutter windows to capture reflected light.

The ToF ∆T is estimated from the measured values corresponding to the different shutter windows according to the following equation:

Equation 9

The distance can then be calculated by replacing ∆T with the expression in Equation 9 in Equation 1, which gives us Equation 10:

Equation 10

It should be noted that these equations are predicated on the assumption that the pulses are perfect rectangular pulses, which is impossible, considering the limitations of the hardware. Additionally, in practical situations, several hundred to several thousand illumination pulses need to be accumulated in order to get a sufficient signal-to-noise ratio (SNR) for measurement.

Advantages/Disadvantages of the CW and Pulsed ToF Technology Systems

Both approaches to ToF have their advantages and disadvantages relative to the application’s use case. Issues such as the distances being measured, the environment in which the system is being used, accuracy requirements, thermal/power dissipation restrictions, form factor, and power supply issues need to be considered. It should be noted that the vast majority of CW ToF systems that have been implemented and are currently on the market use CMOS sensors, while pulsed ToF systems use non-CMOS sensors (notably CCDs). For this reason, the advantages/disadvantages listed as follows will be based on those assumptions:

Advantages of Continuous-Wave Systems:

  • For applications that do not have high precision requirements, a CW system may be easier to implement than a pulse-based system in that the light source does not have to be extremely short, with fast rising/falling edges, though it is difficult to reproduce a perfect sinusoidal wave in practice. However, if precision requirements become more stringent, higher frequency modulation signals will become necessary and may be difficult to implement in practice.
  • Due to the periodicity of the illumination signal, any phase measurement from a CW system measurement will wrap around every 2π, meaning that there will be an aliasing distance. For a system with only one modulation frequency, the aliasing distance will also be the maximum measurable distance. To counteract this limitation, multiple modulation frequencies can be used to perform phase unwrapping, wherein the true distance of an object can be determined if two (or more) phase measurements with different modulation frequencies agree on the estimated distance. This multiple modulation frequency scheme can also be useful in reducing multipath error, which occurs when the reflected light from an object hits another object (or reflects internally within the lens) before returning to the sensor, resulting in measurement errors.
  • As in all CMOS imager systems, standard power supply rails (+5 V, +3.3 V, +1.2 V) can be used, unlike CCDs which may require higher negative (–9 V) and positive (+14 V) supply rails.
  • Depending on their configuration, CMOS ToF imagers tend to have more flexibility and faster readout speed, so functions such as region-of-interest (RoI) output are possible.
  • Calibrating a CW ToF system over temperature may be easier than that of a pulsed ToF system. As the temperature of a system increases, the demodulation signal and the illumination will shift with respect to each other because of the temperature variation, but this shift will only affect the measured distance with an offset error that is constant over the entire range, with the depth linearity remaining essentially stable.

Disadvantages of Continuous-Wave Systems:

  • Though CMOS sensors have higher output data rates compared to that of other sensors, CW sensors require four samples of the correlation function at multiple modulation frequencies, as well as multiframe processing, to calculate depth. The longer exposure time can potentially limit the overall frame rate of the system, or can cause motion blur, which can limit its use to certain types of applications. This higher processing complexity can necessitate an external application processor, which may be beyond the application’s requirements.
  • For longer distance measurements or environments with high levels of ambient light, higher continuous optical power (compared to that of pulsed ToF) will become necessary; this continuous illumination from the laser could cause thermal and reliability issues.

Advantages of Pulse-Based ToF Technology Systems:

  • Pulse-based ToF technology systems often rely on high energy light pulses emitted in very short bursts during a short integration window. This offers the following advantages:
    • It makes it easier to design a system that is robust to ambient light, therefore more conducive to outdoor applications.
    • The shorter exposure time minimizes the effect of motion blur.
  • The duty cycle of the illumination in a pulse-based ToF system is generally much lower than that of a comparable CW system, thereby offering the following benefits:
    • It lowers the overall power dissipation of the system in longer range applications.
    • It avoids interference from other pulsed ToF systems by placing the pulse bursts in a different location in the frame from that of the other systems. This can be done by coordinating the placement of the pulses in the frame of the various systems or by using an external photodetector to determine the location of the other system’s pulses. Another method is to dynamically randomize the location of the pulse bursts, which will eliminate the need to coordinate the timing between the various systems, but it will not completely eliminate the interference.
  • Since the pulse timing and width do not need to be uniform, different timing schemes can be implemented to enable functions such as wider dynamic range and auto-exposure.

Disadvantages of Pulse-Based ToF Technology Systems:

  • Since the pulse width of the transmitted light pulse and the shutter need to be the same, the timing control of the system needs to be highly precise, and a picosecond level of precision may be required depending on the application.
  • For maximum efficiency, the illumination pulse width needs to be very short, but with very high power. For this reason, very fast rising/falling edges (<1 ns) are required from the laser driver.
  • Temperature calibration may be more complicated, compared to that of CW systems since a variation in temperature will affect individual pulse widths, affecting not only the offset and gain, but also its linearity.
  • As previously indicated, the majority of the pulsed systems do not use CMOS sensors. As such:
    • Pulsed ToF systems will almost always require an external analog front end that can digitize and output the depth data (though the CW system may also require an external processor, depending on the complexity of the back-end processing).
    • The system configuration—especially the power supply requirements of the ToF sensor—will require more components and power supply rails.

Other Depth Sensing Technologies

It is instructive to be familiar with other depth mapping technologies to understand the trade-offs; as previously mentioned, all depth detection systems have advantages and disadvantages depending on the use case and the application requirement.

Stereo Vision

Stereo vision for depth sensing works by using more than one camera separated by a certain amount of distance from one another (Figure 6). Like the human eye, a given reference point in space will be in separate positions in each camera, allowing the system to calculate that point’s position in space if the correspondence of that point is made between the two cameras. Determination of this correspondence involves computationally intense and complex algorithms.

Figure 6. 3D depth sensing using stereo vision.


  • No active illumination is needed.
  • Generally cheaper because it only involves two cameras to acquire the data (though a sophisticated application processor may be required to find the corresponding point and develop a 3D image).


  • If there is no contrast to make the corresponding point between the two cameras, distance cannot be calculated. This is especially prevalent in white walls, where there is no contrast between what the two cameras are seeing, and in environments where there is insufficient ambient light.
  • At longer distances, the two cameras need to be located farther apart in order for the corresponding point to be positioned in different locations in the two cameras. This has obvious form factor issues for applications that require the measurement of longer distances.

Structured Light

The structured light method works by projecting a known reference pattern of dots onto an object. The 3D object distorts this reference pattern, and a 2D camera captures this distortion. This distortion is then compared to the reference pattern that was projected, and then calculates a depth map based on the distortion level.

Figure 7. Illustration of depth mapping using the structured light method.


  • Able to achieve very high spatial resolution as well as very high accuracy at close distances (<2 m).


  • Several projections are necessary to extract one frame of information that may lead to slower frame rates, which makes it difficult to extract distance information from moving objects.
  • At longer depth ranges, the camera would need to have its illumination source far from the lens, since the distortion of the pattern may not be discernible if the illumination source is near. This would not be practical to implement in an application that requires a small form factor. For this reason, structured light systems are not commonly used in depth measurement applications longer than 2 m.
  • Outdoor ambient light may also interfere with the distorted pattern, which is the reason why it is better suited for indoor applications.

ADI Depth Sensing (ToF) Technology

ADI’s ToF technology is a pulse-based ToF CCD system (Figure 8) that utilizes a high performance ToF CCD and the ADDI9036, a complete ToF signal processing device that integrates a 12-bit ADC, the depth processor (that processes the raw image data from the CCD into depth/pixel data), as well as a high precision clock generator that generates the timing for both the CCD and the laser. The precision timing core of the timing generator allows the adjustment of clocks and LD output with approximately 174 ps resolution at a clock frequency of 45 MHz.

Figure 8. Block diagram of the ADI ToF system.

ADI’s ToF system differentiates itself from other solutions by:

  • Utilizing a ToF sensor that is 640 × 480 resolution, which is a 4× higher resolution than most other ToF solutions in the market.
  • Using a sensor that has high sensitivity to light at the wavelength of 940 nm. As previously mentioned, ambient light will significantly reduce the SNR of a reflected signal, especially in very high ambient light conditions. 940 nm lasers have become prevalent since this wavelength occupies a space in the spectrum of sunlight where the magnitude of the photon flux is relatively low (Figure 9). The ADI ToF system uses a ToF CCD that is sensitive to 940 nm light, allowing it to capture more data in outdoor environments or areas with strong ambient light.
Figure 9. Photon flux vs. wavelength of sunlight.

A pseudo-randomization algorithm combined with special image processing integrated in the depth processor enables interference cancellation (as previously mentioned). This enables multiple ToF systems to operate in the same environment.

Figure 10. Depth map comparison of an outdoor image.

Figure 10 shows an example where three different depth measurement systems were used outdoors to measure distance. Note that while a CMOS ToF system that used an 850 nm light source has difficulty distinguishing both the person and the tripod, the ToF CCD system was able to distinguish both more clearly.

Which Applications Are Leveraging ToF Technology?

As mentioned in the introduction, the addition of depth information to a 2D image allows useful information to be extracted and can greatly improve the quality of scene information. For example, 2D sensing cannot distinguish between a real person and a photograph. Extracting depth information allows better classification of people, tracking both their facial and body features. ToF depth sensing can provide high quality and reliable face recognition for security authentication. The higher the resolution and depth accuracy, the better the classification algorithm. This can be used for simple functions such as allowing access to mobile devices/our personal home space, or for high end use cases such as secure door access control in commercially sensitive areas.

Figure 11. Digital face recognition.

As depth sensing technologies achieve higher resolutions and depth accuracies, the classification and tracking of people will become easier. The use of artificial intelligence will allow for classification with very high confidence, which in turn will create new and emerging application areas. One use case is for a commercial automatic door opening function, especially in areas of strong sunlight. Ensuring a door only opens for a person and not anything else can provide efficiencies in building management, as well as in security and safety.

Figure 12. People classification for automatic door opening.

As 3D algorithms mature further, data analytics will be leveraged to gather lots of meaningful information about people’s behavior. The first wave of this is likely to happen in building control applications such as door entry/exit systems for people-counting. The addition of depth information from a vertically mounted sensor means that people can be counted with very high accuracy. Another use case is a smart automatic door opening (Figure 13) where people can be classified and the door only opens when a real person is detected. ADI is developing software algorithms in the area of people-counting and classification.

Depth information allows high accuracy classification of people under many challenging conditions, such as environments with low or no ambient light, in areas where there is a significant density of people, and where people are well camouflaged (for example, with hats, scarves, etc.). Most importantly, the false triggering of people-counting is virtually eliminated. Today, stereo cameras can be used for entry/exit detection, but due to the limitations of mechanical size (two sensors) and high processor needs, stereo tends to be expensive and of a large form factor. As the ADI ToF technology directly outputs a depth map and only has onesensor, both the form factor and processing needs are much reduced.

Figure 13. People-tracking algorithm using depth sensing technology.

An important application of depth sensing will be in the industrial, manufacturing, and construction process. The ability to accurately dimension and classify objects in real time through a production process is not trivial. Accurate depth sensing can determine space utilization of warehouse bays. Products that come off a production line need to be dimensioned quickly for transfer. High resolution depth sensing will allow edges and lines of target objects to be determined in real time, and fast volume calculations made. A neural network approach is already being used in such volume determination.

Figure 14. 3D dimensioning.

The autonomous transfer of product within a factory continues to increase. Autonomous vehicles like AGVs (autonomous guided vehicles) will need to self-navigate faster through the factory and warehouse. High accuracy depth sensing technology will allow sensors to map out their environment in real time, localize themselves within that map, and then plot the most efficient navigation path. One of the biggest challenges with the deployment of such technology in factory automation is the interference from other sensors that may operate in the same area. ADI’s interference cancellation IP will allow many such sensors to operate directly in each other’s line of sight without affecting the performance.

Figure 15. Depth sensing use cases in manufacturing.

How Can I Evaluate, Prototype, and Design with ToF Technology?

ADI has developed an optical sensor board (AD-96TOF1-EBZ) that is compatible with the Arrow 96 application processor platform. This 96TOF1 board’s optical specifications are listed in Table 1.

Figure 16. ADI’s 96TOF optical depth sensing board.
Table 1. ADI’s 96TOF Optical Board Specifications
Range <6 m
FOV 90° × 69.2°
Wavelength 940 nm
Frame Rate 30 fps maz
Resolution 640 × 480 pixels

This board can interface directly to Arrow’s 96Boards family. The 96Boards family is a range of hardware processor platforms that make the latest ARM®-based processors available to developers at a reasonable cost. Boards produced to the 96Boards specifications are suitable for rapid prototyping, Qualcomm® SnapdragonTM, and NXP and NVIDIA® processors are all supported within the 96Boards platform.

ToF depth sensing is a complex technology. There is significant optical expertise needed to achieve the highest performance of the VGA sensor. Optical calibration, high speed pulse timing patterns, temperature drifts, and compensation will all affect depth accuracy. Design cycles can be long to achieve the desired performance. While ADI can support chip-down designs for qualified customer opportunities, many customers are looking for an easier, faster, and more efficient approach to get to market.

Many customers are interested in a simple demonstration module to first evaluate the performance of the technology, before moving forward with a real project. ADI has worked with various hardware partners to provide different levels of hardware offerings. The DCAM710 demonstration module, offered by one of our hardware partners (Pico) supports USB depth image streaming directly to a PC.

Figure 17. DCAM710 VGA depth sensing and RGB camera.

DCAM710 Module Specifications

The specifications of the ToF camera DCAM710 module are:

  • Based on ADI’s ToF signal chain products and technology
  • Can output depth map and (710 version) ToF + RGB image (can disable)
  • FOV 70 × 54
  • Depth camera support image size: up to 640 × 480 at 30 FPS
  • RGB camera support image size: up to 1920 × 1080 at 30 FPS
  • USB 2.0 interface
  • Support OS: runs on Android®, Linux®, and Windows® 7/8/10
  • Pico depth sensor SDK, sample code and tools (OpenNI SDK compatible)
  • Sample application algorithms available from ADI in Python®

The Pico SDK software platform can support Windows and Linux operating systems and supports several software functions. A point cloud, which can generate a set of data points in space around an object, and which are often used to generate 3D models, is easily generated through the SDK.

Figure 18. Depth sensing point cloud.

As the demonstration platform streams raw data to a computer via USB, it is easy to develop simple software application algorithms to get customers developing code quickly.

Figure 19. Depth sensing point cloud.
Figure 19. VGA depth sensing streaming to PC via USB.

ADI provides simple sample code in Python to support customer evaluations. The example below shows a screen shot of real-time Python source code being used to detect and classify a person, and then apply a depth measurement to determine where the person is relative to the sensor. Other algorithms that are available include edge detection, object tracking, and a 3D safety curtain.

Figure 20. People classification and range detection.

How Do I Go to Production with ToF?

While the ADI 96TOF reference design is useful for customers doing a chip-down design and the DCAM710 demonstration platform is a cost efficient way to evaluate the technology, in many cases, customers may need a different or more customized solution when they go to production. For example, in AGV systems, a GigE or Ethernet output from an edge node sensing module is often desirable. This gives a robust method of sending high speed, raw depth data from an edge node sensing module to a centralized CPU/GPU controller.

Figure 21. Depth sensing in industrial AGVs (navigation/collision avoidance).

In other applications, customers may want to implement some edge node processing and only send metadata back to the controller. In this case, having a small form factor depth node module with an integrated edge node processor supporting ARM or an FPGA is more desirable. ADI has developed a set of third-party ecosystem partners that can meet all customers’ requirements. These third parties offer a range of capabilities, from complete camera products to small optical modules without external housing that can be integrated into bigger systems. The diagram below shows a tiny MIPI module without external housing that can easily be mechanically integrated into bigger systems. ADI’s partner network can also offer customization of hardware, optics, and application processors if needed. An example of the modules our partners offer today are USB, Ethernet, Wi-Fi, and MIPI, with a range of integrated edge node processors available.

Both ADI and our hardware partners also work with external software partners who bring algorithm expertise in depth processing at the application level.


The advantages of high resolution depth imaging to solve difficult and complex tasks in new and emerging application areas are forcing our customers to quickly drive its adoption. The fastest, lowest risk, and cheapest path to get to market is through affordable, small form factor, high precision modules that can be integrated into larger systems. ADI’s 96TOF reference design platform provides a complete embedded evaluation platform, enabling customers to immediately evaluate the technology and start developing application code. Please contact ADI for more information about ADI’s ToF technology, hardware, or our hardware partners.


Colm Slattery

Colm Slattery

Colm Slattery graduated from the University of Limerick with a bachelor of electronics degree. He joined ADI in 1998 and has held various roles including test development and product and system applications. Colm has also spent 3 years on assignment in a field role in China. Colm is currently a marketing manager in the industrial business unit, focusing on new sensor technologies and new business models.

Yuzo Shida

Yuzo Shida

Yuzo Shida is product marketing manager for ToF products in the Healthcare, Consumer, and Industrial Group at Analog Devices. After graduating from the University of Illinois Urbana-Champaign with an EE degree in 1990, Yuzo joined ADI, where he has been involved in the product and business development of a variety of consumer, automotive, and industrial products.