Overview Of Depth Cameras

There are different types of 3D sensors, that can differ between different methods to acquire world data or how that data is processed in order to represent it in a convenient way. These sensors can vary in many ways such as acquisition method, resolution, range, etc. This page entails a brief overview of the different 3D sensor types and a small survey of what sensors are currently available.

João Alves, Aivero

Last updated:

Know a camera that should be included? Let us know using this form.

The Aivero team and Azure Kinect
The Aivero team and Azure Kinect

What are depth cameras used for?

Depth-sensing cameras enable machines to perceive the environment in three dimensions. Because of the richer information the depth cameras provide, they have gained importance in applications such as vision-guided robotics, inspection and monitoring. 3D data is less susceptible to environmental disturbances such as changing light conditions. Thus, machines become more reliable and precise when performing tasks, such as picking up an object and placing it elsewhere.

How Do Cameras Detect Depth?

Depth Cameras Detect depth mainly using four different methods:

  • Stereo Sensor
  • Time-of-Flight
  • Structured Light
  • Light Detection Ranging

Stereo sensors

Stereo sensors try to mimic human vision by using two cameras facing the scene with some distance between them (baseline, see Figure 1). The images from these cameras are acquired and then used to perform visual feature (relevant visual information) extraction and matching to obtain a disparity map between the cameras’ views.

The disparity information is inversely proportional to depth and can easily be used to obtain the depth map. This also means that some sensors of this type usually only work in feature-rich environments and encounter problems in featureless environments (e.g. a white wall). However, this type of sensor usually allows for greater depth image resolutions than other cameras.

These feature detection and matching computations are done onboard the camera, meaning they usually require more processing power than other methods. Also, the stereo working distance is limited by the baseline between the cameras as the error increases quadratically with objects’ distance to the sensor. Usual applications entail both indoor and outdoor, feature-rich environments.

Figure 1: Stereo Vision

Time-of-Flight

Time of Flight (ToF) sensors flood the scene with light and calculate depth using each photon's time to return to the sensor. This means that each pixel relates to one beam of light projected by the device, which provides more data density, narrower shadows behind objects and more straightforward calibration (no stereo matching).

Since they use light projection, these sensors are sensitive to different types of surfaces, such as very reflective or very dark ones. In these cases, invalid data usually appears in the image. However, they are much more robust to low lighting or dim conditions than stereo sensors which depend on the scene lighting. These sensors can be used both indoors and outdoors.

Structured Light

Structured light (SL) sensors, on the other hand, use a known pattern projected by the IF sensor onto the scene. The way the pattern deforms is used to construct the depth map.

This type of sensor does not require an external light source and is mainly used in indoor environments due to its low robustness to sunlight since interference may occur with the projected light pattern.

Figure 2: Structured Light versus Time of Flight illustration
Figure 2: Structured Light versus Time of Flight illustration

Light Detection Ranging

Similar to ToF but with scanning, light detection and ranging (LiDAR) sensors are more accurate than ToF since they use multiple IR pulses, whereas ToF sensors usually use one light pulse to obtain their image. By mapping the environment point-per-point instead of a burst of points (as in ToF), the overall noise for each point is reduced, producing more accurate results. Furthermore, LiDAR can usually obtain measurements at a greater distance than traditional ToF or Stereo sensors.

However, LiDAR sensors are usually more expensive than other 3D sensors and, since they perform scanning, can be affected by highly dynamic environments (motion blur).

Figure 3: LiDAR Sensor Operation

List of Depth Cameras

Camera V2

Raspberry Pi

Type: Time-of-Flight
Price: $ 77
Dimensions: 2D
Range: From 0 cm to 0 cm.
Depth Res. 0 x 0 px @ 0 fps
Color Res.: 1920 x 1080 px @ 30 fps
Physical Dim. : 0 x 0 x 0 cm.

Helios2

Lucid Vision Labs

Type: Time-of-Flight
Price: $ 0
Dimensions: 2D3D
Range: From 30 cm to 830 cm.
Depth Res. 640 x 480 px @ 0 fps
Color Res.: 0 x 0 px @ 0 fps
Physical Dim. : 0 x 0 x 0 cm.

Helios 2+

Lucid Vision Labs

Type: Time-of-Flight
Price: $ 0
Dimensions: 2D3D
Range: From 30 cm to 600 cm.
Depth Res. 640 x 480 px @ 30 fps
Color Res.: 640 x 480 px @ 30 fps
Physical Dim. : 55 x 55 x 77.5 cm.

Blaze-101

Basler

Type: Time-of-Flight
Price: $ 0
Dimensions: 3D
Range: From 0 cm to 100 cm.
Depth Res. 640 x 480 px @ 30 fps
Color Res.: 640 x 480 px @ 30 fps
Physical Dim. : 100 x 81 x 64 cm.

Zivid One+

Zivid

Type: Structured Light
Price: $ 11850
Dimensions: 2D3D
Range: From 30 cm to 300 cm.
Depth Res. 1920 x 1200 px @ 12 fps
Color Res.: 1920 x 1200 px @ 12 fps
Physical Dim. : 226 x 165 x 86 cm.

Zivid Two

Zivid

Type: Structured Light
Price: $ 12225
Dimensions: 2D3D
Range: From 30 cm to 150 cm.
Depth Res. 1944 x 1200 px @ 30 fps
Color Res.: 1944 x 1200 px @ 30 fps
Physical Dim. : 169 x 56 x 122 cm.