Blog

What is flying pixel and how can it be mitigated in 3D imaging for ToF cameras?

27th June 2024
Sheryl Miles
0

3D imaging technologies have revolutionised various industries, providing advanced capabilities in depth perception and spatial understanding. Among these technologies, Time-of-Flight (ToF) cameras stand out for their ability to capture depth information in real-time.

Leveraging the principle of measuring the time taken for light to travel to and from objects in a scene, ToF cameras provide depth data with remarkable speed and accuracy. With features like high frame rates, wide dynamic range, and suitability for both indoor and outdoor environments, ToF cameras are well-suited for applications ranging from robotics and automotive to augmented reality and biometrics. However, despite their numerous advantages, ToF cameras are susceptible to certain artifacts, such as flying pixels, which can affect the quality of depth images. Understanding the factors contributing to flying pixels and implementing effective mitigation strategies are essential for optimising the performance of ToF cameras and ensuring reliable depth sensing in various applications.

In this blog, e-con Systems will uncover the concept of flying pixels, explore their impact on 3D imaging quality, and discuss how to mitigate their effects in Time-of-Flight camera.

What is flying pixel?

Flying pixels are erroneous depth readings that manifest as false pixels in the 3D point cloud generated by ToF cameras. They often occur at depth discontinuities in the scene, particularly at object boundaries. The light from foreground and background may combine at the edges to produce intermediate erroneous depth values, leading to the appearance of ghost pixels. While not always evident in depth-mapped images, they become noticeable in 3D point clouds.

Factors contributing to flying pixels

Aperture size

The size of the lens aperture determines the quantity of light reaching the sensor. This factor is pivotal in imaging system design. For instance, in low-light conditions, a larger lens aperture is recommended as it allows more photons to reach the sensor, thereby improving image brightness and quality. However, increasing the lens aperture also reduces the depth-of-field of a camera. It may also increase the likelihood of artifacts such as flying pixels, especially near object boundaries.

Conversely, a narrower aperture may limit the amount of light reaching the sensor, potentially reducing SNR and image quality. However, it may also mitigate the effects of flying pixels by reducing the influence of reflected IR light from certain angles.

Exposure or integration time

The pixels of Time-of-Flight (ToF) cameras require a certain amount of time to gather sufficient light from the observed scene, a parameter referred to as exposure or integration time (texp). During this period, the camera’s position and orientation, as well as the targets observed in the scene, ideally remain constant. However, any relative movement between the camera and the scene can result in the observation of light paths from different targets for each pixel during texp. This phenomenon can lead to an inaccurate estimation of depth at object boundaries, giving rise to motion artifacts and flying pixels.

Interference from other IR sources

Flying pixels may occur due to interference from other infrared (IR) sources, such as sunlight, IR rays emitted by other cameras operating at the same wavelength and frequency, or external devices emitting IR light of similar wavelengths. This interference can disrupt the operation of Time-of-Flight (ToF) cameras, impacting their ability to accurately perceive depth information and resulting in the occurrence of flying pixels.

Pixel size

Another reason for flying pixels is that each pixel in the ToF sensor has a certain physical size. When the edge of an object is measured, a single pixel receives light reflected from both the foreground and background simultaneously. This results in the depth information from different distances being superimposed, leading to raw sensor data containing intermediate depths.

Impact of flying pixels on embedded vision applications

Flying pixels, erroneous depth readings in 3D point clouds generated by Time-of-Flight (ToF) cameras, can significantly impact embedded vision applications. These applications encompass a wide range of systems and devices that utilise cameras and vision sensors for diverse tasks such as object detection, recognition, tracking, navigation, and more.

Autonomous Mobile Robots (AMRs), among the embedded vision applications, play critical roles in tasks ranging from delivery and patrol duties to telepresence and automated weeding autonomously. These robots often rely on embedded cameras to capture images of their surroundings, aiding in tasks such as obstacle detection and object recognition. The captured images and videos are typically processed by an edge processor to analyse them for intelligent decision-making.

The false depth data points or flying pixels can create inaccurate perception of the environment. Consequently, an AMR may perceive flying pixels as obstacles and make decisions based on this incorrect information, potentially avoiding paths that are actually free of obstacles.

While the removal of flying pixels may appear to result in a loss of data, it is crucial for ensuring the accuracy of AMR decision-making processes. Filtering out these erroneous pixels helps prevent AMRs from making incorrect decisions based on false obstacle detections. Thus, despite the potential loss of some data, the filtering of flying pixels is essential for maintaining the reliability and effectiveness of AMR operations in various environments.

Different approaches to remove flying pixels

This article has discussed several factors leading to inaccuracies in depth perception and potentially impacting the performance of applications relying on ToF camera data. It’s important to note that flying pixels are absent or minimal in scenarios characterised by uniform surfaces, such as a plain wall or floor.

Diverse filtering and post-processing techniques allow a ToF camera to reliably generate a 3D point cloud and improve performance in the presence of flying pixels. Here, e-con Systems discuss some of the approaches that are employed to remove flying pixels and improve the reliability of 3D imaging systems:

1. Post-processing filters

Post-processing filters are software-based techniques applied to 2D raw depth images to remove flying pixels. One such approach is the Depth Discontinuity Filter, developed by e-con Systems, alongside other commonly used filters like the Median Filter.

a. Depth Discontinuity Filter

The Depth Discontinuity Filter, developed by e-con Systems, works based on the depth discontinuity of a pixel from its neighbours. If a pixel has a large variation in depth from its neighbouring pixels, it is considered a flying pixel. This filter identifies and removes such pixels, improving the overall quality of the depth map.

The left image demonstrates the depth map generated without the Depth Discontinuity Filter, showcasing the flying pixels. In contrast, the right image illustrates the impact of applying the Depth Discontinuity Filter, resulting in a more accurate depth map with reduced flying pixels.

b. The Median Filter

The Median Filter analyses the depth values of neighbouring pixels and excludes those that deviate significantly from the median value. This filter is based on a user-defined threshold, excluding pixels if the difference between the median value and the nearest neighbour exceeds the threshold. By mitigating outliers, the Median Filter effectively suppresses the influence of flying pixels on the final depth map.

Benefits of post-processing filters

  • Faster processing compared to hardware-based approaches.
  • Effective suppression of flying pixels while retaining valid depth data.
  • Can provide high quality images with less irrelevant data.

Challenges of Post-Processing Filters

  • The frame rate of these filters depends on the host PC configuration.
  • May not completely eliminate all flying pixels in complex scenes.

2. Hardware-Based Approaches

As flying pixels originate in the optical pipeline of ToF cameras, which includes various optical components and processes responsible for light collection and processing, modifying these components requires careful consideration to maintain overall light throughput and signal-to-noise ratio.

Hardware-based approaches involve modifications to the camera’s optical system or sensor design to reduce flying pixels. One such approach is the Mask ToF method.

Mask ToF

Mask ToF augments the camera pipeline by inserting a microlens mask between the sensor pixel and its microlens. This mask selectively blocks incident light paths, allowing each pixel to have its own custom aperture. By reducing stray light paths, Mask ToF effectively reduces flying pixels and improves depth sensing accuracy.

Advantages of Mask ToF

  • Provides custom aperture configurations for each pixel, reducing flying pixels.
  • Offers improved depth sensing accuracy without sacrificing overall light throughput.

Challenges of Mask ToF

  • Requires hardware modifications, making it less practical for some applications.
  • Filtered results may not be as effective as software-based approaches in all scenarios.

Both software-based and hardware-based approaches play a crucial role in removing flying pixels from ToF camera data. While post-processing filters offer fast and efficient removal of flying pixels, hardware-based approaches provide more tailored solutions but may require significant modifications to camera hardware.

Featured products

Product Spotlight

Upcoming Events

View all events
Newsletter
Latest global electronics news
© Copyright 2024 Electronic Specifier