Posted in | News | Light / Image Sensor

IDTechEx Outlines Emerging Sensor Types for Embedded Vision

Anyone who has used a ‘smart lens’ app on their smartphone is familiar with machine vision. Rather than simply reproducing and storing a picture for later viewing, machine vision applies an image processing algorithm to obtain additional insight, such as locating edges or object identification. While few would regard a ‘smart lens’ app as essential, machine vision is a critical component of many industrial processes, such as material sorting and quality control. It is also crucial for ADAS (advanced driver assistance systems) and, ultimately, autonomous vehicles.

Image Credit: IDTechEx

Embedded vision brings these computational capabilities to the ‘edge’. Rather than images being sent from a sensor to a central processor, initial analysis is performed adjacent to the sensor on a dedicated, often application-specific, processor. This greatly reduces data transmission requirements since rather than sending all the acquired information (i.e., each pixel’s intensity over time), only the conclusions (e.g., object locations) are transmitted. The combination of reduced data transmission and application-specific processing also reduces latency, enabling quicker system responses.

Minimizing Size, Weight, and Power (SWAP)

While in many image sensing applications maximizing performance metrics such as resolution and dynamic range is the priority, sensors for embedded vision are typically designed with other priorities in mind. Collected data needs to be good enough to meet the requirements of the image processing algorithm, but since the picture will not be seen, then maximizing image quality is somewhat redundant. Instead, the aim is to make the system as light and compact as possible while minimizing power requirements. This SWAP reduction enables more sensors to be deployed in devices with size and weight constraints, such as drones, while also reducing associated costs. As such, small image sensors are typically deployed, often those originally developed for smartphones. Over time, expect to see greater integration, such as stacking the sensing and processing functionalities.

Miniaturized Spectral Sensing

Embedded vision isn’t restricted to conventional image sensors that detect RGB pixels in the visible range. Monochromatic sensing will likely suffice for simpler algorithms, such as edge detection for object location, reducing sensor cost. Alternatively, using more sophisticated sensors with additional capabilities arguably supports the essence of embedded vision by minimizing subsequent processing requirements.

Adding spectral resolution to image sensors can expedite subsequent processing for applications such as material identification, since the additional spectral dimension generally means a smaller training data set is required. However, meeting the SWAP requirements of most embedded vision systems is challenging for many spectral sensors due to their bulky architectures housing diffractive optics. Emerging approaches aiming to resolve this challenge include MEMS (micro electromechanical systems) spectrometers, increasing optical path length using photonic chips, and adding multiple narrow bandwidth spectral filters to image sensors using conventional semiconductor manufacturing techniques.

Event-Based Vision

Another sensing approach that aims to minimize subsequent processing requirements is event-based vision. Rather than acquiring images at a constant frame rate, with event-based vision, each pixel reports timestamps that correspond to intensity changes. As such, these sensors combine greater temporal resolution of rapidly changing regions with far less data from static background regions, thus reducing data transfer and subsequent processing requirements. Furthermore, because each pixel acts independently, the dynamic range is increased. Crucially, this alternative acquisition approach is occurring within the sensing chip, not with post-processing algorithms, enabling an associated embedded vision system to be simpler since it has less data to handle.

Further Insights

The adoption of embedded vision systems will continue to grow as more devices gain autonomous capabilities, with robotics, industrial vision, and vehicles as the dominant markets. This represents an opportunity not just for compact image sensors, optics, and processing ICs, but also for emerging sensor types such as spectral imaging and event-based vision that can reduce processing requirements and thus support embedded vision’s value proposition of reducing size, weight, power, and latency.

IDTechEx’s report “Emerging Image Sensor Technologies 2023-2033: Applications and Markets” explores a diverse range of image sensing technologies capable of resolutions and wavelength detection far beyond what is currently attainable. This includes the spectral and event-based sensing outlined here, but also InGaAs alternatives for short-wave infrared (SWIR) imaging, thin film photodetectors, perovskite photodetectors, wavefront imaging, and more. Many of these emerging technologies are expected to gain traction across sectors, including robotics, industrial imaging, healthcare, biometrics, autonomous driving, agriculture, chemical sensing, and food inspection.


Please use one of the following formats to cite this article in your essay, paper or report:

  • APA

    IDTechEx Ltd.. (2023, June 08). IDTechEx Outlines Emerging Sensor Types for Embedded Vision. AZoSensors. Retrieved on July 23, 2024 from

  • MLA

    IDTechEx Ltd.. "IDTechEx Outlines Emerging Sensor Types for Embedded Vision". AZoSensors. 23 July 2024. <>.

  • Chicago

    IDTechEx Ltd.. "IDTechEx Outlines Emerging Sensor Types for Embedded Vision". AZoSensors. (accessed July 23, 2024).

  • Harvard

    IDTechEx Ltd.. 2023. IDTechEx Outlines Emerging Sensor Types for Embedded Vision. AZoSensors, viewed 23 July 2024,

Tell Us What You Think

Do you have a review, update or anything you would like to add to this news story?

Leave your feedback
Your comment type

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.