Image of the IDS EVS camera and a visualization of the change data with motion vectors

Groundbreaking sensor technology

Event-based imaging is also referred to as neuromorphic sensor technology - sensors with the ability to process information in a similar way to the human neural system. To do this, we need to realize that the evolutionary advantage of our brain is not just that it can efficiently process enormous amounts of data from continuous light stimuli from the photoreceptors in our eyes. The ability to react to changes such as differences in brightness, contrasts and movements is crucial, while uniform stimuli are largely ignored. This means that we focus primarily on movements in a scene instead of constantly re-capturing every static detail of our surroundings. This allows our brain to process relevant information quickly without being flooded with unnecessary data.

To replicate this ability, Prophesee has developed special pixel electronics in cooperation with Sony, whose main task is to detect and record only changes. If the contrast of a pixel value changes over a certain threshold value, a so-called "change event" is triggered. Each pixel acts independently of its pixel neighbors and in real time. This means that it is not bound to a fixed time interval (cf. frame rate). The minimum time span between two pixel events is an important property of this sensor and is referred to as "temporal resolution". Sony specifies this as 1 microsecond for the IMX636 sensor. This enables ultra-fast and almost "gapless" scanning of movements. To realize such a fast rate of change with image-based sensors, this would correspond to frame rates of more than 10,000 images per second!

No pictures! Changes only

While image-based sensors always transmit the complete amount of data from the entire sensor surface at regular intervals, an event-based camera often only generates a very small amount of data in the same period of time. This means that application developers do not have to compromise between high frame rates and large amounts of redundant data to accurately capture fast events. This is because the amount of data generated by EVS cameras depends on the activity in the field of view and adapts automatically if the scene conditions change. In contrast to image-based sensors with a fixed frame rate, EVS pixels only transmit information and generate data traffic when changes occur in the field of view.

From motion blur to motion sharpness

Due to their technology, classic sensors can exhibit motion blur during fast movements. It occurs when contrast boundaries (e.g. due to object edges) move across several neighboring pixels while they are exposed. Each pixel picks up light from different positions of the moving object. The faster the movement or the longer the exposure time, the more difficult it is to obtain a clear image without distortion. EVS pixels, on the other hand, continuously evaluate the incident light and only register the increase or decrease in the amount of light in a comparator. If they exceed the set threshold values, they generate ON or OFF change events with a time accuracy of around one microsecond. Even the fastest movements are scanned pixel by pixel using EVS technology. The result is a high-resolution sequence (stream) of independent pixel events that describe a motion path without any motion blur.

Diagram of the light intensity shows when ON and OFF events are generated.
Each EVS pixel continuously records the incident light and generates a "change event" whenever the light intensity exceeds a certain threshold value, either upwards or downwards.

Less data - more efficient information

The native output data of EVS cameras (pixel position X/Y, ON/OFF polarity of the event, timestamp T), i.e. the information content of the stream of change events, is extremely compact and efficient, but does not provide classic images. This makes them ideally suited for processing by machines and algorithms, but less intuitive or directly usable for humans. If we still want to visualize the result stream in images, these are reminiscent of a 2D camera image after edge detection. The reason for this is that changes in contrast during movement are less noticeable on evenly illuminated surfaces and more noticeable on object edges.

Since only relevant and therefore significantly less data is recorded, storage requirements and processing effort are considerably reduced. In addition, the event information already supports the recognition of movement patterns and directions. The time gaps between the recorded events can also be used to directly calculate how fast a pixel or object is moving without having to process a large number of images, for example to separate relevant information from unnecessary static background data.

Due to the small amount of data, many processes can be analyzed almost in real time. Multi-camera systems are also much easier to implement, as they require significantly less technical effort. Both the image processing performance of the host PCs and the peripherals consisting of cabling, power supply and the like can be dimensioned smaller and more cost-effectively.

Time as information

Based on the microsecond-precise time stamp and the position of each individual pixel event, completely new application possibilities arise. The change events already contain valuable information from which further important information can be derived. Conventional cameras with a fixed frame rate cannot capture these due to their constant sampling rate or they are lost in a large amount of redundant data due to the type of output.

The event data for creating slow-motion recordings offers an exciting analysis option. By accumulating the captured pixel events in a temporal grid and generating complete sensor images from them, slow-motion videos with a variable "exposure time" are created. The playback speed remains variable due to the selected accumulation time and display frame rate. It ranges from real time (super slow motion with one frame per event) to actual movement speed (at approx. 1 frame per 33 ms) to a still image. If all recorded events are combined (in terms of time), the complete movement history becomes visible.

Speed and direction information can also be extracted for precise, numerical analysis of object movements. No more complex image processing is required for this. If, on the other hand, the location and time of several pixel events are accumulated over a certain time range in a 3D visualization, the result is a qualitative representation of the movement path. This in turn helps to understand how and on which paths objects move in a (time) domain. This procedure is used, for example, in flow analysis for high-precision detection of the movement of liquids and gases.

Two visualizations show a qualitative and a quantitative representation of the particle flows around an object.
The accumulation of event-based data creates ideal analysis data for flow visualization and quantification.

"For applications that require exceptional frame rates of 1000 Hertz, such as in flow visualization, the implementation with image-based cameras is often very complex and expensive. With the help of event-based camera technology, we achieve comparable frame rates of 10,000 frames per second and more. To transfer the significantly reduced amount of data, however, we only need standard PC interfaces such as USB. This makes this innovative camera technology particularly interesting for smaller teaching and research facilities."

— Dr. André Brunn, Head of Development for Fluid Mechanics at the Aachen-based company iLA_5150 GmbH —

You can read more about event-based flow visualization in our application story: To the case study "only the changes count"

New data - new processing approaches

However, in order to be able to use this new sensor information, developers need to find a different way to handle the previously cyclical image-based processing sequences. Of course, several event data can be combined in classic frames in order to even process them like conventional images with a constant frame rate. However, this method is not necessarily optimal, as it leaves the advantages of data dynamics unused. For example, the high temporal precision for fast movements and the efficient processing of less data at once, which can also reduce energy consumption. Only with the appropriate functions, tools and algorithms can patterns, movements, times and structures be extracted and processed quickly and efficiently from the event data. Today, this is not found in any of the known (image-based) standard vision frameworks.

However, Prophesee and Sony, the manufacturers of the new sensor technology, have already developed corresponding processing methods and made helpful functions available in a software development kit, the Metavision SDK, together with detailed documentation and numerous samples. This means that users can start using the new possibilities of this innovative technology immediately.

To operate the IDS EVS camera "uEye XCP-E", only the IDS HAL plug-in needs to be installed on the host PC. The camera is then immediately ready for use in Prophesee's Metavision SDK. Watch the how-to video now:

In this video, we show you how to get your event-based camera up and running quickly and efficiently.

High precision in real time - EVS in quality assurance?

The capabilities of neuromorphic sensors can also play an important role in quality assurance and improvement. Especially in applications where accuracy, speed and efficiency in fault detection are required. The added value of being able to record the smallest object and material changes in pixel size and in real time is evident, for example, in the monitoring of machines and processes. Thanks to the high temporal resolution, which extends into the low microsecond range, even high-frequency movements such as vibrations or acoustic signals can be visualized. Analyses reveal unusual patterns (e.g. due to wear, malfunctions) at an early stage, which can lead to damage or production downtime.

As they only perceive movements or contrasts, neuromorphic sensors are much less sensitive to changes in light, which makes them far superior to conventional image processing systems in strongly varying lighting conditions (e.g. reflections, shadows). When it comes to rapid fault detection, process monitoring or inspections under difficult conditions, quality assurance processes can only benefit from the capabilities of neuromorphic sensors.

EVS - trend or must-have?

Event-based sensors do not capture complete images, but only pixel changes over time. However, these can be used to dynamically compile very different visualizations, providing applications with significantly more movement information than cameras with conventional image sensors alone could. Therefore, the technologies do not work in competition! Event-based sensors are therefore not a general replacement for classic image-based cameras or even AI-based image processing, but rather a complementary technology. It opens up new additional possibilities when it comes to recording movements. In various applications, a single sensor type or type of result data is not sufficient. A combination of different information and therefore different camera categories is often necessary to provide an optimum solution to a customer requirement. Event-based cameras are therefore interesting and worthwhile components for fast motion analysis, industrial quality assurance tasks, robotics and autonomous systems in general.

Note from IDS
"We already have a number of concrete application examples for analyzing gases, liquids and vibrations for which this technology is ideally suited for the start of series production of the uEye EVS camera."