Technology

Machine/Computer Vision Challenge

Conventional cameras work to reconstruct images for human consumption by measuring the intensity of light on a pixel array at a fixed frame rate (typically 10-60 frames a second) . As a result these cameras have a single output mode, a continuous stream of full frame images. While this has historically been done for the accurate communication and reproduction in electronic and printed media, it makes these cameras inefficient and limiting in computer vision. This is largely due to the fact that this architecture continually captures redundant/irrelevant information while causing a combination of latency, blurring, and inaccurate signal information.

Redundant information is continuously captured even when the scene remains static. Consider the example where there is only motion in 1% of the total scene. In that scenario 99% of the pixels contain no new information yet they are continually read off of the camera and blindly processed.

Latency is incurred because of the tremendous amount of data that needs to be read off of the sensor and it must be completed within a frame time. Consider a camera that is operating at 30 fps. Waiting for the previous frame to be transferred can account for up to 33 ms of delay depending on the exposure time and that doesn't include its own data transfer or processing delays.

Blurring occurs when either an object is in motion or the camera itself is in motion (even when the scene is static). This happens because a point in the scene ends up being captured by multiple pixels during the exposure time.

 

While the exposure time can be shortened to limit the effects of motion blur, it negatively effects the accuracy of the signal. For example, with an exposure time of 1 ms and a fixed frame rate of 30 fps, there are now 32 ms between every frame where no data is captured. That makes the sensor blind for 97% of the time. Signal inaccuracy occurs when the sensor is no longer measuring light (looking at the scene) so it cannot capture anything happening during that time and the signal fidelity suffers.

When used in computer or machine vision, these fundamental limitations result in substantial signal inaccuracies and substantial inefficiencies in processing, power, bandwidth, and latency.

OCULI SPU™: Integrated Neuromorphic Sensing & Processing

Oculi S41 SPU.png

Oculi makes the OCULI SPU™, Sensing and Processing Unit, the only  single-chip vision solution combining sensing + pre-processing in the pixel to deliver real-time vision intelligence (VI) at the edge.

The OCULI SPU™ provides a disruptive architecture inspired by the eye:

  • Integrated sensing + processing

  • Parallel sensing + processing

  • Saliency/features (smart events) output

  • Sparse processing

  • Bi-directional communication

Colored Symbol - Copyright S down - With transparent background- used alone.png

OCULI SPU™ is the first practical silicon that closely mimics biology in selectivity, parallel processing, and efficiency but outperforms in speed.

OCULI SPU™: Advantages

Artboard 46.png
Captures high speed dynamics

High temporal resolution

Artboard 46 copy.png
Manages extreme lighting
 

140dB dynamic range

​​

Artboard 46 copy_4.png
Wavelength and color agnostic

Demonstrated in both visible and IR

Artboard 46 copy_8.png
Real time image statistics at the edge

On-chip Image Signal Processor

Artboard 46 copy_1.png
Low latency signal to information

us vs 10's ms

Artboard 46 copy_3.png
Lower total size, weight, and power

mW vs 100's mW

​​

Artboard 46 copy_7.png
Post-processing agnostic
 

Compatible with standard AI algorithms and general-purpose processors

Artboard 46 copy_2.png
Low bandwidth and low processing

Reduced bandwidth and external processing up to 99% with zero loss of relevant data.

Artboard 46 copy_5.png
Versatile multimodal technology

event, smart event, frame, actionable signal

Artboard 46 copy_6.png
Software-defined architecture

Programmable features on-the-fly

Subscribe to Our Site

Thanks for subscribing!