Silicon Retina: History, Live Demo, and Whiteboard Pixel Design
Rockwood Memorial Lecture 2023: Tobi Delbruck, Institute of Neuroinformatics, UZH-ETH Zürich
Event Camera Silicon Retina; History, Live Demo, and Whiteboard Circuit Design
Rockwood Memorial Lecture 2023 (11/20/23)
https://inc.ucsd.edu/events/rockwood/
Hosted by: Terry Sejnowski, Ph.D. and Gert Cauwenberghs, Ph.D.
Organized by: Institute for Neural Computation, https://inc.ucsd.edu
Abstract: Event cameras electronically model spike-based sparse output from biological eyes to reduce latency, increase dynamic range, and sparsify activity in comparison to conventional imagers. Driven by the need for more efficient battery powered, always-on machine vision in future wearables, event cameras have emerged as a next step in the continued evolution of electronic vision. This lecture will have 3 parts: 1. A brief history of silicon retina development starting from Fukushima’s Neocognitron and Mahowald and Mead’s earliest spatial retinas; 2: A live demo of a contemporary frame-event DAVIS camera that includes an inertial measurement unit (IMU) vestibular system, 3: (targeted for neuromorphic analog circuit design students in the BENG 216 class), a whiteboard discussion about event camera pixel design at the transistor level, highlighting design aspects of event camera pixels which endow them with fast response even under low lighting, precise threshold matching even under large transistor mismatch, and temperature-independent event threshold.
Is here scientific grade event cameras ?
ReplyDeleteYes, at around 45:30 he talks about SciDVS.
ReplyDeletethank you
DeleteWhat are principal applications for which event sensors are must to have please? thanks
ReplyDeleteYou might want to watch the lecture to answer that!
DeleteI mean the real commercial applications ...
ReplyDeleteI think Car industry main driver. May be military little bit. Someone who have very high dynamic range scenes, need low power consumption, automatic movement detection without complex programming
DeleteThe ability to stream visual information at milli-second increments and at very low power opens many potential applications in which Frame based approaches are limited. Blurr correction is an imediate application that could be applied on Mobile cameras, high speed imaging in automation is another. In the future, event cameras may be the most awaited breaktrough for AR/VR to interface correctly with human eyes and also the outside world. The same could be said in the context of Automotive and Autonomous Driving.
DeleteWonderful to see the link to the Toby Delbruck lecture here. I would additionally suggest "The Silicon Eye" by George F. Gilder - well written book that allows great insights into the neuromorphic imagers development and people connected to it. It has always been marvelous, how early on the key components have been invented and how long it takes mainstream products to develop from those inventions.
ReplyDeleteMaybe the MP won't come until spectacles finally come to MP.... we'll see. But obviously visual prosthetics for the blind that have to have low latency, HDR and low power consumption along with a sparse output that can drive inference hardware that can exploit this sparsity and a form of data that is closer to what the brain uses already.
ReplyDeleteBut main Problems eyes implants not sensors. Rather cost and connection to neurones.
DeleteI want Eric to comment on what I said about his definitions of the "perfect image sensor" around slide 4... what do you say Eric?
ReplyDeletehi Tobi, I think the main thing I would like to say is that I think it is wonderful that after all these years working on this technology, it is going niche-mainstream. Congratulations! Regarding off-hand silly comments on a "perfect image sensor," we can add zero power and zero latency to the list. I appreciate just being mentioned in your talk! Take care Tobi. Looking forward to our next meeting. We missed you in Scotland.
Delete50:00,cancel the offset of comparator with preamp. But why not to use the Auto zero method?
ReplyDeletethis should be for reducing kick-off
DeleteCan this technology complement ToF sensors such as SPAD? I think if algorithms allow it for a low lattency but more reliable object identification will be a good application. Does anyone has a reference to such applications today? any products maybe on the market?
ReplyDeleteThere already was a paper where an EVS sensor determined ROI to guide a scanning Lidar (https://ieeexplore.ieee.org/document/9665844). That Lidar, of course, can be SPAD based. Some SPAD sensors are actually read-out event-based. But that's very different. So, yes. Event-sensors can complement ToF sensors and they can inspire ToF sensor designs. But they can't replace ToF sensors (SPADs have ns response, sub-ns jitter, and single-electron sensitivity - in comparison Event-based vision sensors have us-ms latency - but they can be very low-power).
DeleteKey players are iniVation, Prophesee, Samsung, OMNIVISION and Sony. And there's a new player Ruisi Zhixin (AlpsenTek) but they haven't published anything peer-reviewed yet (to my knowledge). You can buy cameras and dev kits from iniVation and Prophesee. Samsung was the first one to target a mass product (security camera). OMNIVISION, Sony and AlpsenTek appear targeted at mobile phones. And Sony, Prophesee and iniVation made announcements/publications directed at AR/VR.