DARPA looks to unconventional processors for ISR data analysis
DARPA is looking to contractors for ways break the CMOS status quo in processing the massive amounts of sensor data collected by intelligence, surveillance and reconnaissance (ISR) platforms, noting the capability is critical to security and defense.
It said on Aug. 14, that not only has the volume of sensor data increased exponentially, there has also been a dramatic increase in the complexity of analysis required for applications such as target identification and tracking. The digital processors used for ISR data analysis are limited by power requirements, potentially limiting the speed and type of data analysis that can be done. DARPA, the Defense Department’s research operation, said a new, ultra-low power processing method may enable faster, mission critical analysis of ISR data.
The agency said its Unconventional Processing of Signals for Intelligent Data Exploitation (UPSIDE) program seeks to break the status quo of digital processing with methods of video and imagery analysis based on the physics of nanoscale devices. It said UPSIDE processing will be non-digital and fundamentally different from current digital processors and the power and speed limitations associated with them.
Instead of traditional complementary metal–oxide–semiconductor (CMOS)-based electronics, DARPA said UPSIDE envisions arrays of physics-based devices (nanoscale oscillators may be one example) performing the “processing. The arrays would self-organize and adapt to inputs and won’t need to be programmed as digital processors are. Unlike traditional digital processors that operate by executing specific instructions to compute, DARPA envisions UPSIDE arrays will relying on a higher level computational element based on probabilistic inference embedded within a digital system.
Probabilistic inference is the fundamental computational model for the UPSIDE program. An inference process uses energy minimization to determine a probability distribution to find the object that is the most likely interpretation of the sensor data. It can be implemented directly in approximate precision by traditional semiconductors as well as by new kinds of emerging devices.
“Redefining the fundamental computation as inference could unlock processing speeds and power efficiency for visual data sets that are not currently possible,” explained Dan Hammerstrom, DARPA program manager. “DARPA hopes that this type of technology will not only yield faster video and image analysis, but also lend itself to being scaled for increasingly smaller platforms.”
The agency said it expects an interdisciplinary approach for interested performer teams that addresses three tasks set forth in its UPSIDE solicitation. Task 1, it said, forms the foundation for the program and involves the development of the computational model and the image processing application that will be used for demonstration and benchmarking. Tasks 2 and 3 will build on the results of Task 1 to demonstrate the inference module implemented in mixed signal CMOS in Task 2 and with non-CMOS emerging nano-scale devices in Task 3. The ability to successfully address all three tasks will require close collaboration within the proposer’s team and will be an important aspect of any successful UPSIDE effort.
“Leveraging the physics of devices to perform computations is not a new idea, but it is one that has never been fully realized,” added Hammerstrom. “However, digital processors can no longer keep up with the requirements of the Defense mission. We are reaching a critical mass in terms of our understanding of the required algorithms, of probabilistic inference and its role in sensor data processing, and the sophistication of new kinds of emerging devices. At DARPA, we believe that the time has come to fund the development of systems based on these ideas and take computational capabilities to the next level.”