Despite the significant progress made in deep learning on digital computers, their energy consumption and computational speed still fall short of meeting the standards for brain-like computing. To address these limitations, reservoir computing (RC) has been gaining increasing attention across communities of electronic devices, computing systems, and machine learning, notably with its in-memory or in-sensor implementation on the hardware–software co-design. Hardware regarded, in-memory or in-sensor computers leverage emerging electronic and optoelectronic devices for data processing right where the data are stored or sensed. This technology dramatically reduces the energy consumption from frequent data transfers between sensing, storage, and computational units. Software regarded, RC enables real-time edge learning thanks to its brain-inspired dynamic system with massive training complexity reduction. From this perspective, we survey recent advancements in in-memory/in-sensor RC, including algorithm designs, material and device development, and downstream applications in classification and regression problems, and discuss challenges and opportunities ahead in this emerging field.
I. INTRODUCTION
In recent years, significant progress has been made in the fields of bio-inspired computing and artificial intelligence (AI) by drawing inspiration from biological systems to create innovative algorithms and computational models. This interdisciplinary field has given rise to revolutionary techniques such as artificial neural networks (ANNs)1 and spiking neural networks (SNNs),2 which excel at tackling complex learning and optimization challenges. These advancements have been applied to various fields, including computer vision,3 natural language processing,4 robotics,5 and healthcare,6 laying the foundation for a new generation of intelligent systems and providing immense potential for achieving artificial general intelligence (AGI).7
Despite recent advancements, AI models operating on traditional digital computers have yet to match the energy efficiency and rapid learning capabilities of the human brain. For instance, adapting or fine-tuning a large language model (LLM) toward a particular task or domain can consume a large carbon footprint,8 while the human brain requires only about 20 W of power to learn something new in just a few hours.9–12 This limitation in energy efficiency primarily stems from the physical separation between sensing, memory, and computing units in digital computers, known as the von Neumann bottleneck.13–20 This issue is exacerbated by the slowing down of Moore’s law as transistor size approaches its physical limit. The slow learning of AI models can be partially attributed to the cumbersome stochastic gradient descent optimization,21 which is particularly challenging to manage on resource-constrained edge computers, especially considering the data explosion in the Internet of Things (IoT) era. Consequently, energy-efficient and fast-learning AI in intelligent edge systems has been sought-after.
A potential solution to this challenge is in-memory and in-sensor reservoir computing (RC).20,22 From a hardware perspective, these systems leverage emerging and scalable electronic and optoelectronic devices, enabling them to store, process, and sense data all in one place. These nanoscale devices23–28 not only address the transistor scaling limit but also substantially decrease the cost of transferring data between sensing, memory, and processing components, thereby resolving the von Neumann bottleneck. On the software side, RC deviates from conventional deep neural networks (DNNs)1 that require large datasets to optimize numerous model parameters using stochastic gradient descent21 and error backpropagation.29 Instead, RC primarily relies on a complex, black-box dynamic system, requiring only a limited number of trainable parameters for optimization.30–32 This considerably reduces the training cost. The synergy between hardware and software components offers a new solution for edge AI hardware that demands both high efficiency and real-time learning.
Therefore, this paper first surveys the new solution from the software perspective. We discuss the three major RC models: echo state networks (ESNs),30 liquid state machines (LSMs),31 and single dynamic nodes with delayed feedback32 and their differences. Afterward, we introduce two representative RC-associated hardware architectures, the in-memory and in-sensor architectures,33–35 and discuss the emerging materials and devices that constitute the circuits of such novel computing systems. Then, from an application perspective, we categorize RC systems into two major categories: classification and regression problems. Within each category, various applications using different input modalities, such as images, sounds, and graphs, are summarized. Finally, we discuss novel perspectives on the future development of RC systems in three aspects: materials, architecture, and applications. For instance, multimodality and the neuron architecture search (NAS) reservoir are discussed.
II. RESERVOIR COMPUTING MODELS
RC is a framework originally proposed in the early 2000s for training recurrent neural networks (RNNs). Currently, it is a vibrant and emerging AI research domain that has recently gained wide attention due to its extremely low training complexity. RC is an overarching concept that encompasses different RNN models, such as ESN,30 LSM,31 and the single node delayed feedback dynamical model.32
A. Echo state network
ESNs were first introduced by Jaeger in 200130 and have gained popularity in various applications, such as time series prediction,30 speech recognition,36 and control tasks.37
B. Liquid state machine
The SNN variant of the RC, known as the LSM, was introduced by Maass et al. in 2002.31 The term “liquid” refers to the idea that the network’s activity resembles the constantly changing patterns in a liquid, where input signals can create various activity patterns. LSMs are particularly useful in dealing with event-type data, such as those acquired by dynamic vision and audio sensors.38–40
C. Single node delayed feedback dynamical model
III. IN-MEMORY AND IN-SENSOR COMPUTING ARCHITECTURES
Traditional digital computers are primarily based on von Neumann and Harvard architectures, which comprise input/output (e.g., sensors), memory, processing, and control units.42 In this configuration, data are first collected from sensors and then undergo digital–analog conversion before being stored in memory. Following this, the control unit manages the data transfer from memory to the processing units.43 Despite the von Neumann architecture remaining the dominant model for most general-purpose computers at present, it has raised significant concerns regarding the “von Neumann bottleneck,” or the energy and latency costs associated with data movement between physically separated sensors, memory, and processing units.35,44–46 To address these challenges, in-memory computing and in-sensor computing architectures have been proposed. RC systems with emerging memory naturally fall into these categories.
A. In-memory reservoir computing
In-memory computing architecture embeds computational capabilities directly within the memory, thus eliminating the frequent data shuttling between memory and processing units. Figures 2(a) and 2(b) show the block diagram of a representative in-memory RC system. The data sampled by sensors are first passed to the in-memory computing module implementing the reservoir, followed by the readout map module.
The in-memory reservoir module consists of emerging memory, peripherals, and input/output interfaces. The reservoir model and the associated circuit depend on the choice of emerging memory. Volatile electronic memory (e.g., discrete volatile resistive switches) functions as a single nonlinear node with delayed feedback, given the inherent dynamics, which can be described by a delay differential equation.47–50 Non-volatile memory (e.g., crossbar arrays of non-volatile resistive switches) can physically implement ESNs and LSMs, primarily capitalizing on the intrinsic programming stochasticity to produce fixed random weights of ESNs51,52 and LSMs,38 which significantly reduces training overhead.
The readout map module can either be implemented in digital (i.e., CMOS51) or analog (i.e., emerging non-volatile memory24). The former digitizes reservoir outputs, featuring precise and fast readout map weight updates. The latter benefits the readout map in terms of inference energy efficiency via in-memory computing. The trade-off depends on the readout weight population and the emerging memory programming energy.24,26,53,54
B. In-sensor reservoir computing
In-sensor computing architecture physically integrates data acquisition, memory, and processing, notably all in the analog domain, thus eliminating the need for analog-to-digital conversion and mitigating the hardware challenges in heterogeneous integration of sensors, memory, and logic circuits. Figure 2(c) illustrates the block diagram of a representative in-sensor RC system, consisting of an in-sensor reservoir module followed by the readout map module.
The in-sensor reservoir module mostly incorporates discrete nonlinear nodes with delayed feedback. Recent implementations of such nodes leverage emerging sensors, which are paired with peripheral circuits (analog–digital conversion), input/output interfaces, and buffers. These emerging sensors can work with various input modalities, including optical images,55–57 electrical signals such as electroencephalogram (EEG) and electrocardiogram (ECG) signals,27,58,59 mechanical signals like tactile signals,60 and chemical signals such as odor.61 Moreover, recent efforts have been made to develop multimodal in-sensor reservoirs to fuse the information carried by individual modalities.60 The underlying ionic or electronic dynamics of the sensors are essentially governed by the delay differential equation, thereby implementing sensing, memory, and processing within the same device.
The in-sensor reservoir outputs are typically digitized before being sent to the digital readout map for post-processing in a manner similar to that of in-memory reservoirs.
C. Summary
Finally, to illustrate the advantages of the in-memory/in-sensor RC paradigm, we have summarized the energy consumption overhead of inference per input for RC systems compared to traditional digital hardware and the training cost ratio compared to trainable DNN models in Table I. Overall, both in-memory and in-sensor RC systems can save several times to hundreds of times the energy consumption overhead for ESN, LSM, or Single Node Delayed Feedback. Moreover, thanks to the untrained random weights in RC systems, the training cost can be reduced by up to hundreds of times compared to trainable DNNs.
RC types . | Computing paradigm . | Energy efficiency . | Training cost (compare to trainable DNNs) . |
---|---|---|---|
ESN51 | In-memory | About 2–40 times | Reduce more |
reservoir computing | (compare to digital hardware) | than 90% | |
LSM38 | In-memory | About 23–150 times | More than |
reservoir computing | (compare to digital hardware) | 100 reduction | |
Single node | In-memory/in-sensor | 3–6 nJ per input | Achieve about |
delayed feedback47,48,57 | reservoir computing | 5–15 reduction |
RC types . | Computing paradigm . | Energy efficiency . | Training cost (compare to trainable DNNs) . |
---|---|---|---|
ESN51 | In-memory | About 2–40 times | Reduce more |
reservoir computing | (compare to digital hardware) | than 90% | |
LSM38 | In-memory | About 23–150 times | More than |
reservoir computing | (compare to digital hardware) | 100 reduction | |
Single node | In-memory/in-sensor | 3–6 nJ per input | Achieve about |
delayed feedback47,48,57 | reservoir computing | 5–15 reduction |
IV. DEVICE AND MATERIALS FOR IN-MEMORY AND IN-SENSOR RESERVOIR
The in-memory and in-sensor reservoir modules in Fig.2 comprise analog circuits using emerging devices, which are also widely used for implementing synapses and neurons in neuromorphic computing.62–65 Such devices and materials are not only energy-efficient due to their inherent in-sensor and in-memory computing abilities, but they also feature potentially higher integration density and smaller RC system footprints compared to digital alternatives. Moreover, these emerging devices streamline the in-memory and in-sensor chip fabrications by eliminating the heterogeneous integration of sensing, memory, and processing units. The RC module is characterized by nonlinear dynamics and fading memory. Therefore, dynamic nonlinearity and fading memory have become the most important prerequisites for emerging memory device-based analog circuits for hardware reservoirs.24,66,67 In this section, we will review the devices and materials that meet these requirements and are used in both in-memory and in-sensor reservoirs, focusing on their mechanisms and strengths.
A. Materials/devices for in-memory reservoirs
Two types of emerging memory devices are used for in-memory RC. For ESNs and LSMs, their weight matrices are random and fixed. In such cases, non-volatile memory (e.g., redox resistive switches, ferroelectric tunneling junctions) crossbar arrays are utilized to encode the weight matrices and perform analog matrix-vector multiplication using Ohm’s law and Kirchhoff’s law. This in-memory computing scheme co-locates both memory and processing, featuring high energy efficiency.68–70 Furthermore, the inherent programming variation of these devices is no longer a disadvantage in neural network training but instead offers a unique opportunity to provide truly random and high-density weight matrices at a low cost compared to digital random number generation.71–73 Regarding the single nonlinear node with time-delayed feedback, emerging memory devices exhibiting short-term memory can physically implement the delay differential equation and thus represent the dynamics of the reservoir. Such memory can be ion-driven, including redox oxide and peroxide, or result from electronic effects such as charge trapping or polarization, as exhibited by ferroelectric devices. As shown in Fig. 3, four different types of in-memory devices are introduced.
1. Oxide redox resistive switches
Redox resistive switches [Figs. 3(a) and 3(c)] are capacitor-like. They operate on the formation or dissolution of conductive filaments in the dielectric layer due to redox reactions and ion migration. Depending on whether the filaments spontaneously rupture, redox oxide resistive switches are categorized into volatile and non-volatile types, which are employed by single nodes with delayed feedback and ESNs/LSMs, respectively.
Volatile redox oxide devices, which are widely reported for their short-term memory, are used to implement delay differential equations of single nonlinear nodes with time-delayed feedback. Such short-term memory can be attributed to cations,24,34,74,75 such as Ag and Cu ions. When Ag and Cu serve as electrodes of redox resistive switches, they are oxidized to Ag+ and Cu+ and migrate under an external electric field. They are subsequently reduced to atoms, gradually forming conductive paths. Upon voltage removal, the metallic channels coalesce into particles, causing the conductance to spontaneously decay. This behavior of input history-dependent filament growth and decay meets the definition of a general memristor and the delay differential equation.
Short-term memory can also be attributed to anions, as reported in TiOx,47,53,76 WOx,48,49 and AlOx.77 Taking a Ti/TiOx/TaOy/Pt resistive switch as an example,47 oxygen ions or vacancies migrate in the TaOx layer which are primarily driven by the applied electric field. The TiOx layer serves as a reservoir for oxygen ions, creating a chemical potential difference that induces oxygen ion diffusion. A positive voltage on the top electrode causes oxygen ions to drift away from the Pt/TaOx interface due to the electric field, reducing the Schottky-like barrier and increasing conductance. When the external bias is removed, oxygen ions diffuse back because of the chemical gradient, and the device gradually relaxes to its high-resistance state. Like the cation, such conductance evolution meets the general definition of a memristor and the delay differential equation.
Non-volatile redox resistive switches can physically implement the random and fixed weights of ESNs and LSMs.23 Anion-based redox oxides like TaOx and HfOx are widely reported for their stable and analog conductance due to the formation of oxygen vacancy conductive filaments. The inherent programming stochasticity of non-volatile redox resistive switches, such as those based on TaOx, implements an echo state graph neural network.51 Specifically, a fixed voltage is applied to an array of TaOx resistive switches. The resultant conductance of the array exhibits a normal distribution due to the device-to-device variation in breakdown voltage, which implements the fixed random weights of the echo state graph layer and offers a significant energy efficiency boost due to in-memory computing.
2. Perovskite resistive switch
Perovskite [Figs. 3(b) and 3(d)] with its ABX3 structure features rich ionic dynamics,26,78–81 which are divided into cation-dominated or cation–anion co-driven. These dynamics can be tailored for RC, such as the single nonlinear node with delayed feedback. In terms of cation-driven ionic dynamics, perovskite resistive switches with Ag electrodes exhibit tunable volatility, enabling single nonlinear nodes with delayed feedback. The reaction between Ag+ and the X ions in perovskites is the key to tunable data retention.78 Ag+ and X ions migrate under an external electric field, thus changing the perovskite conductivity. Applying weak stimulation results in volatile resistive switching due to the formation of an unstable AgX layer caused by weak drift behavior. This meets the general definition of a memristor and the delay differential equation. As the external stimulation increases, the AgX layer strengthens, and thick filaments form via ion drift in addition to ion diffusion. As a result, volatile and non-volatile perovskite resistive switches can simultaneously implement single nonlinear nodes with delayed feedback and RC readout maps, respectively. In terms of cation–anion co-driven dynamics, perovskite resistive switches with inert electrodes have exhibited volatile resistive switching for the single nonlinear node with delayed feedback. This stems from the interface effect under voltage bias.82 When external voltages are applied, positive and negative ions with extremely low activation energy within the perovskite accumulate at the device’s upper and lower interfaces. This changes the interfacial barriers, enhances carrier injection, and alters device conductance. Upon voltage removal, the interfacial barriers gradually recover, yielding volatile resistive switching that meets the general definition of a memristor and the delay differential equation.
3. Ferroelectric tunneling junctions
Ferroelectric tunneling junctions (FTJs) [Fig. 3(e)] comprise two metal electrodes separated by a thin ferroelectric layer. When a voltage is applied across the electrodes, the ferroelectric domains switch polarization directions, which modulate the tunneling barrier, causing a change in the tunneling current through the junction. As the associated lattice distortion is very limited, FTJs feature fast switching at ultra-low switching energy.25,83,84 The tunable switching energy also yields different data retention for different types of RC. As for non-volatile FTJs, multi-domain FTJs feature multi-bit data storage per cell as well as intrinsic device variations. Crossbar arrays of FTJs with controllable device-to-device variation have also been proposed to implement fixed random weights for ESNs and LSMs.85 As for volatile FTJs, their short-term memory is due to ferroelectric degradation.50 This is because the energy barrier between states scales with the ferroelectric domain volume, so degradation is frequently spotted in FTJs of ferroelectric layers less than 7 nm.86 Furthermore, by applying special electrical operation schemes that impose a small pulse sequence with opposite polarity after a polarization pulse, the conductance decay time constants could be modulated for a reconfigurable single nonlinear node with delayed feedback.50
4. Nanowire
The nanowire network [Fig. 3(f)] resembles the random and complex neural network of the brain. It is a collection of nonlinear dynamic nodes (i.e., junctions between two nanowires) interconnected in a random topology like that of ESN and LSM.28,87–90 Like neuron cells interacting through synaptic intersections, nanowires interact with each other via intersections of different core/shell materials. For example, volatile redox resistive switches are formed at the junction between Ag nanowires coated with polyvinylpyrrolidone (PVP).54 When a voltage is applied across intersecting nanowires, it triggers the anodic dissolution of Ag into Ag+ ions. These ions then travel through the PVP-insulated nanowire shell layer, forming a conductive bridge. This bridge adjusts the conductivity of the junction, creating a node of nonlinear dynamics. Such a network of coupled nonlinear dynamic nodes with a random coupling topology produces very complicated nonlinear dynamics and short-term memory of the reservoir.
B. In-sensor reservoir implementation
In-sensor RC seeks to integrate data acquisition, memory, and computation within a single unit, maximizing the energy and area efficiency of reservoirs. In-sensor reservoirs are mostly single nonlinear nodes with delayed feedback. They employ novel optoelectronic devices, electronic memory, and tactile sensors for vision,55–57 bio-electrical signals (e.g., EEG and ECG),27,58,59 and tactile signals,60 respectively. The dynamic responses of such sensors feature a short-term memory, matching the requirement of delay differential equations and thus implementing the nonlinear nodes with delayed feedback. Figure 4 shows different types of in-sensor devices.
1. Organic electrochemical transistor
The organic electrochemical transistors function by moving ions from the gate electrolyte into the channel and vice versa, enabling doping of the channel and, consequently, adjusting the conductance. It responds to both light and voltage stimulation while featuring short-term memory.57,91–94 When it comes to light stimulation, charges are generated throughout the channel during the photoexcitation process. The majority of mobile carriers are electrons, which contribute to the rise of photocurrent, whereas holes are localized immediately after being generated. Upon switching off the light, the channel current first experiences a rapid decrease due to the recombination of electrons with holes in shallow traps. This is followed by a gradual decay due to inefficient recombination of electrons with holes in deeper traps. This leads to short-term memory and implements a single nonlinear node with delayed feedback. In terms of voltage stimulation, organic electrolyte transistors respond to EEG and ECG signals while displaying short-term memory. This short-term memory is based on ion injection into and diffusion back from the channel. Furthermore, the characteristic time of the short-term memory, which depends on the ion trapping energy, can be tunable in a vertical traverse structure,27 as shown in Fig. 4(a), where a small (large) gate voltage can inject ions into the amorphous (crystalline) region, leading to a volatile (non-volatile) behavior.
2. 2D materials
2D materials are crystalline solids composed of a single or few layers of atoms that often exhibit unique optoelectronic properties. Various 2D material-based field-effect transistors have been reported for in-sensor computing, such as MoS2,95–98 h-BN,99 In2Se3,60,100 WSe2,101,102 SnS,56 and layered black phosphorus.103,104 As illustrated in Fig. 4(b), using SnS as an example, the channel conductance of an SnS device exhibits nonlinear short-term memory. This is attributed to charge trapping and detrapping in defect states associated with Sn and S vacancies. The device displays synaptic depression and facilitation under electrical and optical stimulation, respectively. This implements a single nonlinear node with delayed feedback.
3. Perovskite
In addition to in-memory computing, perovskites are also widely used in solar cells. Their photoresponse makes perovskite optoelectronic cells contenders for in-sensor RC.105–107 For instance, as shown in Fig. 4(c), a self-powered Au/P(VDF − TrFE)/Cs2AgBiBr6/ITO worked as a nonlinear node with delayed feedback for in-sensor RC.108 Under optical stimulation, the separation of photon-generated carriers by the built-in electric field in the Schottky barrier results in an increase in photocurrent. After the stimulus is removed, the high binding energy of the valence electrons in the designed P(VDF-TrFE) layer forms a potential well for hole carriers at the P(VDF − TrFE)/Cs2AgBiBr6 interface. This potential well hinders the migration of photon-generated hole carriers toward the Au electrode, leading to a gradual current decay that meets the definition of delay differential equations.
4. Oxide
Metal oxide photodetectors cover a wide spectrum of electromagnetic radiation due to the numerous possibilities of bandgaps55,109 and hence in-sensor reservoirs. For instance, a-GaOx photo-synapses [Fig. 4(d)] have been reported to implement in-sensor reservoirs due to the persistent photoconductivity effect, which involves the generation, trapping, and detrapping of non-equilibrium carriers within the a-GaOx. Under ultraviolet light, the number of photogenerated electron–hole pairs increases in the a-GaOx layer, while the holes drift toward the electrode. When the light is turned off, electrons and holes gradually recombine, thus exhibiting short-term memory for the nonlinear node with delayed feedback.
V. APPLICATIONS
RC has been widely used in statistical classification and regression, as illustrated in Table II. For classification, RC has been employed across different data modalities, such as images, audio, events, general graphs, spatiotemporal sequences, and multimodal fusion. In the case of regression problems, RC is mainly utilized for time series datasets to forecast future trends in time series signals.
Task . | Data type . | Dataset . |
---|---|---|
Regression | Time series | Hénon map,47 nonlinear autoregressive |
moving average (NARMA)110,111 | ||
Mackey–Glass time series,48,110,111 Santa Fe laser time series110 | ||
Classification | Image | Handwritten digit,24,49,50,54,57,60,111 self-made pattern images,54,56 |
self-made noisy images,49 FVC 2002 database55 | ||
Sound | NIST TI46 database47,48 | |
Event-data | N-MNIST,38 N-TIDIGITS,38 | |
DVS Gesture128,57 neural firing patterns26 | ||
Graph | MUTAG, COLLAB and CORA51 | |
Sequence | MIT-BIH heart arrhythmia database,53,91 | |
PTB-XL,27 self-made gesture data25,53 | ||
Multimodal | Tactile and visual digit,60 audio and | |
visual digit,38 audio and electrophysiological signal112 |
Task . | Data type . | Dataset . |
---|---|---|
Regression | Time series | Hénon map,47 nonlinear autoregressive |
moving average (NARMA)110,111 | ||
Mackey–Glass time series,48,110,111 Santa Fe laser time series110 | ||
Classification | Image | Handwritten digit,24,49,50,54,57,60,111 self-made pattern images,54,56 |
self-made noisy images,49 FVC 2002 database55 | ||
Sound | NIST TI46 database47,48 | |
Event-data | N-MNIST,38 N-TIDIGITS,38 | |
DVS Gesture128,57 neural firing patterns26 | ||
Graph | MUTAG, COLLAB and CORA51 | |
Sequence | MIT-BIH heart arrhythmia database,53,91 | |
PTB-XL,27 self-made gesture data25,53 | ||
Multimodal | Tactile and visual digit,60 audio and | |
visual digit,38 audio and electrophysiological signal112 |
A. Classification
RC has been extensively explored for classification problems using both in-memory and in-sensor implementations. These approaches have been categorized into six major classes based on the modality of input signals, as shown in Fig. 5.
Images are originally spatial signals. Converting spatial images into spatiotemporal signals makes them compatible with RC. In terms of in-memory computing, oxide redox resistive switches and FTJs implementing nonlinear nodes with delayed feedback are benchmarked on datasets such as MNIST.24,49,50,111 For instance, a crossbar array of redox resistive switches serves as the nonlinear nodes of a delayed feedback RC, while a digital computer is used for supervised training of a readout layer to classify the MNIST dataset.49 In addition, redox resistive switches, rather than digital hardware, have been used to implement the readout map, achieving an accuracy of 83% on the same dataset.24 As for in-sensor computing, the current approach mainly involves using optoelectronic devices to achieve delayed feedback RC to process images, such as handwritten digit images from the MNIST dataset,57 optical garment images sampled from the Fashion MNIST dataset,57 and letter images from the E-MNIST dataset57 or the Korean letter images.56 Moreover, to enhance the performance of image recognition, novel reservoir architectures have also been proposed. For instance, the rotation-based architecture employs an ensemble of nonlinear nodes and dynamically rotates the links between the input channels and the nonlinear nodes.111
Audio signals are temporal in nature, and RC can be utilized to extract features from them. Large parallel RC systems have been built using in-memory computing by connecting multiple single nonlinear nodes with delayed feedback in parallel. These systems have achieved a low word error rate of 0.4% on spoken-digit datasets.47 Similarly, dynamic tungsten oxide (WOx) memristor-based reservoirs have exhibited comparable performance.48
Event data, which are sparse representations generated by dynamic vision or audio sensors, mimic the signals received by human eyes or ears. This data type is intrinsically sparse, offering significant energy savings while preserving privacy and security. Due to their inherent temporal dimension, event data are well-suited for RC. In-memory RC implementations, such as LSMs, have been used for zero-shot learning feature alignments between the N-MNIST and N-TIDIGITS datasets.38 In-sensor computing approaches have employed organic nonlinear dynamic nodes to construct reservoir-in-pixel architectures, which can classify DVS128 datasets.57 Additionally, reconfigurable halide perovskite-based nonlinear dynamic nodes have achieved ∼90% classification accuracy for four event-type neural firing patterns.26
General graphs, comprising nodes and edges, are natural representations of molecules and social networks. Messages passing on graphs can be treated temporally, making them compatible with RC. In in-memory reservoirs, each graph node can be associated with an ESN. These ESNs share the same weights and are coupled according to the graph topology. ESN-based graph embeddings have demonstrated significant energy efficiency improvements when classifying MUTAG molecule datasets and COLLAB citation network datasets.51
Spatiotemporal signals generated by IoT devices, such as wearable sensors, can be processed by RC due to their temporal dimension. In-memory computing approaches have utilized single node delayed feedback RC for recognizing health categories in the MIT-BIH arrhythmia database and self-collected gesture data.53 High-density 3D stacked redox resistive switches have been employed for dynamic gesture data classification.25 In-sensor computing methods have implemented organic electrochemical transistors as nonlinear nodes with delayed feedback for diagnosing cardiac diseases.27
Multimodal data fusion has been explored in RC systems due to its rich representation and strong generalization capabilities. The recognition and perception of multimodal data are more biologically plausible, as they resemble the information acquisition process in the human brain. Recent research has focused on equipping RC with multimodal inputs, such as tactile, visual, and auditory signals. For instance, tactile and visual signal combinations have been employed in RC systems for digit recognition, resulting in significant improvements in forward inference performance.60 RC systems have also been applied to touchless user interfaces for virtual reality, leveraging their acoustic and electrophysiological perception capabilities to provide a more immersive experience resistant to interference.112 Furthermore, LSMs implemented with redox resistive switches have demonstrated multimodal zero-shot transfer learning using event visual and audio datasets.38
B. Regression
Regression aims to establish the relationship between independent and dependent variables. A representative problem in regression is fitting or predicting chaotic systems, as shown in Fig. 6. This challenge is primarily due to the positive Lyapunov exponent characterizing chaotic systems, which leads to exponential growth in the separation between closely related trajectories. As a result, even minor prediction errors can quickly cause significant divergence from the ground truth.48,113 In-memory reservoirs based on nonlinear nodes with delayed feedback have been used for regression tasks such as the Hénon map,47 NARMA,110,111 Mackey–Glass time series, and Santa Fe laser time series.48,110,111 These reservoirs capitalize on the short-term memory and nonlinear properties of redox resistive switches.47,48,110 Additionally, an RC system entirely built on redox resistive switches, possessing a novel rotating reservoir architecture, is used for Mackey–Glass time series prediction. In this way, the readout layer is implemented on non-volatile redox resistive switches.111
VI. PERSPECTIVE
At present, although in-memory and in-sensor RC with emerging memory have shown promising results in certain tasks, they still cannot parallel traditional DNNs on digital hardware for the rest of the tasks. This paper provides an outlook on potential directions for the future RC systems in three aspects: hardware, architecture, and application, as illustrated in Fig. 7.
A. Hardware
Hardware-wise, we summarize challenges with both single devices and systems, as shown in the bottom row of Fig. 7.
At the device level, no current emerging devices simultaneously provide large-scale integration, CMOS compatibility, and multimodal sensing capabilities. In-memory devices, such as redox resistive switches and FTJs, feature high integration density and CMOS compatibility. However, they frequently lack multimodal sensing capabilities. For example, monolithic integration of 4M redox resistive switches and CMOS has been reported.114 The CMOS compatibility also endorses the yield of the redox resistive switches. However, for 2D material-based optoelectronic devices, as a representative in-sensor RC solution, large-scale integration is difficult due to challenges with wafer-scale transfer or CVD growth for heterostructures. As such, the 2D material-based crossbar array still shows lower integration density and a relatively poor yield.115,116 In-sensor devices favoring functional materials, like perovskite, 2D materials, and organic materials, offer a wealth of multimodal sensing opportunities. Nevertheless, they face substantial challenges in large-scale array fabrication and homogeneous integration.
Along with Table III, we examine and compare several electrical properties of different material-device combinations for RC, namely the variations, switching speed, and endurance. Device variations, including cycle-to-cycle variation and device-to-device variation, negatively impact the performance of RC models. Compared with redox resistive switches, homogeneous ion motion in OECTs and perovskite resistive switches may outperform filamentary switching in achieving reduced cycle-to-cycle variations.27,38,81 The device-to-device variation is similar among different material-device pairs, and it could likely be improved using advanced CMOS-compatible manufacturing.27,51,81,85,95 Regarding switching, it is essential to achieve application-specific switching incubation and relaxation time for nonlinear nodes with delayed feedback RC. Redox resistive switches and ferroelectric devices provide options for fast operation,53 while perovskite materials and OECTs are more suitable for relatively large time-scale operations.26,27,82,99 The switching speed also scales with the energy efficiency for nonlinear nodes with delayed feedback RC, as it underpins the duration to process given time sequences. Finally, in terms of endurance, greater endurance benefits all types of RC applications. Due to the continuous dynamic memory behavior, the requirement for switching endurance of single-node delayed feedback RC would be much larger than that of ESN and LSM. So far, redox resistive switches and perovskite materials have demonstrated endurance greater than 106,26,51 owing to the relatively stable host materials.
. | Max. cycle-to-cycle repeatability (%) . | Min. device-to-device variance (%) . | Min. switching incubation/relaxation time . | Max. endurance . |
---|---|---|---|---|
Redox | <1638 | <1551 | 1 ns53/⋯ | 3 × 106 cycles51 |
Ferroelectric | ⋯ | 1985 | ⋯ | ⋯ |
Perovskite | 2.581 | 14.481 | 2 ms82/>5 ms26 | 2 × 106 cycles26 |
2D material | ⋯ | 12.895 | 1 ms99/⋯ | 2 × 103 cycles98 |
OECT | 0.4927 | 1727 | 0.82 ms27/1.399 ns57 | 8 × 103 cycles27 |
Nanowire | ⋯ | ⋯ | 10 μs54/⋯ | 3 × 103 cycles90 |
. | Max. cycle-to-cycle repeatability (%) . | Min. device-to-device variance (%) . | Min. switching incubation/relaxation time . | Max. endurance . |
---|---|---|---|---|
Redox | <1638 | <1551 | 1 ns53/⋯ | 3 × 106 cycles51 |
Ferroelectric | ⋯ | 1985 | ⋯ | ⋯ |
Perovskite | 2.581 | 14.481 | 2 ms82/>5 ms26 | 2 × 106 cycles26 |
2D material | ⋯ | 12.895 | 1 ms99/⋯ | 2 × 103 cycles98 |
OECT | 0.4927 | 1727 | 0.82 ms27/1.399 ns57 | 8 × 103 cycles27 |
Nanowire | ⋯ | ⋯ | 10 μs54/⋯ | 3 × 103 cycles90 |
At the system level, efficiency and footprint are closely related to the partition between analog and digital components. There are two popular ways to partition the RC system. The first partition features a digital readout layer and an analog reservoir consisting of emerging in-memory or in-sensor devices.53 Hybrid analog–digital systems demonstrate balanced efficiency and programmability, where the efficiency (programmability) comes from analog in-memory or in-sensor computing (digital readout map). However, the overall performance is limited by analog–digital conversion. The second partition is an entirely analog system, a fully analog RC system to capitalize on the efficiency advantages of in-memory and in-sensor computing.117 Specifically, current research on reconfigurable volatile and non-volatile devices suggests that a homogeneously integrated physical RC system could deliver significant performance enhancement. However, current analog memory devices are less precise and slower when serving as the weights of the readout map, which requires a time-consuming tuning process.
Furthermore, such partitioning can be tailored toward particular applications or hardware–software co-design, a general method for improving overall system performance.51 On the hardware side, it is expected to design more efficient circuit mapping for the algorithm. On the software side, the algorithm’s tolerance for limited hardware nonidealities should be improved. The design goal is to concurrently maximize efficiency, footprint, and dataset performance by adjusting both hardware and software design parameters. Such design parameters include but are not limited to, signal design, the number of hardware neurons, analog-to-digital conversion bitwidth, RC model structure, and readout weight optimization. An example is the co-design for emerging memory-based readout map training. As emerging memory suffers from programming energy and stochasticity, lightweight first-order and controlled error (FORCE) learning could be employed,118 which minimizes emerging memory programming compared to alternative training protocols.
B. Architecture
At present, there are three key aspects worth exploring at the architecture level for both in-memory and in-sensor RC.
First, there is a trend to shift from single modality to multiple modalities. Most current work emphasizes individual modalities, such as visual or auditory. However, achieving human-like intelligence requires integrating multiple modalities, including vision, taste, smell, pain, touch, and hearing. This would result in more information and context for a given task or problem, improving RC system performance.
Second, the in-sensor architecture is gaining popularity on the edge. However, current in-sensor RC primarily employs nonlinear nodes with delayed feedback, where nodes are somewhat independent, limiting functionality. Moreover, the depth of in-sensor reservoirs is shallow since they convert signals from one type to another. This limitation can be addressed by cascading in-sensor and in-memory reservoirs to form a deep RC system. Additionally, dynamic reservoir structures, such as rotating RC systems, may benefit complex tasks.111
Third, the architectural design overlaps with the hardware–software co-design. As such, it can adopt hardware-aware neural architecture search (NAS).119,120 Using a reward function to benchmark both hardware and software performance, a search agent for hardware-aware NAS explores the design space to identify the best design parameters that maximize the reward. This scheme helps to explore novel software/hardware architectural designs, such as feedback-based RC systems.121
C. Application
Thanks to their low training complexity and high efficiency, in-memory and in-sensor RC systems using emerging memory have already found wide applications in classification and regression. RC can be used for new applications such as generation (e.g., chatbots), cybersecurity (privacy and model protection), as well as system control (robotics).
Generative models, comprising the most recent LLMs, drive chatbots capable of simulating human conversations through text or voice.122 These chatbots deliver information and services to users in a variety of fields, such as customer service, entertainment, education, and social interaction. RNNs, including LSTMs, have been widely utilized for machine translation123 and text generation.124 As a result, generative models present a promising application opportunity for RC.
From a security perspective, RC may have implications for data privacy and machine learning model protection. Data privacy is garnering increasing attention as cloud computing (e.g., LLMs) exposes users’ personal privacy information to risk. Potential solutions could include homomorphic encryption125,126 or differential privacy techniques,127 which are computationally expensive for edge devices. The RC system may serve as an alternative solution to encrypt or encode edge data, leveraging the same capability that has been used for PUF.
Another important field of cybersecurity is model protection. This is because the development of machine learning models requires significant investment in data and training, thus making model protection indispensable.128–133 In-memory and in-sensor RC might offer novel solutions. For instance, in-memory ESN/LSM exploits the intrinsic stochasticity of emerging memory during the manufacturing process, which is difficult for replication. Therefore, it is suggested that it might likely protect intellectual property.
In recent years, the field of robotics has witnessed a surge in the application and study of RC systems, primarily due to their low training cost and impressive performance in tackling complex and practical tasks. For instance, by leveraging multimodal input LSM and digital neuromorphic chips, intelligent robots can be designed to excel in place recognition while consuming less power and exhibiting reduced latency compared to traditional mobile robot processors, such as the Jetson Xavier NX.134 Furthermore, LSM has been employed for flight navigation control in drones, showcasing superior generalization capabilities in contrast to conventional algorithms and delivering excellent performance in novel environments.135 Despite these advancements, the current robotic applications of these algorithms predominantly depend on traditional digital hardware. Investigating the potential of memristive devices to implement compute-storage integrated RC systems represents a crucial avenue for future research.
ACKNOWLEDGMENTS
This research is supported by the Hong Kong Research Grant Council (Grant Nos. 27206321, 17205922, 17212923)and is also partially supported by ACCESS – AI Chip Center for Emerging Smart Systems, sponsored by the Innovation and Technology Fund (ITF), Hong Kong SAR.
AUTHOR DECLARATIONS
Conflict of Interest
The authors have no conflicts to disclose.
Author Contributions
N.L., J.C., and R.Z. contributed equally to this work.
Ning Lin: Conceptualization (equal); Visualization (equal); Writing – original draft (equal). Jia Chen: Conceptualization (equal); Visualization (equal); Writing – original draft (equal). Ruoyu Zhao: Writing – review & editing (equal). Yangu He: Supervision (equal); Writing – review & editing (equal). Kwunhang Wong: Conceptualization (equal); Supervision (equal); Writing – review & editing (equal). Qinru Qiu: Conceptualization (equal); Supervision (equal); Writing – review & editing (equal). Zhongrui Wang: Conceptualization (equal); Supervision (equal); Writing – review & editing (equal). J. Joshua Yang: Conceptualization (equal); Supervision (equal); Writing – review & editing (equal).
DATA AVAILABILITY
Data sharing is not applicable to this article as no new data were created or analyzed in this study.