An Area - and Energy-Efficient Spiking Neural Network With Spike-Time-Dependent Plasticity Realized With SRAM Processing-in-Memory Macro and On-Chip Unsupervised Learning
An Area - and Energy-Efficient Spiking Neural Network With Spike-Time-Dependent Plasticity Realized With SRAM Processing-in-Memory Macro and On-Chip Unsupervised Learning
1, FEBRUARY 2023
Abstract—In this article, we present a spiking neural net- level of bionics. Compared with other neural networks, e.g.,
work (SNN) based on both SRAM processing-in-memory (PIM) convolutional neural networks (CNNs), SNNs work more likely
macro and on-chip unsupervised learning with Spike-Time- to the human brain. By replacing continuous values with discrete
Dependent Plasticity (STDP). Co-design of algorithm and hard-
ware for hardware-friendly SNN and efficient STDP-based learn- spiking events over time, the hardware and energy consumed in
ing methodology is used to improve area and energy efficiency. processing algorithms are greatly reduced. However, efficiently
The proposed macro utilizes charge sharing of capacitors to implementing SNNs is still one of the major challenges cur-
perform fully parallel Reconfigurable Multi-bit PIM Multiply- rently. Realization of SNNs includes two main processes: 1)
Accumulate (RMPMA) operations. A thermometer-coded Pro-
efficient SNN training for discretized and non-derivable data;
grammable High-precision PIM Threshold Generator (PHPTG)
is designed to achieve low differential non-linearity (DNL) and and 2) design of energy-efficient and high-throughput SNN
high linearity. In the macro, each column of PIM cells and a architecture adapted to artificial intelligence (AI) applications,
comparator act as a neuron to accumulate membrane potential such as edge computing, wearable smart devices, Internet of
and fire spikes. A simplified Winner Takes All (WTA) mecha- Things (IoT), etc. [1], [2], [3].
nism is used in the proposed hardware-friendly architecture. By
The training of SNNs can be divided into two categories:
combining the hardware-friendly STDP algorithm as well as the
parallel Word Lines (WLs) and Processing Bit Lines (PBLs), we supervised learning and unsupervised learning. One of the su-
realize unsupervised learning and recognize the Modified National pervised learning methods is to train SNNs using improved
Institute of Standards and Technology (MNIST) dataset. The chip back propagation (BP algorithms) [4], [5], [6]. These methods
for the hardware implementation was fabricated with a 55 nm rely on high-precision derivation and multiplication/division
CMOS process. The measurement shows that the chip achieves a
operations, which consume extra computing resources and
learning efficiency of 0.47 nJ/pixel, with a learning energy efficiency
of 70.38 TOPS/W. This work paves a pathway for the on-chip energy. For example, BP algorithms usually utilize 32/64-bit
learning algorithm in PIM with lower power consumption and floating-point digits for computation [7]. Besides, although
fewer hardware resources. supervised learning can realize higher accuracy, it requires to
Index Terms—MNIST, on-chip unsupervised learning, process- label training data, which can take a lot of effort and time.
ing-in-memory (PIM), spiking neural network (SNN), SRAM, On the other hand, compared with supervised learning, which
spike-time-dependent plasticity (STDP). relies on labelled input and output training data, unsupervised
learning processes unlabelled data, which is becoming more
I. INTRODUCTION attractive due to its adaptability to various environments and
applications [8].
MONG many artificial neural networks (ANNs), spiking
A neural networks (SNNs) have achieved an unprecedented
As an extension to Hebb’s learning rules [9], spike-time-
dependent plasticity (STDP) is considered to be one of the
outstanding candidates to provide an energy-efficient and low-
Manuscript received 27 October 2022; revised 9 December 2022 and 2 January
2023; accepted 23 January 2023. Date of publication 6 February 2023; date of cost solution for unsupervised on-chip learning. Recently, some
current version 23 March 2023. This work was supported by the NSFC under works based on STDP unsupervised training were reported [10],
Grant 92064004. This paper was recommended by Associate Editor C. Frenkel. [11], [12], [13], [14], [15], [16], [17]. For example, Kim
(Corresponding author: J. J. Wang.)
Shuang Liu, J. J. Wang, J. T. Zhou, S. G. Hu, Q. Yu, and Y. Liu are with et al. [12] presented a stochastic-STDP-based SNN, in which
the State Key Laboratory of Thin Solid Films and Integrated Devices, Univer- linear feedback shift register (LSFR) is used to generate the
sity of Electronic Science and Technology of China, Chengdu 610054, China updated weight. More recently, unsupervised learning based on
(e-mail: [email protected]; [email protected]; [email protected];
[email protected]; [email protected]). STDP using resistive random-access memory (ReRAM) has
T. P. Chen is with Nanyang Technological University, Singapore 639798 (e- been reported by several groups [18], [19], [20], [21]. Zhao
mail: [email protected]). et al. [18] proposed an SNN using memristor-based inhibitory
Color versions of one or more figures in this article are available at
https://doi.org/10.1109/TBCAS.2023.3242413. synapses to realize the mechanisms of lateral inhibition and
Digital Object Identifier 10.1109/TBCAS.2023.3242413 homeostasis with low hardware complexity.
1932-4545 © 2023 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: AREA- AND ENERGY-EFFICIENT SPIKING NEURAL NETWORK 93
Fig. 2. (a) Circuit schematic, top view, and (b) waveforms during the PIM
operation of the proposed 9T1C SRAM cell.
in the 1-bit mode, both S1 and S2 are turned on. Each column Q= c0 × (V DD − wi · Vtracei ) (2)
of the PBL in Fig. 3(a) is connected to 64 MOM capacitors. The i=1
charges on the four columns of PBL exhibit the same weight. where n is the row number of the 9T1C array and n equals to
The RMPMA unit can be regarded as a full connection between 64 in the 1-bit weighted design; c0 is the MOM capacitance of
64 inputs and 4 neurons with 1-bit weight. In 2-bit mode, S1 and each SRAM PIM cell; wi and Vtracei are the weight and the
S2 are turned on at first; after each SRAM PIM unit completing input of the ith row, respectively. Thereby the final voltage on
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: AREA- AND ENERGY-EFFICIENT SPIKING NEURAL NETWORK 95
V DD × m
Fig. 4. (a) The architecture and (b) the charge operation of PHPTG. VV BL = (4)
n
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.
96 IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS, VOL. 17, NO. 1, FEBRUARY 2023
Fig. 5. (a) Transfer function, and (b) voltage variability of Programmable Fig. 6. (a) The circuit schematic and (b) the Monte Carlo SPICE simulation
High-precision PIM Threshold Generation (PHPTG). result of the proposed area- and energy-efficient dynamic comparator.
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: AREA- AND ENERGY-EFFICIENT SPIKING NEURAL NETWORK 97
comparator (SOUT ) is reset to VSS. Then, SEN (SENB ) approach A. Hardware-Friendly STDP
to VDD (VSS), and the voltage difference between the compara-
tor inputs, VIP and VIN, determine whether PP/PN (NP/NN) The conventional STDP algorithm is shown in Fig. 8(a).
is pulled up (down) fastest. The following latch structure can The update of weight depends on the time interval and sequence
distinguish the small offset between PP (NP) and PN (NN) and between the pre-neuron and post-neuron. If a pre-neuron’s spike
rapidly generate a rail-to-rail output through positive feedback. arrives before a post-neuron’s firing, the pre-neuron can be
Fig. 6(b) shows the Monte Carlo SPICE simulation result of the considered to have a facilitative effect on post-neuron’s fir-
proposed area- and energy-efficient dynamic comparator, given ing, known as long-term potentiation (LTP). As a result, the
1000 points, the offset voltage distribution of the comparator has weights of pre-to-post neurons increase. On the other hand, if
a sigma value of 3.794 mV, which shows a 1.98 × improvement a pre-neuron fires after a post-neuron, this pre-neuron has an
in offset voltage compared to [36]. inhibitory effect on the post-neuron, which is called long-term
depression (LTD). Thus, the weight is decreased. Each update
value of weight is a function of the pulse interval. Therefore,
III. STDP LEARNING under the conventional design, when a spike arrives, there should
In this section, we introduce a co-design of algorithm and be a calculation to update the weights, which is not friendly to
hardware for hardware-friendly SNN and efficient STDP-based hardware design and consumes a lot of energy.
learning methodology. As shown in Fig. 7, the amount of input To optimize hardware overhead and achieve low power con-
neurons is the same as that of excitatory neurons and the two sumption, two techniques are proposed. The first one is to use a
types of neurons are fully connected. Similarly to [18], the trace technique to track the time-varying trajectory of the input
inhibitory layer of the proposed SNN has only a neuron, which layer in SNN. During the tracking process, when an input spike
is used to realize the winner-takes-all (WTA) mechanism [37]. arrives, the trace potential increases by a fixed value. When there
When an excitatory neuron fires, the excitation is transmitted to is no spike input, the trace potential decays over time, as shown
the inhibitory layer, which inhibits the firing of other neurons in in the right panel of Fig. 8(b).
the excitatory layer. This mechanism is achieved by resetting the The second technique is to simplify the LTD process. In
comparators of the remaining unfired excitatory neurons through STDP algorithms, LTP is always triggered by the post-neuron.
the STDP/WTA Controller in the WUM shown in Fig. 1. Since the pre-neurons have been traced, the value of LTP can be
The mapping of the SNN is also shown in Fig. 7. The obtained only by sampling the trace of each input neuron when
Spike Trace Module is responsible for the operations of the the post-neuron fires. On the other hand, LTD is triggered by
input layer. The full-connected weights between the input layer pre-neurons, which is not easy to achieve the post-neuron trace
and excitatory layer and the threshold voltage of each neuron for each input neuron. Therefore, in our design, the weights
are stored in 9T1C-based RMPMA and PHPTG, respectively. only need to be subtracted over a fixed value when neurons
The comparator array, RMPMA, PHPTG, and some controlling trigger LTD, as shown in the right panel of Fig. 8(b). This
logic in System Controller constitute the excitatory layer. The method can reduce the complexity of the circuit, while within
inhibition layer is realized by WTA Controller in STDP-based acceptable accuracy. The detailed comparison will be discussed
WUM. During unsupervised learning, the STDP-based WUM in the evaluation section. In the proposed STDP algorithm, each
can update the weights and threshold voltages in RMPMA and trigger of the LTP process triggers an LTD process, and the LTD
PHPTG through writing/reading operations. needs one clock cycle to update.
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.
98 IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS, VOL. 17, NO. 1, FEBRUARY 2023
Fig. 10. Decay behavior driven by the input spike trace as well as waveforms
at different terminals.
T race2) from the Spike Trace Module, and the synapse weights
Fig. 9. (a) Illustration of the spike trace circuit that can minimize the analog-
nonlinearity and (b) the trace potential of the output as a function of the input of T race1 and T race2 are both fixed at 1, Fig. 10 shows the
spikes. decay behavior caused by input. On the other hand, the output
membrane potential on PBL is also illustrated in Fig. 10. As the
PIM is driven by clock, the output membrane potential waveform
In the aforementioned LTP procedure, when an input spike on PBL is pulse-like as shown in Fig. 10.
arrives, the trace potential of the pre-neuron increases by a fixed
value. To increase analog linearity, two clamp amplifiers are B. Parallel WLs and PBLs
used in the Spike Trace Module in the PIM-based LIFNM, as
In conventional PIM, WL is vertical to the PBL, meaning that
shown in Fig. 9(a). The feedback of the amplifier makes the drain
SRAM should be accessed row by row, as shown in Fig. 11(a).
voltages of transistors N M 1 and N M 2 tend to be the same.
For STDP, the weight updating occurs between all of the input
When the input spike is enabled, the drain and source voltages
neurons and the firing post-neuron. In other words, the updated
of transistors P M 1 and P M 2 are approximately equal. Thus,
STDP weights are stored in columns, which are synchronized
no matter what the value of the trace potential is, the charging
in the same direction as PBL. This may lead to multiple writ-
current of capacitor C is approximately equal to the reference
ing/reading cycles during the STDP updating. A disadvantage
current Iref. It should be noted that a suitable decay factor
may exist that unrelated cells may be written/read. In order to
(Vdecay ) is needed to let the trace potential decay over time,
avoid the above issue, a parallel direction of WL and PBL is
as shown in Fig. 9(b). The capacitor C in Fig. 9(a) is composed
adopted to accommodate the column-by-column updating char-
of a row of 9T1C capacitors in RMPMA. The decay time is set
acteristic of STDP, as shown in Fig. 11(b). This topology only
to 160 ns, with Vdecay biased at around 300 mV by external
takes one writing/reading cycle to complete updating/loading all
DAC. In charge-sharing PIMs [25], [38], [39], the input drivers
STDP weights.
are necessary because of parallel connection of capacitors in
the bit cells. In our design, the proposed Input Trace Module
C. STDP Learning Timing Diagram
is responsible not only for formatting the inputs, but also for
driving the RMPMA and reflecting the decay behavior of LIF The timing diagram for 4-bit-weight STDP learning is shown
neurons. To reduce hardware overhead, we designed an Input in Fig. 12. In LIF operation, the membrane potential start to leaky
Trace Module with full driving capability rather than a cascade integrate and fire including four stages, i.e., three PIM phases
of a low-power Input Trace Module accompanied with a strong and one phase which contains precision config and comparator
PIM driver, which results in additional capacitor and amplifier. sensing. These four stages are periodically performed every two
In LIF neurons, leakage behavior is necessary to avoid clocks. Once the voltage on VBL exceeds the voltage on PBL,
membrane potential saturation. To further facilitate the area e.g., VBLm and PBLm , the output of mth comparator (Soutm )
efficiency, the neuron membrane potential decay behavior is will be pulled up. This leads to WTA and the disability of SEN .
realized by the decay of the input spike trace, as shown in Fig. 10. Meanwhile, the trace voltages of the spike trace circuit in Fig.
Assuming that the RMPMA module has two inputs (T race1 and 9(a) are sampled by ADCs in WUM and reset to VSS later. In
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: AREA- AND ENERGY-EFFICIENT SPIKING NEURAL NETWORK 99
Fig. 11. (a) Conventional memory access direction (blue) which is perpendic-
ular to the direction of PBLs (red); and (b) the proposed memory access direction Fig. 14. Layout of the proposed 9T1C bit cell.
(blue) which is in parallel with the direction of PBLs (red).
IV. EVALUATION
The proposed PIM-based SRAM macro with the capability
of on-chip unsupervised STDP learning was manufactured with
a 55 nm CMOS process. Fig. 13 shows a micrograph of the
chip with an area of 0.21 mm2 (excluding IO pads). In LIFNM,
the on-chip SRAM macro consists of 64 × 128 9T1C SRAM
cells in RMPMA and 64 × 128 9T1C SRAM cells in PHPTG,
respectively. We also built a larger SRAM macro with STDP
algorithm based on the Cadence Advanced Mixed-signal Sim-
Fig. 12. Timing diagram including LIF, read SRAM, STDP update, and write ulator (AMS) and PyTorch framework to evaluate the accuracy
SRAM.
and weight distribution.
The layout of the proposed 9T1C cell is shown in Fig. 14.
For high-density integration, the MOM capacitor formed by the
read operation, WLs in RMPMA (WLm [63:0](RMPMA)) and third-to-sixth metals in the 6M1T (6 metal and 1 top metal)
PHPTG (WLm [63:0](PHPTG)) are activated to read the 64 4- process is fabricated on the top of the 9 T SRAM cell without
bit weights (QPBLm [63:0]) and threshold value (QVBLm [63:0]) the requirement of extra area. The MOM capacitor, occupying
into WUM. an area of 3.68 μm2 , has a capacitance of ∼6 fF. The proposed
In STDP operation, two clocks are needed. At the first clock, 9T1C cell is 26% larger in area than the 7T1R cell [40] owing
QPBLm [63:0] and the results of 64 ADCs are individually to the two additional transistors and is 0.86 × that of the 8T1C
summed to achieve LTP. Each value in QVBLm [63:0] is summed cell [25].
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.
100 IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS, VOL. 17, NO. 1, FEBRUARY 2023
Fig. 16. Reconstructed weight matrices of 10 neurons before learning (a) and
(b) after on-chip unsupervised learning; and (c) distributions of PHPTG voltage
Fig. 15. Flowchart of the on-chip unsupervised learning of the proposed SNN. before and after learning.
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: AREA- AND ENERGY-EFFICIENT SPIKING NEURAL NETWORK 101
TABLE I
COMPARISON OF THE PROPOSED SNN MODEL WITH THE SNN MODEL IN [16]
Fig. 19. (a) Measured and (b) simulated reconstructed weight matrices using
(c) an original 8 × 8 face image; and simulated reconstructed weight matrices
under (d) 1-bit, (e) 2 b, (f) 4 b weight precisions using (g) an original 19 × 19
face image.
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.
102 IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS, VOL. 17, NO. 1, FEBRUARY 2023
TABLE II
COMPARISON OF THE SNN CHIP OF THIS WORK WITH OTHERS REPORTED IN THE LITERATURE
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.
LIU et al.: AREA- AND ENERGY-EFFICIENT SPIKING NEURAL NETWORK 103
[11] G. K. Chen, R. Kumar, H. E. Sumbul, P. C. Knag, and R. K. Krishnamurthy, [34] J. Gowda, J. Bharadwaj, N. H. Sastry, C. R. Patel, and B. Sangeetha, “4-Bit
“A 4096-neuron 1M-synapse 3.8-pJ/SOP spiking neural network with on- R-2R DAC in 18nm FinFET technology,” in Proc. IEEE Int. Conf. Circuits,
chip STDP learning and sparse weights in 10-nm FinFET CMOS,” IEEE Controls Commun., 2021, pp. 1–5.
J. Solid-State Circuits, vol. 54, no. 4, pp. 992–1002, Apr. 2019. [35] B. Razavi, “The strongARM Latch [a circuit for all seasons],” IEEE Solid-
[12] D. Kim, X. She, N. M. Rahman, V. C. K. Chekuri, and S. Mukhopad- State Circuits Mag., vol. 7, no. 2, pp. 12–17, 2015.
hyay, “Processing-in-memory-based on-chip learning with spike-time- [36] Y. Chen, L. Lu, B. Kim, and T. T.-H. Kim, “Reconfigurable 2T2R ReRAM
dependent plasticity in 65-nm CMOS,” IEEE Solid-State Circuits Lett., architecture for versatile data storage and computing in-memory,” IEEE
vol. 3, pp. 278–281, 2020. Trans. Very Large Scale Integration Syst., vol. 28, no. 12, pp. 2636–2649,
[13] G. Kim, K. Kim, S. Choi, H. J. Jang, and S.-O. Jung, “Area-and energy- Dec. 2020.
efficient STDP learning algorithm for spiking neural network SoC,” IEEE [37] J. Wang et al., “Winner-takes-all mechanism realized by memristive neural
Access, vol. 8, pp. 216922–216932, 2020. network,” Appl. Phys. Lett., vol. 115, no. 24, 2019, Art. no. 243701.
[14] H. Kim, H. Tang, W. Choi, and J. Park, “An energy-quality scalable STDP [38] E. J. Choi et al., “SRAM-based computing-in-memory macro with fully
based sparse coding processor with on-chip learning capability,” IEEE parallel one-step multibit computation,” IEEE Solid-State Circuits Lett.,
Trans. Biomed. Circuits Syst., vol. 14, no. 1, pp. 125–137, Feb. 2020. vol. 5, pp. 234–237, 2022.
[15] J. Park and S.-D. Jung, “Presynaptic spike-driven spike timing-dependent [39] B. Zhang et al., “PIMCA: A programmable in-memory computing accel-
plasticity with address event representation for large-scale neuromorphic erator for energy-efficient dnn inference,” IEEE J. Solid-State Circuits, to
systems,” IEEE Trans. Circuits Syst. I: Regular Papers, vol. 67, no. 6, be published, doi: 10.1109/JSSC.2022.3211290.
pp. 1936–1947, Jun. 2020. [40] Z. Lin, Y. Wang, C. Peng, X. Wu, X. Li, and J. Chen, “Multiple sharing
[16] G. Qiao et al., “A neuromorphic-hardware oriented bio-plausible 7T1R nonvolatile SRAM with an improved read/write margin and reliable
online-learning spiking neural network model,” IEEE Access, vol. 7, restore yield,” IEEE Trans. Very Large Scale Integrat. Syst., vol. 28, no. 3,
pp. 71730–71740, 2019. pp. 607–619, 2019.
[17] J. Wu et al., “Efficient design of spiking neural network with STDP learning [41] “CBCL Face Database.” Accessed: Feb. 01, 2023. [Online]. Available:
based on fast CORDIC,” IEEE Trans. Circuits Syst. I: Regular Papers, https://github.com/galeone/face-miner/tree/master/datasets/mitcbcl
vol. 68, no. 6, pp. 2522–2534, Jun. 2021.
[18] Z. Zhao et al., “A memristor-based spiking neural network with high
scalability and learning efficiency,” IEEE Trans. Circuits Syst. II: Exp.
Briefs, vol. 67, no. 5, pp. 931–935, May 2020.
[19] Y. Zhou et al., “Complementary memtransistor-based multilayer neural
networks for online supervised learning through (anti-) spike-timing- Shuang Liu received the B.S. degree in microelec-
dependent plasticity,” IEEE Trans. Neural Netw. Learn. Syst., vol. 33, tronics from the University of Electronic Science
no. 11, pp. 6640–6651, Nov. 2022. and Technology of China, Chengdu, China, where
[20] Y. Guo, H. Wu, B. Gao, and H. Qian, “Unsupervised learning on resistive he is currently working toward the Ph.D. degree.
memory array based spiking neural networks,” Front. Neurosci., vol. 13, His research interests include processing-in-memory
2019, Art. no. 812. circuits and neuromorphic systems.
[21] D. Ielmini, “Brain-inspired computing with resistive switching memory
(RRAM): Devices, synapses and neural networks,” Microelectronic Eng.,
vol. 190, pp. 44–53, 2018.
[22] J. Backus, “Can programming be liberated from the von Neumann style?
a functional style and its algebra of programs,” Commun. ACM, vol. 21,
no. 8, pp. 613–641, 1978.
[23] X. Huang, C. Liu, Y.-G. Jiang, and P. Zhou, “In-memory comput-
ing to break the memory wall,” Chin. Phys. B, vol. 29, no. 7, 2020, J. J. Wang received the Ph.D. degree in microelec-
Art. no. 0 78504. tronics from the University of Electronic Science
[24] D. Saito et al., “Analog in-memory computing in FeFET-based 1T1R and Technology of China, Chengdu, China. He is
array for edge AI applications,” in Proc. IEEE Symp. VLSI Technol., 2021, currently a Postdoctoral Researcher with the Univer-
pp. 1–2. sity of Electronic Science and Technology of China.
[25] Z. Jiang, S. Yin, J.-S. Seo, and M. Seok, “C3SRAM: An in-memory- His research interests include digital circuit design,
computing SRAM macro based on robust capacitive coupling computing nonvolatile memory devices, and their applications
mechanism,” IEEE J. Solid-State Circuits, vol. 55, no. 7, pp. 1888–1897, in artificial intelligence.
Jul. 2020.
[26] X. Si et al., “A local computing cell and 6T SRAM-based computing-
in-memory macro with 8-b MAC operation for edge AI chips,” IEEE J.
Solid-State Circuits, vol. 56, no. 9, pp. 2817–2831, Sep. 2021.
[27] H. Valavi, P. J. Ramadge, E. Nestler, and N. Verma, “A 64-Tile 2.4-
Mb in-memory-computing CNN accelerator employing charge-domain J. T. Zhou received the B.S. degree in microelec-
compute,” IEEE J. Solid-State Circuits, vol. 54, no. 6, pp. 1789–1799, tronics from the University of Electronic Science
Jun. 2019. and Technology of China, Chengdu, China, where
[28] X. Si et al., “A dual-split 6t SRAM-based computing-in-memory unit- he is currently working toward the M.S. degree. His
macro with fully parallel product-sum operation for binarized DNN edge research interests include neuromorphic computation
processors,” IEEE Trans. Circuits Syst. I: Regular Papers, vol. 66, no. 11, and digital IC design.
pp. 4172–4185, Nov. 2019.
[29] H. Cho, H. Son, K. Seong, B. Kim, H.-J. Park, and J.-Y. Sim, “An on-
chip learning neuromorphic autoencoder with current-mode transposable
memory read and virtual lookup table,” IEEE Trans. Biomed. Circuits
Syst., vol. 12, no. 1, pp. 161–170, Feb. 2018.
[30] S. Li, Z. Zhang, R. Mao, J. Xiao, L. Chang, and J. Zhou, “A fast and
energy-efficient SNN processor with adaptive clock/event-driven compu-
tation scheme and online learning,” IEEE Trans. Circuits Syst. I: Regular S. G. Hu received the Ph.D. degree in microelec-
Papers, vol. 68, no. 4, pp. 1543–1552, Apr. 2021. tronics from the University of Electronic Science
[31] L. F. Abbott, “Lapicque’s introduction of the integrate-and-fire model and Technology of China, Chengdu, China. His
neuron (1907),” Brain Res. Bull., vol. 50, no. 5/6, pp. 303–304, 1999. research interests include thin-film transistor, non-
[32] R. O. Topaloglu, “Process variation-aware multiple-fault diagnosis of volatile memory devices, and their applications in
thermometer-coded current-steering DACs,” IEEE Trans. Circuits Syst. artificial intelligence. Since 2016, he has been an
II: Exp. Briefs, vol. 54, no. 2, pp. 191–195, Feb. 2007. Associate Professor with the University of Electronic
[33] D. Marche and Y. Savaria, “Modeling R−2R segmented-ladder dacs,” Science and Technology of China.
IEEE Trans. Circuits Syst. I: Regular Papers, vol. 57, no. 1, pp. 31–43,
Jan. 2010.
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.
104 IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS, VOL. 17, NO. 1, FEBRUARY 2023
Q. Yu received the Ph.D. degree from the Univer- Y. Liu received the B.Sc. degree in microelectronics
sity of Electronic Science and Technology of China from Jilin University, Changchun, China, in 1998, and
(UESTC), Chengdu, China, in 2010. He is currently the Ph.D. degree from Nanyang Technological Uni-
a Professor and the Vice Dean with the School of Mi- versity, Singapore, in 2005. From May 2005 to July
croelectronics and Solid-State electronics, UESTC. 2006, he was a Research Fellow with Nanyang Tech-
nological University. In 2008, he joined the School
of Microelectronics, University of Electronic Science
and Technology of China, Chengdu, China, as a Full
Professor. He is the author or coauthor of more than
130 peer-reviewed journal papers and more than 100
conference papers. His research includes memristor
neural network system, neuromorphic computing ICs, and AI-RFICs. In 2006,
T. P. Chen received the Ph.D. degree from The he was the recipient of the prestigious Singapore Millennium Foundation Fel-
University of Hong Kong, Hong Kong, in 1994. He lowship, and one US patent and more than 30 China patents.
is currently an Associate Professor with the School
of Electrical and Electronic Engineering, Nanyang
Technological University, Singapore.
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. Downloaded on December 22,2023 at 10:30:36 UTC from IEEE Xplore. Restrictions apply.