Introduction

In the past few decades, neuromorphic computing, mimicking the human brain’s architecture and operation with electronic devices, has attracted great interest due to its high biomimetic and high-energy efficiency1,2,3. Artificial neurons are the core components of neuromorphic computing implementation, emulating biological neurons functions of potential accumulation and firing4,5. For the hardware implementation of neurons, hardware overhead, energy efficiency, and reliability are the critical evaluation criteria5,6. Yet, current hardware demonstrations of neurons struggle to satisfy these key metrics simultaneously.

Generally, the complementary metal-oxide-semiconductor (CMOS) circuit is the most mature and stable scheme for emulating biological neurons. Nevertheless, due to the lack of intrinsic biological resemblance and the complexity of circuits, CMOS neurons face many challenges in density or energy efficiency7,8,9. Recently, various emerging devices have been extensively explored to emulate biological neurons benefiting from their biological resemblance and scalability. Memristive neurons have trigged the most interest among them, including redox memristors10,11,12,13, Mott memristors14,15,16,17,18,19, phase-change memristors (PCM)20,21,22, magnetic random access memory (MRAM)23,24, etc. These neurons utilize the gradual switching of conductance to mimic membrane potential evolution, successfully emulating essential biological neuron functions with low hardware cost. However, high electroforming voltage and limited reproducibility due to temporal and spatial variations are still open questions25,26. In addition, capacitors are usually needed to realize the integration in memristive neurons, which limits their practical applications in large-scale neuromorphic computing systems10,17,27. In the very recent research, novel ferroelectric polarization-based neurons are proposed and experimentally demonstrated28,29,30,31,32. They utilize gradual polarization to mimic the integration process of biological neurons without additional capacitors33. Moreover, polarization is the intrinsic property of ferroelectric materials, which is recognized to be reproducible, reliable, and energy-efficient29,34. These features are promising to implement neurons. However, ferroelectric devices are nonvolatile, and thus need a feedback path28,29,30 or a special design of ferroelectric layer31,32 to achieve spontaneous reset after firing. The feedback path will increase the hardware cost and energy consumption of neuron implementation. In addition, it will increase the complexity of the operation, as each new input must wait for the completion of the previous reset process, especially in a system with a rate coding scheme. Thus, demonstrating an ideal electronic device that processes advanced and balanced neuronal performance without additional capacitors and reset feedback path deserves more attention.

In this work, we report a leaky integrate-and-fire (LIF) neuron based on a CMOS-compatible anti-ferroelectric field-effect transistor (AFeFET). The intrinsic polarization/depolarization processes of the Hf0.2Zr0.8O2 AFeFET successfully emulate the integrate/leaky neuronal functions without any capacitors and reset peripheral circuits. Furthermore, attributing to the plentiful merits of ferroelectric materials, AFeFET neuron exhibits many superiorities: electroforming-free, ultra-low-energy consumption (37 fJ/spike), high endurance (>1012), small cycle-to-cycle variation (as low as 3.93%) and device-to-device variations (7.57%). Also, we present that the temporal integration speed in such an AFeFET neuron depends on the intensity of postsynaptic potential, illustrating the fundamental features for performing classification tasks. Subsequently, we demonstrate a two-layer spiking neural network (SNN) with full-ferroelectric architecture for learning and recognizing the Modified National Institute of Standards and Technology (MNIST) datasets by simulation, obtaining the maximum recognition accuracy 96.8% comparable to ideal neurons. These results demonstrate that the proposed AFeFET neuron is a competitive candidate for constructing neuromorphic systems.

Results

Volatile AFeFET as LIF neurons

Figure 1a shows the architecture and processing model of biological neurons. In specific, neuronal dendrites receive input spike information from pre-neurons and transmit it to soma. Then soma integrates information and triggers an action potential when the membrane potential reaches a threshold value. The axon transmits the generated action potentials to post-neurons, and the membrane potential depolarizes to a resting state5. The increase/decrease of membrane potential corresponds to the opening or shutting of Na+/K+ channels, corresponding to three stages in Fig. 1b. Here, the dynamic process of membrane potential can be mimicked vividly by the intrinsic polarization/depolarization of anti-ferroelectric (AFE) materials. Under a silent state, AFE materials have spontaneous polarizations, but the orientations of adjacent diploes are opposite, resulting in zero net macroscopic remanent polarization, as shown in the inset of Fig. 1c. However, the diploes can be aligned by the sufficient electric field, and the phase switches from AFE to ferroelectric (FE)35. Usually, the electric field-induced FE phase is not stable, which will recover to AFE phase when the electric field is released. Thus, AFE materials exhibit volatile characteristics and representative double hysteresis (Pr ≈ 0 at 0 MV/cm) as shown in Fig. 1c. We establish the dynamic relation between the intrinsic volatile characteristic of AFE and integrate/leaky neuronal functions by constructing an AFeFET, which is integrated by an AFE capacitor (TiN/Hf0.2Zr0.8O2/TiN) and a MOSFET (see “Methods” for the details of fabrication processes), as shown in the inset of Fig. 1d. Figure 1d shows that the typical transfer characteristic of AFeFET exhibits volatility, which differs from that of nonvolatile FeFET at Vg = 0 (Supplementary Fig. 1).

Fig. 1: Leaky integrate-and-fire dynamic of biological neuron vs AFeFET artificial neurons.
figure 1

a The architecture and processing model of biological neurons. External information is perceived by dendrites and then transmitted to the soma for processing. When the membrane potential surpasses a threshold value, an action potential can be generated and transmitted to post-neurons through the axon. b The integrate-and-fire dynamics of a biological neuron. With the signal received by dendrites and transmitted to soma, few Na+ channels are activated, and the membrane potential goes up gradually (stage 1). As further signal is received, more Na+ channels are activated quickly, and a large amount of Na+ flows inside the membrane, inducing the membrane potential goes up rapidly (stage 2). Once the membrane potential exceeds a certain threshold, it decreases due to the inactivation of Na+ channels and the opening of K+ channels (stage 3). c The representative double hysteresis of AFE materials. The zero net macroscopic remanent polarization indicates the volatile characteristics of AFE materials. d The typical transfer curves of AFeFET, exhibiting counterclockwise hysteresis and volatility. The arrows indicate the switching sequences. e The continuous firing events of AFeFET neuron. After several milliseconds of free time, the AFeFET neuron can restart the LIF process and fire again.

The volatility of AFeFET emulates the self-recovery of biological neurons, which helps avoid external peripheral reset circuits. Furthermore, the AFeFET device is electroforming-free, which saves an additional high-voltage forming circuit. In order to investigate the LIF function of AFeFET, continuous gate pulses (100 μs width, 100 μs interval, 1.5 V amplitude) representing postsynaptic potentials are applied to the gate of the AFeFET. The corresponding drain current (Id) representing the membrane potential demonstrates replicable LIF behavior under gate pulse trains (Fig. 1e). The Id increases under the excitatory spike trains and the neuron fires when Id reaches a threshold (1 μA). After firing, Id decreases spontaneously in a millisecond of free time eventually, which means the AFeFET neuron recovers and gets prepared for the next firing. These merits make AFeFET suitable for emulating the integration and recovery process of neurons.

Device characteristics and mechanism of the AFeFET

In this work, the volatile characteristics of the AFeFET neuron are dominated by the composition of zirconium in HfxZr1-xO2 film. The HfxZr1-xO2 exhibits paraelectric-FE-AFE transition with increasing the concentration of zirconium element (0–100%) (Supplementary Fig. 2). Actually, HfO2 and ZrO2 exhibit very similar physical and chemical properties, such as crystal phases, lattice parameters except the dielectric properties. The pure HfO2 shows linear dielectric characteristics under electric field due to the centrosymmetric monoclinic structure36. The FE properties occur with the increasing zirconium content in doped HfO2, which is induced by the existence of a non-centrosymmetric o-phase structure. The hafnium-rich ferroelectric HfxZr1-xO2 oxides exhibit nonvolatility37,38,39, thus generally serve as memory materials. With further increasing the zirconium content, the volatile AFE properties occur in zirconium-rich HfxZr1-xO2 oxides. The polarization of AFE can be triggered by an electric field and increases under a higher electric field. But it can still revert to the initial state as the applied electric field is removed (Supplementary Fig. 3). Usually, the polarization of AFE can be ascribed to the phase transition from AFE to FE phase under the influence of electric field. The electric field-induced phase transition is always accompanied by a large-volume change40. When the electric field is released, the induced FE phase will recover to the AFE phase due to the strains resulting from volume expansion38,41,42. As a result, the zirconium-rich HfxZr1-xO2 oxides exhibit intrinsic volatility. This is the charm of AFE materials used for constructing artificial neurons.

Figure 2a, b shows the plane structure and the detailed cross-sectional image of the AFeFET neuron. According to these images, the structure of the AFeFET neuron can be observed clearly, in which an AFE capacitor (yellow square) integrates on the gate of a conventional MOSFET. The energy-dispersive X-ray spectroscopy (EDS) mapping and line scan EDS were performed to further identify the elements and structure of AFeFET (Fig. 2c, d). The thickness of TiN/HfxZr1-xO2/TiN is 40 nm/10 nm/40 nm, and the interfaces of all layers are clean and flat. In addition, the Hf, Zr, Ti, N, W elements distribute uniform and are free of inter-diffused. Then we focus on investigating the characteristics of AFE layer due to its dominant role in neuronal behavior. The composition of AFE layer is controlled by alternate deposition (one cycle HfO2 and four cycles ZrO2), and is confirmed (hafnium:zirconium ≈1:4) by the peak areas and the relative sensitivity factors in X-ray photoelectron spectroscopy (XPS) results (Supplementary Fig. 4). The high-resolution transmission electron microscopy (HRTEM) image for the details of the AFE films is presented in Fig. 2e. The polycrystalline nature and the lattice fringes of different crystals can be observed clearly. Figure 2f, g depicts the crystal structure and corresponding fast Fourier transform (FFT) image of the white square area in Fig. 2e. The relative angle and distance between two lattice planes and diffraction spots indicate the existence of [0-10]-oriented AFE tetragonal P42/nmc phase. In addition, the arrangement of zirconium atoms (green dots) is very regular, and the relative angle and lattice constants are measured directly as 55.6°, 3.49 Å, and 3.2 Å, respectively. These zirconium atoms parameters match the atomic model of [0-10]-oriented plane of tetragonal P42/nmc phase exactly in Supplementary Fig. 5. These results confirm the existence of t-phase in AFE films, which is the foundation of the AFeFET neuron.

Fig. 2: Device characteristics and mechanism of AFeFET neurons.
figure 2

a Planer structure of AFeFET device. b The cross-sectional image of AFeFET device. c The elemental mapping of the materials in the system for Hf, Zr, Ti, N, and W, respectively. d The line scan EDS of the device cross-section, which corresponds to the red arrow in (b). e The HRTEM image of the AFE films. f The crystal structure of the domain, which corresponds to the white square area in (e). g The FFT image of the crystalline domain in (f). The measured angles and distances between relative crystal faces, and diffraction spots suggest the existence of [0-10] oriented tetragonal P42/nmc structure. h The integrate-and-fire dynamics of AFeFET neuron. At stage 1, the electric field-induced FE domain nucleates, and the electrons begin to accumulate in the channel of AFeFET. With the further gate pulse applied (stage 2), the electric field-induced FE domains grow and expand, resulting in more attracted electrons accumulate in the channel of AFeFET. At stage 3, the electric field-induced FE domains transform back into the AFE domains when the gate pulse electric field is removed. The channel AFeFET switches off, and the neuron backs to resting potential. i The tendencies of Id under continuous gate pulses with different amplitudes (1.3–1.8 V amplitude, fixed 100-μs interval, fixed 100 μs width).

In order to demonstrate the dynamic of AFeFET neuron, the mechanism is shown in Fig. 2h, which is related to the transformation between AFE and FE domains. At stage 1, several input pulses as postsynaptic signals are applied to the gate of AFeFET neuron, and the electric field-induced FE orthorhombic phase (o-phase) domains nucleate, which transform from AFE tetragonal phase (t-phase) domains under the gate pulse stimuli. The polarized charges accumulate in the AFE layer and modulate the channel resistance of MOSFET, resulting in the Id begins stepping up gradually. This process is just as the small portion of Na+ channels opening. With further applying gate pulses, it comes to stage 2, at which the electric field-induced FE o-phase domains grow and expand. As a result, more attracted electrons accumulate in the channel of AFeFET, and the Id increases greatly, corresponding more Na+ channels opening. Once the Id surpasses the threshold, the AFeFET neuron would fire. After firing, the electric field-induced FE o-phase domains transform back to AFE t-phase domains due to the release of gate pulse. Consequently, the attracted electrons discharge and the channel of AFeFET switches off, indicating the AFeFET neuron returns to resting potential (stage 3). This process corresponds to the opening of K+ channels in biological neurons. Then, the AFeFET neuron fires again under another gate pulse stimuli. This repeatable and stable electric field-induced phase transition accounts for the intrinsic neuronal resemblance of AFeFET. The atomic-scale phase transition between t-phase and o-phase under the influence of electric field has been observed clearly by Lombardo et al. via in situ HRTEM42.

To present the gradual electric field-induced phase transition of AFE t-phase, we investigate the tendencies of Id under continuous gate pulses. Figure 2i shows the tendencies of Id under different gate pulse amplitude, while the gate pulse interval and width are fixed to 100 μs, respectively. Under the first 20 continuous gate pulse stimuli, an obvious integration process of Id can be observed. This resulted from the gradual formation of electric field-induced o-phase, which induces the electrons accumulation in the channel of AFeFET. The gate pulses with larger amplitude result in more o-phase e domains formation and quicker growth of Id. With further gate pulse stimuli, the reversible domains tend to reach a saturation regime. This represents the dynamic balance between the electric field-induced phase transition and the recovery of the AFE t-phase, and the Id does not increase anymore. Compared between stimuli with different amplitudes, it is clearly that the higher pulse amplitude will lead to faster growth speed and a larger saturation value of Id, which is because that more AFE t-phase domains can be switched to FE o-phase domains. Noting that a similar tendency of Id can be observed under different gate pulse intervals and widths, as illustrated in Supplementary Fig. 6. Input stimuli pulses with shorter pulse intervals or wider widths induce faster integration speed and larger saturation value of Id. In all cases, the Id increases gradually and then tends to saturate corresponds to the gradual electric field-induced phase transition and saturation processes of AFE t-phase. In addition, the tendencies of Id also illustrate the intrinsic plasticity of neurons43, demonstrating that the AFeFET has high potentiality for hardware implementation of artificial LIF neurons.

Neuronal characteristics of the AFeFET

To investigate the strength-modulated integration process of AFeFET neurons, we apply gate pulse stimuli with different amplitudes and widths to implement LIF neuron functions, as shown in Fig. 3a, b. As the input pulse intensity (amplitude/width) increasing, the AFeFET neuron needs fewer input spikes to reach the threshold (1 μA), which indicates a higher firing rate under stronger stimuli strength. This is because more electric field-induced o-phases are formed under stronger pulse intensity, resulting in faster charge integration speed. Correspondingly, a higher Id of the AFeFET neuron needs longer time to leak (Supplementary Fig. 7), which could be clarified as adaptive recovery. To further study how the leaky behavior influences the integration process, we measure the Id under gate pulse stimuli with different intervals (50–600 μs), as shown in Fig. 3c. As the interval increasing, more input pulses are needed to integrate the Id to reach the threshold value. This is because that more charges are released during the free interval time, and more input pulses are required to compensate for that. It is worth noting that, when the interval time is wider enough, the Id cannot reach the threshold anyway. This feature represents the filtering capability of the neuron for weaker input signals, which is important in biological systems and neuromorphic systems44. To further evaluate the stability of the AFeFET neuron, we extract the statistical data of input spike numbers for firing as a function of input amplitudes, as shown in Fig. 3d. The firing event needs fewer input spikes and tends to be more stable as the stimuli intensity increasing. This phenomenon exhibits that the AFeFET neuron performs high-precision computation under enough stimuli intensity, which is favorable for performing high-precision tasks. A similar relationship between input spike numbers for firing and pulse widths (or intervals) is observed, as shown in Fig. 3e, f, respectively. To directly present the stability of the AFeFET neuron, we calculate the standard deviation (σ) of integration pulse number under each stimuli condition and label them out in Fig. 3d–f. Furthermore, the cycle-to-cycle variation is calculated by dividing standard deviation (σ) by mean value (μ). The lowest variation (3.93%) between cycles is obtained under 1.7 V gate pulse amplitude, 50-μs interval, and 100 μs width. It should be noted that optimizing the gate pulse parameters may further enhance the uniformity between cycles. These results demonstrate that the AFeFET neurons can successfully emulate the strength-modulated spike frequency characteristics of biological neurons with high stability, making the AFeFET neurons firstcapable of carrying out the classification tasks11,45.

Fig. 3: AFeFET LIF neuron device characteristics.
figure 3

The dynamic of the strength-modulated integration process of AFeFET neuron under (a) different gate pulse amplitudes (1.2–1.7 V amplitude, fixed 100 μs interval, fixed 100 μs width), b different gate pulse widths (fixed 1.7 V amplitude, fixed 100 μs interval, 10 μs–150 μs width), c different gate pulse intervals (fixed 1.7 V amplitude, 50–600 μs interval, fixed 100 μs width). When the input pulse intensity is not enough, the AFeFET neuron cannot fire. Statistical data of input pulse numbers of AFeFET neuron firing as a function of d amplitudes, e widths, f intervals. The firing event needs fewer input spikes and tends to be more stable as the stimuli intensity increasing. Each data point is collected from 50 firing activities. g The energy consumption dependence on pulse parameters and threshold values of AFeFET neurons. The energy consumption decreases remarkably as the threshold and pulse-width (interval) decrease. The experimental lowest energy consumption is 37 fJ per spike and can be further reduced as decreasing pulse parameters (threshold). h The AFeFET neuron can fire more than 1012 cycles stably without significant deterioration. The results are extracted from polarization-voltage (PV) hysteresis loops @0 V forward field, 1.5 V forward field, 1.5 V backward field, respectively. i Statistical data of input pulse numbers for AFeFET neuron firing. Data are collected from 100 firing activities for each AFeFET neuron.

To further access the compatibility of the AFeFET neuron for implementing unsupervised learning, we investigated the lateral inhibitory property. During the accumulation of AFeFET neuron membrane potential, the excitability will be inhibited immediately when the AFeFET neuron receives inhibitory stimuli from adjacent neurons, as shown in Supplementary Fig. 8a,b. Moreover, the AFeFET neuron needs more excitatory inputs for the next firing under stronger lateral inhibition intensity (Supplementary Fig. 8c). This behavior is similar to the suppressive phenomenon in biological neurons between each other, which is valuable for performing competitive learning tasks.

For artificial neurons, low-energy consumption, high endurance, and high reliability are critical merits. To investigate the energy consumption of the AFeFET neuron, we performed a systematic analysis under different input pulse parameters and threshold values, as shown in Fig. 3g (extracted from Supplementary Fig. 9). The lowest energy consumption of 37 fJ/per spike can be obtained under the 50 nA threshold, 1 μs pulse width, and 1-μs pulse interval. Furthermore, the energy consumption decreases remarkably as the threshold and pulse-width (interval) decrease. Thus, it is reasonable to infer that the energy consumption can be further reduced by decreasing the threshold and pulse width (interval). Moreover, the AFeFET neurons demonstrate considerable repeatability. Supplementary Fig. 10a shows 5 × 105 stable firing cycles of the AFeFET neuron without any significant deterioration. In order to speed up the measurement, the AFE MIM structure, which is the endurance bottleneck of AFeFET, is measured for higher endurance (1012 cycles) (Supplementary Fig. 10b). Based on the endurance measurements above, it is reasonable to believe the AFeFET neuron could support more than 1012 firing events (Fig. 3h). Figure 3i shows the histograms of input spike numbers for firing, which are collected from 100 firing activities of each AFeFET neuron. The required pulse number for firing is concentrated nearby 12, and the device-to-device variation is calculated to be as low as 7.57%. This variation is extracted from 500 firing activities in five AFeFET neurons. As we claimed before, the uniformity could be further enhanced by optimizing the gate pulse parameters. These results indicate that the AFeFET neuron has high uniformity and great potential in large-scale applications.

The basic integration and fire functionality of the proposed neuron can be achieved by only one AFeFET, while the examination of current threshold, the generation of output spike, and controllable refractory period need additional circuits as shown in Supplementary Fig. 11. The detail of this circuit design is described in the supporting information. For clearly presenting the comprehensive merits of our AFeFET neuron, a benchmark comparison with other typical neurons based on emerging devices is summarized in Table 1. Considering the energy consumption, endurance, and hardware overhead, which are the critical evaluation criteria of the artificial neurons, the AFeFET neuron exhibits attractive performances.

Table 1 Comparison of various hardware implementations of spiking neurons

Network-level performance of AFeFET neuron

We have demonstrated that the AFeFET neurons can provide better energy efficiency and higher uniformity compared to the other neurons based on emerging devices. It is also essential to evaluate the network-level performance using the AFeFET neuron for the hardware implementation of SNNs. Subsequently, we construct a two-layer SNN (784 × 400 × 10) for classifying MNIST datasets, as shown in Fig. 4a. In this network, we adopt a time-to-first-spike coding method, in which all input neurons fire exactly one spike per stimulus, but the firing order carries information. A larger input corresponds to an earlier spike of the neuron. In the output layer, the first fired neuron determines the class of stimulus. As soon as one of the output neurons fires, the network assigns the corresponding category to the input, and the inference process stop. Thus, such a coding scheme is much more suitable for hardware implementation. The right panel of Fig. 4a presents the proposed hardware implementation of the network based on ferroelectric field-effect transistor (FeFET) synapses and AFeFET neurons. During inference, the input signal is applied to the drains of FeFET synapses (BLs), pulse-width modulators (PWMs) collect current on source lines (SLs) and convert to pulses with fixed amplitude and various widths. The outputs of PWMs serve as postsynaptic potentials and be applied to the gates of AFeFET neurons for performing the integration process. Then, we train such a network to learn MNIST datasets for illustrating the feasibility. During training, we adopt a supervised temporal backpropagation algorithm proposed by Kheradpisheh46. Figure 4b shows the training results with the AFeFET neurons under 1 μA threshold, achieving ~95% recognition accuracy. These results demonstrate that our AFeFET neurons have great potential to be used for fabricating SNNs chips.

Fig. 4: Network-level performance of AFeFET neuron.
figure 4

a Schematic of two-layer fully ferroelectric SNN (784 × 400 × 10) for classifying MNIST datasets and the proposed hardware implementation of the network based on FeFET synapses and AFeFET neurons. b The inference accuracy as a function of training epochs with ideal LIF neurons and AFeFET neurons, 95% recognition accuracy can be achieved for AFeFET neurons under the 1 μA threshold. c The speed-accuracy trade-off. The network is trained under different thresholds for all hidden and output neurons, the recognition accuracy and inference time increases with the threshold increasing. d The average spike number and energy consumption of the network under different threshold values. The spike number in the hidden layer increases with decreasing the threshold while the energy consumption on neurons decreases (threshold >62.5 nA). This indicates the energy-friendly feature under a lower threshold. e The robustness of the network. The network could maintain a high accuracy of ~94.1% even the device variation increases to ±16.7%.

During training, the threshold value determines the number of integrated inputs (the number of membrane states) and thus affects the network performance. Then, the relation between the threshold value and network performance is further investigated as shown in Fig. 4c. It should be noted that the recognition accuracy is related to the threshold, with the highest 96.8% accuracy under the 2 μA threshold, which is nearly identical to ideal IF neurons. The pulse number for firing is equivalent to the number of the membrane potential during training. When increasing the threshold, the number of potential membranes increases, corresponding to the increasing precision of the membrane potential. Thus, with increasing the threshold, the recognition accuracy increases. Nonetheless, the inference time increases with increasing the threshold value because more integration number is required to trigger neuron firing at higher threshold cases. Thus, there should be a trade-off between recognition accuracy and inference time, and an appropriate threshold value should be selected to balance the network performance in practical applications. Fortunately, the network can still achieve high accuracy (>86%) even at a threshold value low to 62.5 nA, which is favorable for applications that need faster inference time but not rigorous accuracy. As we claimed before, the threshold also affects the energy consumption of the neuron, which is the key parameter for SNNs chip applications. We extract the spike number and energy consumption of the neurons in the system with different training thresholds to evaluate this feature, as shown in Fig. 4d. When the threshold is higher than 0.125 μA, the spike number in the hidden layer (2nd layer) decreases with increasing the threshold. This is because the hidden neurons with a higher threshold are hard to fire. On the contrary, we observe that the total energy consumption of neurons decreases as the threshold decreases. When the threshold reduces to 62.5 nA, the spike number abruptly decreases to be less than 100. This is because under a low threshold value, the winner neuron in the output layer fires earlier, and the network could finish the inference process faster. In that case, only a minority of neurons in the hidden layers fire, and thus the total neurons consume less energy. These results demonstrate that decreasing the threshold supports fast inference speed and could greatly decrease the energy consumption of hardware neurons. In addition, device variation is another important parameter in practical applications. Figure 4e presents the inference accuracy under various device variations after training under the 1 μA threshold. As the red dots are shown, the inference accuracy only decreases 1% even the variation increases to ±16.7%, illustrating nearly no network performance degradation. These results further demonstrate that the proposed AFeFET neuron is suitable for performing SNNs tasks and has great potential for the hardware implementation of SNNs chips.

Discussion

SNNs, inspired by the human brain, are powerful platforms for enabling low-power event-driven neuromorphic hardware. In SNNs, spiking neurons are the key units that enable spikes, which exchange information through connected plastic synapses. With rich physical dynamics, memristive devices are considered promising devices to emulate spiking neurons. However, the high-energy consumption or limited reliability hinders the applications of memristive neurons in neuromorphic computing.

In this work, we demonstrated a leaky integrate-and-fire neuron based on an AFeFET. The dynamic relationship between the intrinsic polarization/depolarization process of the Hf0.2Zr0.8O2 AFeFET and integrate/leaky neuronal functions are successfully built. The AFeFET neuron features CMOS-compatible, tunable firing frequency, ultra-low hardware cost (no capacitance and additional reset circuit), ultra-low-energy consumption (37 fJ/spike), high endurance (>1012), and high uniformity among different cycles and devices, showing advanced overall performances compared with emerging devices-based neurons in literature. To verify the feasibility of the neuron, we constructed a two-layer SNN combined with FeFET synapses, achieving high recognition accuracy (96.8%), low-energy consumption, and high robustness on MNIST datasets. These results demonstrate that the AFeFET neuron is a promising candidate for constructing high-efficient SNN systems and may promote the industrial landing of neuromorphic machines based on anti-ferroelectric materials.

Methods

Sample fabrication

The fabrication processes of AFeFET neuron devices are as follows: (1) After ultraviolet lithography and lift-off process, the bottom electrode TiN (40 nm) was deposited on the gate terminal of the NMOS transistor by ion beam sputtering. The NMOS transistor was fabricated by 0.18 μm CMOS technology. The W/L of NMOS is 10 μm/1 μm and its dielectric thickness is 4 nm. (2) Then, 10 nm Hf0.2Zr0.8O2 AFE thin films were deposited on 40 nm-thick TiN bottom electrode by atomic layer deposition (ALD) process at 280 °C substrate temperature. The Hf[N(C2H5)CH3]4, Zr[N(C2H5)CH3]4, and H2O were used as hafnium precursor, Zr precursor and oxygen source, respectively. The hafnium/zirconium ratio was controlled by alternate deposition of one cycle HfO2 and four cycles ZrO2. (3) Then after the ultraviolet lithography process, 40 nm-thick TiN was grown by an ion beam top electrode was released. The two-terminal metal-insulator–metal structure was integrated on the gate of the NMOS transistor. (4) The fabricated device was annealed for 30 s at 500 °C in nitrogen atmosphere to crystallize.

Measurement method

The element ratio is confirmed by X-ray photoelectron spectroscopy (ESCALAB 250Xi). The cross-section TEM, high-resolution TEM, energy-dispersive spectroscopy and crystal structure were analyzed by transmission electron microscopy (FEI Tecnai TF-20, UK). The DC mode is measured by Agilent B1500 semiconductor parameter analyzer. A B1530A fast measurement unit module was used for generating the voltage pulse and measure the response current at the same time. Capacitance-electric field (C-E) tests are performed with 10 kHz AC probing frequency and 30 mV amplitude by Agilent B1500.

SNN simulation method

In this work, we use time-to-first-spike time coding to encode the input image into a sparse spike train. Each pixel of the input image is encoded into a single spike whose spiking time is inversely proportional to its pixel value. The dense input corresponds to earlier spiking time. And each input pixel will only generate one spike, resulting in sparser spike train than the rate coding method and significantly reducing energy consumption in hardware implementation. We constructed a 784 × 400 × 10 fully ferroelectric SNN for MNIST recognition based on such a coding method. FeFET synapses were considered during training, whose conductance was between 5 μS and 60 μS according to the experimental data in ref. 47. The pulse number (64) for firing under 10 μs pulse width is used due to the highest number counts that correspond to the bits number (6 bits) of the neuron’s membrane potentials. The 9% cycle-to-cycle variation is extracted from the statistical data of 10 μs pulse width in Fig. 3e. In addition, the leaky time constant (800 μs) is extracted from the integrate-and-fire process under 10 μs pulse width (Supplementary Fig. 7b). The energy consumption was calculated according to the data of 10 μs pulse width in Fig. 3g. When the neuron reaches the firing threshold, it will emit a spike to the subsequent layer. After emitting a spike, neurons will remain resting state until the end of the time window. In the output layer, the first spiking neuron determines the network decision. Before training, the synapse weights are initialized randomly. In the training process, the target firing time of the correct output neuron is the earliest time that all neurons fire, and the target firing time of other neurons is set to be later than the earliest firing time. According to the defined error function, synapses of the fired neuron before the actual firing time will be updated. In order to update the weights of the hidden layer, a backpropagation algorithm is used to calculate the error of the hidden layer46. During training, all time steps need to calculate the error function. In the inference process, the recognition result is obtained when the first spike is generated in the output layer, so there is no need to perform the later time step. Fewer time steps mean lower recognition latency and less energy consumption. By adjusting the neuron’s threshold, the spike generation time can be adjusted, resulting in the adjustable recognition speed and energy consumption with acceptable accuracy loss.