[1]
[2]
[3]
[4]
[5]
[6]
Fig. 7. Improved DHNN system.
[7]
𝑉𝑏 are the total bias current and the dc supply voltage (Vb of 2.5
mV is used in our cell library [16, 17]), respectively. Dynamic
power dissipation at the frequency of 50 GHz and static power
dissipation of the whole designed circuit is 11.2 W and 1.46
mW, respectively.
According to the digital simulation results, computation time
of the whole system for two storage patterns with 8 pixels is
746.1 ps. The time to calculate the hi per storage pattern is 238.0
ps and the time to calculate each pixel is 29.8 ps. Computation
time of the whole system is proportional to the number of pixels
and storage patterns.
The designed DHNN circuit could not operate ℎ𝑖 in parallel,
because the ℎ𝑖 values are calculated by the PEs sequentially. To
resolve this, the bit count circuit used a trigger flip-flop (TFF)
cell (T1 cell) that can accumulate the binary stream of information and read and store the state. The result of the accumulation is passed to the register and the next PE; as a result, ℎ𝑖 is
calculated in parallel, as illustrated in Fig. 7. The new architecture reduces latency by a factor of number of storage pattern
over the original architecture.
IV. CONCLUSION
We investigated the hardware architecture for the SFQ
DHNN system that realizes designing the circuit with a small
footprint. We added registers to PE to parallelize the algorithm.
Digital circuit simulations indicates that the retrieve pattern is
converged to the storage pattern. This means the designed SFQ
DHNN system can be applied to image recognition. We designed the DHNN circuit with two storage patterns and eight elements. The circuit area was 6.5 mm × 2.3 mm and the circuit
was numerically simulated at up to 50 GHz clock frequencies,
assuming the use of the AIST-ADP2.
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
D.A. et al., “Deep Speech 2: End-to-End Speech Recognition in English
and Mandarin,” Proc. 33rd Int’l Conf. Machine Learning, 2016, pp. 173–
182.
K. K. Likharev and V. K. Semenov, “RSFQ logic/memory family: a new
Josephson-junction technology for sub-terahertz-clock-frequency digital
systems,” IEEE Trans. Appl. Supercond., vol. 1, no. 1, pp. 3–28,
Mar.1991.
D. E. Kirichenko, S. Sarwana and A. F. Kirichenko, “Zero static power
dissipation biasing of RSFQ circuits, ” IEEE Trans. Appl. Supercond., vol.
21, no. 3, pp. 776–779, Jun.2011.
I. Nagaoka, M. Tanaka, K. Inoue, and A. Fujimaki, “A 48GHz 5.6mW
gate-level-pipelined multiplier using single-flux quantum logic,” in IEEE
International Solid- State Circuits Conference (ISSCC) 2019, pp. 460–
462.
I. Nagaoka, M. Tanaka, K. Sano, T. Yamashita, A. Fujimaki, and K. Inoue, “Demonstration of an energy-efficient, gate-level-pipelined 100
TOPS/W arithmetic logic unit based on low-voltage rapid single-fluxquantum logic,” in IEEE International Superconductive Electronics Conference (ISEC) 2019.
Schneider, M.L., Donnelly, C.A., Haygood, I.W. et al. “Synaptic
weighting in single flux quantum neuromorphic computing,” Sci Rep., vol.
10, 934, Jan. 2020.
R. Cai, A. Ren, O. Chen, N. Liu, C. Ding, X. Qian, J. Han, W. Luo, N.
Yoshikawa, and Y. Wang, “A stochastic-computing based deep learning
framework using adiabatic quantum-flux-parametron superconducting
technology,” In Proceedings of the 46th International Symposium on
Computer Architecture (ISCA '19). New York, NY, USA, pp. 567–578,
2019.
K. Ishida et al., "SuperNPU: An Extremely Fast Neural Processing Unit
Using Superconducting Logic Devices," in 53rd Annual IEEE/ACM
International Symposium on Microarchitecture (MICRO) 2020, pp. 5872, doi: 10.1109/MICRO50266.2020.00018.
M. Rastegari, V. Ordonez, J. Redmon, and A. Farhadi, “Xnor-net:
Imagenet classification using binary convolutional neural networks,” in
European conference on computer vision, pp. 525–542. 2016.
J. J. Hopfield and D. W. Tank, “Computing with neural circuits- A model.”
Science, vol. 233, no. 4764, pp. 625–633, Aug. 1986.
D. W. Tank and J. J. Hopfield, “Collective computation in neuronlike
circuits,” Scientific American, vol. 257, no. 6 , pp. 104–114, Dec. 1987.
N. Singh and A. Kapoor, “Cloud Hopfield neural network: Analysis and
simulation,” in International Conference on Advances in Computing,
Communications and Informatics (ICACCI) 2015, pp. 203-209, doi:
10.1109/ICACCI.2015.7275610.
N. Soni, N. Singh, A. Kapoor and E. K. Sharma, “Face recognition using
cloud Hopfield neural network,” in International Conference on Wireless
Communications, Signal Processing and Networking (WiSPNET) 2016,
pp. 416-419, doi: 10.1109/WiSPNET.2016.7566167.
S. Haykin, “Neural Network: A comprehensive foundation,” Prentice
Hall ,1998.
S. Nagasawa, K. Hinode, T. Satoh, M. Hidaka, H. Akaike, A. Fujimaki,
N. Yoshikawa, K. Takagi, and N. Takagi, “Nb 9-layer fabrication process
for superconducting large-scale SFQ circuits and its process evaluation,”
IEICE Trans. Electron., vol. E97-C, no. 3, pp. 132–140, Mar. 2014
Yorozu S, Kameda Y, Terai H, Fujimaki A, Yamada T and Tahara, “A
single flux quantum standard logic cell library,” Physica C, Vol. 378–381,
no. 2, pp. 1471–1474, Oct. 2002
H. Akaike et al., “Design of single flux quantum cells for a 10-Nb-layer
process,” Physica C, vol. 469, no. 15–20, pp. 1670–1673, Oct. 2009.
Y. Yamanashi et al., “100 GHz Demonstrations Based on the Single-Flux
Quantum Cell Library for the 10 kA/cm2 Nb Multi-Layer Process,” IEICE
Trans. Electron., vol. E93-C, no. 4, pp. 440–444, Apr. 2010.
...