![PDF] A 0.3–2.6 TOPS/W precision-scalable processor for real-time large-scale ConvNets | Semantic Scholar PDF] A 0.3–2.6 TOPS/W precision-scalable processor for real-time large-scale ConvNets | Semantic Scholar](https://d3i71xaburhd42.cloudfront.net/f2dd73ae127c5ee3713a92e1057eddea92fbf207/2-Figure1-1.png)
PDF] A 0.3–2.6 TOPS/W precision-scalable processor for real-time large-scale ConvNets | Semantic Scholar
![A 161.6 TOPS/W Mixed-mode Computing-in-Memory Processor for Energy-Efficient Mixed-Precision Deep Neural Networks (유회준교수 연구실) - KAIST 전기 및 전자공학부 A 161.6 TOPS/W Mixed-mode Computing-in-Memory Processor for Energy-Efficient Mixed-Precision Deep Neural Networks (유회준교수 연구실) - KAIST 전기 및 전자공학부](https://ee.kaist.ac.kr/wp-content/uploads/2022/08/%EC%9C%A0%ED%9A%8C%EC%A4%8010.png)
A 161.6 TOPS/W Mixed-mode Computing-in-Memory Processor for Energy-Efficient Mixed-Precision Deep Neural Networks (유회준교수 연구실) - KAIST 전기 및 전자공학부
![PDF] A 3.43TOPS/W 48.9pJ/pixel 50.1nJ/classification 512 analog neuron sparse coding neural network with on-chip learning and classification in 40nm CMOS | Semantic Scholar PDF] A 3.43TOPS/W 48.9pJ/pixel 50.1nJ/classification 512 analog neuron sparse coding neural network with on-chip learning and classification in 40nm CMOS | Semantic Scholar](https://d3i71xaburhd42.cloudfront.net/a2e283532b71e9b6af7addb3b3f4f4a1af6e0fb4/2-Figure1-1.png)
PDF] A 3.43TOPS/W 48.9pJ/pixel 50.1nJ/classification 512 analog neuron sparse coding neural network with on-chip learning and classification in 40nm CMOS | Semantic Scholar
![EdgeCortix Announces Sakura AI Co-Processor Delivering Industry Leading Low-Latency and Energy-Efficiency | EdgeCortix EdgeCortix Announces Sakura AI Co-Processor Delivering Industry Leading Low-Latency and Energy-Efficiency | EdgeCortix](https://www.edgecortix.com/hs-fs/hubfs/SAKURA-April-21-22.jpeg?width=603&height=350&name=SAKURA-April-21-22.jpeg)
EdgeCortix Announces Sakura AI Co-Processor Delivering Industry Leading Low-Latency and Energy-Efficiency | EdgeCortix
![A 0.32–128 TOPS, Scalable Multi-Chip-Module-Based Deep Neural Network Inference Accelerator With Ground-Referenced Signaling in 16 nm | Research A 0.32–128 TOPS, Scalable Multi-Chip-Module-Based Deep Neural Network Inference Accelerator With Ground-Referenced Signaling in 16 nm | Research](https://research.nvidia.com/sites/default/files/styles/wide/public/publications/RC18_photo_0_0.jpg?itok=ubFiJEaV)