Home

Awesome

<div align="center"><h1>AI Chip (ICs and IPs)</h1></div> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/AI-chips.png"></div> <br> <div align="center">Editor <a href="https://www.linkedin.com/in/shan-tang-27342510/"><strong>S.T.</strong></a>(Linkedin)</div> <div align="center"><strong>Welcome to My Wechat Blog <a href="[https://mp.weixin.qq.com/s/axfIBbQBDhTJ2Zt7U5WQBw](https://mp.weixin.qq.com/mp/appmsgalbum?action=getalbum&__biz=MzI3MDQ2MjA3OA==&scene=1&album_id=1374108991751782402&count=3#wechat_redirect)">StarryHeavensAbove</a> for more AI chip related articles</strong></div> <div align="center"><strong>欢迎访问我的微信公众号 <a href="[https://mp.weixin.qq.com/s/axfIBbQBDhTJ2Zt7U5WQBw](https://mp.weixin.qq.com/mp/appmsgalbum?action=getalbum&__biz=MzI3MDQ2MjA3OA==&scene=1&album_id=1374108991751782402&count=3#wechat_redirect)">StarryHeavensAbove</a></strong></div> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/qrcode_for_weichat_258.jpg" height="100"></div> <div align="center"><h1> </h1></div> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/AI_Chip_Landscape_v0p7.png"></div> <div align="center"><h1> </h1></div> <div align="center"><h2>Latest updates</h2></div> <HR> <font color="Darkred"> <ul> <li>Add news of <a href="#SambaNova">SambaNova</a>.</li> <li>Add news of <a href="#Groq">Groq</a>.</li> <li>Add news of <a href="#d-matrix">d-Matrix</a>.</li> <li>Add news of <a href="#Neureality">Neureality</a>.</li> <li>Add news of <a href="#Qualcomm">Qualcomm</a>.</li> <li>Add news of <a href="#Nvidia">Nvidia</a>.</li> <li>Add news of <a href="#Cerebras">Cerebras</a>.</li> <li>Add link to <a href="#AIChipBenchmarks">Latest MLPerf Results from MLCommons</a>.</li> <li>Add news of <a href="#IBM">IBM AIU</a>.</li> <li>Add news of <a href="#Tesla">Tesla Dojo</a>.</li> <li>Add link to <a href="#AIChipBenchmarks">Latest MLPerf Results from MLCommons</a>.</li> <li>Add news of <a href="#Cerebras">Cerebras</a>.</li> <li>Add startup <a href="#d-matrix">d-Matrix</a>.</li> <li>Add news of <a href="#Tachyum">Tachyum Prodigy Universal Processor</a>.</li> <li>Add news of <a href="#Habana">Intel Habana Gaudi®2</a>.</li> <li>Add startup <a href="#Modular">Modular AI in AI compiler section</a>.</li> <li>Add startup <a href="#Teramem">TeraMem</a>.</li> <li>Add startup <a href="#Aspinity">Aspinity</a>.</li> <li>Add news of <a href="#Synopsys">Synopsys DesignWare ARC NPX6 NPU IP</a>.</li> <li>Add news of <a href="#Nvidia">Nvidia Hopper</a>.</li> <li>Add news of <a href="#Graphcore">Graphcore</a>.</li> <li>Add startup <a href="#Ceremorphic">Ceremorphic</a>.</li> <li>Add news of <a href="#Lightelligence">Lightelligence</a>.</li> <li>Add link to <a href="#AIChipBenchmarks">Latest MLPerf Results from MLCommons</a>.</li> <li>Add news of <a href="#Cerebras">Cerebras</a>.</li> <li>Add news of <a href="#Habana">Habana</a>.</li> <li>Add news of <a href="#Google">Google Tensor Chip</a>.</li> <li>Add news of <a href="#Intel">Intel Loihi 2</a>.</li> <li>Add news of <a href="#Tesla">Tesla Dojo</a>.</li> <li>Add news of <a href="#Untether">Untether AI</a>.</li> <li>Add startup <a href="#Innatera">Innatera Nanosystems</a>.</li> <li>Add startup <a href="#EdgeQ">EdgeQ</a>.</li> <li>Add startup <a href="#Quadric">Quadric</a>.</li> <li>Add startup <a href="#AnalogInference">Analog Inference</a>.</li> <li>Add news of <a href="#Tenstorrent">Tenstorrent</a>.</li> <li>Add news of <a href="#Google">Google</a>.</li> <li>Add news of <a href="#SiMa">SiMa.ai</a>.</li> <li>Add startup <a href="#Neureality">Neureality</a>.</li> <li>Add news of <a href="#Cerebras">Cerebras</a>.</li> <li>Add news of <a href="#Groq">Groq</a>.</li> <li>Add news of <a href="#Nvidia">Nvidia</a>.</li> <li>Add news of <a href="#SambaNova">SambaNova</a>.</li> </ul> </font> <div align="center"><h1> </h1></div> <div align="center"><h2>Shortcut</h2></div> <HR> <table style="width:100%"> <tr> <th><a href="#IC_Vendors">IC Vendors</a></th><td><a href="#Intel">Intel</a>, <a href="#Qualcomm">Qualcomm</a>, <a href="#Nvidia">Nvidia</a>, <a href="#Samsung">Samsung</a>, <a href="#AMD">AMD</a>,<a href="#IBM">IBM</a>, <a href="#Marvell">Marvell</a></td> </tr> <tr> <th><a href="#Tech_Giants">Tech Giants & HPC Vendors</a></th><td><a href="#Google">Google</a>, <a href="#Amazon_AWS">Amazon_AWS</a>, <a href="#Microsoft">Microsoft</a>, <a href="#Apple">Apple</a>, <a href="#Alibaba">Alibaba Group</a>, <a href="#Tencent_Cloud">Tencent Cloud</a>, <a href="#Baidu">Baidu</a>, <a href="#Fujitsu">Fujitsu</a>, <a href="#Nokia">Nokia</a>, <a href="#Facebook">Facebook</a>, <a href="#Tesla">Tesla</a></td> </tr> <tr> <th><a href="#IP_Vendors">IP Vendors</a></th><td><a href="#ARM">ARM</a>, <a href="#Synopsys">Synopsys</a>, <a href="#Imagination">Imagination</a>, <a href="#CEVA">CEVA</a>, <a href="#Cadence">Cadence</a>, <a href="#VeriSilicon">VeriSilicon</a></td> </tr> <tr> <th><a href="#Startups_Worldwide">Startups</a></th> <td><a href="#Cerebras">Cerebras</a>, <a href="#Graphcore">Graphcore</a>, <a href="#Tenstorrent">Tenstorrent</a>, <a href="#Blaize">Blaize</a>, <a href="#Koniku">Koniku</a>, <a href="#Adapteva">Adapteva</a>, <a href="#Mythic">Mythic</a>, <a href="#Brainchip">BrainChip</a>, <a href="#Leepmind">Leepmind</a>, <a href="#Groq">Groq</a>, <a href="#Kneron">Kneron</a>, <a href="#Esperanto">Esperanto Technologies</a>, <a href="#GTI">Gyrfalcon Technology</a>, <a href="#SambaNova">SambaNova Systems</a>, <a href="#GreenWaves">GreenWaves Technology</a>, <a href="#Lightelligence">Lightelligence</a>, <a href="#Lightmatter">Lightmatter</a>, <a href="#Hailo">Hailo</a>,<a href="#Tachyum">Tachyum</a>,<a href="#Alphaics">AlphaICs</a>,<a href="#Syntiant">Syntiant</a>, <a href="#aiCTX">aiCTX</a>, <a href="#Flexlogix">Flex Logix</a>, <a href="#PFN">Preferred Network</a>, <a href="#Cornami">Cornami</a>, <a href="#Anaflash">Anaflash</a>, <a href="#Optalysys">Optaylsys</a>, <a href="#etacompute">Eta Compute</a>, <a href="#Achronix">Achronix</a>, <a href="#Areanna">Areanna AI</a>, <a href="#Neuroblade">Neuroblade</a>, <a href="#Luminous">Luminous Computing</a>, <a href="#Efinix">Efinix</a>, <a href="#AIstorm">AISTORM</a>, <a href="#SiMa">SiMa.ai</a>,<a href="#Untether">Untether AI</a>, <a href="#GrAI">GrAI Matter Lab</a>, <a href="#Rain">Rain Neuromorphics</a>, <a href="#ABR">Applied Brain Research</a>, <a href="#Xmos">XMOS</a>, <a href="#DinoplusAI">DinoPlusAI</a>, <a href="#Furiosa">Furiosa AI</a>, <a href="#Perceive">Perceive</a>, <a href="#SimpleMachines">SimpleMachines</a>, <a href="#Neureality">Neureality</a>, <a href="#AnalogInference">Analog Inference</a>, <a href="#Quadric">Quadric</a>, <a href="#EdgeQ">EdgeQ</a>, <a href="#Innatera">Innatera Nanosystems</a>, <a href="#Ceremorphic">Ceremorphic</a>, <a href="#Aspinity">Aspinity</a>, <a href="#Teramem">TeraMem, <a href="#d-matrix">d-Matrix</a></a></td> </tr> </table> <div align="center"><h1> </h1></div> <div align="center"><h2><a name="IC_Vendors"></a>I. IC Vendors</h2></div> <HR> <div align="center"><h1> </h1></div> <div align="center"><h3> </h3></div> <a name="Nvidia"></a> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Nvidia_logo.png" height="50"> </div> <div align="center"><h3> </h3></div> <div align="center"><h3>GPU</h3></div> <p><a href="https://nvidianews.nvidia.com/news/nvidia-microsoft-accelerate-cloud-enterprise-ai">NVIDIA Teams With Microsoft to Build Massive Cloud AI Computer</a></p> <blockquote> <p>Tens of Thousands of NVIDIA GPUs, NVIDIA Quantum-2 InfiniBand and Full Stack of NVIDIA AI Software Coming to Azure; NVIDIA, Microsoft and Global Enterprises to Use Platform for Rapid, Cost-Effective AI Development and Deployment</p> </blockquote> <p><strong><a href="https://developer.nvidia.com/blog/nvidia-hopper-architecture-in-depth">NVIDIA Hopper Architecture In-Depth</a></strong></p> <blockquote> <p>Today during the 2022 NVIDIA GTC Keynote address, NVIDIA CEO Jensen Huang introduced the new NVIDIA H100 Tensor Core GPU based on the new NVIDIA Hopper GPU architecture. This post gives you a look inside the new H100 GPU and describes important new features of NVIDIA Hopper architecture GPUs.</p> </blockquote> <p><a href="https://www.anandtech.com/show/16610/nvidia-unveils-grace-a-highperformance-arm-server-cpu-for-use-in-ai-systems">NVIDIA Unveils Grace: A High-Performance Arm Server CPU For Use In Big AI Systems</a></p> <blockquote> <p>Kicking off another busy Spring GPU Technology Conference for NVIDIA, this morning the graphics and accelerator designer is announcing that they are going to once again design their own Arm-based CPU/SoC. Dubbed Grace – after Grace Hopper, the computer programming pioneer and US Navy rear admiral – the CPU is NVIDIA’s latest stab at more fully vertically integrating their hardware stack by being able to offer a high-performance CPU alongside their regular GPU wares. According to NVIDIA, the chip is being designed specifically for large-scale neural network workloads, and is expected to become available in NVIDIA products in 2023.</p> </blockquote> <div align="center"><h3> </h3></div> <a name="Intel"></a> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Intel_logo.png" height="60"></div> <div align="center"><h3> </h3></div>

<a name="Mobileye"></a>

<div align="center"><h3>Mobileye EyeQ</h3></div> > Mobileye is currently developing its fifth generation SoC, the <a href="https://www.mobileye.com/our-technology/evolution-eyeq-chip/">EyeQ®5</a>, to act as the vision central computer performing sensor fusion for Fully Autonomous Driving (Level 5) vehicles that will hit the road in 2020. To meet power consumption and performance targets, EyeQ® SoCs are designed in most advanced VLSI process technology nodes – down to 7nm FinFET in the 5th generation.

<a name="Loihi 2"></a>

<div align="center"><h3>Loihi</h3></div> <p><a href="https://www.intel.com/content/www/us/en/newsroom/news/intel-unveils-neuromorphic-loihi-2-lava-software.html">Intel Advances Neuromorphic with Loihi 2, New Lava Software Framework and New Partners</a></p> <blockquote> <p>Second-generation research chip uses pre-production Intel 4 process, grows to 1 million neurons. Intel adds open software framework to accelerate developer innovation and path to commercialization.</p> </blockquote> <p><a name="Habana"></a></p> <div align="center"><h3>Habana</h3></div> <p><strong><a href="https://www.intel.com/content/www/us/en/newsroom/news/vision-2022-habana-gaudi2-greco.html">Intel’s Habana Labs Launches Second-Generation AI Processors for Training and Inferencing</a></strong></p> <blockquote> <p>Today at Intel Vision, Intel announced that Habana Labs, its data center team focused on AI deep learning processor technologies, launched its second-generation deep learning processors for training and inference: Habana® Gaudi®2 and Habana® Greco™. These new processors address an industry gap by providing customers with high-performance, high-efficiency deep learning compute choices for both training workloads and inference deployments in the data center while lowering the AI barrier to entry for companies of all sizes.</p> </blockquote> <p><a href="https://habana.ai/aws-launches-ec2-dl1-instances/">Habana Gaudi debuts in the Amazon EC2 cloud</a></p> <blockquote> <p>The primary motivation to create this new training instance class was presented by Andy Jassy in the 2020 re:Invent: “To provide our end-customers with up to 40% better price-performance than the current generation of GPU-based instances.”</p> </blockquote> <div align="center"><h3> </h3></div> <a name="Qualcomm"></a> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Qualcomm_logo.png" height="40"></div> <div align="center"><h3> </h3></div>

<a href="https://www-forbes-com.cdn.ampproject.org/c/s/www.forbes.com/sites/karlfreund/2022/11/16/qualcomm-ups-the-snapgragon-ai-game/amp/">Qualcomm Ups The Snapgragon AI Game</a>

<blockquote> <p>The leader in premium mobile SoCs has applied AI across the entire platform.</p> </blockquote>

<strong><a href="https://www.qualcomm.com/products/technology/processors/cloud-artificial-intelligence/cloud-ai-100">Qualcomm Cloud AI 100</a></strong>

<blockquote> <p>The Qualcomm Cloud AI 100, designed for AI inference acceleration, addresses unique requirements in the cloud, including power efficiency, scale, process node advancements, and signal processing—facilitating the ability of datacenters to run inference on the edge cloud faster and more efficiently. Qualcomm Cloud AI 100 is designed to be a leading solution for datacenters who increasingly rely on infrastructure at the edge-cloud.</p> </blockquote> <div align="center"><h3> </h3></div> <a name="Samsung"></a> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Samsung_logo.png" height="35"></div> <div align="center"><h3> </h3></div> <strong><a href="https://news.samsung.com/global/samsung-brings-on-device-ai-processing-for-premium-mobile-devices-with-exynos-9-series-9820-processor">Samsung Brings On-device AI Processing for Premium Mobile Devices with Exynos 9 Series 9820 Processor</a></strong> > Fourth-generation custom core and 2.0Gbps LTE Advanced Pro modem enables enriched mobile experiences including AR and VR applications <br> Samsung resently unveiled “<a href="https://news.samsung.com/global/samsung-optimizes-premium-exynos-9-series-9810-for-ai-applications-and-richer-multimedia-content">The new Exynos 9810 brings premium features with a 2.9GHz custom CPU, an industry-first 6CA LTE modem and deep learning processing capabilities</a>”. <div align="center"><h3> </h3></div> <a name="AMD"></a> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/AMD_logo.png" height="35"></div> <div align="center"><h3> </h3></div> The soon to be released <a href="https://www.amd.com/en/graphics/instinct-server-accelerators">AMD Instinct™ MI Series Accelerators</a> > AMD Instinct™ accelerators are engineered from the ground up for this new era of data center computing, supercharging HPC and AI workloads to propel new discoveries. The AMD Instinct™ family of accelerators can deliver industry leading performance for the data center at any scale from single server solutions up to the world’s largest supercomputers.1 With new innovations in AMD CDNA™ 2 architecture, AMD Infinity Fabric™ technology and packaging technology, the latest AMD Instinct™ accelerators are designed to power discoveries at exascale, enabling scientists to tackle our most pressing challenges. <p><a name="IBM"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/IBM_logo.png" height="40"></div> <div align="center"><h3> </h3></div> <p><a href="https://www.ibm.com/blogs/systems/ibm-telum-processor-the-next-gen-microprocessor-for-ibm-z-and-ibm-linuxone/">Meet the IBM Artificial Intelligence Unit</a></p> <blockquote> <p>It’s our first complete system-on-chip designed to run and train deep learning models faster and more efficiently than a general-purpose CPU.</p> </blockquote> <p><a href="https://www.ibm.com/blogs/systems/ibm-telum-processor-the-next-gen-microprocessor-for-ibm-z-and-ibm-linuxone/">IBM Telum Processor: the next-gen microprocessor for IBM Z and IBM LinuxONE</a></p> <blockquote> <p>The 7 nm microprocessor is engineered to meet the demands our clients face for gaining AI-based insights from their data without compromising response time for high volume transactional workloads. </p> </blockquote> <p><a href="https://www.ibm.com/blogs/research/tag/truenorth/">TrueNorth</a> is IBM's Neuromorphic CMOS ASIC developed in conjunction with the DARPA <a href="https://en.wikipedia.org/wiki/SyNAPSE">SyNAPSE</a> program.</p> <blockquote> <p>It is a manycore processor network on a chip design, with 4096 cores, each one simulating 256 programmable silicon "neurons" for a total of just over a million neurons. In turn, each neuron has 256 programmable "synapses" that convey the signals between them. Hence, the total number of programmable synapses is just over 268 million (228). In terms of basic building blocks, its transistor count is 5.4 billion. Since memory, computation, and communication are handled in each of the 4096 neurosynaptic cores, TrueNorth circumvents the von-Neumann-architecture bottlenecks and is very energy-efficient, consuming 70 milliwatts, about 1/10,000th the power density of conventional microprocessors. <a href="https://en.wikipedia.org/wiki/TrueNorth">Wikipedia</a></p> </blockquote> <p><a href="https://www.research.ibm.com/artificial-intelligence/ai-hardware-center/">AI Hardware Center</a></p> <blockquote> <p>"The IBM Research AI Hardware Center is a global research hub headquartered in Albany, New York. The center is focused on enabling next-generation chips and systems that support the tremendous processing power and unprecedented speed that AI requires to realize its full potential.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Marvell"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Marvell_logo.png" height="60"></div> <div align="center"><h3> </h3></div> <p><a href="https://www.marvell.com/products/data-processing-units.html">Data Processing Units</a></p> <blockquote> <p>Built on seven generations of the industry’s first, most scalable and widely adopted data infrastructure processors, Marvell’s OCTEON™, OCTEON™ Fusion and ARMADA® platforms are optimized for wireless infrastructure, wireline carrier networks, enterprise and cloud data centers.</p> </blockquote> <div align="center"><h3> </h3></div> <div align="center"><h2><a name="Tech_Giants"></a>II. Tech Giants & HPC Vendors</h2></div> <p><HR></p> <div align="center"><h3> </h3></div> <p><a name="Google"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Google_logo.png" height="40"></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.zdnet.com/article/google-tensor-everything-you-need-to-know-about-the-pixel-6-chip/">Google Tensor: Everything you need to know about the Pixel 6 chip</a></strong></p> <blockquote> <p>Google has taken the wraps off its latest Pixel smartphones and, among the changes, the one with the biggest long-term impact is the switch to in-house silicon for the search giant.</p> </blockquote> <p><a href="https://www.hpcwire.com/2021/05/20/google-launches-tpu-v4-ai-chips/">Google Launches TPU v4 AI Chips</a></p> <blockquote> <p>Google CEO Sundar Pichai spoke for only one minute and 42 seconds about the company’s latest TPU v4 Tensor Processing Units during his keynote at the Google I/O virtual conference this week, but it may have been the most important and awaited news from the event.</p> </blockquote> <p><a href="https://cloud.google.com/tpu">Cloud TPU</a></p> <blockquote> <p>Machine learning has produced business and research breakthroughs ranging from network security to medical diagnoses. We built the Tensor Processing Unit (TPU) in order to make it possible for anyone to achieve similar breakthroughs. Cloud TPU is the custom-designed machine learning ASIC that powers Google products like Translate, Photos, Search, Assistant, and Gmail. Here’s how you can put the TPU and machine learning to work accelerating your company’s success, especially at scale.</p> </blockquote> <p><a href="https://cloud.google.com/edge-tpu/">Edge TPU</a></p> <blockquote> <p>AI is pervasive today, from consumer to enterprise applications. With the explosive growth of connected devices, combined with a demand for privacy/confidentiality, low latency, and bandwidth constraints, AI models trained in the cloud increasingly need to be run at the edge. Edge TPU is Google’s purpose-built ASIC designed to run AI at the edge. It delivers high performance in a small physical and power footprint, enabling the deployment of high-accuracy AI at the edge.</p> </blockquote> <p>Other references are:<br> <a href="https://mp.weixin.qq.com/s/b22p26_delWfSpy9kDJKhA">Google TPU3 看点</a><br><br> <a href="https://mp.weixin.qq.com/s/Kf_L4u7JRxJ8kF3Pi8M5iw">Google TPU 揭密</a><br><br> <a href="https://mp.weixin.qq.com/s/lBQyNSNa6-joeLZ_Kq2W8A">Google的神经网络处理器专利</a><br><br> <a href="https://mp.weixin.qq.com/s/g-BDlvSy-cx4AKItcWF7jQ">脉动阵列 - 因Google TPU获得新生</a><br><br> <a href="https://www.linkedin.com/pulse/should-we-all-embrace-systolic-arrays-chien-ping-lu">Should We All Embrace Systolic Arrays?</a><br></p> <div align="center"><h3> </h3></div> <p><a name="Amazon_AWS"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Amazon_AWS.png" height="50"></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://aws.amazon.com/cn/machine-learning/trainium/">AWS Trainium</a></strong></p> <blockquote> <p>AWS Trainium is the second custom machine learning (ML) chip designed by AWS that provides the best price performance for training deep learning models in the cloud. Trainium offers the highest performance with the most teraflops (TFLOPS) of compute power for the fastest ML training in Amazon EC2 and enables a broader set of ML applications. The Trainium chip is specifically optimized for deep learning training workloads for applications including image classification, semantic search, translation, voice recognition, natural language processing and recommendation engines.</p> </blockquote> <p><a href="https://aws.amazon.com/cn/machine-learning/inferentia/">AWS Inferentia. High performance machine learning inference chip, custom designed by AWS.</a></p> <blockquote> <p>AWS Inferentia provides high throughput, low latency inference performance at an extremely low cost. Each chip provides hundreds of TOPS (tera operations per second) of inference throughput to allow complex models to make fast predictions. For even more performance, multiple AWS Inferentia chips can be used together to drive thousands of TOPS of throughput. AWS Inferentia will be available for use with Amazon SageMaker, Amazon EC2, and Amazon Elastic Inference.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Microsoft"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Microsoft_logo.png" height="60"></div> <div align="center"><h3> </h3></div> <div align="center"><h3> </h3></div> <p><a name="Apple"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Apple_logo.png" height="60"></div> <div align="center"><h3> </h3></div> <div align="center"><h3> </h3></div> <p><a name="Alibaba"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/alibaba_logo.png" height="60"></div> <div align="center"><h3> </h3></div> <p><a href="https://medium.com/syncedreview/alibabas-new-ai-chip-can-process-nearly-80k-images-per-second-63412dec22a3">Alibaba’s New AI Chip Can Process Nearly 80K Images Per Second</a></p> <blockquote> <p>At the Alibaba Cloud (Aliyun) Apsara Conference 2019, Pingtouge unveiled its first AI dedicated processor for cloud-based large-scale AI inferencing. The Hanguang 800 is the first semiconductor product in Alibaba’s 20-year history.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Tencent_Cloud"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Tencent_Cloud_logo.png" height="30"></div> <div align="center"><h3> </h3></div> <p><a href="https://www.datacenterdynamics.com/en/news/tencent-reveals-three-data-center-chips-for-ai-video-transcoding-and-networking/">Tencent reveals three data center chips - for AI, video transcoding, and networking</a></p> <blockquote> <p>The company claims that the Zixiao AI chip is twice as good as comparable competing products, video transcoding chip Canghai was 30 percent better, and SmartNIC Xuanling was apparently four times as good. It did not provide external benchmarks or specific product details.</p> </blockquote> <div align="center"><h3> </h3></div> <p><br /> <a name="Baidu"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Baidu_logo.png" height="40"></div> <div align="center"><h3> </h3></div> <p><a href="https://www.reuters.com/technology/baidu-says-2nd-gen-kunlun-ai-chips-enter-mass-production-2021-08-18/">Baidu says 2nd-gen Kunlun AI chips enter mass production</a></p> <blockquote> <p>Chinese tech giant Baidu said on Wednesday it had begun mass-producing second-generation Kunlun artificial intelligence (AI) chips, as it races to become a key player in the chip industry which Beijing is trying to strengthen.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Fujitsu"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Fujitsu_logo.png" height="40"></div> <div align="center"><h3> </h3></div> <blockquote> <p>This <a href="https://www.nextplatform.com/2017/08/09/fujitsu-bets-deep-leaning-hpc-divergence/">DLU that Fujitsu is creating</a> is done from scratch, and it is not based on either the Sparc or ARM instruction set and, in fact, it has its own instruction set and a new data format specifically for deep learning, which were created from scratch. Japanese computing giant Fujitsu. Which knows a thing or two about making a very efficient and highly scalable system for HPC workloads, as evidenced by the K supercomputer, does not believe that the HPC and AI architectures will converge. Rather, the company is banking on the fact that these architectures will diverge and will require very specialized functions. </p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Nokia"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Nokia_logo.png" height="30"></div> <div align="center"><h3> </h3></div> <blockquote> <p>Nokia has developed the <a href="https://networks.nokia.com/5g/reefshark">ReefShark chipsets</a> for its 5G network solutions. AI is implemented in the ReefShark design for radio and embedded in the baseband to use augmented deep learning to trigger smart, rapid actions by the autonomous, cognitive network, enhancing network optimization and increasing business opportunities.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Facebook"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/facebook_logo.png" height="50"></div> <div align="center"><h3> </h3></div> <p><a href="https://www.reuters.com/technology/facebook-developing-machine-learning-chip-information-2021-09-09/">Facebook developing machine learning chip - The Information</a></p> <blockquote> <p>Facebook Inc (FB.O) is developing a machine learning chip to handle tasks such as content recommendation to users, The Information reported on Thursday, citing two people familiar with the project.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Tesla"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Tesla_logo.png" height="60"></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.forbes.com/sites/jamesmorris/2022/10/06/teslas-biggest-news-at-ai-day-was-the-dojo-supercomputer-not-the-optimus-robot/">Tesla’s Biggest News At AI Day Was The Dojo Supercomputer, Not The Optimus Robot</a></strong></p> <blockquote> <p>Elon Musk played AI Day to the crowd with the focus on the Optimus humanoid robot. But while this could have a huge impact on our lives and society if it does enter mass production at the price Musk suggested ($20,000), another part of the presentation will have more immediate effects. That was the status report on the Dojo supercomputer. It could really change the world much more quickly than a bipedal bot.</p> </blockquote> <p><a href="https://semianalysis.com/tesla-dojo-ai-super-computer-unique-packaging-and-chip-design-allow-an-order-magnitude-advantage-over-competing-ai-hardware/">Tesla Dojo – Unique Packaging and Chip Design Allow An Order Magnitude Advantage Over Competing AI Hardware</a></p> <blockquote> <p>Tesla hosted their AI Day and revealed the innerworkings of their software and hardware infrastructure. Part of this reveal was the previously teased Dojo AI training chip. Tesla claims their D1 Dojo chip has a GPU level compute, CPU level flexibility, with networking switch IO. </p> </blockquote> <div align="center"><h3> </h3></div> <div align="center"><h2><a name="IP_Vendors"></a>III. Traditional IP Vendors</h2></div> <p><HR></p> <div align="center"><h3> </h3></div> <p><a name="ARM"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/ARM_logo.png" height="30"></div> <div align="center"><h3> </h3></div> <a href="https://www.arm.com/products/silicon-ip-cpu/ethos/ethos-n78">NPU ETHOS-N78</a></p> <blockquote> <p>Specifically designed for inference at the edge, the ML processor gives an industry-leading performance of 4.6 TOPs, with a stunning efficiency of 3 TOPs/W for mobile devices and smart IP cameras.</p> </blockquote> <p><strong><a href="https://www.anandtech.com/show/12791/arm-details-project-trillium-mlp-architecture">ARM Details "Project Trillium" Machine Learning Processor Architecture</a></strong></p> <blockquote> <p>Arm’s second-generation, highly scalable and efficient NPU, the Ethos-N78 enables new immersive applications with a 2.5x increase in single-core performance now scalable from 1 to 10 TOP/s and beyond through many-core technologies. It provides flexibility to optimize the ML capability with 90+ configurations.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Synopsys"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Synopsys_logo.png" height="40"></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://news.synopsys.com/2022-04-19-Synopsys-Introduces-Industrys-Highest-Performance-Neural-Processor-IP">Synopsys Introduces Industry's Highest Performance Neural Processor IP</a></strong></p> <blockquote> <p>New DesignWare ARC NPX6 NPU IP Delivers Up to 3,500 TOPS Performance for Automotive, Consumer and Data Center Chip Designs</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Imagination"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Imagination_logo.png" height="60"></div> <div align="center"><h3> </h3></div> <p><a href="https://www.imaginationtech.com/products/ai/">AI Processors</a></p> <blockquote> <p>Whether you want smartness residing in the palm of your hand, consumer products or industrial robots, or enabled by powerful servers in the cloud, we can help you achieve your vision. We enable the smartness in your products with our PowerVR Neural Network Accelerators (NNA) and GPUs. Our NC-SDK enables seamless deployment of AI acceleration on either our hardware IP either in isolation or combined. Our NNA provides maximum efficiency with a scalable architecture which enables a wide range of smart edge and end point devices from low performance IoT to high performance RoboTaxi.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="CEVA"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/CEVA_logo.png" height="40"></div> <div align="center"><h3> </h3></div> <p><a href="https://www.ceva-dsp.com/app/deep-learning/">Deep learning for the real-time embedded world</a></p> <blockquote> <p>One solution lies in supplying a dedicated low power AI processor for Deep Learning at the edge, combined with a deep neural network (DNN) graph compiler</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Cadence"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Cadence_logo.png" height="40"></div> <div align="center"><h3> </h3></div> <p><a href="https://www.cadence.com/en_US/home/tools/ip/tensilica-ip/tensilica-ai-platform.html">Tensilica AI Platform</a></p> <div align="center"><h3> </h3></div> <p><a name="VeriSilicon"></a></p> <div align="center"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/VeriSilicon_logo.png" height="40"></div> <div align="center"><h3> </h3></div> <p><a href="https://www.verisilicon.com/en/IPPortfolio/VivanteNPUIP">Vivante® NPU IP</a></p> <blockquote> <p>VeriSilicon's Neural Network Processor (NPU) IP is a highly scalable, programmable computer vision and artificial intelligence processor that supports AI operations upgrades for endpoints, edge devices, and cloud devices. Designed to meet a variety of chip sizes and power budgets, the Vivante NPU IP is a cost-effective, high-quality neural network acceleration engine solution.</p> </blockquote> <div align="center"><h3> </h3></div> <div align="center"><h2><a name="Startups"></a>IV. Startups</h2></div> <p><HR></p> <div align="center"><h3> </h3></div> <p><a name="Cerebras"></a></p> <div align="center"><a href="https://www.cerebras.net/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Cerebras_logo.png" height="50"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.cerebras.net/press-release/cerebras-unveils-andromeda-a-13.5-million-core-ai-supercomputer-that-delivers-near-perfect-linear-scaling-for-large-language-models">Cerebras Unveils Andromeda, a 13.5 Million Core AI Supercomputer that Delivers Near-Perfect Linear Scaling for Large Language Models</a></strong></p> <blockquote> <p>Delivering more than 1 Exaflop of AI compute and 120 Petaflops of dense compute, Andromeda is one of the largest AI supercomputers ever built, and is dead simple to use</p> </blockquote> <p><a href="https://www.cerebras.net/blog/cerebras-sets-record-for-largest-ai-models-ever-trained-on-single-device">Cerebras Sets Record for Largest AI Models Ever Trained on Single Device</a></p> <blockquote> <p>We are announcing the largest models ever trained on a single device. Using the Cerebras Software Platform (CSoft), our customers can easily train state-of-the-art GPT language models (such as GPT-3[i] and GPT-J[ii]) with up to 20 billion parameters on a single CS-2 system. Running on a single CS-2, these models take minutes to set up and users can quickly move between models with just a few keystrokes. With clusters of GPUs, this takes months of engineering work.</p> </blockquote> <p><a href="https://www.anandtech.com/show/17061/cerebras-completes-series-f-funding-another-250m-for-4b-valuation">Cerebras Completes Series F Funding, Another $250M for $4B Valuation</a></p> <blockquote> <p>The new Series F funding round nets the company another $250m in capital, bringing the total raised through venture capital up to $720 million.</p> </blockquote> <p><a href="https://www.anandtech.com/show/16626/cerebras-unveils-wafer-scale-engine-two-wse2-26-trillion-transistors-100-yield">Cerebras Unveils Wafer Scale Engine Two (WSE2): 2.6 Trillion Transistors, 100% Yield</a></p> <blockquote> <p>Two years ago Cerebras unveiled a revolution in silicon design: a processor as big as your head, using as much area on a 12-inch wafer as a rectangular design would allow, built on 16nm, focused on both AI as well as HPC workloads. Today the company is launching its second generation product, built on TSMC 7nm, with more than double the cores and more than double of everything.</p> </blockquote> <p><a href="https://techcrunch.com/2019/11/19/the-cerebras-cs-1-computes-deep-learning-ai-problems-by-being-bigger-bigger-and-bigger-than-any-other-chip/">The Cerebras CS-1 computes deep learning AI problems by being bigger, bigger, and bigger than any other chip</a></p> <blockquote> <p>Today, the company announced the launch of its end-user compute product, the Cerebras CS-1, and also announced its first customer of Argonne National Laboratory.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Graphcore"></a></p> <div align="center"><a href="https://www.graphcore.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Graphcore_logo.png" height="70"></a></div> <p><strong><a href="https://www.eetimes.com/graphcore-supercharges-ipu-with-wafer-on-wafer/">Graphcore Supercharges IPU with Wafer-on-Wafer</a></strong></p> <blockquote> <p>Graphcore unveiled its third-generation intelligence processing unit (IPU), the first processor to be built using 3D wafer-on-wafer (WoW) technology.</p> </blockquote> <p><a href="https://www.graphcore.ai/mk2-benchmarks">MK2 PERFORMANCE BENCHMARKS</a></p> <p><a href="https://techcrunch.com/2020/02/24/graphcore-the-ai-chipmaker-raises-another-150m-at-a-1-95b-valuation/">Graphcore, the AI chipmaker, raises another $150M at a $1.95B valuation</a></p> <blockquote> <p>Graphcore, the Bristol-based startup that designs processors specifically for artificial intelligence applications, announced it has raised another $150 million in funding for R&D and to continue bringing on new customers. It’s valuation is now $1.95 billion.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a href="https://mp.weixin.qq.com/s/CH9h8dUtoNK_2ZfkK5YU0g">解密又一个xPU:Graphcore的IPU</a> give some analysis on its IPU architecture.</p> <p><a href="https://mp.weixin.qq.com/s/AMuqeaShqEv3DnibH3scEA">Graphcore AI芯片:更多分析</a> More analysis.</p> <p><a href="https://mp.weixin.qq.com/s/qP0zsSA7SQWXDqWGEAXmOg">深度剖析AI芯片初创公司Graphcore的IPU</a> In-depth analysis after more information was disclosed.</p> <div align="center"><h3> </h3></div> <p><a name="Tenstorrent"></a></p> <div align="center"><a href="http://tenstorrent.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Tenstorrent_logo.png" height="100"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://www.prnewswire.com/news-releases/tenstorrent-raises-over-200-million-at-1-billion-valuation-to-create-programmable-high-performance-ai-computers-301295913.html">Tenstorrent Raises over $200 million at $1 billion Valuation to Create Programmable, High Performance AI Computers</a></p> <blockquote> <p>TORONTO, May 20, 2021 /PRNewswire/ - Tenstorrent, a hardware start-up developing next generation computers, announced today that it has raised over $200 million in a recent funding round that values the company at $1 billion. The round was led by Fidelity Management and Research Company and includes additional investments from Eclipse Ventures, Epic CG and Moore Capital. </p> </blockquote> <p><a href="https://www.anandtech.com/show/16709/an-interview-with-tenstorrent-ceo-ljubisa-bajic-and-cto-jim-keller">An Interview with Tenstorrent: CEO Ljubisa Bajic and CTO Jim Keller</a></p> <div align="center"><h3> </h3></div> <p><a name="Blaize"></a></p> <div align="center"><a href="https://www.blaize.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Blaize_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://www.eetimes.com/automotive-ai-startup-blaize-closes-71-million-funding-round/">Automotive AI Startup Blaize Closes $71 Million Funding Round</a></p> <blockquote> <p>Blaize, formerly ThinCI, has closed a Series D round of funding at $71 million. New investor Franklin Templeton and existing investor Temasek led the round, along with participation from Denso and other new and existing investors. This round brings Blaize’s total funding to around $155 million total.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Koniku"></a></p> <div align="center"><a href="http://koniku.io/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Koniku_logo.png" height="50"></a></div> <div align="center"><h3> </h3></div> <blockquote> <p>Founded in 2014, Newark, California startup <a href="http://koniku.io/">Koniku</a> has taken in $1.65 million in funding so far to become “the world’s first neurocomputation company“. The idea is that since the brain is the most powerful computer ever devised, why not reverse engineer it? Simple, right? Koniku is actually integrating biological neurons onto chips and has made enough progress that they claim to have AstraZeneca as a customer. Boeing has also signed on with a letter of intent to use the technology in chemical-detecting drones.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Adapteva"></a></p> <div align="center"><a href="http://www.adapteva.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Adapteva_logo.png" height="70"></a></div> <div align="center"><h3> </h3></div> <p><a href="http://www.adapteva.com/">Adapteva</a> has taken in $5.1 million in funding from investors that include mobile giant Ericsson. <a href="http://www.parallella.org/docs/e5_1024core_soc.pdf">The paper "Epiphany-V: A 1024 processor 64-bit RISC System-On-Chip"</a> describes the design of Adapteva's 1024-core processor chip in 16nm FinFet technology. </p> <p><a name="Mythic"></a></p> <div align="center"><a href="https://mythic.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Mythic_logo.png" height="20"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://www.linkedin.com/pulse/era-analog-compute-has-arrived-michael-b-henry/">The Era of Analog Compute has Arrived!</a></p> <blockquote> <p>ResNet-50 in our prototype analog AI processor. Production release will support 900-1000 fps and INT8 accuracy at 3W.</p> </blockquote> <p><a href="https://venturebeat.com/2021/06/07/mythic-launches-analog-ai-processor-that-consumes-10-times-less-power/">Mythic launches analog AI processor that consumes 10 times less power</a></p> <blockquote> <p>Analog AI processor company Mythic launched its M1076 Analog Matrix Processor today to provide low-power AI processing.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Brainchip"></a></p> <div align="center"><a href="http://www.brainchipinc.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Brainchip_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><a href="hhttps://venturebeat.com/2022/01/18/brainchip-launches-neuromorphic-process-for-ai-at-the-edge/">BrainChip launches neuromorphic process for AI at the edge</a> </p> <blockquote> <p>BrainChip today announced the commercialization of its Akida neural networking processor. Aimed at a variety of edge and internet of things (IoT) applications, BrainChip claims to be the first commercial producer of neuromorphic AI chips, which could deliver benefits in ultra-low power and performance over conventional approaches.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Deepvision"></a></p> <div align="center"><a href="https://deepvision.io/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Deepvision_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><a href="AI Processor Chipmaker Deep Vision Raises $35 Million in Series B Funding">AI Processor Chipmaker Deep Vision Raises $35 Million in Series B Funding</a></p> <blockquote> <p>Tiger Global Leads Series B Financing, Enabling Deep Vision to Expand Video Analytics and Natural Language Processing Capabilities in Edge Computing Applications</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Groq"></a></p> <div align="center"><h2><a href="http://groq.com/">Groq</a></h2></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.eetimes.com/groq-demos-fast-llms-on-4-year-old-silicon/">Groq Demonstrates Fast LLMs on 4-Year-Old Silicon</a></strong></p> <blockquote> <p>MOUNTAIN VIEW, CALIF. — Groq has repositioned its first-generation AI inference chip as a language processing unit (LPU), and demonstrated Meta’s Llama-2 70-billion–parameter large language model (LLM) running inference at 240 tokens per second per user. Groq CEO Jonathan Ross told EE Times that the company had Llama-2 up and running on the company’s 10-rack (64-chip) cloud-based dev system in “a couple of days.” This system is based on the company’s first gen AI silicon, released four years ago.</p> </blockquote> <p><a href="https://www.forbes.com/sites/amyfeldman/2021/04/14/ai-chip-startup-groq-founded-by-ex-googlers-raises-300-million-to-power-autonomous-vehicles-and-data-centers/">AI Chip Startup Groq, Founded By Ex-Googlers, Raises $300 Million To Power Autonomous Vehicles And Data Centers</a></p> <blockquote> <p>Jonathan Ross left Google to launch next-generation semiconductor startup Groq in 2016. Today, the Mountain View, California-based firm said that it had raised $300 million led by Tiger Global Management and billionaire investor Dan Sundheim’s D1 Capital as it officially launched into public view. </p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Kneron"></a></p> <div align="center"><a href="http://www.kneron.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Kneron_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://www.prnewswire.com/news-releases/kneron-to-accelerate-edge-ai-development-with-more-than-10-million-usd-series-a-financing-300556674.html">Kneron to Accelerate Edge AI Development with more than 10 Million USD Series A Financing</a></p> <div align="center"><h3> </h3></div> <p><a name="GTI"></a></p> <div align="center"><a href="https://www.gyrfalcontech.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/GTI_Logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p>According to this article, <a href="https://www.prnewswire.com/news-releases/gyrfalcon-offers-automotive-ai-chip-technology-300860069.html">"Gyrfalcon offers Automotive AI Chip Technology"</a></p> <blockquote> <p>Gyrfalcon Technology Inc. (GTI), has been promoting matrix-based application specific chips for all forms of AI since offering their production versions of AI accelerator chips in September 2017. Through the licensing of its proprietary technology, the company is confident it can help automakers bring highly competitive AI chips to production for use in vehicles within 18 months, along with significant gains in AI performance, improvements in power dissipation and cost advantages.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="SambaNova"></a></p> <div align="center"><a href="https://sambanovasystems.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/SambaNova_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://venturebeat.com/ai/sambanova-unveils-new-ai-chip-to-power-full-stack-ai-platform/">SambaNova unveils new AI chip to power full-stack AI platform</a></strong></p> <blockquote> <p>Today Palo-Alto-based SambaNova Systems unveiled a new AI chip, the SN40L, which will power its full-stack large language model (LLM) platform, the SambaNova Suite, that helps enterprises go from chip to model — building and deploying customized generative AI models.</p> </blockquote> <p><a href="https://techcrunch.com/2021/04/13/sambanova-raises-676m-at-a-5-1b-valuation-to-double-down-on-cloud-based-ai-software-for-enterprises/">SambaNova raises $676M at a $5.1B valuation to double down on cloud-based AI software for enterprises</a></p> <blockquote> <p>SambaNova — a startup building AI hardware and integrated systems that run on it that only officially came out of three years in stealth last December — is announcing a huge round of funding today to take its business out into the world. The company has closed on $676 million in financing, a Series D that co-founder and CEO Rodrigo Liang has confirmed values the company at $5.1 billion.</p> </blockquote> <p><a href="https://sambanova.ai/articles/introducing-sambanova-systems-datascale-a-new-era-of-computing/">Introducing SambaNova Systems DataScale: A New Era of Computing</a></p> <blockquote> <p>SambaNova has been working closely with many organizations the past few months and has established a new state of the art in NLP. This advancement in NLP deep learning is illustrated by a GPU-crushing, world record performance result achieved on SambaNova Systems’ Dataflow-optimized system. </p> </blockquote> <p><a href="https://sambanova.ai/a-new-state-of-the-art-in-nlp-beyond-gpus/">A New State of the Art in NLP: Beyond GPUs</a></p> <blockquote> <p>SambaNova has been working closely with many organizations the past few months and has established a new state of the art in NLP. This advancement in NLP deep learning is illustrated by a GPU-crushing, world record performance result achieved on SambaNova Systems’ Dataflow-optimized system. </p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="GreenWaves"></a></p> <div align="center"><a href="https://greenwaves-technologies.com/en/greenwaves-technologies-2/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/GreenWaves_logo.png" height="50"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://www.eetimes.eu/greenwaves-shows-off-advanced-audio-demos/">GreenWaves Shows Off Advanced Audio Demos</a></p> <blockquote> <p>The Gap9 processor, a successor to Gap8 which targets computer vision in IoT devices, is an ultra-low power neural network processor suitable for battery-powered devices. GreenWaves’ vice president of marketing Martin Croome told EE Times Europe that the company decided to focus Gap9 on the hearables market after receiving traction from this sector for Gap8.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Lightelligence"></a></p> <div align="center"><a href="https://www.lightelligence.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Lightelligence_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.eetimes.com/optical-computing-chip-runs-hardest-math-problems-100x-faster-than-gpus/">Optical Chip Solves Hardest Math Problems Faster than GPUs</a></strong></p> <blockquote> <p>Optical computing startup Lightelligence has demonstrated a silicon photonics accelerator running the Ising problem more than 100 times faster than a typical GPU setup.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Lightmatter"></a></p> <div align="center"><a href="https://www.lightmatter.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Lightmatter_logo.png" height="50"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://www.eetimes.com/lightmatter-raises-more-funding-for-photonic-ai-chip/">Lightmatter Raises More Funding for Photonic AI Chip</a></p> <blockquote> <p>ightmatter, the MIT spinout building AI accelerators with a silicon photonics computing engine, announced a Series B funding round, raising an additional $80 million. The company’s technology is based on proprietary silicon photonics technology which manipulates coherent light inside a chip to perform calculations very quickly while using very little power</p> </blockquote> <p><a name="Hailo"></a></p> <div align="center"><a href="https://www.hailotech.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Hailo_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://www.eetimes.com/unicorn-ai-chipmaker-hailo-raises-136-million/">‘Unicorn’ AI Chipmaker Hailo Raises $136 Million</a></p> <blockquote> <p>Israeli AI chip startup Hailo has raised $136 million in a Series C funding round, bringing the company’s total to $224 million. The company has also reportedly reached “unicorn” status.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Tachyum"></a></p> <div align="center"><a href="http://www.tachyum.com"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Tachyum_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://www.hpcwire.com/off-the-wire/tachyum-launches-prodigy-universal-processor/">Tachyum Launches Prodigy Universal Processor</a></p> <blockquote> <p>May 11, 2021 — Tachyum today launched the world’s first universal processor, Prodigy, which unifies the functionality of a CPU, GPU and TPU in a single processor, creating a homogeneous architecture, while delivering massive performance improvements at a cost many times less than competing products</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Alphaics"></a></p> <div align="center"><a href="https://www.alphaics.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Alphaics_logo.png" height="50"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://www.eetimes.com/alphaics-begins-sampling-its-deep-learning-co-processor/">AlphaICs Begins Sampling Its Deep Learning Co-Processor</a></p> <blockquote> <p>AlphaICs, a startup developing edge AI and learning silicon aimed at smart vision applications, is sampling its deep learning co-processor, Gluon, that also comes with a software development kit.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="Syntiant"></a></p> <div align="center"><a href="https://www.syntiant.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Syntiant_logo.png" height="30"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://semiengineering.com/syntiant-analog-deep-learning-chips/">Syntiant: Analog Deep Learning Chips</a></p> <blockquote> <p>Startup Syntiant Corp. is an Irvine, Calif. semiconductor company led by former top Broadcom engineers with experience in both innovative design and in producing chips designed to be produced in the billions, according to company CEO Kurt Busch.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="aiCTX"></a></p> <div align="center"><a href="https://aictx.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/aiCTX_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.eetimes.com/document.asp?doc_id=1333983">Baidu Backs Neuromorphic IC Developer</a></strong></p> <blockquote> <p>MUNICH — Swiss startup aiCTX has closed a $1.5 million pre-A funding round from Baidu Ventures to develop commercial applications for its low-power neuromorphic computing and processor designs and enable what it calls “neuromorphic intelligence.” It is targeting low-power edge-computing embedded sensory processing systems.</p> </blockquote> <p><a name="Flexlogix"></a></p> <div align="center"><a href="http://www.flex-logix.com/nmax"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/flexlogix_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.zdnet.com/article/flex-logix-has-two-paths-to-making-a-lot-of-money-challenging-nvidia-in-ai/">Flex Logix has two paths to making a lot of money challenging Nvidia in AI</a></strong></p> <blockquote> <p>The programmable chip company scores $55 million in venture backing, bringing its total haul to $82 million</p> </blockquote> <p><a name="PFN"></a></p> <div align="center"><a href="https://projects.preferred.jp/mn-core/en/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/PFN_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.preferred-networks.jp/en/news">Preferred Networks develops a custom deep learning processor MN-Core for use in MN-3, a new large-scale cluster, in spring 2020</a></strong></p> <blockquote> <p>Dec. 12, 2018, Tokyo Japan – Preferred Networks, Inc. (“PFN”, Head Office: Tokyo, President & CEO: Toru Nishikawa) announces that it is developing MN-Core (TM), a processor dedicated to deep learning and will exhibit this independently developed hardware for deep learning, including the MN-Core chip, board, and server, at the SEMICON Japan 2018, held at Tokyo Big Site. </p> </blockquote> <p><a name="Cornami"></a></p> <div align="center"><a href="http://cornami.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Cornami_logo.jpg" height="30"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.zdnet.com/article/ai-startup-cornami-reveals-details-of-neural-net-chip/">AI Startup Cornami reveals details of neural net chip</a></strong></p> <blockquote> <p>Stealth startup Cornami on Thursday revealed some details of its novel approach to chip design to run neural networks. CTO Paul Masters says the chip will finally realize the best aspects of a technology first seen in the 1970s. </p> </blockquote> <p><a name="Anaflash"></a></p> <div align="center"><a href="http://anaflash.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Anaflash_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.smart2zero.com/news/ai-chip-startup-offers-new-edge-computing-solution">AI chip startup offers new edge computing solution</a></strong></p> <blockquote> <p>Anaflash Inc. (San Jose, CA) is a startup company that has developed a test chip to demonstrate analog neurocomputing taking place inside logic-compatible embedded flash memory. </p> </blockquote> <p><a name="Optalysys"></a></p> <div align="center"><a href="https://www.optalysys.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Optalysys_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.globenewswire.com/news-release/2019/03/07/1749510/0/en/Optalysys-launches-world-s-first-commercial-optical-processing-system-the-FT-X-2000.html">Optalysys launches world’s first commercial optical processing system, the FT:X 2000</a></strong></p> <blockquote> <p>Optalysys develops Optical Co-processing technology which enables new levels of processing capability delivered with a vastly reduced energy consumption compared with conventional computers. Its first coprocessor is based on an established diffractive optical approach that uses the photons of low-power laser light instead of conventional electricity and its electrons. This inherently parallel technology is highly scalable and is the new paradigm of computing. </p> </blockquote> <p><a name="etacompute"></a></p> <div align="center"><a href="https://etacompute.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/etacompute_logo.png" height="80"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://spectrum.ieee.org/tech-talk/semiconductors/processors/lowpower-ai-startup-eta-compute-delivers-first-commercial-chips">Low-Power AI Startup Eta Compute Delivers First Commercial Chips</a></p> <blockquote> <p>The firm pivoted away from riskier spiking neural networks using a new power management scheme</p> </blockquote> <p><a href="https://spectrum.ieee.org/tech-talk/semiconductors/processors/eta-compute-debuts-spiking-neural-network-chip-for-edge-ai">Eta Compute Debuts Spiking Neural Network Chip for Edge AI</a></p> <blockquote> <p>Chip can learn on its own and inference at 100-microwatt scale, says company at Arm TechCon.</p> </blockquote> <p><a name="Achronix"></a></p> <div align="center"><a href="https://www.achronix.com/product/speedster7t/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Achronix_logo.png" height="30"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.eetimes.com/document.asp?doc_id=1334717">Achronix Rolls 7-nm FPGAs for AI</a></strong></p> <blockquote> <p>Achronix is back in the game of providing full-fledged FPGAs with a new high-end 7-nm family, joining the Gold Rush of silicon to accelerate deep learning. It aims to leverage novel design of its AI block, a new on-chip network, and use of GDDR6 memory to provide similar performance at a lower cost than larger rivals Intel and Xilinx.</p> </blockquote> <p><a name="Areanna"></a></p> <div align="center"><a href="https://areanna-ai.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Areanna_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.eetimes.com/document.asp?doc_id=1334947#">Startup Runs AI in Novel SRAM</a></strong></p> <blockquote> <p>Areanna is the latest example of an explosion of new architectures spawned by the rise of deep learning. The debut of a whole new approach to computing has fired imaginations of engineers around the industry hoping to be the next Hewlett and Packard.</p> </blockquote> <p><a name="Neuroblade"></a></p> <div align="center"><a href="https://www.neuroblade.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Neuroblade_logo.png" height="120"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.eetasia.com/news/article/NeuroBlade-Preps-Inference-Chip">NeuroBlade Preps Inference Chip</a></strong></p> <blockquote> <p>Add NeuroBlade to the dozens of startups working on AI silicon. The Israeli company just closed a $23 million Series A, led by the founder of Check Point Software and with participation from Intel Capital.</p> </blockquote> <p><a name="Luminous"></a></p> <div align="center"><a href="https://www.luminouscomputing.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Luminous_logo.png" height="90"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.technologyreview.com/s/613668/ai-chips-uses-optical-semiconductor-machine-learning/">Bill Gates just backed a chip startup that uses light to turbocharge AI</a></strong></p> <blockquote> <p>Luminous Computing has developed an optical microchip that runs AI models much faster than other semiconductors while using less power.</p> </blockquote> <p><a name="Efinix"></a></p> <div align="center"><a href="https://www.efinixinc.com"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Efinix_logo.png" height="25"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.zdnet.com/article/chip-startup-efinix-hopes-to-bootstrap-ai-efforts-in-iot/">Chip startup Efinix hopes to bootstrap AI efforts in IoT</a></strong></p> <blockquote> <p>Six-year-old startup Efinix has created an intriguing twist on the FPGA technology dominated by Intel and Xiliinx; the company hopes its energy-efficient chips will bootstrap the market for embedded AI in the Internet of Things.</p> </blockquote> <p><a name="AIstorm"></a></p> <div align="center"><a href="https://aistorm.ai"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/AIstorm_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://venturebeat.com/2019/02/11/aistorm-raises-13-2-million-for-ai-edge-computing-chips/">AIStorm raises $13.2 million for AI edge computing chips</a></strong></p> <blockquote> <p>David Schie, a former senior executive at Maxim, Micrel, and Semtech, thinks both markets are ripe for disruption. He — along with WSI, Toshiba, and Arm veterans Robert Barker, Andreas Sibrai, and Cesar Matias — in 2011 cofounded AIStorm, a San Jose-based artificial intelligence (AI) startup that develops chipsets that can directly process data from wearables, handsets, automotive devices, smart speakers, and other internet of things (IoT) devices. </p> </blockquote> <p><a name="SiMa"></a></p> <div align="center"><a href="http://sima.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/SiMa_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.businesswire.com/news/home/20200512005313/en/SiMa.ai-Raises-30-Million-Series-Investment-Led">SiMa.ai Raises $30 Million in Series A Investment Round Led by Dell Technologies Capital</a></strong></p> <blockquote> <p>SAN JOSE, Calif.--(BUSINESS WIRE)--SiMa.ai, the company enabling high performance machine learning to go green, today announced its Machine Learning SoC (MLSoC) platform – the industry’s first unified solution to support traditional compute with high performance, lowest power, safe and secure machine learning inference. Delivering the highest frames per second per watt, SiMa.ai’s MLSoC is the first machine learning platform to break the 1000 FPS/W barrier for ResNet-501. In customer engagements, the company has demonstrated 10-30x improvement in FPS/W through its automated software flow across a wide range of embedded edge applications, over today’s competing solutions. The platform will provide machine learning solutions that range from 50 TOPs@5W to 200 TOPs@20W, delivering an industry first of 10 TOPs/W for high performance inference.</p> </blockquote> <p><a href="https://www.businesswire.com/news/home/20191022005079/en/SiMa.ai%E2%84%A2-Introduces-MLSoC%E2%84%A2">SiMa.ai™ Introduces MLSoC™ – First Machine Learning Platform to Break 1000 FPS/W Barrier with 10-30x Improvement over Alternative Solutions</a></p> <blockquote> <p>SiMa.ai, the company enabling high performance machine learning to go green, today announced its Machine Learning SoC (MLSoC) platform – the industry’s first unified solution to support traditional compute with high performance, lowest power, safe and secure machine learning inference. Delivering the highest frames per second per watt, SiMa.ai’s MLSoC is the first machine learning platform to break the 1000 FPS/W barrier for ResNet-501. In customer engagements, the company has demonstrated 10-30x improvement in FPS/W through its automated software flow across a wide range of embedded edge applications, over today’s competing solutions. The platform will provide machine learning solutions that range from 50 TOPs@5W to 200 TOPs@20W, delivering an industry first of 10 TOPs/W for high performance inference.</p> </blockquote> <p><a name="Untether"></a></p> <div align="center"><a href="https://untether.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Untether_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://venturebeat.com/2021/07/20/untether-ai-nabs-125m-for-ai-acceleration-chips/">Untether AI nabs $125M for AI acceleration chips</a></strong></p> <blockquote> <p>Untether AI, a startup developing custom-built chips for AI inferencing workloads, today announced it has raised $125 million from Tracker Capital Management and Intel Capital. The round, which was oversubscribed and included participation from Canada Pension Plan Investment Board and Radical Ventures, will be used to support customer expansion.</p> </blockquote> <p><a name="GrAI"></a></p> <div align="center"><a href="https://www.graimatterlabs.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/GrAI_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://venturebeat.com/2019/09/18/grai-matter-labs-reveals-neuronflow-technology-and-announces-graiflow-sdk/">GrAI Matter Labs Reveals NeuronFlow Technology and Announces GrAIFlow SDK</a></strong></p> <blockquote> <p>GrAI Matter Labs (aka GML), a neuromorphic computing pioneer today revealed NeuronFlow – a new programmable processor technology – and announced an early access program to its GrAIFlow software development kit.</p> </blockquote> <p><a name="Rain"></a></p> <div align="center"><a href="http://rain-neuromorphics.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Rain_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.crunchbase.com/organization/rain-neuromorphics">Rain Neuromorphics on Crunchbase</a></strong></p> <blockquote> <p>We build artificial intelligence processors, inspired by the brain. Our mission is to enable brain-scale intelligence.</p> </blockquote> <p><a name="ABR"></a></p> <div align="center"><a href="https://appliedbrainresearch.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/ABR_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.crunchbase.com/organization/applied-brain-research">Applied Brain Research on Crunchbase</a></strong></p> <blockquote> <p>ABR makes the world's most advanced neuromoprhic compiler, runtime and libraries for the emerging space of neuromorphic computing.</p> </blockquote> <p><a name="Xmos"></a></p> <div align="center"><a href="https://www.xmos.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Xmos_logo.png" height="40"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.eetimes.com/xmos-adapts-xcore-into-aiot-crossover-processor/">XMOS adapts Xcore into AIoT ‘crossover processor’</a></strong></p> <blockquote> <p>EE Times exclusive! The new chip targets AI-powered voice interfaces in IoT devices — “the most important AI workload at the endpoint.”</p> </blockquote> <p><a href="https://venturebeat.com/2020/02/12/xmos-unveils-xcore-ai-a-powerful-chip-designed-for-ai-processing-at-the-edge/">XMOS unveils Xcore.ai, a powerful chip designed for AI processing at the edge</a></p> <blockquote> <p>The latest xcore.ai is a crossover chip designed to deliver high-performance AI, digital signal processing, control, and input/output in a single device with prices from $1.</p> </blockquote> <p><a name="DinoplusAI"></a></p> <div align="center"><a href="http://dinoplus.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/DinoplusAI_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <blockquote> <p>We design and produce AI processors and the software to run them in data centers. Our unique approach optimizes for inference with the focus on performance, power efficiency, and ease of use; and at the same time our approach enables cost-effective training. </p> </blockquote> <p><a name="Furiosa"></a></p> <div align="center"><a href="https://www.furiosa.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Furiosa_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <blockquote> <p>We build high-performance AI inference coprocessors that can be seamlessly integrated into various computing platforms including data centers, servers, desktops, automobiles and robots. </p> </blockquote> <p><a name="Corerain"></a></p> <div align="center"><a href="http://www.corerain.com/en"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Corerain_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <blockquote> <p>Corerain provides ultra-high performance AI acceleration chips and the world's first streaming engine-based AI development platform.</p> </blockquote> <p><a name="Perceive"></a></p> <div align="center"><a href="https://perceive.io/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Perceive_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://venturebeat.com/2020/03/31/perceive-emerges-from-stealth-with-ergo-edge-ai-chip/">Perceive emerges from stealth with Ergo edge AI chip</a></p> <blockquote> <p>On-device computing solutions startup Perceive emerged from stealth today with its first product: the Ergo edge processor for AI inference. CEO Steve Teig claims the chip, which is designed for consumer devices like security cameras, connected appliances, and mobile phones, delivers “breakthrough” accuracy and performance in its class.</p> </blockquote> <p><a name="SimpleMachines"></a></p> <div align="center"><a href="https://www.simplemachines.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/SimpleMachines_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><a href="https://www.design-reuse.com/news/49012/simplemachines-ai-chip-tsmc-16nm.html">SimpleMachines, Inc. Debuts First-of-its-Kind High Performance Chip</a></p> <blockquote> <p>As traditional chip makers struggle to embrace the challenges presented by the rapidly evolving AI software landscape, a San Jose startup has announced it has working silicon and a whole new future-proof chip paradigm to address these issues.

The SimpleMachines, Inc. (SMI) team – which includes leading research scientists and industry heavyweights formerly of Qualcomm, Intel and Sun Microsystems – has created a first-of-its-kind easily programmable, high-performance chip that will accelerate a wide variety of AI and machine-learning applications. </p>

</blockquote> <p><a name="Neureality"></a></p> <div align="center"><a href="https://www.neureality.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Neureality_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://techcrunch.com/2022/12/06/neureality-ai-accelerator-chips-startup-raises-35m/">NeuReality lands $35M to bring AI accelerator chips to market</a></strong></p> <blockquote> <p>NeuReality, a startup developing AI inferencing accelerator chips, has raised $35 million in new venture capital.</p> </blockquote> <p><a href="https://www.electronicsmedia.info/2021/05/06/neureality-unveiled-nr1-p-a-novel-ai-centric-inference-platform/">NeuReality unveiled NR1-P, A novel AI-centric inference platform</a></p> <blockquote> <p>NeuReality has unveiled NR1-P, a novel AI-centric inference platform. NeuReality has already started demonstrating its AI-centric platform to customers and partners. NeuReality has redefined today’s outdated AI system architecture by developing an AI-centric inference platform based on a new type of System-on-Chip (SoC). </p> </blockquote> <p><a href="https://techcrunch.com/2021/02/10/neureality-raises-8m-for-its-novel-ai-inferencing-platform/">NeuReality raises $8M for its novel AI inferencing platform</a></p> <blockquote> <p>NeuReality, an Israeli AI hardware startup that is working on a novel approach to improving AI inferencing platforms by doing away with the current CPU-centric model, is coming out of stealth today and announcing an $8 million seed round. </p> </blockquote> <p><a name="AnalogInference"></a></p> <div align="center"><a href="https://www.analog-inference.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/AnalogInference_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.eenewsanalog.com/news/analog-inference-startup-raises-106-million">Analog inference startup raises $10.6 million</a></strong></p> <blockquote> <p>The company is backed by Khosla Ventures and is developing its first generation of products for AI computing at the edge. The company raised $4.5 million shortly after its formation in March 2018, so the latest tranche brings the total raised to-date to $15.1 million</p> </blockquote> <p><a name="Quadric"></a></p> <div align="center"><a href="https://www.quadric.io/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Quatric_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.hpcwire.com/off-the-wire/quadric-announces-unified-silicon-and-software-platform-optimized-for-on-device-ai/">Quadric Announces Unified Silicon and Software Platform Optimized for On-Device AI</a></strong></p> <blockquote> <p>BURLINGAME, Calif., June 22, 2021 — Quadric (quadric.io), an innovator in high-performance edge processing, has introduced a unified silicon and software platform that unlocks the power of on-device AI. </p> </blockquote> <p><a name="EdgeQ"></a></p> <div align="center"><a href="https://edgeq.io/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/EdgeQ_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://techcrunch.com/2021/01/26/edgeq-reveals-more-details-behind-its-next-gen-5g-ai-chip/">EdgeQ reveals more details behind its next-gen 5G/AI chip</a></strong></p> <blockquote> <p>5G is the current revolution in wireless technology, and every chip company old and new is trying to burrow their way into this ultra-competitive — but extremely lucrative — market. One of the most interesting new players in the space is EdgeQ, a startup with a strong technical pedigree via Qualcomm that we covered last year after it raised a nearly $40 million Series A.</p> </blockquote> <p><a name="Innatera"></a></p> <div align="center"><a href="http://www.innatera.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Innatera_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.eetimes.com/innatera-unveils-neuromorphic-ai-chip-to-accelerate-spiking-networks/">Innatera Unveils Neuromorphic AI Chip to Accelerate Spiking Networks</a></strong></p> <blockquote> <p>Innatera, the Dutch startup making neuromorphic AI accelerators for spiking neural networks, has produced its first chips, gauged their performance, and revealed details of their architecture.</p> </blockquote> <p><a name="Ceremorphic"></a></p> <div align="center"><a href="https://ceremorphic.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Ceremorphic_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.eetimes.com/redpine-founder-launches-ai-processor-startup/">Redpine Founder Launches AI Processor Startup</a></strong></p> <blockquote> <p>Ceremorphic, an AI chip startup emerging from stealth mode this week, is readying a heterogeneous AI processor aimed at model training in data centers, automotive, high-performance computing, robotics and other emerging applications.</p> </blockquote> <p><a name="Aspinity"></a></p> <div align="center"><a href="https://www.aspinity.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Aspinity_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://embeddedcomputing.com/technology/analog-and-power/analog-semicundoctors-sensors/aspinity-analog-ml-chip-allows-battery-powered-always-on">Aspinity Analog ML Chip Allows Battery-Powered “Always On”</a></strong></p> <blockquote> <p>Machine learning (ML) is all about massive amounts of processing, DSP, etc., right? Maybe not, according to the team at Aspinity. The company continues to push ahead on the analog front. The latest member of the company’s analogML family, the AML100, operates completely in the analog domain. As a result, it can reduce always-on system power by 95% (for the record, we had to walk through this a couple of times before I believed them).</p> </blockquote> <p><a name="Teramem"></a></p> <div align="center"><a href="https://www.tetramem.com/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/Teramem_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="https://www.tetramem.com/posts/TetraMem-Technology-Debut-at-Linley">TetraMem enjoyed an exciting public debut of our analog in-memory compute technology at the Linley Spring 2022 Processor Conference.</a></strong></p> <blockquote> <p></p> </blockquote> <p><a name="d-matrix"></a></p> <div align="center"><a href="https://www.d-matrix.ai/"><img src="https://github.com/basicmi/Deep-Learning-Processor-List/raw/master/resource/d-matrix_logo.png" height="60"></a></div> <div align="center"><h3> </h3></div> <p><strong><a href="www.reuters.com/technology/ai-chip-startup-d-matrix-raises-110-mln-with-backing-microsoft-2023-09-06/">Exclusive: AI chip startup d-Matrix raises $110 million with backing from Microsoft</a></strong></p> <blockquote> <p>Sept 6 (Reuters) - Silicon Valley-based artificial intelligence chip startup d-Matrix has raised $110 million from investors that include Microsoft Corp (MSFT.O) at a time when many chip companies are struggling to raise cash.</p> </blockquote> <p><a href="https://www.forbes.com/sites/karlfreund/2022/06/21/d-matrix-ai-chip-promises-efficient-transformer-processing/">D-Matrix AI chip promises efficient transformer processing</a></p> <blockquote> <p>The startup combines digital in-memory compute and chiplet implementations for data-center-grade inference.</p> </blockquote> <div align="center"><h3> </h3></div> <p><a name="AIChipCompilers"></a></p> <div align="center"><h2>AI Chip Compilers</h2></div> <p><HR> 1. <a href="https://github.com/pytorch/glow">pytorch/glow</a><br> 2. <a href="https://tvm.ai/">TVM:End to End Deep Learning Compiler Stack</a><br> 3. <a href="https://www.tensorflow.org/xla">Google Tensorflow XLA</a><br> 4. <a href="https://developer.nvidia.com/tensorrt">Nvidia TensorRT</a><br> 5. <a href="https://github.com/plaidml/plaidml">PlaidML</a><br> 6. <a href="https://github.com/NervanaSystems/ngraph">nGraph</a><br> 7. <a href="https://github.com/Tiramisu-Compiler/tiramisu">MIT Tiramisu compiler</a><br> 8. <a href="https://onnc.ai/">ONNC (Open Neural Network Compiler)</a><br> 9. <a href="https://mlir.llvm.org/">MLIR: Multi-Level Intermediate Representation</a><br> 10. <a href="http://tensor-compiler.org/">The Tensor Algebra Compiler (taco)</a><br> 11. <a href="https://facebookresearch.github.io/TensorComprehensions/">Tensor Comprehensions</a><br> 12. <a href="https://www.polymagelabs.com//">PolyMage Labs</a><br> 13. <a href="https://octoml.ai/">OctoML</a><br> 14. <a href="https://www.modular.com/">Modular AI</a><br> <div align="center"><h3> </h3></div> <p><a name="AIChipBenchmarks"></a></p> <div align="center"><h2>AI Chip Benchmarks</h2></div> <p><HR>
  1. <a href="https://dawn.cs.stanford.edu/benchmark/index.html">DAWNBench:An End-to-End Deep Learning Benchmark and Competition Image Classification (ImageNet)</a><br>

  2. <a href="https://github.com/rdadolf/fathom">Fathom:Reference workloads for modern deep learning methods</a><br>

  3. <a href="https://mlperf.org/">MLPerf:A broad ML benchmark suite for measuring performance of ML software frameworks, ML hardware accelerators, and ML cloud platforms</a>. <strong>You can find latest MLPerf results: training 2.1, HPC 2.0, inference tiny 1.0 <a href="https://mlcommons.org/en/news/mlperf-training-4q2022/">here.</a></strong>. <br> <strong>You can find MLPerf inference results v2.1 <a href="https://mlcommons.org/en/news/mlperf-inference-v21/">here.</a></strong>. <br> <strong>You can find MLPerf training results v1.0 <a href="https://mlcommons.org/en/news/mlperf-training-2q2022/">here.</a></strong>. <br>

  4. <a href="https://aimatrix.ai/en-us/index.html">AI Matrix</a><br>

  5. <a href="http://ai-benchmark.com/index.html">AI-Benchmark</a><br>

  6. <a href="https://github.com/AIIABenchmark/AIIA-DNN-benchmark">AIIABenchmark</a><br>

  7. <a href="https://www.eembc.org/mlmark/">EEMBC MLMark Benchmark</a><br>

<div align="center"><h3> </h3></div> <p><a name="Reference"></a></p> <div align="center"><h2>Reference</h2></div> <p><HR> <div align="center"><h3> </h3></div>
  1. <a href="https://meanderful.blogspot.jp/2017/06/fpgas-and-ai-processors-dnn-and-cnn-for.html">FPGAs and AI processors: DNN and CNN for all</a><br>
  2. <a href="http://www.nanalyze.com/2017/05/12-ai-hardware-startups-new-ai-chips/">12 AI Hardware Startups Building New AI Chips</a><br>
  3. <a href="http://eyeriss.mit.edu/tutorial.html">Tutorial on Hardware Architectures for Deep Neural Networks</a><br>
  4. <strong><a href="https://nicsefc.ee.tsinghua.edu.cn/projects/neural-network-accelerator/">Neural Network Accelerator Comparison</a></strong><br>
  5. "White Paper on AI Chip Technologies 2018". You can download it from <a href="https://cloud.tsinghua.edu.cn/f/9aa0a4f0a5684cc48495/?dl=1">here</a>, or <a href="https://drive.google.com/open?id=1ieDm0bpjVWl5MnSESRs92EcmoSzG5vcm">Google drive.</a><br>
  6. <strong>"What We Talk About When We Talk About AI Chip". <a href="https://mp.weixin.qq.com/s/SbX5yz5d3GXaLcl15DO6OQ">#1</a>, <a href="https://mp.weixin.qq.com/s/zvgDgKpIMIRLFUEW0fFOeg">#2</a>, <a href="https://mp.weixin.qq.com/s/CKHs5yblcMur4h2BwUBICw">#3</a>, <a href="https://mp.weixin.qq.com/s/hFnHhaWWYTFRUsD3HlMbLw">#4</a></strong><br>
  7. <strong><a href="https://birenresearch.github.io/AIChip_Paper_List/">AI Chip Paper List</a></strong><br>
  8. <strong><a href="https://khairy2011.medium.com/tpu-vs-gpu-vs-cerebras-vs-graphcore-a-fair-comparison-between-ml-hardware-3f5a19d89e38">TPU vs GPU vs Cerebras vs Graphcore: A Fair Comparison between ML Hardware</a></strong><br>
<div align="center"> <a href="http://www.reliablecounter.com" target="_blank"><img src="http://www.reliablecounter.com/count.php?page=https://basicmi.github.io/Deep-Learning-Processor-List/&digit=style/plain/3/&reloads=1" alt="laptop" title="laptop" border="0"></a> </div>