In the absence of a new transistor technology to replace CMOS, design specialization has emerged as one of the most immediate options for achieving high-performance computing. One notable example of purpose-built architecture for inference workloads is the Google tensor processing unit (TPU). The TPU has demonstrated significantly higher efficiency compared to using a general-purpose chip. In the field of artificial intelligence, there have been numerous successful implementations of application-specific designs and accelerators in industry. For instance, Nervana’s AI architecture, Facebook’s “Big Sur”, and various forms of computer-in-network acceleration for large data centers, such as Microsoft’s FPGA Configurable Cloud and Project Catapult for FPGA-accelerated search. The objective of this Special Issue is to explore a wide range of research and demonstrations on computation-intensive applications for high-performance computing, focusing on the various specialized designs that have been developed.
Specifically, Sha et al. (reference [1]) present a design structure that integrates a CPU-based control plane and an FPGA-based data plane. Their aim is to support multiple network functions while achieving a high performance at 100 Gbps. In the deep learning field, Xu et al. (reference [2]) propose a low-power design for the YOLOv4-tiny model using an FPGA. Their design utilizes 16-bit fixed-point operators, which trade precision for the achievement of over 10 times and 3 times the power dissipation compared to CPU and GPU, respectively. Another paper in the neural network design field is from Xie et al. (reference [3]), who demonstrate an efficient accelerator for N:M sparse convolutional neural networks (CNNs) with layer-wise sparse patterns. Their implementation of FPGA validates the acceleration of classical CNNs such as Alexnet, VGG-16, and ResNet-50. Madineni et al. (reference [4]) present a parameterized design of a CNN network using Chisel, an open-source hardware construction language developed at UC Berkeley. This design allows for flexible implementation options, supporting 16-bit, 32-bit, 64-bit, and 128-bit configurations on FPGA. Popovici et al. (reference [5]) introduce a real-time RISC-V-based CAN-FD bus diagnosis tool and make the design publicly available. Wang et al. (reference [6]) propose a TCP offload engine (TOE) prototype system on an FPGA to support a 100 Gbps high-performance throughput. Their design enables the concurrent processing of from hundreds to 250,000 TCP connection state hardware maintenance on a single network node, thereby improving the overall performance of the network system.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Sha, M.; Guo, Z.; Guo, Y.; Zeng, X. A High-Performance and Flexible Architecture for Accelerating SDN on the MPSoC Platform. Micromachines 2022, 13, 1854. [Google Scholar] [CrossRef] [PubMed]
- Xu, S.; Zhou, Y.; Huang, Y.; Han, T. YOLOv4-Tiny-Based Coal Gangue Image Recognition and FPGA Implementation. Micromachines 2022, 13, 1983. [Google Scholar] [CrossRef] [PubMed]
- Xie, X.; Zhu, M.; Lu, S.; Wang, Z. Efficient Layer-Wise N:M Sparse CNN Accelerator with Flexible SPEC: Sparse Processing Element Clusters. Micromachines 2023, 14, 528. [Google Scholar] [CrossRef] [PubMed]
- Madineni, M.; Vega, M.; Yang, X. Parameterizable Design on Convolutional Neural Networks Using Chisel Hardware Construction Language. Micromachines 2023, 14, 531. [Google Scholar] [CrossRef] [PubMed]
- Popovici, C.; Stan, A. Real-Time RISC-V-Based CAN-FD Bus Diagnosis Tool. Micromachines 2023, 14, 196. [Google Scholar] [CrossRef] [PubMed]
- Wang, K.; Guo, Y.; Guo, Z. Highly Concurrent TCP Session Connection Management System on FPGA Chip. Micromachines 2023, 14, 385. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).