Stars
High-Performance Sparse Linear Algebra on HBM-Equipped FPGAs Using HLS
(Minimalism Style) Powered by Jekyll, based on the Minimal Mistakes theme and Jason Ansel's website
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Shuhai is a benchmarking-memory tool that allows FPGA programmers to demystify all the underlying details of memories, e.g., HBM and DDR4, on a Xilinx FPGA [FCCM 20]
Representation and Reference Lowering of ONNX Models in MLIR Compiler Infrastructure
A collection of pre-trained, state-of-the-art models in the ONNX format
HeteroCL: A Multi-Paradigm Programming Infrastructure for Software-Defined Heterogeneous Computing
AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.
Windows Calculator: A simple yet powerful calculator that ships with Windows
pku-liang / TensorLib
Forked from kirliavc/tensorlibA Spatial Accelerator Generation Framework for Tensor Algebra.
CIRCT-based HLS compilation flows, debugging, and cosimulation tools.
A collection of gcc tips. 100 maybe just mean many here.
Flexible Intermediate Representation for RTL
A hardware synthesis framework with multi-level paradigm
PyTorch model to RTL flow for low latency inference
A scalable High-Level Synthesis framework on MLIR
An out-of-tree MLIR dialect template.
NeuroKit2: The Python Toolbox for Neurophysiological Signal Processing