Universal LLM Deployment Engine with ML Compilation
-
Updated
Mar 9, 2026 - Python
Universal LLM Deployment Engine with ML Compilation
High-performance In-browser LLM Inference Engine
Open Machine Learning Compiler Framework
Bringing stable diffusion models to web browsers. Everything runs inside the browser with no server support.
TVM Documentation in Chinese Simplified / TVM Zhong Wen Wen Dang
AutoKernel Shi Yi Ge Jian Dan Yi Yong ,Di Men Jian De Zi Dong Suan Zi You Hua Gong Ju ,Ti Gao Shen Du Xue Xi Suan Fa Bu Shu Xiao Lu .
yolort is a runtime stack for yolov5 on specialized accelerators such as tensorrt, libtorch, onnxruntime, tvm and ncnn.
TON Foundation invites talent to imagine and realize projects that have the potential to integrate with the daily lives of users.
This repository lists some awesome public CUDA, cuda-python, cuBLAS, cuDNN, CUTLASS, TensorRT, TensorRT-LLM, Triton, TVM, MLIR, PTX and High Performance Computing (HPC) projects.
Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.
Open, Modular, Deep Learning Accelerator
Optimizing Mobile Deep Learning on ARM GPU with TVM
Solidity compiler for TVM
A home for the final text of all TVM RFCs.
Add a description, image, and links to the tvm topic page so that developers can more easily learn about it.
To associate your repository with the tvm topic, visit your repo's landing page and select "manage topics."