Developers
Learn
Company
Chat
Chat with and directly compare LLM endpoints
Benchmarks
Compare LLM endpoints with live performance benchmarks
Documentation
Learn how to use the Unify API
Blog
Read about LLM deployment infrastructure
Newsletter
Stay up to date with the latest in AI
Paper Readings
Join our discussions around cuttin-edge AI research
Talks
Dive deep with us into the AI landscape
Careers
Join our team and letβs Unify AI!
Contact
Reach out to our team
Privacy & Cookies
How we treat your navigation data
Terms Of Service
General requirements for using our Service
Socials
Follow us through our social accounts:
Chat
Chat with and directly compare LLM endpoints
Benchmarks
Compare LLM endpoints with live performance benchmarks
Documentation
Learn how to use the Unify API
Blog
Read about LLM deployment infrastructure
Newsletter
Stay up to date with the latest in AI
Paper Readings
Join our discussions around cuttin-edge AI research
Talks
Dive deep with us into the AI landscape
Careers
Join our team and letβs Unify AI!
Contact
Reach out to our team
Privacy & Cookies
How we treat your navigation data
Terms Of Service
General requirements for using our Service
Socials
Follow us through our social accounts:
143 sets
Contribute
Category
compilers
compression
hardware
serving
supported-hardware
eco-system
compilers
mlir
inference-optimizer
llvm
Show hot cards π₯
Name - Ascending
Name - Descending
Sort by name...
π₯
gptq
apache-2.0
compression
quantization
π₯
TensorLy
bsd-3-clause
compression
jax
mxnet
hugginface/optimum
apache-2.0
compression
distillation
open-source
AIMET
bsd-3-clause
compression
onnx
open-source
π₯
Torch-Pruning
compression
mit
open-source
pruning
π₯
only_train_once
compression
mit
open-source
pruning
π₯
bitsandbytes
compression
mit
quantization
π₯
TensorLy-Torch
bsd-3-clause
compression
pytorch
tensorization
π₯
Built In Pytorch Compression
bsd-3-clause
compression
pruning
quantization
Trailmet
binarization
compression
distillation
mit
SqueezeLLM
compression
mit
open-source
pytorch
π₯
Neural Compressor
apache-2.0
compression
distillation
mxnet
π₯
Pruna-AI
compression
distillation
proprietary
pruning
π₯
torchdistill
compression
distillation
mit
open-source
π₯
SmoothQuant
compression
mit
open-source
pytorch
Model Compression Toolkit
apache-2.0
compression
open-source
quantization
π₯
llm-awq
compression
llm
mit
open-source
Neural Network Intelligence
bsd-3-clause
compression
open-source
pruning
1
2
3
4
5
6
7
8