#

quantization

UFund-Me/Qbot
https://static.github-zh.com/github_avatars/UFund-Me?size=40

#区块链#[🔥updating ...] AI 自动量化交易机器人(完全本地部署) AI-powered Quantitative Investment Research Platform. 📃 online docs: https://ufund-me.github.io/Qbot ✨ news qbot-mini: https://github.com/Charmve/iQuant

Jupyter Notebook 13.98 k
2 个月前
bitsandbytes-foundation/bitsandbytes
https://static.github-zh.com/github_avatars/bitsandbytes-foundation?size=40
Python 7.59 k
1 天前
https://static.github-zh.com/github_avatars/kornelski?size=40
C 5.44 k
2 个月前
https://static.github-zh.com/github_avatars/AutoGPTQ?size=40
Python 4.95 k
5 个月前
https://static.github-zh.com/github_avatars/IntelLabs?size=40

Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller

Jupyter Notebook 4.4 k
2 年前
https://static.github-zh.com/github_avatars/huawei-noah?size=40

Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.

Python 3.14 k
2 年前
https://static.github-zh.com/github_avatars/huggingface?size=40

🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools

Python 3.09 k
3 天前
https://static.github-zh.com/github_avatars/nunchaku-tech?size=40

[ICLR2025 Spotlight] SVDQuant: Absorbing Outliers by Low-Rank Components for 4-Bit Diffusion Models

Python 3.01 k
2 天前
IntelLabs/nlp-architect
https://static.github-zh.com/github_avatars/IntelLabs?size=40

#自然语言处理#A model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks

Python 2.94 k
3 年前
https://static.github-zh.com/github_avatars/aaron-xichen?size=40

Base pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet)

Python 2.69 k
3 年前
https://static.github-zh.com/github_avatars/stochasticai?size=40

#大语言模型#Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJ...

Python 2.66 k
14 小时前
https://static.github-zh.com/github_avatars/intel?size=40

SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime

Python 2.49 k
1 天前
https://static.github-zh.com/github_avatars/quic?size=40

#计算机科学#AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.

Python 2.45 k
13 小时前
https://static.github-zh.com/github_avatars/thu-ml?size=40

#大语言模型#Quantized Attention achieves speedup of 2-5x and 3-11x compared to FlashAttention and xformers, without lossing end-to-end metrics across language, image, and video models.

Cuda 2.4 k
1 个月前
https://static.github-zh.com/github_avatars/pytorch?size=40
Python 2.36 k
3 小时前
loading...
Website
Wikipedia