GitHub 中文社区
回车: Github搜索    Shift+回车: Google搜索
论坛
排行榜
趋势
登录

©2025 GitHub中文社区论坛GitHub官网网站地图GitHub官方翻译

  • X iconGitHub on X
  • Facebook iconGitHub on Facebook
  • Linkedin iconGitHub on LinkedIn
  • YouTube iconGitHub on YouTube
  • Twitch iconGitHub on Twitch
  • TikTok iconGitHub on TikTok
  • GitHub markGitHub’s organization on GitHub
集合主题趋势排行榜
#

model-quantization

Website
Wikipedia
https://static.github-zh.com/github_avatars/Efficient-ML?size=40
Efficient-ML / Awesome-Model-Quantization

#计算机科学#A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...

深度学习quantizationAwesome Listsmodel-compressionefficient-deep-learningmodel-quantization
2.17 k
5 个月前
https://static.github-zh.com/github_avatars/horseee?size=40
horseee / Awesome-Efficient-LLM

#大语言模型#A curated list for Efficient Large Language Models

compressionknowledge-distillationlanguage-model大语言模型model-quantizationefficient-llm
Python 1.8 k
1 个月前
https://static.github-zh.com/github_avatars/datawhalechina?size=40
datawhalechina / awesome-compression

模型压缩的小白入门教程,PDF下载地址 https://github.com/datawhalechina/awesome-compression/releases

knowledge-distillationmodel-compressionmodel-pruningquantizationcompressionmodel-quantizationneural-architecture-searchsvd
309
2 个月前
https://static.github-zh.com/github_avatars/inferflow?size=40
inferflow / inferflow

Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).

llama2llamacppllm-inferencemodel-quantizationmulti-gpu-inferencemixture-of-expertsmoegemmafalconminicpmmistralbloomdeepseekinternlmbaichuan2mixtralqwen
C++ 245
1 年前
https://static.github-zh.com/github_avatars/Efficient-ML?size=40
Efficient-ML / Awesome-Efficient-AIGC

#大语言模型#A list of papers, docs, codes about efficient AIGC. This repo is aimed to provide the info for efficient AIGC research, including language and vision, we are continuously improving the project. Welcom...

aigcdiffusion-modelsdistillationefficient-deep-learninggenerative-modellarge-language-models大语言模型model-compressionmodel-quantizationpruningAwesome Lists
186
6 个月前
https://static.github-zh.com/github_avatars/sayakpaul?size=40
sayakpaul / Adventures-in-TensorFlow-Lite

This repository contains notebooks that show the usage of TensorFlow Lite for quantizing deep neural networks.

tensorflow-2tensorflow-liteon-device-mlmodel-quantizationpost-training-quantizationquantization-aware-trainingpruninginference
Jupyter Notebook 171
3 年前
https://static.github-zh.com/github_avatars/RodolfoFerro?size=40
RodolfoFerro / psychopathology-fer-assistant

[WINNER! 🏆] Psychopathology FER Assistant. Because mental health matters. My project submission for #TFWorld TF 2.0 Challenge at Devpost.

Python树莓派google-colabmodel-quantizationTensorflowtflitefirebase-realtime-databaseFlaskdash
Jupyter Notebook 77
2 年前
https://static.github-zh.com/github_avatars/htqin?size=40
htqin / BiBench

[ICML 2023] This project is the official implementation of our accepted ICML 2023 paper BiBench: Benchmarking and Analyzing Network Binarization.

benchmarkmodel-compressionbinary-neural-networksmodel-quantization
Python 56
1 年前
https://static.github-zh.com/github_avatars/htqin?size=40
htqin / QuantSR

[NeurIPS 2023 Spotlight] This project is the official implementation of our accepted NeurIPS 2023 (spotlight) paper QuantSR: Accurate Low-bit Quantization for Efficient Image Super-Resolution.

model-quantizationsuper-resolution
Python 50
1 年前
https://static.github-zh.com/github_avatars/nbasyl?size=40
nbasyl / OFQ

The official implementation of the ICML 2023 paper OFQ-ViT

icmlmodel-compressionmodel-quantizationvision-transformervision-transformers
Python 33
2 年前
https://static.github-zh.com/github_avatars/seonglae?size=40
seonglae / llama2gptq

#大语言模型#Chat to LLaMa 2 that also provides responses with reference documents over vector database. Locally available model using GPTQ 4bit quantization.

langchainquantizationtransformersmodel-quantizationCUDA聊天机器人question-answeringChatGPTgptllama-2llama2
Python 30
2 年前
https://static.github-zh.com/github_avatars/HaoranREN?size=40
HaoranREN / TensorFlow_Model_Quantization

#计算机科学#A tutorial of model quantization using TensorFlow

model-quantizationTensorflowtensorflow-litetflite机器学习quantization-aware-training
Python 12
4 年前
https://static.github-zh.com/github_avatars/cantbebetter2?size=40
cantbebetter2 / Awesome-Diffusion-Quantization

A list of papers, docs, codes about diffusion quantization.This repo collects various quantization methods for the Diffusion Models. Welcome to PR the works (papers, repositories) missed by the repo.

Awesome Listsdiffusion-modelsmodel-compressionmodel-quantization
7
13 天前
https://static.github-zh.com/github_avatars/frickyinn?size=40
frickyinn / BiDense

PyTorch implementation of "BiDense: Binarization for Dense Prediction," A binary neural network for dense prediction tasks.

model-compressionmodel-quantization
Python 6
8 个月前
https://static.github-zh.com/github_avatars/dcarpintero?size=40
dcarpintero / ai-engineering

AI Engineering: Annotated NBs to dive into Self-Attention, In-Context Learning, RAG, Knowledge-Graphs, Fine-Tuning, Model Optimization, and many more.

bertchunkingembeddingsfine-tuninggenerative-aihuggingface-transformersin-context-learningknowledge-graphlangchainlarge-language-modelsllama3-1model-quantizationretrieval-augmented-generationself-attentiontransformerweights-and-biasesai-engineering
Jupyter Notebook 6
4 个月前
https://static.github-zh.com/github_avatars/NANEXLABS?size=40
NANEXLABS / Nanex-AI

Enterprise multi-agent framework for secure, borderless data collaboration with zero-trust and federated learning-lightweight edge-ready.

人工智能aiagentedge-computingfederated-learninggrpc-webiot-securitymodel-quantizationmqtt-protocolonnx-runtimetensorflow-litezero-trust-security
Python 5
4 个月前
https://static.github-zh.com/github_avatars/medoidai?size=40
medoidai / model-quantization-blog-notebooks

#计算机科学#Notebook from "A Hands-On Walkthrough on Model Quantization" blog post.

人工智能深度学习机器学习model-quantization
Jupyter Notebook 4
1 年前
https://static.github-zh.com/github_avatars/SRDdev?size=40
SRDdev / Model-Quantization

Quantization is a technique to reduce the computational and memory costs of running inference by representing the weights and activations with low-precision data types like 8-bit integer (int8) instea...

机器学习model-quantizationquantization
Jupyter Notebook 4
2 年前
https://static.github-zh.com/github_avatars/BjornMelin?size=40
BjornMelin / local-llm-workbench

🧠 A comprehensive toolkit for benchmarking, optimizing, and deploying local Large Language Models. Includes performance testing tools, optimized configurations for CPU/GPU/hybrid setups, and detailed...

cpu-inferenceCUDAgpu-accelerationinference-optimizationllama-cpplocal-llmmodel-managementmodel-quantization
Shell 2
4 个月前
https://static.github-zh.com/github_avatars/dwain-barnes?size=40
dwain-barnes / LLM-GGUF-Auto-Converter

#大语言模型#Automated Jupyter notebook solution for batch converting Large Language Models to GGUF format with multiple quantization options. Built on llama.cpp with HuggingFace integration.

batch-processingCUDAggufhuggingfaceJupyter Notebookllama-cpp大语言模型model-quantization
Jupyter Notebook 2
6 个月前
loading...