GitHub 中文社区
回车: Github搜索    Shift+回车: Google搜索
论坛
排行榜
趋势
登录

©2025 GitHub中文社区论坛GitHub官网网站地图GitHub官方翻译

  • X iconGitHub on X
  • Facebook iconGitHub on Facebook
  • Linkedin iconGitHub on LinkedIn
  • YouTube iconGitHub on YouTube
  • Twitch iconGitHub on Twitch
  • TikTok iconGitHub on TikTok
  • GitHub markGitHub’s organization on GitHub
集合主题趋势排行榜
#

mixture-of-experts

Website
Wikipedia
https://static.github-zh.com/github_avatars/deepspeedai?size=40
deepspeedai / DeepSpeed

#计算机科学#DeepSpeed Chat: 一键式RLHF训练,让你的类ChatGPT千亿大模型提速省钱15倍

深度学习PyTorchgpu机器学习billion-parametersdata-parallelismmodel-parallelisminferencepipeline-parallelismcompressionmixture-of-expertstrillion-parameterszero
Python 38.87 k
16 小时前
https://static.github-zh.com/github_avatars/codelion?size=40
codelion / optillm

#大语言模型#Optimizing inference proxy for LLMs

agentagentic-aiagentic-workflowagentsAPIgenailarge-language-models大语言模型llm-inferencellmapimixture-of-expertsmoaopenaiopenai-apioptimizationproxy-serveragentic-frameworkchain-of-thoughtmonte-carlo-tree-searchprompt-engineering
Python 2.52 k
2 天前
https://static.github-zh.com/github_avatars/dvmazur?size=40
dvmazur / mixtral-offloading

#大语言模型#Run Mixtral-8x7B models in Colab or consumer desktops

colab-notebook深度学习google-colablanguage-model大语言模型mixture-of-expertsoffloadingPyTorchquantization
Python 2.31 k
1 年前
https://static.github-zh.com/github_avatars/learning-at-home?size=40
learning-at-home / hivemind

#计算机科学#Decentralized deep learning in PyTorch. Built to train models on thousands of volunteers across the world.

深度学习PyTorchvolunteer-computingmixture-of-expertsdistributed-trainingdistributed-systemsasynchronous-programmingasynciodhtneural-networks机器学习
Python 2.21 k
1 个月前
https://static.github-zh.com/github_avatars/PKU-YuanGroup?size=40
PKU-YuanGroup / MoE-LLaVA

Mixture-of-Experts for Large Vision-Language Models

large-vision-language-modelmixture-of-expertsmoemulti-modal
Python 2.18 k
6 个月前
https://static.github-zh.com/github_avatars/davidmrau?size=40
davidmrau / mixture-of-experts

PyTorch Re-Implementation of "The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer et al. https://arxiv.org/abs/1701.06538

moemixture-of-expertsPyTorch
Python 1.12 k
1 年前
https://static.github-zh.com/github_avatars/rhymes-ai?size=40
rhymes-ai / Aria

Codebase for Aria - an Open Multimodal Native MoE

mixture-of-expertsmultimodalvision-and-language
Jupyter Notebook 1.05 k
5 个月前
https://static.github-zh.com/github_avatars/pjlab-sys4nlp?size=40
pjlab-sys4nlp / llama-moe

#大语言模型#⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training (EMNLP 2024)

llama大语言模型mixture-of-expertsmoe
Python 967
6 个月前
https://static.github-zh.com/github_avatars/microsoft?size=40
microsoft / Tutel

#大语言模型#Tutel MoE: Optimized Mixture-of-Experts Library, Support DeepSeek FP8/FP4

PyTorchmoemixture-of-expertsdeepseek大语言模型
C 839
8 天前
https://static.github-zh.com/github_avatars/SMTorg?size=40
SMTorg / smt

#计算机科学#Surrogate Modeling Toolbox

derivativesamplingmixture-of-expertspredictive-modeling机器学习
Jupyter Notebook 773
10 天前
https://static.github-zh.com/github_avatars/lucidrains?size=40
lucidrains / mixture-of-experts

#计算机科学#A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models

人工智能深度学习transformermixture-of-experts
Python 763
2 年前
https://static.github-zh.com/github_avatars/AviSoori1x?size=40
AviSoori1x / makeMoE

#大语言模型#From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)

large-language-models大语言模型mixture-of-experts深度学习neural-networksPyTorchpytorch-implementation
Jupyter Notebook 719
8 个月前
https://static.github-zh.com/github_avatars/drawbridge?size=40
drawbridge / keras-mmoe

#计算机科学#A TensorFlow Keras implementation of "Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts" (KDD 2018)

机器学习深度学习数据科学深度神经网络KerasTensorflowmulti-task-learningmixture-of-experts
Python 713
2 年前
https://static.github-zh.com/github_avatars/ymcui?size=40
ymcui / Chinese-Mixtral

#自然语言处理#中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs)

large-language-models大语言模型mixtralmixture-of-expertsmoe自然语言处理
Python 605
1 年前
https://static.github-zh.com/github_avatars/Leeroo-AI?size=40
Leeroo-AI / mergoo

#自然语言处理#A library for easily merging multiple LLM experts, and efficiently train the merged LLM.

generative-ai大语言模型mergemixture-of-experts自然语言处理fine-tuninglarge-language-modelslora人工智能transformersmulti-modelOpen Source
Python 480
10 个月前
https://static.github-zh.com/github_avatars/lucidrains?size=40
lucidrains / st-moe-pytorch

#计算机科学#Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch

人工智能深度学习mixture-of-experts
Python 340
1 年前
https://static.github-zh.com/github_avatars/lucidrains?size=40
lucidrains / soft-moe-pytorch

#计算机科学#Implementation of Soft MoE, proposed by Brain's Vision team, in Pytorch

人工智能深度学习mixture-of-expertstransformers
Python 297
2 个月前
https://static.github-zh.com/github_avatars/Luodian?size=40
Luodian / Generalizable-Mixture-of-Experts

#计算机科学#GMoE could be the next backbone model for many kinds of generalization task.

深度学习domain-generalizationPyTorchpytorch-implementationmixture-of-experts
Python 272
2 年前
https://static.github-zh.com/github_avatars/SkyworkAI?size=40
SkyworkAI / MoH

MoH: Multi-Head Attention as Mixture-of-Head Attention

attentiondit大语言模型mixture-of-expertsmoetransformervit
Python 251
8 个月前
https://static.github-zh.com/github_avatars/inferflow?size=40
inferflow / inferflow

Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).

llama2llamacppllm-inferencemodel-quantizationmulti-gpu-inferencemixture-of-expertsmoegemmafalconminicpmmistralbloomdeepseekinternlmbaichuan2mixtralqwen
C++ 242
1 年前
loading...