GitHub 中文社区
回车: Github搜索    Shift+回车: Google搜索
论坛
排行榜
趋势
登录

©2025 GitHub中文社区论坛GitHub官网网站地图GitHub官方翻译

  • X iconGitHub on X
  • Facebook iconGitHub on Facebook
  • Linkedin iconGitHub on LinkedIn
  • YouTube iconGitHub on YouTube
  • Twitch iconGitHub on Twitch
  • TikTok iconGitHub on TikTok
  • GitHub markGitHub’s organization on GitHub
集合主题趋势排行榜
#

triton-inference-server

Website
Wikipedia
https://static.github-zh.com/github_avatars/NVIDIA?size=40
NVIDIA / GenerativeAIExamples

#大语言模型#Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

gpu-accelerationlarge-language-models大语言模型llm-inference微服务nemoragretrieval-augmented-generationtensorrttriton-inference-server
Jupyter Notebook 3.18 k
5 天前
https://static.github-zh.com/github_avatars/CoinCheung?size=40
CoinCheung / BiSeNet

Add bisenetv2. My implementation of BiSeNet

cityscapesPyTorchtensorrtncnnopenvinotriton-inference-serverade20k
Python 1.52 k
6 个月前
https://static.github-zh.com/github_avatars/isarsoft?size=40
isarsoft / yolov4-triton-tensorrt

#计算机科学#This repository deploys YOLOv4 as an optimized TensorRT engine to Triton Inference Server

triton-inference-serverobject-detectiontensorrtyolov4yolov4-tinyDocker深度学习
C++ 285
3 年前
https://static.github-zh.com/github_avatars/npuichigo?size=40
npuichigo / openai_trtllm

#大语言模型#OpenAI compatible API for TensorRT LLM triton backend

langchain大语言模型openai-apitensorrt-llmtriton-inference-server
Rust 209
10 个月前
https://static.github-zh.com/github_avatars/NetEase-Media?size=40
NetEase-Media / grps

Deep Learning Deployment Framework: Supports tf/torch/trt/trtllm/vllm and other NN frameworks. Support dynamic batching, and streaming modes. It is dual-language compatible with Python and C++, offeri...

Tensorflowtensorrttorchvllmservingtriton-inference-servertensorrt-llm
C++ 165
1 个月前
https://static.github-zh.com/github_avatars/torchpipe?size=40
torchpipe / torchpipe

Serving Inside Pytorch

部署inferencepipeline-parallelismservingtensorrttriton-inference-serverrayPyTorchtorch2trtservellm-serving
C++ 160
7 天前
https://static.github-zh.com/github_avatars/triton-inference-server?size=40
triton-inference-server / onnxruntime_backend

The Triton backend for the ONNX Runtime.

triton-inference-server后端onnx-runtimeinference
C++ 150
4 天前
https://static.github-zh.com/github_avatars/clearml?size=40
clearml / clearml-serving

#计算机科学#ClearML - Model-Serving Orchestration and Repository Solution

机器学习mlopsDevOps深度学习Kubernetes人工智能model-servingservingtritontriton-inference-server
Python 150
5 个月前
https://static.github-zh.com/github_avatars/kamalkraj?size=40
kamalkraj / stable-diffusion-tritonserver

#计算机科学#Deploy stable diffusion model with onnx/tenorrt + tritonserver

DockerNvidiastablediffusiontransformersdeployonnxPythontriton-inference-serverinference机器学习PyTorchtensorrt
Jupyter Notebook 123
2 年前
https://static.github-zh.com/github_avatars/NVIDIA-ISAAC-ROS?size=40
NVIDIA-ISAAC-ROS / isaac_ros_dnn_inference

#计算机科学#NVIDIA-accelerated DNN model inference ROS 2 packages using NVIDIA Triton/TensorRT for both Jetson and x86_64 with CUDA-capable GPU

rosdnntensorrttritontriton-inference-server深度学习Nvidia人工智能ros2-humbleros2gpujetson
C++ 115
4 个月前
https://static.github-zh.com/github_avatars/notAI-tech?size=40
notAI-tech / fastDeploy

#计算机科学#Deploy DL/ ML inference pipelines with minimal extra code.

深度学习PyTorchservingfalcongeventDockermodel-deploymentmodel-servinghttp-servergunicorntriton-inference-serverPythontritoninference-serverstreaming-audioWebSocket
Python 98
7 个月前
https://static.github-zh.com/github_avatars/Koldim2001?size=40
Koldim2001 / TrafficAnalyzer

Анализ трафика на круговом движении с использованием компьютерного зрения

bytetrackDocker ComposeGrafanagrafana-dashboardhydramultiprocessingoop-principlesPostgreSQLtraffic-analysisyolov8multiple-object-trackingobject-detectionFlasknginxtriton-inference-serverinfluxdbDocker
Python 90
3 个月前
https://static.github-zh.com/github_avatars/trinhtuanvubk?size=40
trinhtuanvubk / Diff-VC

Diffusion Model for Voice Conversion

diffusion-modelsgradiotriton-inference-servervoice-conversion
Jupyter Notebook 55
1 年前
https://static.github-zh.com/github_avatars/bug-developer021?size=40
bug-developer021 / YOLOV5_optimization_on_triton

Compare multiple optimization methods on triton to imporve model service performance

gpuinferencetensorrttriton-inference-serveryolov5
Jupyter Notebook 51
1 年前
https://static.github-zh.com/github_avatars/akiragy?size=40
akiragy / recsys_pipeline

Build Recommender System with PyTorch + Redis + Elasticsearch + Feast + Triton + Flask. Vector Recall, DeepFM Ranking and Web Application.

elasticsearchPythonPyTorchrecommender-systemRedisinverted-indexrankingretrievalrecommendationtriton-inference-servervector-databaseFlask
Python 50
2 年前
https://static.github-zh.com/github_avatars/rtzr?size=40
rtzr / tritony

Tiny configuration for Triton Inference Server

inferencemlopstriton-inference-server
Python 45
5 个月前
https://static.github-zh.com/github_avatars/chiehpower?size=40
chiehpower / Setup-deeplearning-tools

#计算机科学#Set up CI in DL/ cuda/ cudnn/ TensorRT/ onnx2trt/ onnxruntime/ onnxsim/ Pytorch/ Triton-Inference-Server/ Bazel/ Tesseract/ PaddleOCR/ NVIDIA-docker/ minIO/ Supervisord on AGX or PC from scratch.

tensorrtCUDAcudnninstallationonnxruntimePyTorchtesseract-ocrtriton-inference-serverNvidia深度学习持续集成Dockerpaddleocrminio
Python 43
2 年前
https://static.github-zh.com/github_avatars/omarabid59?size=40
omarabid59 / yolov8-triton

Provides an ensemble model to deploy a YoloV8 ONNX model to Triton

部署triton-inference-serverultralyticsyolov8
Python 39
2 年前
https://static.github-zh.com/github_avatars/k9ele7en?size=40
k9ele7en / Triton-TensorRT-Inference-CRAFT-pytorch

Advanced inference pipeline using NVIDIA Triton Inference Server for CRAFT Text detection (Pytorch), included converter from Pytorch -> ONNX -> TensorRT, Inference pipelines (TensorRT, Triton server -...

triton-inference-servertensorrtonnxPyTorchnvidia-dockerinference-engineinference-serverinferencetext-detection
Python 33
4 年前
https://static.github-zh.com/github_avatars/Bobo-y?size=40
Bobo-y / triton_ensemble_model_demo

triton server ensemble model demo

triton-inference-serverpipeline
Python 30
3 年前
loading...