GitHub 中文社区
回车: Github搜索    Shift+回车: Google搜索
论坛
排行榜
趋势
登录

©2025 GitHub中文社区论坛GitHub官网网站地图GitHub官方翻译

  • X iconGitHub on X
  • Facebook iconGitHub on Facebook
  • Linkedin iconGitHub on LinkedIn
  • YouTube iconGitHub on YouTube
  • Twitch iconGitHub on Twitch
  • TikTok iconGitHub on TikTok
  • GitHub markGitHub’s organization on GitHub
集合主题趋势排行榜
#

inference-server

Website
Wikipedia
https://static.github-zh.com/github_avatars/containers?size=40
containers / ramalama

#大语言模型#RamaLama is an open-source developer tool that simplifies the local serving of AI models from any source and facilitates their use for inference in production, all through the familiar language of con...

人工智能containersCUDADockerhipinference-serverintelllamacpp大语言模型podmanvllm
Python 1.76 k
5 天前
roboflow/inference
https://static.github-zh.com/github_avatars/roboflow?size=40
roboflow / inference

#计算机科学#Turn any computer or edge device into a command center for your computer vision projects.

机器视觉inference-apiinference-servervityolov5yolov8jetsontensorrtclassificationinstance-segmentationobject-detectiononnx部署Dockerinference机器学习Pythonyolo11agents
Python 1.74 k
2 天前
https://static.github-zh.com/github_avatars/basetenlabs?size=40
basetenlabs / truss

#计算机科学#The simplest way to serve AI/ML models in production

机器学习人工智能easy-to-useinference-apiinference-servermodel-servingOpen Sourcepackagingfalconstable-diffusionWhisperwizardlm
Python 1 k
5 天前
https://static.github-zh.com/github_avatars/pipeless-ai?size=40
pipeless-ai / pipeless

#计算机科学#An open-source computer vision framework to build and deploy apps in minutes

人工智能机器视觉multimediamultimedia-applicationscloud深度学习机器学习object-detectionVideoyoloFFmpeggstreamerinference-serverPythonvision-frameworkinferenceperceptionpipeline-frameworkstream-processingvideo-processing
Rust 756
1 年前
https://static.github-zh.com/github_avatars/underneathall?size=40
underneathall / pinferencia

#自然语言处理#Python + Inference - Model Deployment library in Python. Simplest model inference server ever.

人工智能inference-serverpredictinference深度学习机器学习Pythonservingmodel-deploymenthuggingfacePyTorchTensorflowtransformers数据科学model-serving机器视觉自然语言处理paddlepaddle
Python 556
2 年前
https://static.github-zh.com/github_avatars/NVIDIA?size=40
NVIDIA / gpu-rest-engine

#计算机科学#A REST API for Caffe using Docker and Go

caffegpuinferenceinference-serverDocker深度学习
C++ 419
7 年前
https://static.github-zh.com/github_avatars/BMW-InnovationLab?size=40
BMW-InnovationLab / BMW-YOLOv4-Inference-API-GPU

#计算机科学#This is a repository for an nocode object detection inference API using the Yolov3 and Yolov4 Darknet framework.

yolov3inferencegpuAPI深度学习机器视觉bounding-boxesinference-serverDockerREST APIyolo神经网络Dockerfileyolov4无代码
Python 280
3 年前
https://static.github-zh.com/github_avatars/containers?size=40
containers / podman-desktop-extension-ai-lab

Work with LLMs on a local environment using containers

人工智能containersinference-server大语言模型localpodman
TypeScript 229
5 天前
https://static.github-zh.com/github_avatars/BMW-InnovationLab?size=40
BMW-InnovationLab / BMW-YOLOv4-Inference-API-CPU

#计算机科学#This is a repository for an nocode object detection inference API using the Yolov4 and Yolov3 Opencv.

yolov3inferenceAPIcpu深度学习机器视觉OpenCVobject-detectionDocker深度神经网络神经网络REST APIinference-serverbounding-boxesyolov4无代码
Python 220
3 年前
https://static.github-zh.com/github_avatars/BMW-InnovationLab?size=40
BMW-InnovationLab / BMW-TensorFlow-Inference-API-CPU

#计算机科学#This is a repository for an object detection inference API using the Tensorflow framework.

TensorflowinferenceAPIcpu深度学习object-detection机器视觉Dockerbounding-boxesDocker Imagedocker-ceinference-engineinference-serverREST API
Python 183
3 年前
https://static.github-zh.com/github_avatars/kibae?size=40
kibae / onnxruntime-server

#计算机科学#ONNX Runtime Server: The ONNX Runtime Server is a server that provides TCP and HTTP/HTTPS REST APIs for ONNX inference.

人工智能机器学习onnxonnxruntime深度学习inference-servernueral-networksCUDAcontributions-welcome
C++ 158
1 个月前
https://static.github-zh.com/github_avatars/autodeployai?size=40
autodeployai / ai-serving

Serving AI/ML models in the open standard formats PMML and ONNX with both HTTP (REST API) and gRPC endpoints

onnxinference-serveronnx-modelsinference
Scala 158
8 个月前
https://static.github-zh.com/github_avatars/vertexclique?size=40
vertexclique / orkhon

#计算机科学#Orkhon: ML Inference Framework and Server Runtime

inference-server机器学习PythonTensorflowasyncmultiprocessingdata-parallelism
Rust 149
4 年前
https://static.github-zh.com/github_avatars/kf5i?size=40
kf5i / k3ai

K3ai is a lightweight, fully automated, AI infrastructure-in-a-box solution that allows anyone to experiment quickly with Kubeflow pipelines. K3ai is perfect for anything from Edge to laptops.

kubeflow-pipelinesKubernetesk3s机器学习datascience人工智能Edgekubeflowinference-server
PowerShell 101
4 年前
https://static.github-zh.com/github_avatars/notAI-tech?size=40
notAI-tech / fastDeploy

#计算机科学#Deploy DL/ ML inference pipelines with minimal extra code.

深度学习PyTorchservingfalcongeventDockermodel-deploymentmodel-servinghttp-servergunicorntriton-inference-serverPythontritoninference-serverstreaming-audioWebSocket
Python 98
7 个月前
https://static.github-zh.com/github_avatars/RubixML?size=40
RubixML / Server

#计算机科学#A standalone inference server for trained Rubix ML estimators.

机器学习http-serverinfrastructureAPImodel-deployment微服务JSON:APIPHPREST APIinferenceinference-engineml-infrastructureinference-server
PHP 62
3 个月前
https://static.github-zh.com/github_avatars/friendliai?size=40
friendliai / friendli-client

#大语言模型#Friendli: the fastest serving engine for generative AI

generative-ai大语言模型llm-inferencellmopsservinggptgpt3inferencellama2llm-servinginference-engineinference-server人工智能llm-opsmistral机器学习mlopsstable-diffusion
Python 47
5 个月前
https://static.github-zh.com/github_avatars/curtisgray?size=40
curtisgray / wingman

#下载器#Wingman is the fastest and easiest way to run Llama models on your PC or Mac.

人工智能聊天机器人ChatGPTLinuxllamallamacpp大语言模型localmacOSWindowsdownload下载器openaigpugpu-accelerationgpu-monitoringinferenceinference-engineinference-server
TypeScript 41
1 年前
https://static.github-zh.com/github_avatars/k9ele7en?size=40
k9ele7en / Triton-TensorRT-Inference-CRAFT-pytorch

Advanced inference pipeline using NVIDIA Triton Inference Server for CRAFT Text detection (Pytorch), included converter from Pytorch -> ONNX -> TensorRT, Inference pipelines (TensorRT, Triton server -...

triton-inference-servertensorrtonnxPyTorchnvidia-dockerinference-engineinference-serverinferencetext-detection
Python 33
4 年前
https://static.github-zh.com/github_avatars/haicheviet?size=40
haicheviet / fullstack-machine-learning-inference

#计算机科学#Fullstack machine learning inference template

Amazon Web ServicescloudformationFastAPIfull-stackinference-serverInfrastructure as code机器学习
Jupyter Notebook 30
2 年前
loading...