Awesome Open Source
Search
Programming Languages
Languages
All Categories
Categories
About
Search results for triton inference server
triton-inference-server
x
24 search results found
Bisenet
⭐
1,130
Add bisenetv2. My implementation of BiSeNet
Generativeaiexamples
⭐
458
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
Yolov4 Triton Tensorrt
⭐
184
This repository deploys YOLOv4 as an optimized TensorRT engine to Triton Inference Server
Clearml Serving
⭐
115
ClearML - Model-Serving Orchestration and Repository Solution
Onnxruntime_backend
⭐
103
The Triton backend for the ONNX Runtime.
Torchpipe
⭐
91
Boosting DL Service Throughput 1.5-4x by Ensemble Pipeline Serving with Concurrent CUDA Streams for PyTorch/LibTorch Frontend and TensorRT/CVCUDA, etc., Backends
Fastdeploy
⭐
90
Deploy DL/ ML inference pipelines with minimal extra code.
Isaac_ros_dnn_inference
⭐
88
Hardware-accelerated DNN model inference ROS 2 packages using NVIDIA Triton/TensorRT for both Jetson and x86_64 with CUDA-capable GPU
Stable Diffusion Tritonserver
⭐
49
Deploy stable diffusion model with onnx/tenorrt + tritonserver
Yolov5_optimization_on_triton
⭐
41
Compare multiple optimization methods on triton to imporve model service performance
Setup Deeplearning Tools
⭐
41
Set up CI in DL/ cuda/ cudnn/ TensorRT/ onnx2trt/ onnxruntime/ onnxsim/ Pytorch/ Triton-Inference-Server/ Bazel/ Tesseract/ PaddleOCR/ NVIDIA-docker/ minIO/ Supervisord on AGX or PC from scratch.
Openai_trtllm
⭐
34
OpenAI compatible API for TensorRT LLM triton backend
Tritony
⭐
32
Tiny configuration for Triton Inference Server
Recsys_pipeline
⭐
16
Build Recommender System with PyTorch + Redis + Elasticsearch + Feast + Triton + Flask. Vector Recall, DeepFM Ranking and Web Application.
Triton_ensemble_model_demo
⭐
15
triton server ensemble model demo
Serving Compare Middleware
⭐
12
FastAPI middleware for comparing different ML model serving approaches
Yolov8 Triton
⭐
11
Provides an ensemble model to deploy a YoloV8 ONNX model to Triton
Tensorrt Triton Magface
⭐
10
Magface Triton Inferece Server Using Tensorrt
Serving Codegen Gptj Triton
⭐
9
Serving Example of CodeGen-350M-Mono-GPTJ on Triton Inference Server with Docker and Kubernetes
Novelist Triton Server
⭐
8
Deploy KoGPT with Triton Inference Server
Diff Vc
⭐
7
Diffusion Model for Voice Conversion
Redis Feast Gcp
⭐
5
A demo of Redis Enterprise as the Online Feature Store deployed on GCP with Feast and NVIDIA Triton Inference Server.
Armnn_tflite_backend
⭐
5
TensorFlow Lite backend with ArmNN delegate support for Nvidia Triton
Openvino Model Server Wrapper
⭐
5
Python wrapper class for OpenVINO Model Server. User can submit inference request to OVMS with just a few lines of code.
Related Searches
Python Triton Inference Server (12)
Pytorch Triton Inference Server (11)
Deep Learning Triton Inference Server (8)
Tensorrt Triton Inference Server (7)
C Plus Plus Triton Inference Server (7)
Jupyter Notebook Triton Inference Server (6)
Machine Learning Triton Inference Server (6)
Docker Triton Inference Server (5)
Tensorflow Serving Triton Inference Server (4)
Onnx Triton Inference Server (4)
1-24 of 24 search results
Privacy
|
About
|
Terms
|
Follow Us On Twitter
Copyright 2018-2024 Awesome Open Source. All rights reserved.