{"payload":{"pageCount":2,"repositories":[{"type":"Public","name":"tutorials","owner":"triton-inference-server","isFork":false,"description":"This repository contains tutorials and examples for Triton Inference Server","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":14,"issueCount":8,"starsCount":528,"forksCount":90,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,4,2,0,2,2,3,1,1,0,0,1,0,1,0,0,2,0,0,0,0,0,0,2,0,0,0,0,0,0,1,0,2,0,0,1,0,1,0,1,1,0,0,2,1,2,1,3,1,1,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-20T00:54:49.579Z"}},{"type":"Public","name":"perf_analyzer","owner":"triton-inference-server","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":8,"issueCount":4,"starsCount":12,"forksCount":2,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[4,3,4,2,1,1,2,0,2,2,2,2,1,0,0,1,0,0,0,0,0,0,4,19,34,12,19,5,6,4,11,9,10,12,6,7,5,10,4,4,10,5,5,8,5,5,4,3,13,7,3,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-20T00:25:26.507Z"}},{"type":"Public","name":"server","owner":"triton-inference-server","isFork":false,"description":"The Triton Inference Server provides an optimized cloud and edge inferencing solution. ","allTopics":["machine-learning","cloud","deep-learning","gpu","inference","edge","datacenter"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":57,"issueCount":537,"starsCount":8071,"forksCount":1451,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[2,4,13,10,6,9,12,11,11,1,4,9,5,2,4,7,6,7,8,2,6,3,3,5,9,5,3,3,14,7,7,3,6,6,7,9,10,6,2,1,2,8,2,2,6,6,1,7,6,3,3,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-20T00:15:35.594Z"}},{"type":"Public","name":"fil_backend","owner":"triton-inference-server","isFork":false,"description":"FIL backend for the Triton Inference Server","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":3,"issueCount":49,"starsCount":68,"forksCount":35,"license":"Apache License 2.0","participation":[0,0,2,0,0,0,1,0,0,0,0,0,0,0,2,3,1,0,0,1,0,0,0,0,1,0,0,2,0,0,0,0,2,0,0,2,0,0,0,1,1,1,0,0,1,0,0,0,0,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-20T00:08:34.495Z"}},{"type":"Public","name":"vllm_backend","owner":"triton-inference-server","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":6,"issueCount":0,"starsCount":170,"forksCount":16,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,8,21,1,2,1,2,0,1,0,1,0,1,0,0,0,1,0,0,1,1,0,0,3,0,0,0,0,2,1,0,2,0,0,0,2,1,0,0,0,0,1,0,1,1,2,0,2,1,1,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T22:38:28.835Z"}},{"type":"Public","name":"onnxruntime_backend","owner":"triton-inference-server","isFork":false,"description":"The Triton backend for the ONNX Runtime.","allTopics":["inference","backend","triton-inference-server","onnx-runtime"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":4,"issueCount":69,"starsCount":122,"forksCount":54,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[1,1,1,0,0,0,0,0,1,0,0,3,1,1,0,1,0,0,1,1,1,0,0,1,1,1,0,0,0,0,0,0,0,0,0,2,1,0,0,0,0,0,1,0,1,0,0,0,1,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T16:25:51.663Z"}},{"type":"Public","name":"triton_cli","owner":"triton-inference-server","isFork":false,"description":"Triton CLI is an open source command line interface that enables users to create, deploy, and profile models served by the Triton Inference Server.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":48,"forksCount":2,"license":null,"participation":[0,0,0,1,0,1,0,0,1,0,0,5,5,0,0,5,3,5,1,3,0,0,0,6,2,1,1,0,0,1,0,0,3,3,2,1,3,2,0,0,1,0,2,0,1,1,0,1,0,1,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-18T19:41:03.288Z"}},{"type":"Public","name":"python_backend","owner":"triton-inference-server","isFork":false,"description":"Triton backend that enables pre-process, post-processing and other logic to be implemented in Python.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":11,"issueCount":0,"starsCount":523,"forksCount":142,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[1,0,6,0,4,2,2,0,1,0,1,1,3,0,0,3,1,1,1,1,0,0,0,2,2,1,2,0,2,2,0,0,1,0,0,0,1,2,1,0,0,0,0,0,2,2,0,1,1,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-18T12:54:32.050Z"}},{"type":"Public","name":"core","owner":"triton-inference-server","isFork":false,"description":"The core library and APIs implementing the Triton Inference Server. ","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":18,"issueCount":0,"starsCount":102,"forksCount":97,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,1,2,1,1,3,4,2,3,0,1,3,0,0,1,4,4,4,3,0,0,2,1,1,2,1,0,1,1,0,1,1,4,1,3,0,6,4,0,0,2,2,3,1,1,0,0,4,0,2,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-18T12:51:02.957Z"}},{"type":"Public","name":"client","owner":"triton-inference-server","isFork":false,"description":"Triton Python, C++ and Java client libraries, and GRPC-generated client examples for go, java and scala.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":25,"issueCount":24,"starsCount":547,"forksCount":225,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[4,4,6,4,3,3,2,2,3,3,2,2,1,0,1,5,3,0,1,0,1,0,5,20,35,12,19,5,7,8,12,9,13,12,7,7,5,10,4,4,10,5,5,8,2,2,0,1,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-18T12:50:04.961Z"}},{"type":"Public","name":"backend","owner":"triton-inference-server","isFork":false,"description":"Common source, scripts and utilities for creating Triton backends.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":2,"issueCount":0,"starsCount":280,"forksCount":84,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,2,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,2,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-18T12:49:56.597Z"}},{"type":"Public","name":"tensorrtllm_backend","owner":"triton-inference-server","isFork":false,"description":"The Triton TensorRT-LLM Backend","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":14,"issueCount":246,"starsCount":656,"forksCount":93,"license":"Apache License 2.0","participation":[4,0,4,8,0,2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,0,3,1,1,1,1,1,1,1,1,1,1,1,1,1,1,2,1,1,1,1,1,1,3,3,1,1,1,1,1,1,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-18T01:26:21.090Z"}},{"type":"Public","name":"developer_tools","owner":"triton-inference-server","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":5,"issueCount":0,"starsCount":16,"forksCount":10,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-16T15:30:05.738Z"}},{"type":"Public","name":"tensorflow_backend","owner":"triton-inference-server","isFork":false,"description":"The Triton backend for TensorFlow.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":2,"issueCount":0,"starsCount":44,"forksCount":19,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-11T23:30:38.270Z"}},{"type":"Public","name":"tensorrt_backend","owner":"triton-inference-server","isFork":false,"description":"The Triton backend for TensorRT.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":1,"issueCount":0,"starsCount":59,"forksCount":28,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,1,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,1,0,0,0,1,1,1,2,0,1,0,0,0,1,0,0,0,1,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-11T23:28:37.264Z"}},{"type":"Public","name":"pytorch_backend","owner":"triton-inference-server","isFork":false,"description":"The Triton backend for the PyTorch TorchScript models.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":3,"issueCount":0,"starsCount":118,"forksCount":43,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-11T23:28:07.563Z"}},{"type":"Public","name":"dali_backend","owner":"triton-inference-server","isFork":false,"description":"The Triton backend that allows running GPU-accelerated data pre-processing pipelines implemented in DALI's python API.","allTopics":["python","deep-learning","gpu","image-processing","dali","data-preprocessing","nvidia-dali","fast-data-pipeline"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":5,"issueCount":20,"starsCount":122,"forksCount":28,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-10T16:14:25.834Z"}},{"type":"Public","name":"model_navigator","owner":"triton-inference-server","isFork":false,"description":"Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.","allTopics":["deep-learning","gpu","inference"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":3,"starsCount":178,"forksCount":25,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-10T12:10:37.722Z"}},{"type":"Public","name":"third_party","owner":"triton-inference-server","isFork":false,"description":"Third-party source packages that are modified for use in Triton.","allTopics":[],"primaryLanguage":{"name":"C","color":"#555555"},"pullRequestCount":5,"issueCount":0,"starsCount":6,"forksCount":50,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-07T00:32:35.164Z"}},{"type":"Public","name":"square_backend","owner":"triton-inference-server","isFork":false,"description":"Simple Triton backend used for testing.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":4,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-07T00:28:38.150Z"}},{"type":"Public","name":"repeat_backend","owner":"triton-inference-server","isFork":false,"description":"An example Triton backend that demonstrates sending zero, one, or multiple responses for each request. ","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":7,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-07T00:28:36.795Z"}},{"type":"Public","name":"redis_cache","owner":"triton-inference-server","isFork":false,"description":"TRITONCACHE implementation of a Redis cache","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":2,"starsCount":11,"forksCount":4,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-07T00:28:35.110Z"}},{"type":"Public","name":"openvino_backend","owner":"triton-inference-server","isFork":false,"description":"OpenVINO backend for Triton.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":3,"issueCount":4,"starsCount":29,"forksCount":15,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,2,0,0,1,3,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-07T00:28:32.570Z"}},{"type":"Public","name":"model_analyzer","owner":"triton-inference-server","isFork":false,"description":"Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Server models.","allTopics":["deep-learning","gpu","inference","performance-analysis"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":19,"starsCount":419,"forksCount":74,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-07T00:28:30.676Z"}},{"type":"Public","name":"local_cache","owner":"triton-inference-server","isFork":false,"description":"Implementation of a local in-memory cache for Triton Inference Server's TRITONCACHE API","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":1,"starsCount":4,"forksCount":1,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-07T00:28:29.089Z"}},{"type":"Public","name":"identity_backend","owner":"triton-inference-server","isFork":false,"description":"Example Triton backend that demonstrates most of the Triton Backend API.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":6,"forksCount":12,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-07T00:28:28.418Z"}},{"type":"Public","name":"common","owner":"triton-inference-server","isFork":false,"description":"Common source, scripts and utilities shared across all Triton repositories.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":3,"issueCount":0,"starsCount":62,"forksCount":73,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-07T00:28:25.930Z"}},{"type":"Public","name":"checksum_repository_agent","owner":"triton-inference-server","isFork":false,"description":"The Triton repository agent that verifies model checksums.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":10,"forksCount":7,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-07T00:28:23.209Z"}},{"type":"Public","name":"pytriton","owner":"triton-inference-server","isFork":false,"description":"PyTriton is a Flask/FastAPI-like interface that simplifies Triton's deployment in Python environments.","allTopics":["gpu","deep-learning","inference"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":9,"starsCount":716,"forksCount":48,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-06T18:19:25.092Z"}},{"type":"Public","name":"contrib","owner":"triton-inference-server","isFork":false,"description":"Community contributions to Triton that are not officially supported or maintained by the Triton project.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":8,"forksCount":7,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T22:13:16.062Z"}}],"repositoryCount":35,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"triton-inference-server repositories"}