-
multi-lora-trtllm-triton Public
Example of using multiple Loras using TRT-LLM + Triton
-
nim-simple-demo Public
A simple getting started guide for NVIDIA LLM NIMs. It walks you through the docs and gives more of an intuition.
Shell UpdatedDec 3, 2024 -
docs Public
Forked from mezerotm/docsGlobally distributed compute in the cloud built for production.
JavaScript UpdatedMay 20, 2024 -
tutorials Public
Forked from triton-inference-server/tutorialsThis repository contains tutorials and examples for Triton Inference Server
Python BSD 3-Clause "New" or "Revised" License UpdatedApr 11, 2023 -
models Public
Forked from NVIDIA-Merlin/modelsMerlin Models is a collection of deep learning recommender system model reference implementations
Python Apache License 2.0 UpdatedMar 30, 2023 -
model_analyzer Public
Forked from triton-inference-server/model_analyzerTriton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Server models.
Python Apache License 2.0 UpdatedMar 10, 2023 -
nvidia-triton-poc Public
Repo with instructions that can be leveraged to get NVIDIA Triton Inference Server up and running. Mostly it is a conglomeration of already available Triton tutorials
UpdatedMar 7, 2023 -
merlin-install-gcp Public
Instruction for installing Merlin on GCP, not using Vertex AI endpoints
UpdatedFeb 13, 2023 -
NVTabular Public
Forked from NVIDIA-Merlin/NVTabularNVTabular is a feature engineering and preprocessing library for tabular data designed to quickly and easily manipulate terabyte scale datasets used to train deep learning based recommender systems.
Python Apache License 2.0 UpdatedJan 30, 2023 -
prompt-engineering-resources Public
Repo to collect public information for prompt engineering
UpdatedJan 30, 2023 -
server Public
Forked from triton-inference-server/serverThe Triton Inference Server provides an optimized cloud and edge inferencing solution.
Python BSD 3-Clause "New" or "Revised" License UpdatedJul 25, 2022 -
TensorRT Public
Forked from NVIDIA/TensorRTTensorRT is a C++ library for high performance inference on NVIDIA GPUs and deep learning accelerators.
C++ Apache License 2.0 UpdatedOct 19, 2021 -
txtai Public
Forked from neuml/txtaiAI-powered search engine
Python Apache License 2.0 UpdatedAug 20, 2020 -
-
-
GrassGraphs Public
Code for the GrassGraphs algorithm for PAMI 2017 paper.
-
Traffic_Sign_Classifier Public
Traffic Sign Classifier Udacity Self-Driving Car Nanodegree
HTML UpdatedMar 24, 2018 -
pattern_classification Public
Forked from rasbt/pattern_classificationA collection of tutorials and examples for solving and understanding machine learning and pattern classification tasks
Jupyter Notebook GNU General Public License v3.0 UpdatedFeb 7, 2017 -
-
Covariance_ObjectDetector Public
Implementation of Tuzel's Fast Covariance Object Detector