TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.
Import incomplete
This repository was imported from GitHub but the migration didn't fully complete. You can retry the import to fix it.
Quick setup
Get started by creating a new file or uploading an existing file. We recommend every repository include a README, LICENSE, and .gitignore.
https://gitmorph.com/NVIDIA/TensorRT-LLM.git CREATE A NEW REPOSITORY ON THE COMMAND LINE
touch README.md git init git checkout -b main git add README.md git commit -m "first commit" git remote add origin https://gitmorph.com/NVIDIA/TensorRT-LLM.git git push -u origin main
PUSH AN EXISTING REPOSITORY
git remote add origin https://gitmorph.com/NVIDIA/TensorRT-LLM.git
git push -u origin main