A RetroSearch Logo

Home - News ( United States | United Kingdom | Italy | Germany ) - Football scores

Search Query:

Showing content from https://github.com/triton-inference-server/server/releases/latest below:

Release Release 2.59.1 corresponding to NGC container 25.07 · triton-inference-server/server · GitHub

Triton Inference Server

The Triton Inference Server provides a cloud inferencing solution optimized for both CPUs and GPUs. The server provides an inference service via an HTTP or GRPC endpoint, allowing remote clients to request inferencing for any model being managed by the server. For edge deployments, Triton Server is also available as a shared library with an API that allows the full functionality of the server to be included directly in an application.

New Features and Improvements Known Issues Client Libraries and Examples

Ubuntu 24.04 builds of the client libraries and examples are included in this release in the attached v2.59.1_ubuntu2404.clients.tar.gz file. The SDK is also available for as an Ubuntu 24.04 based NGC Container. The SDK container includes the client libraries and examples, Performance Analyzer and Model Analyzer. Some components are also available in the tritonclient pip package. See Getting the Client Libraries for more information on each of these options.

Windows Support

[!NOTE]
There is no Windows release for 25.07, the latest release is 25.01.

Jetson iGPU Support

A release of Triton for IGX is provided in the attached tar file: tritonserver2.59.1-igpu.tar.

The tar file contains the Triton server executable and shared libraries and also the C++ and Python client libraries and examples. For more information on how to install and use Triton on JetPack refer to jetson.md.

The wheel for the Python client library is present in the tar file and can be installed by running the following command:

python3 -m pip install --upgrade clients/python/tritonclient-2.59.0-py3-none-manylinux2014_aarch64.whl[all]
Triton TRT-LLM Container Support Matrix

The Triton TensorRT-LLM container is built from the 25.04 image nvcr.io/nvidia/tritonserver:25.04-py3-min. Please refer to the support matrix and compatibility.md for all dependency versions related to 25.04. However, the packages listed below have different versions than those specified in the support matrix.

Dependency Version TensorRT-LLM 0.20.0 TensorRT 10.10.0.31

RetroSearch is an open source project built by @garambo | Open a GitHub Issue

Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo

HTML: 3.2 | Encoding: UTF-8 | Version: 0.7.4