Hermes Agent · Optional

tensorrt-llm

Optimizes LLM inference with NVIDIA TensorRT for maximum throughput and lowest latency. Use for production deployment on NVIDIA GPUs (A100/H100), when you need 10-100x faster inference than PyTorch, or for serving models with quantization (FP8/INT4), in-flight batching, and multi-GPU scaling.

MlopsOptionalv1.0.0MIT

What this skill is

This directory page tracks a Hermes-compatible skill reference and links back to the original source for install instructions, files, and updates.

Tags and platforms

Inference ServingTensorRT-LLMNVIDIAInference OptimizationHigh ThroughputLow LatencyProductionFP8INT4In-Flight BatchingMulti-GPU

Featured

Your product here

Show your offer to OpenClaw operators and AI builders across every page and blog.

Advertise