Skip to content

Commit 04a2a26

Browse files
Fixed typos for GPT-OSS Deployment Guide
1 parent e492a59 commit 04a2a26

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

docs/source/blogs/tech_blog/blog9_Deploying_GPT_OSS_on_TRTLLM.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ We have a forthcoming guide for getting great performance on H100, however the b
2222

2323
</br>
2424

25-
## Launching the TRTLLM docker container
25+
## Launching the TensorRT-LLM docker container
2626

2727
The container image that you will use will be pulled from NVIDIA's NGC. This container is multi-platform and will run on both x64 and arm64 architectures: `nvcr.io/nvidia/tensorrt-llm/release:gpt-oss-dev`
2828

@@ -83,7 +83,7 @@ EOF
8383

8484
#### Launching TensorRT-LLM Serve
8585

86-
To launch the TensorRT-LLM Server to serve the model with the **low latency** config, run the following command. Commands for different GPU configures are provided (1xGPU, 8xGPU, 4xGPU):
86+
To launch the TensorRT-LLM Server to serve the model with the **low latency** config, run the following command. Commands for different GPU configurations are provided (1xGPU, 8xGPU, 4xGPU):
8787

8888
<details open> <summary>1x B200/GB200/H200</summary>
8989

@@ -160,7 +160,7 @@ EOF
160160
161161
#### Launching TensorRT-LLM Serve
162162

163-
To launch the TensorRT-LLM Server to serve the model with the **max throughput** config, run the following command. Commands for different GPU configures are provided (1xGPU, 8xGPU, 4xGPU):
163+
To launch the TensorRT-LLM Server to serve the model with the **max throughput** config, run the following command. Commands for different GPU configurations are provided (1xGPU, 8xGPU, 4xGPU):
164164

165165
<details open> <summary>1x B200/GB200/H200</summary>
166166

0 commit comments

Comments
 (0)