Skip to content

Commit

Permalink
Minor fix
Browse files Browse the repository at this point in the history
  • Loading branch information
nv-hwoo committed Sep 27, 2023
1 parent 8537564 commit cacdcac
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions src/c++/perf_analyzer/docs/inference_load_modes.md
Original file line number Diff line number Diff line change
Expand Up @@ -57,14 +57,14 @@ concurrent requests received M number of responses back from the server.
The user can also specify custom parameters to the model using
`--request-parameter <name:value:type>` option.
For instance, passing `--request-parameter max_tokens:256:uint` will set an
additional parameter `max_tokens` of type `uint` to 256 as part of the request.
additional parameter `max_tokens` of type `int` to 256 as part of the request.

```bash
perf_analyzer -m <model_name> -i grpc --async --streaming \
--profile-export-file profile.json \
--periodic-concurrency-range 10:100:30 \
--request-period 10 \
--request-parameter max_tokens:256:uint
--request-parameter max_tokens:256:int
```

> **Note**
Expand Down

0 comments on commit cacdcac

Please sign in to comment.