Skip to content

Commit

Permalink
Rewriting LLM config search paragraph for clarity
Browse files Browse the repository at this point in the history
  • Loading branch information
nv-braf committed Apr 9, 2024
1 parent ad399c0 commit 1c1df45
Showing 1 changed file with 3 additions and 1 deletion.
4 changes: 3 additions & 1 deletion docs/config_search.md
Original file line number Diff line number Diff line change
Expand Up @@ -310,7 +310,9 @@ _This mode has the following limitations:_

- Summary/Detailed reports do not include the new metrics

LLMs can be optimized using either Quick or Brute search mode by setting `--model-type LLM`. You can specify CLI options to the GenAI-Perf tool using `genai_perf_flags`. See the [GenAI-Perf CLI](https://github.com/triton-inference-server/client/blob/main/src/c%2B%2B/perf_analyzer/genai-perf/README.md#cli) documentation for a list of the flags that can be specified.
In order to profile LLMs you must tell MA that the model type is LLM by setting `--model-type LLM` in the CLI/config file. You can specify CLI options to the GenAI-Perf tool using `genai_perf_flags`. See the [GenAI-Perf CLI](https://github.com/triton-inference-server/client/blob/main/src/c%2B%2B/perf_analyzer/genai-perf/README.md#cli) documentation for a list of the flags that can be specified.

LLMs can be optimized using either Quick or Brute search mode.

_An example model analyzer YAML config for a LLM:_

Expand Down

0 comments on commit 1c1df45

Please sign in to comment.