From 05fb4eecf766bf22cf5b107f418eaada76c452ac Mon Sep 17 00:00:00 2001 From: Chen Yu Date: Sun, 20 Oct 2024 17:14:29 +0800 Subject: [PATCH] Update README.md --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index b53d118..5aa7116 100644 --- a/README.md +++ b/README.md @@ -85,7 +85,7 @@ Options: - `--experimental-structured-mode array` Structures the input and output into a plain array format. This option is more concise compared to base mode, though it uses slightly more tokens per batch. - `--experimental-structured-mode object` Structures both the input and output into a dynamically generated object schema based on input values. This option is even more concise and uses fewer tokens, but requires smaller batch sizes and can be slow and unreliable. Due to its unreliability, it may lead to more resubmission retries, potentially wasting more tokens in the process. - `--experimental-use-full-context` - Include the full context of translated data to work well with [prompt caching](https://openai.com/index/api-prompt-caching/). The length of the chunks is defined by `--history-prompt-length`. May risk running into the model's context window limit, typically `128K`, which should be sufficient for most cases. + Include the full context of translated data to work well with [prompt caching](https://openai.com/index/api-prompt-caching/). The translated lines per user and assistant message pairs are sliced as defined by `--history-prompt-length`. May risk running into the model's context window limit, typically `128K`, which should be sufficient for most cases. - `--log-level ` Log level (default: `debug`, choices: `trace`, `debug`, `info`, `warn`, `error`, `silent`) - `--silent`