Fix (#12390)
This commit is contained in:
parent
cb7b08948b
commit
8331875f34
1 changed files with 2 additions and 0 deletions
|
|
@ -54,6 +54,8 @@ llm = LLM(model="YOUR_MODEL",
|
|||
disable_async_output_proc=True,
|
||||
distributed_executor_backend="ray",
|
||||
max_model_len=2000,
|
||||
trust_remote_code=True,
|
||||
block_size=8,
|
||||
max_num_batched_tokens=2000)
|
||||
# Generate texts from the prompts. The output is a list of RequestOutput objects
|
||||
# that contain the prompt, generated text, and other information.
|
||||
|
|
|
|||
Loading…
Reference in a new issue