Fix (#12390)
This commit is contained in:
parent
cb7b08948b
commit
8331875f34
1 changed files with 2 additions and 0 deletions
|
|
@ -54,6 +54,8 @@ llm = LLM(model="YOUR_MODEL",
|
||||||
disable_async_output_proc=True,
|
disable_async_output_proc=True,
|
||||||
distributed_executor_backend="ray",
|
distributed_executor_backend="ray",
|
||||||
max_model_len=2000,
|
max_model_len=2000,
|
||||||
|
trust_remote_code=True,
|
||||||
|
block_size=8,
|
||||||
max_num_batched_tokens=2000)
|
max_num_batched_tokens=2000)
|
||||||
# Generate texts from the prompts. The output is a list of RequestOutput objects
|
# Generate texts from the prompts. The output is a list of RequestOutput objects
|
||||||
# that contain the prompt, generated text, and other information.
|
# that contain the prompt, generated text, and other information.
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue