Replies: 2 comments
-
|
@apepkuss You can just ask a model to generate you n-number of responses, |
Beta Was this translation helpful? Give feedback.
0 replies
-
|
We don't support that for now, but you can have the same result by passing an array of prompts to For example, this is equivalent to Remember to set a high temperature |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment

Uh oh!
There was an error while loading. Please reload this page.
-
Hi guys,
OpenAI defines a param
nin the chat completion request, which is used to specify how many chat completion choices to generate for each input message. Does llama.cpp provide support for multiple chat completions? Thanks a lot!Beta Was this translation helpful? Give feedback.
All reactions