move the model to the endpoint as well
This commit is contained in:
parent
23f5e0274b
commit
b91e6a7ca5
@ -183,8 +183,7 @@ async def on_message(ctx):
|
|||||||
"role": "user",
|
"role": "user",
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
"model": "llama2-chat-7b",
|
"model": requests.get(llm_rule_endpoint).json()["model"],
|
||||||
# "model": "mistral-ins-7b-q4",
|
|
||||||
"stream": False,
|
"stream": False,
|
||||||
"max_tokens": 4096,
|
"max_tokens": 4096,
|
||||||
"stop": ["goon"],
|
"stop": ["goon"],
|
||||||
|
Loading…
x
Reference in New Issue
Block a user