Do not incorrectly give temp parameter to expert model.
This commit is contained in:
parent
5240fb2617
commit
54fdebfc3a
|
|
@ -122,9 +122,10 @@ def create_llm_client(
|
||||||
if not config:
|
if not config:
|
||||||
raise ValueError(f"Unsupported provider: {provider}")
|
raise ValueError(f"Unsupported provider: {provider}")
|
||||||
|
|
||||||
# Handle temperature for expert mode
|
# Only pass temperature if it's explicitly set and not in expert mode
|
||||||
if is_expert:
|
temp_kwargs = {}
|
||||||
temperature = 0
|
if not is_expert and temperature is not None:
|
||||||
|
temp_kwargs = {"temperature": temperature}
|
||||||
|
|
||||||
if provider == "deepseek":
|
if provider == "deepseek":
|
||||||
return create_deepseek_client(
|
return create_deepseek_client(
|
||||||
|
|
@ -145,26 +146,26 @@ def create_llm_client(
|
||||||
return ChatOpenAI(
|
return ChatOpenAI(
|
||||||
api_key=config["api_key"],
|
api_key=config["api_key"],
|
||||||
model=model_name,
|
model=model_name,
|
||||||
**({"temperature": temperature} if temperature is not None else {}),
|
**temp_kwargs,
|
||||||
)
|
)
|
||||||
elif provider == "anthropic":
|
elif provider == "anthropic":
|
||||||
return ChatAnthropic(
|
return ChatAnthropic(
|
||||||
api_key=config["api_key"],
|
api_key=config["api_key"],
|
||||||
model_name=model_name,
|
model_name=model_name,
|
||||||
**({"temperature": temperature} if temperature is not None else {}),
|
**temp_kwargs,
|
||||||
)
|
)
|
||||||
elif provider == "openai-compatible":
|
elif provider == "openai-compatible":
|
||||||
return ChatOpenAI(
|
return ChatOpenAI(
|
||||||
api_key=config["api_key"],
|
api_key=config["api_key"],
|
||||||
base_url=config["base_url"],
|
base_url=config["base_url"],
|
||||||
temperature=temperature if temperature is not None else 0.3,
|
**temp_kwargs if temp_kwargs else {"temperature": 0.3},
|
||||||
model=model_name,
|
model=model_name,
|
||||||
)
|
)
|
||||||
elif provider == "gemini":
|
elif provider == "gemini":
|
||||||
return ChatGoogleGenerativeAI(
|
return ChatGoogleGenerativeAI(
|
||||||
api_key=config["api_key"],
|
api_key=config["api_key"],
|
||||||
model=model_name,
|
model=model_name,
|
||||||
**({"temperature": temperature} if temperature is not None else {}),
|
**temp_kwargs,
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
raise ValueError(f"Unsupported provider: {provider}")
|
raise ValueError(f"Unsupported provider: {provider}")
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue