load model lazily in expert tool to avoid model instantiation when module is imported
This commit is contained in:
parent
c1217912f1
commit
4de776c8e9
|
|
@ -8,7 +8,13 @@ from langchain_openai import ChatOpenAI
|
||||||
from .memory import get_memory_value, get_related_files
|
from .memory import get_memory_value, get_related_files
|
||||||
|
|
||||||
console = Console()
|
console = Console()
|
||||||
model = ChatOpenAI(model_name="o1-preview")
|
_model = None
|
||||||
|
|
||||||
|
def get_model():
|
||||||
|
global _model
|
||||||
|
if _model is None:
|
||||||
|
_model = ChatOpenAI(model_name="o1-preview")
|
||||||
|
return _model
|
||||||
|
|
||||||
# Keep track of context globally
|
# Keep track of context globally
|
||||||
expert_context = []
|
expert_context = []
|
||||||
|
|
@ -168,7 +174,7 @@ def ask_expert(question: str) -> str:
|
||||||
query = "\n".join(query_parts)
|
query = "\n".join(query_parts)
|
||||||
|
|
||||||
# Get response
|
# Get response
|
||||||
response = model.invoke(query)
|
response = get_model().invoke(query)
|
||||||
|
|
||||||
# Format and display response
|
# Format and display response
|
||||||
console.print(Panel(
|
console.print(Panel(
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue