init
This commit is contained in:
@@ -74,7 +74,7 @@ DEFAULT_CONFIGS = {
|
||||
"openai": ModelConfig(
|
||||
provider="openai",
|
||||
model_name="gpt-4o-mini",
|
||||
temperature=0.1
|
||||
temperature=0.1,
|
||||
),
|
||||
"openai_light": ModelConfig(
|
||||
provider="openai",
|
||||
@@ -117,13 +117,19 @@ def get_llm(config: ModelConfig) -> BaseChatModel:
|
||||
from langchain_google_genai import ChatGoogleGenerativeAI
|
||||
|
||||
api_key = config.api_key or os.getenv("GOOGLE_API_KEY")
|
||||
print("Using GOOGLE_API_KEY:", api_key)
|
||||
if not api_key:
|
||||
raise ValueError("GOOGLE_API_KEY required for Gemini. Set via env or config.api_key")
|
||||
|
||||
return ChatGoogleGenerativeAI(
|
||||
model=config.model_name,
|
||||
temperature=config.temperature,
|
||||
google_api_key=api_key
|
||||
google_api_key=api_key,
|
||||
version="v1",
|
||||
additional_headers={
|
||||
"User-Agent": "PostmanRuntime/7.43.0",
|
||||
"Accept": "*/*"
|
||||
}
|
||||
)
|
||||
|
||||
elif provider == "openai":
|
||||
@@ -136,7 +142,8 @@ def get_llm(config: ModelConfig) -> BaseChatModel:
|
||||
return ChatOpenAI(
|
||||
model=config.model_name,
|
||||
temperature=config.temperature,
|
||||
api_key=api_key
|
||||
api_key=api_key,
|
||||
base_url=config.base_url or None
|
||||
)
|
||||
|
||||
else:
|
||||
|
||||
Reference in New Issue
Block a user