Fix: Update Gemini model to stable gemini-1.5-flash

This commit is contained in:
Marcos
2026-03-22 11:26:57 -03:00
parent 0d774f7486
commit b787cb7baa

View File

@@ -9,7 +9,7 @@ def get_llm_response(prompt: str, provider: str, cfg: dict) -> str:
"""Invoca o provedor de LLM configurado.""" """Invoca o provedor de LLM configurado."""
if provider == "gemini": if provider == "gemini":
api_key = cfg.get("gemini_api_key") or os.getenv("GEMINI_API_KEY") api_key = cfg.get("gemini_api_key") or os.getenv("GEMINI_API_KEY")
url = f"https://generativelanguage.googleapis.com/v1beta/models/gemini-2.0-flash:generateContent?key={api_key}" url = f"https://generativelanguage.googleapis.com/v1beta/models/gemini-1.5-flash:generateContent?key={api_key}"
payload = {"contents": [{"parts": [{"text": prompt}]}]} payload = {"contents": [{"parts": [{"text": prompt}]}]}
res = requests.post(url, json=payload) res = requests.post(url, json=payload)
if res.status_code == 200: if res.status_code == 200: