malek-messaoudii
commited on
Commit
Β·
ce133a0
1
Parent(s):
e2251fd
refactor: Update GROQ_TOPIC_MODEL to the latest production version and adjust fallback models in TopicService for improved model selection and stability.
Browse files- config.py +2 -1
- services/topic_service.py +4 -5
config.py
CHANGED
|
@@ -43,7 +43,7 @@ GROQ_TTS_FORMAT = "wav"
|
|
| 43 |
GROQ_CHAT_MODEL = "llama3-70b-8192"
|
| 44 |
|
| 45 |
# **Topic Extraction Model**
|
| 46 |
-
GROQ_TOPIC_MODEL = "
|
| 47 |
|
| 48 |
# ============ SUPABASE ============
|
| 49 |
SUPABASE_URL = os.getenv("SUPABASE_URL", "")
|
|
@@ -90,6 +90,7 @@ logger.info(f" HF Label Model : {HUGGINGFACE_LABEL_MODEL_ID}")
|
|
| 90 |
logger.info(f" GROQ STT Model : {GROQ_STT_MODEL}")
|
| 91 |
logger.info(f" GROQ TTS Model : {GROQ_TTS_MODEL}")
|
| 92 |
logger.info(f" GROQ Chat Model : {GROQ_CHAT_MODEL}")
|
|
|
|
| 93 |
logger.info(f" Google API Key : {'β Configured' if GOOGLE_API_KEY else 'β Not configured'}")
|
| 94 |
logger.info(f" Supabase URL : {'β Configured' if SUPABASE_URL else 'β Not configured'}")
|
| 95 |
logger.info("="*60)
|
|
|
|
| 43 |
GROQ_CHAT_MODEL = "llama3-70b-8192"
|
| 44 |
|
| 45 |
# **Topic Extraction Model**
|
| 46 |
+
GROQ_TOPIC_MODEL = "llama-3.3-70b-versatile" # Latest production model, fallback: "llama3-70b-8192"
|
| 47 |
|
| 48 |
# ============ SUPABASE ============
|
| 49 |
SUPABASE_URL = os.getenv("SUPABASE_URL", "")
|
|
|
|
| 90 |
logger.info(f" GROQ STT Model : {GROQ_STT_MODEL}")
|
| 91 |
logger.info(f" GROQ TTS Model : {GROQ_TTS_MODEL}")
|
| 92 |
logger.info(f" GROQ Chat Model : {GROQ_CHAT_MODEL}")
|
| 93 |
+
logger.info(f" GROQ Topic Model: {GROQ_TOPIC_MODEL}")
|
| 94 |
logger.info(f" Google API Key : {'β Configured' if GOOGLE_API_KEY else 'β Not configured'}")
|
| 95 |
logger.info(f" Supabase URL : {'β Configured' if SUPABASE_URL else 'β Not configured'}")
|
| 96 |
logger.info("="*60)
|
services/topic_service.py
CHANGED
|
@@ -22,14 +22,13 @@ class TopicService:
|
|
| 22 |
|
| 23 |
def __init__(self):
|
| 24 |
self.llm = None
|
| 25 |
-
# Use valid Groq model - defaults from config, fallback to
|
| 26 |
self.model_name = GROQ_TOPIC_MODEL if GROQ_TOPIC_MODEL else "llama3-70b-8192"
|
| 27 |
# Fallback models to try if primary fails (using current/available Groq models)
|
| 28 |
self.fallback_models = [
|
| 29 |
-
"llama3-70b-8192",
|
| 30 |
-
"llama-3.1-8b-instant",
|
| 31 |
-
"
|
| 32 |
-
"gemma2-9b-it" # Alternative smaller model
|
| 33 |
]
|
| 34 |
self.initialized = False
|
| 35 |
|
|
|
|
| 22 |
|
| 23 |
def __init__(self):
|
| 24 |
self.llm = None
|
| 25 |
+
# Use valid Groq model - defaults from config, fallback to stable model
|
| 26 |
self.model_name = GROQ_TOPIC_MODEL if GROQ_TOPIC_MODEL else "llama3-70b-8192"
|
| 27 |
# Fallback models to try if primary fails (using current/available Groq models)
|
| 28 |
self.fallback_models = [
|
| 29 |
+
"llama3-70b-8192", # Stable production model (same as chat)
|
| 30 |
+
"llama-3.1-8b-instant", # Faster, smaller alternative
|
| 31 |
+
"openai/gpt-oss-20b" # Alternative OpenAI OSS model
|
|
|
|
| 32 |
]
|
| 33 |
self.initialized = False
|
| 34 |
|