malek-messaoudii commited on
Commit
ce133a0
Β·
1 Parent(s): e2251fd

refactor: Update GROQ_TOPIC_MODEL to the latest production version and adjust fallback models in TopicService for improved model selection and stability.

Browse files
Files changed (2) hide show
  1. config.py +2 -1
  2. services/topic_service.py +4 -5
config.py CHANGED
@@ -43,7 +43,7 @@ GROQ_TTS_FORMAT = "wav"
43
  GROQ_CHAT_MODEL = "llama3-70b-8192"
44
 
45
  # **Topic Extraction Model**
46
- GROQ_TOPIC_MODEL = "llama3-70b-8192" # Using same model as chat for consistency
47
 
48
  # ============ SUPABASE ============
49
  SUPABASE_URL = os.getenv("SUPABASE_URL", "")
@@ -90,6 +90,7 @@ logger.info(f" HF Label Model : {HUGGINGFACE_LABEL_MODEL_ID}")
90
  logger.info(f" GROQ STT Model : {GROQ_STT_MODEL}")
91
  logger.info(f" GROQ TTS Model : {GROQ_TTS_MODEL}")
92
  logger.info(f" GROQ Chat Model : {GROQ_CHAT_MODEL}")
 
93
  logger.info(f" Google API Key : {'βœ“ Configured' if GOOGLE_API_KEY else 'βœ— Not configured'}")
94
  logger.info(f" Supabase URL : {'βœ“ Configured' if SUPABASE_URL else 'βœ— Not configured'}")
95
  logger.info("="*60)
 
43
  GROQ_CHAT_MODEL = "llama3-70b-8192"
44
 
45
  # **Topic Extraction Model**
46
+ GROQ_TOPIC_MODEL = "llama-3.3-70b-versatile" # Latest production model, fallback: "llama3-70b-8192"
47
 
48
  # ============ SUPABASE ============
49
  SUPABASE_URL = os.getenv("SUPABASE_URL", "")
 
90
  logger.info(f" GROQ STT Model : {GROQ_STT_MODEL}")
91
  logger.info(f" GROQ TTS Model : {GROQ_TTS_MODEL}")
92
  logger.info(f" GROQ Chat Model : {GROQ_CHAT_MODEL}")
93
+ logger.info(f" GROQ Topic Model: {GROQ_TOPIC_MODEL}")
94
  logger.info(f" Google API Key : {'βœ“ Configured' if GOOGLE_API_KEY else 'βœ— Not configured'}")
95
  logger.info(f" Supabase URL : {'βœ“ Configured' if SUPABASE_URL else 'βœ— Not configured'}")
96
  logger.info("="*60)
services/topic_service.py CHANGED
@@ -22,14 +22,13 @@ class TopicService:
22
 
23
  def __init__(self):
24
  self.llm = None
25
- # Use valid Groq model - defaults from config, fallback to common models
26
  self.model_name = GROQ_TOPIC_MODEL if GROQ_TOPIC_MODEL else "llama3-70b-8192"
27
  # Fallback models to try if primary fails (using current/available Groq models)
28
  self.fallback_models = [
29
- "llama3-70b-8192",
30
- "llama-3.1-8b-instant",
31
- "mixtral-8x7b-32768",
32
- "gemma2-9b-it" # Alternative smaller model
33
  ]
34
  self.initialized = False
35
 
 
22
 
23
  def __init__(self):
24
  self.llm = None
25
+ # Use valid Groq model - defaults from config, fallback to stable model
26
  self.model_name = GROQ_TOPIC_MODEL if GROQ_TOPIC_MODEL else "llama3-70b-8192"
27
  # Fallback models to try if primary fails (using current/available Groq models)
28
  self.fallback_models = [
29
+ "llama3-70b-8192", # Stable production model (same as chat)
30
+ "llama-3.1-8b-instant", # Faster, smaller alternative
31
+ "openai/gpt-oss-20b" # Alternative OpenAI OSS model
 
32
  ]
33
  self.initialized = False
34