from groq import Groq def get_model(): client = Groq(api_key="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") # Replace with your actual key def generate(text): try: # Ensure text doesn't exceed token limits (approx 4 chars per token) if len(text) > 24000: # Conservative limit (6000 tokens * 4 chars) text = text[:24000] + "...[content truncated due to length]" response = client.chat.completions.create( model="llama3-8b-8192", messages=[{"role": "user", "content": text}], max_tokens=2000 # Set a max output token limit ) return response.choices[0].message.content except Exception as e: return f"Error generating response: {str(e)}" return generate