Meta’s Llama 3.1 8B instruct-tuned model, designed for fast and efficient dialogue. It performs strongly in human evaluations and is ideal for applications requiring a balance of speed and quality.
from openai import OpenAI client = OpenAI( base_url="https://api.naga.ac/v1", api_key="YOUR_API_KEY", ) resp = client.chat.completions.create( model="llama-3.1-8b-instruct", messages=[ {"role": "user", "content": "What's 2+2?"} ], temperature=0.2, ) print(resp.choices[0].message.content)