Skip to content

TanzirR/llm_engineering

Repository files navigation

#OPENAI API

 response = openai.chat.completions.create(
        model=MODEL,
        messages=[
            {"role": "system", "content": "system_prompt"},
            {"role": "user", "content": "user_prompt"}
      ],
        response_format={"type": "json_object"}
    )
 result = response.choices[0].message.content
 json.loads(result)
# Claude 4.0 Sonnet
# API needs system message provided separately from user prompt
# Also adding max_tokens

message = claude.messages.create(
    model="claude-sonnet-4-20250514",
    max_tokens=200,
    temperature=0.7,
    system=system_message,
    messages=[
        {"role": "user", "content": user_prompt},
    ],
)

print(message.content[0].text)
google.generativeai.configure(api_key=google_api_key)
gemini = google.generativeai.GenerativeModel(
    model_name='gemini-2.0-flash',
    system_instruction="system_prompt"
)
response = gemini.generate_content("user_prompt")
print(response.text)

About

LLM based projects

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors