Set Up
Installation
You can install the azllm
package via pip:
pip install azllm
Prerequisites
Python 3.11+
.env
file to store your API keysOPENAI_API_KEY=your_openai_api_key DEEPSEEK_API_KEY=your_deepseek_api_key XAI_API_KEY=your_xai_api_key GEMINI_API_KEY=your_gemini_api_key ANTHROPIC_API_KEY=your_anthropic_api_key FIREWORKS_API_KEY=your_fireworks_api_key
Ollama must be installed and running locally to use Ollama models.
The package expects the Ollama server to be available at
http://localhost:11434/v1
(default)