0.5.1 • Published 2 years ago
@skadefro/litellm v0.5.1
Usage
npm install litellmimport { completion } from 'litellm';
process.env['OPENAI_API_KEY'] = 'your-openai-key';
const response = await completion({
model: 'gpt-3.5-turbo',
messages: [{ content: 'Hello, how are you?', role: 'user' }],
});
// or stream the results
const stream = await completion({
model: "gpt-3.5-turbo",
messages: [{ content: "Hello, how are you?", role: "user" }],
stream: true
});
for await (const part of stream) {
process.stdout.write(part.choices[0]?.delta?.content || "");
}Features
We aim to support all features that LiteLLM python package supports.
- Standardised completions
- Standardised embeddings
- Caching ❌
- Proxy ❌
Supported Providers
| Provider | Completion | Streaming |
|---|---|---|
| openai | ✅ | ✅ |
| cohere | ✅ | ✅ |
| anthropic | ✅ | ✅ |
| ollama | ✅ | ✅ |
| ai21 | ✅ | ✅ |
| replicate | ✅ | ✅ |
| huggingface | ❌ | ❌ |
| together_ai | ❌ | ❌ |
| openrouter | ❌ | ❌ |
| vertex_ai | ❌ | ❌ |
| palm | ❌ | ❌ |
| baseten | ❌ | ❌ |
| azure | ❌ | ❌ |
| sagemaker | ❌ | ❌ |
| bedrock | ❌ | ❌ |
| vllm | ❌ | ❌ |
| nlp_cloud | ❌ | ❌ |
| aleph alpha | ❌ | ❌ |
| petals | ❌ | ❌ |
| deepinfra | ❌ | ❌ |
Development
Clone the repo
git clone https://github.com/zya/litellmjs.gitInstall dependencies
npm installRun unit tests
npm tRun E2E tests
First copy the example env file.
cp .example.env .envThen fill the variables with your API keys to be able to run the E2E tests.
OPENAI_API_KEY=<Your OpenAI API key>
....Then run the command below to run the tests
npm run test:e2e