Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/sinedied/ollamazure
Emulates Azure OpenAI API on your local machine using Ollama and open-source models.
https://github.com/sinedied/ollamazure
ai azure emulator llm ollama openai
Last synced: 2 months ago
JSON representation
Emulates Azure OpenAI API on your local machine using Ollama and open-source models.
- Host: GitHub
- URL: https://github.com/sinedied/ollamazure
- Owner: sinedied
- License: mit
- Created: 2024-07-03T11:58:02.000Z (6 months ago)
- Default Branch: main
- Last Pushed: 2024-08-10T18:02:25.000Z (5 months ago)
- Last Synced: 2024-09-30T10:49:04.441Z (3 months ago)
- Topics: ai, azure, emulator, llm, ollama, openai
- Language: TypeScript
- Homepage:
- Size: 440 KB
- Stars: 19
- Watchers: 3
- Forks: 4
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- Changelog: CHANGELOG.md
- License: LICENSE
Awesome Lists containing this project
README
# ollamazure
[![NPM version](https://img.shields.io/npm/v/ollamazure.svg?style=flat-square)](https://www.npmjs.com/package/ollamazure)
[![Build Status](https://img.shields.io/github/actions/workflow/status/sinedied/ollamazure/ci.yml?style=flat-square&label=Build)](https://github.com/sinedied/ollamazure/actions)
![Node version](https://img.shields.io/node/v/ollamazure?style=flat-square)
[![XO code style](https://img.shields.io/badge/code_style-XO-5ed9c7?style=flat-square)](https://github.com/sindresorhus/xo)
[![TypeScript](https://img.shields.io/badge/TypeScript-blue?style=flat-square&logo=typescript&logoColor=white)](https://www.typescriptlang.org)
[![License](https://img.shields.io/badge/license-MIT-blue?style=flat-square)](LICENSE)⭐ If you like this tool, star it on GitHub — it helps a lot!
[Overview](#overview) • [Usage](#usage) • [Azure OpenAI compatibility](#azure-openai-compatibility) • [Sample code](#sample-code)
## Overview
**ollamazure** is a local server that emulates [Azure OpenAI](https://learn.microsoft.com/azure/ai-services/openai/overview) API on your local machine using [Ollama](https://ollama.com) and open-source models.
Using this tool, you can run your own local server that emulates the Azure OpenAI API, allowing you to test your code locally without incurring costs or being rate-limited. This is especially useful for development and testing purposes, or when you need to work offline.
By default, [`phi3`](https://ollama.com/library/phi3) is used as the model for completions, and [`all-minilm:l6-v2`](https://ollama.com/library/all-minilm:l6-v2) for embeddings. You can change these models using the [configuration options](#configuration-options).
> [!NOTE]
> This tool use different models than Azure OpenAI, so you should expect differences in results accuracy and performance. However, the API is compatible so you can use the same code to interact with it.## Usage
You need [Node.js v20+](https://nodejs.org) and [Ollama](https://ollama.com) installed on your machine to use this tool.
You can start the emulator directly using `npx` without installing it:
```bash
npx ollamazure
```Once the server is started, leave it open in a terminal window and you can use the Azure OpenAI API to interact with it. You can find sample code for different languages and frameworks in the [sample code](#sample-code) section.
For example, if you have an existing project that uses the Azure OpenAI SDK, you can point it to your local server by setting the `AZURE_OPENAI_ENDPOINT` environment variable to `http://localhost:4041` without changing the rest of your code.
### Installation
```bash
npm install -g ollamazure
```Once the installation is completed, start the emulator by running the following command in a terminal:
```bash
ollamazure
# or use the shorter alias `oaz`
```### Configuration options
```txt
ollamazure --help
Usage: ollamazure [options]Emulates Azure OpenAI API on your local machine using Ollama and open-source models.
Options:
--verbose show detailed logs
-y, --yes do not ask for confirmation (default: false)
-m, --model model to use for chat and text completions (default: "phi3")
-e, --embeddings model to use for embeddings (default: "all-minilm:l6-v2")
-d, --use-deployment use deployment name as model name (default: false)
-h, --host host to bind to (default: "localhost")
-p, --port port to use (default: 4041)
-o, --ollama-url ollama base url (default: "http://localhost:11434")
-v, --version show the current version
--help display help for command
```## Azure OpenAI compatibility
| Feature | Supported / with streaming |
| ------- | -------------------------- |
| [Completions](https://learn.microsoft.com/azure/ai-services/openai/reference#completions) | ✅ / ✅ |
| [Chat completions](https://learn.microsoft.com/azure/ai-services/openai/reference#chat-completions) | ✅ / ✅ |
| [Embeddings](https://learn.microsoft.com/azure/ai-services/openai/reference#embeddings) | ✅ / - |
| [JSON mode](https://learn.microsoft.com/azure/ai-services/openai/how-to/json-mode?tabs=python) | ✅ / ✅ |
| [Function calling](https://learn.microsoft.com/azure/ai-services/openai/how-to/function-calling) | ⛔ / ⛔ |
| [Reproducible outputs](https://learn.microsoft.com/azure/ai-services/openai/how-to/reproducible-output?tabs=pyton) | ✅ / ✅ |
| [Vision](https://learn.microsoft.com/azure/ai-services/openai/how-to/gpt-with-vision?tabs=rest%2Csystem-assigned%2Cresource) | ⛔ / ⛔ |
| [Assistants](https://learn.microsoft.com/azure/ai-services/openai/how-to/assistant) | ⛔ / ⛔ |Unimplemented features are currently not supported by Ollama, but are being worked on and may be added in the future.
## Sample code
See all code examples in the [samples](samples) folder.
### JavaScript
Azure OpenAI SDK
```typescript
import { AzureOpenAI } from 'openai';const openai = new AzureOpenAI({
// This is where you point to your local server
endpoint: 'http://localhost:4041',// Parameters below must be provided but are not used by the local server
apiKey: '123456',
apiVersion: '2024-02-01',
deployment: 'gpt-4',
});const chatCompletion = await openai.chat.completions.create({
messages: [{ role: 'user', content: 'Say hello!' }],
});console.log('Chat completion: ' + chatCompletion.choices[0]!.message?.content);
```Alternatively, you can set the `AZURE_OPENAI_ENDPOINT` environment variable to `http://localhost:4041` instead of passing it to the constructor. Everything else will work the same.
If you're using managed identity, this will work as well unless you're in a local container. In that case, you can use a dummy function `() => '1'` for the `azureADTokenProvider` parameter in the constructor.
LangChain.js
```typescript
import { AzureChatOpenAI } from '@langchain/openai';// Chat completion
const model = new AzureChatOpenAI({
// This is where you point to your local server
azureOpenAIBasePath: 'http://localhost:4041/openai/deployments',// Parameters below must be provided but are not used by the local server
azureOpenAIApiKey: '123456',
azureOpenAIApiVersion: '2024-02-01',
azureOpenAIApiDeploymentName: 'gpt-4'
});const completion = await model.invoke([{ type: 'human', content: 'Say hello!' }]);
console.log(completion.content);
```Alternatively, you can set the `AZURE_OPENAI_BASE_PATH` environment variable to `http://localhost:4041/openai/deployments` instead of passing it to the constructor. Everything else will work the same.
If you're using managed identity this will work the same unless you're in a local container. In that case, you can use a dummy function `() => '1'` for the `azureADTokenProvider` parameter in the constructor.
LlamaIndex.TS
```typescript
import { OpenAI } from "llamaindex";// Chat completion
const llm = new OpenAI({
azure: {
// This is where you point to your local server
endpoint: 'http://localhost:4041',// Parameters below must be provided but are not used by the local server
apiKey: '123456',
apiVersion: '2024-02-01',
deployment: 'gpt-4'
}
});const chatCompletion = await llm.chat({
messages: [{ role: 'user', content: 'Say hello!' }]
});console.log(chatCompletion.message.content);
```Alternatively, you can set the `AZURE_OPENAI_ENDPOINT` environment variable to `http://localhost:4041` instead of passing it to the constructor. Everything else will work the same.
If you're using managed identity, this will work as well unless you're in a local container. In that case, you can use a dummy function `() => '1'` for the `azureADTokenProvider` parameter in the constructor.
### Python
Azure OpenAI SDK
```python
from openai import AzureOpenAIopenai = AzureOpenAI(
# This is where you point to your local server
azure_endpoint="http://localhost:4041",# Parameters below must be provided but are not used by the local server
api_key="123456",
api_version="2024-02-01"
)# Chat completion
chat_completion = openai.chat.completions.create(
# Model must be provided but is not used by the local server
model="gpt-4",
messages=[
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Say hello!"}
]
)print(chat_completion.choices[0].message.content)
```Alternatively, you can set the `AZURE_OPENAI_ENDPOINT` environment variable to `http://localhost:4041` instead of passing it to the constructor. Everything else will work the same.
If you're using managed identity, this will work as well unless you're in a local container. In that case, you can use a dummy function `lambda:"1"` for the `azure_ad_token_provider` parameter in the constructor.
LangChain
```python
from langchain_openai import AzureChatOpenAI# Chat completion
model = AzureChatOpenAI(
# This is where you point to your local server
azure_endpoint="http://localhost:4041",# Parameters below must be provided but are not used by the local server
api_key="123456",
api_version="2024-02-01",
azure_deployment="gpt-4"
)chat_completion = model.invoke([{"type": "human", "content": "Say hello!"}])
print(chat_completion.content)
```Alternatively, you can set the `AZURE_OPENAI_ENDPOINT` environment variable to `http://localhost:4041` instead of passing it to the constructor. Everything else will work the same.
If you're using managed identity, this will work as well unless you're in a local container. In that case, you can use a dummy function `lambda:"1"` for the `azure_ad_token_provider` parameter in the constructor.
LlamaIndex
```python
from llama_index.core.llms import ChatMessage
from llama_index.llms.azure_openai import AzureOpenAI# Chat completion
llm = AzureOpenAI(
# This is where you point to your local server
azure_endpoint="http://localhost:4041",# Parameters below must be provided but are not used by the local server
api_key="123456",
api_version="2024-02-01",
engine="gpt-4"
)chat_completion = llm.chat([ChatMessage(role="user", content="Say hello!")])
print(chat_completion.message.content)
```Alternatively, you can set the `AZURE_OPENAI_ENDPOINT` environment variable to `http://localhost:4041` instead of passing it to the constructor. Everything else will work the same.
If you're using managed identity, this will work as well unless you're in a local container. In that case, you can use a dummy function `lambda:"1"` for the `azure_ad_token_provider` parameter in the constructor.
### C#
Azure OpenAI SDK
```csharp
using Azure;
using Azure.AI.OpenAI;
using OpenAI.Chat;// Chat completion
AzureOpenAIClient azureClient = new(
new Uri("http://localhost:4041"),
// Must be provided but are not used by the local server
new AzureKeyCredential("123456"));ChatClient chatClient = azureClient.GetChatClient("gpt-4");
ChatCompletion completion = chatClient.CompleteChat([new UserChatMessage("Say hello!")]);
Console.WriteLine(completion.Content[0].Text);
```Semantic Kernel
```csharp
using Microsoft.SemanticKernel;
using Microsoft.SemanticKernel.Connectors.OpenAI;var builder = Kernel.CreateBuilder();
// Chat completion
builder.AddAzureOpenAIChatCompletion(
"gpt-4", // Deployment Name (not used by the local server)
"http://localhost:4041", // Azure OpenAI Endpoint
"123456"); // Azure OpenAI Key (not used by the local server)var kernel = builder.Build();
var chatFunction = kernel.CreateFunctionFromPrompt(@"{{$input}}");var chatCompletion = await kernel.InvokeAsync(chatFunction, new() { ["input"] = "Say hello!" });
Console.WriteLine(chatCompletion);
```### Java
Azure OpenAI SDK
```csharp
package com.sample.azure;import java.util.ArrayList;
import java.util.List;import com.azure.ai.openai.*;
import com.azure.ai.openai.models.*;public class App
{
public static void main( String[] args )
{
// Chat completion
OpenAIClient client = new OpenAIClientBuilder()
// This is where you point to your local server
.endpoint("http://localhost:4041")
.buildClient();List chatMessages = new ArrayList<>();
chatMessages.add(new ChatRequestUserMessage("Say hello!"));
ChatCompletions chatCompletions = client.getChatCompletions("gpt-4",
new ChatCompletionsOptions(chatMessages));
System.out.println( chatCompletions.getChoices().get(0).getMessage().getContent());
}
}
```If you have the error `Key credentials require HTTPS to prevent leaking the key`, this means that you have to avoid setting the `.credential()` option on the `OpenAIClientBuilder`. This is currently a limitation of the Azure OpenAI SDK.