https://github.com/peopleworks/semantickernelfromlocaltocloud
Semantic Kernel using local (Ollama, phi3.5, llama3.2 ) and Remote (Azure, OpenAI, Google Gemini, Groq)
https://github.com/peopleworks/semantickernelfromlocaltocloud
Last synced: 6 months ago
JSON representation
Semantic Kernel using local (Ollama, phi3.5, llama3.2 ) and Remote (Azure, OpenAI, Google Gemini, Groq)
- Host: GitHub
- URL: https://github.com/peopleworks/semantickernelfromlocaltocloud
- Owner: peopleworks
- Created: 2024-10-16T23:02:16.000Z (8 months ago)
- Default Branch: master
- Last Pushed: 2024-10-21T14:37:23.000Z (8 months ago)
- Last Synced: 2024-10-22T00:19:11.855Z (8 months ago)
- Language: C#
- Size: 15.6 KB
- Stars: 2
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Semantic Kernel From Local To Cloud
[](#)
Integrate Semantic Kernel services across local and remote providers using a unified codebase. Seamlessly switch between providers like Ollama, phi3.5, llama3.2, Azure, OpenAI, Google Gemini, and Groq by simply changing the service name. Leverage function calling and memory with all supported models.
## Table of Contents
- [About the Project](#about-the-project)
- [Features](#features)
- [Getting Started](#getting-started)
- [Prerequisites](#prerequisites)
- [Installation](#installation)
- [Usage](#usage)
- [Roadmap](#roadmap)
- [Contributing](#contributing)
- [License](#license)
- [Contact](#contact)
- [Acknowledgments](#acknowledgments)---
## About the Project
This project showcases how to use Microsoft's Semantic Kernel with both **local** (Ollama, phi3.5, llama3.2) and **remote** services (Azure, OpenAI, Google Gemini, Groq) using the same codebase. By simply changing the service name, you can switch between different AI providers without altering your core logic.
## Features
- **Unified Kernel Creation**: Integrate all Semantic Kernel services using a single kernel setup.
- **Multi-Provider Support**: Seamlessly switch between local and remote AI models.
- **Function Calling**: Utilize function calling capabilities across all supported models.
- **Memory Integration**: Implement memory functions with ease.
- **Extensible Architecture**: Easily expand to support more providers and functionalities.## Getting Started
### Prerequisites
- **Development Environment**: Visual Studio or Visual Studio Code
- **.NET SDK**: Ensure you have the latest .NET SDK installed
- **API Keys**: Obtain API keys for the services you wish to use### Installation
1. **Clone the Repository**
```bash
git clone https://github.com/peopleworks/SemanticKernelFromLocalToCloud.git
```
2. **Open the Project**
- Navigate to the project directory and open it in Visual Studio or Visual Studio Code.
3. **Update API Keys**
- Locate the `appsettings.json` file.
- Update the `ApiKeys` section with your respective API keys.
4. **Run the Application**
- Use your IDE's build and run feature to start the application.## Usage
- **Switch Providers**
- Change the service name in the configuration to switch between different AI models.
- **Function Calling**
- Utilize the function calling feature as documented in Semantic Kernel.
- **Memory Features**
- Implement and test memory functionalities across different models.## Roadmap
- **Upcoming Features**
- Adding Retrieval-Augmented Generation (RAG) for both local and remote models.
- Implementing additional functions and enhancements in the next version.Stay tuned for more exciting updates!
## Contributing
Contributions are what make the open-source community such an amazing place to learn, inspire, and create. Any contributions you make are **greatly appreciated**.
1. **Fork the Project**
2. **Create your Feature Branch** (`git checkout -b feature/AmazingFeature`)
3. **Commit your Changes** (`git commit -m 'Add some AmazingFeature'`)
4. **Push to the Branch** (`git push origin feature/AmazingFeature`)
5. **Open a Pull Request**## License
Distributed under the **MIT License**. See `LICENSE` file for more information.
## Contact
- **Email**: [[email protected]](mailto:[email protected])
## Acknowledgments
- **Microsoft® Semantic Kernel Team**: For their outstanding work.
- **Ollama**: For providing an excellent tool.
- **Google**: For their fast and well-crafted models.
- **Groq**: For being a part of the developer community.
- **Meta**: For supporting open-source initiatives.---
*This README was generated to provide a comprehensive overview of the project, aiming to assist developers and contributors in understanding and utilizing the Semantic Kernel across various platforms.*