https://github.com/maklonfr/deepseek-r1-flask
Creating a Deepssek Model chatbot with Ollama using Flask
https://github.com/maklonfr/deepseek-r1-flask
css deepseek-r1 flask html ollama
Last synced: 6 months ago
JSON representation
Creating a Deepssek Model chatbot with Ollama using Flask
- Host: GitHub
- URL: https://github.com/maklonfr/deepseek-r1-flask
- Owner: MaklonFR
- Created: 2025-02-01T17:31:16.000Z (8 months ago)
- Default Branch: main
- Last Pushed: 2025-02-02T02:53:11.000Z (8 months ago)
- Last Synced: 2025-02-02T03:27:10.870Z (8 months ago)
- Topics: css, deepseek-r1, flask, html, ollama
- Language: HTML
- Homepage:
- Size: 29.3 KB
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Chatbot DeepSeek
A simple Flask-based web application for interacting with a chatbot. This project uses HTML and CSS for the frontend and Flask for the backend.
## Features
- User-friendly interface for interacting with the chatbot.
- Responsive design using HTML and CSS.
- Flask backend for handling requests and responses.## Prerequisites
Before running the application, ensure you have the following installed:
- Python (preferably version 3.7 or higher)
- Flask (for creating the web server)
- Ollama (for running DeepSeek locally)
- DeepSeek Model (deepseek-r1:1.5b)
- Text Editor/IDE (like VSCode, PyCharm, or Sublime Text)## Installation
1. **Download Ollama**
**For Windows:**
Visit the official Ollama website https://ollama.com/download. Download the Windows installer.After installation, verify it by running the following command in PowerShell:
2. **Pulling and Running the DeepSeek-r1:1.5b Model**
With Ollama installed, it's time to pull the DeepSeek-r1:1.5b model and run it locally.
How to Pull the DeepSeek Model
- Open your terminal or command prompt.
- Type the following command to pull the model:
```
ollama pull deepseek-r1:1.5b
```
This command will download the model, which may take some time depending on your internet speed and system resources.
4. **Checking Available Models Using ollama list**
```
ollama list
```
This will display all the models you have pulled and installed, including their version numbers. You can use this to verify that DeepSeek-r1:1.5b is available and ready to be used.5. **Clone the repository**:
```bash
git clone https://github.com/yourusername/deepseek-R1-flask.git
cd deepseek-R1-flask
6. **Set up a virtual environment:**
```
python -m venv venv
source venv/bin/activate # On Windows, use `venv\Scripts\activate`
```
7. **Install dependencies:**
```
pip install -r requirements.txt
```
8. **Run the Flask application:**
```
flask run
```
9. **Access the application:**Open your browser and go to
```
http://127.0.0.1:5000.
```## Result
