https://github.com/eternalflame02/single-node-finetuning-of-tiny-llama-using-intel-xeon-spr
The project was undertaken as part of the Intel Unnati Industrial Training program for the year 2024. The primary objective of this project aligns with Problem Statement PS-04: Introduction to GenAI LLM Inference on CPUs and subsequent LLM Model Finetuning for the development of a Custom Chatbot.
https://github.com/eternalflame02/single-node-finetuning-of-tiny-llama-using-intel-xeon-spr
intel-unnati llm-finetuning llm-inference python tinyllama
Last synced: about 2 months ago
JSON representation
The project was undertaken as part of the Intel Unnati Industrial Training program for the year 2024. The primary objective of this project aligns with Problem Statement PS-04: Introduction to GenAI LLM Inference on CPUs and subsequent LLM Model Finetuning for the development of a Custom Chatbot.
- Host: GitHub
- URL: https://github.com/eternalflame02/single-node-finetuning-of-tiny-llama-using-intel-xeon-spr
- Owner: eternalflame02
- Created: 2024-07-15T10:51:36.000Z (10 months ago)
- Default Branch: main
- Last Pushed: 2024-07-15T15:55:31.000Z (10 months ago)
- Last Synced: 2025-02-06T00:27:07.602Z (3 months ago)
- Topics: intel-unnati, llm-finetuning, llm-inference, python, tinyllama
- Language: Jupyter Notebook
- Homepage:
- Size: 12.2 MB
- Stars: 1
- Watchers: 1
- Forks: 1
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
19 BITS
Mar Baselios College of Engineering and Technology (Autonomous)
Single-Node-Finetuning-of-Tiny-LLama-using-Intel-Xeon-SPR
This repository contains the implementation of Single-Node Finetuning of the Tiny LLaMA language model utilizing Intel Xeon Scalable Processors (SPR). The project was undertaken as part of the Intel Unnati Industrial Training program for the year 2024. The primary objective of this project aligns with Problem Statement PS-04: Introduction to Generative Artificial Intelligence (GenAI) Basic Large Language Model (LLM) Inference on CPUs and subsequent LLM Model Finetuning for the development of a Custom Chatbot.
🛠️ Installation Steps:
1. Environment Setup(Execute them in terminal):
Set up a Python environment with the necessary dependencies.
1.1. Create and activate a Conda environment:
```
conda create -n itrex-1 python=3.10 -y
conda activate itrex-1
```This creates a new Conda environment named itrex-1 with Python 3.10 and activates it.
1.2. Install required Python packages:
```
pip install intel-extension-for-transformers
```2. Cloning the Repository
2.1. Clone the repository:
```
git clone https://github.com/eternalflame02/Single-Node-FInetuning-of-Tiny-LLama-using-Intel-Xeon-SPR.git
```2.2. Navigate to the fine-tuning directory:
```
cd ./Single-Node-FInetuning-of-Tiny-LLama-using-Intel-Xeon-SPR/Fine Tuning/
```3. Installing Additional Dependencies
Install additional dependencies required for fine-tuning.
3.1. Install dependencies from the requirements.txt file:
```
pip install -r requirements.txt
```3.2. Install Jupyter and IPython kernel:
```
python3 -m pip install jupyter ipykernel
python3 -m ipykernel install --name neural-chat--user
```4. Setting Up Hugging Face Authentication
Authenticate with Hugging Face to access and download models.
4.1. Login to Hugging Face:
```
huggingface-cli login
```Create a token in https://huggingface.co/settings/tokens insert them in the huggingface login interface.
5. Downloading Data
Download the dataset required for fine-tuning.
5.1. Download the Alpaca dataset:
```
Curl -O https://github.com/tatsu-lab/stanford_alpaca/raw/main/alpaca_data.json
```💻 Built with
Public Repositories:
* [intel-extension-for-transformers](https://github.com/intel/intel-extension-for-transformers)
* [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0)
* [stanford_alpaca](https://github.com/tatsu-lab/stanford_alpaca)🍰 Contributors:
[@Rohith NS](https://github.com/eternalflame02)
[@Joshua Sunny Ninan](https://github.com/Joshualostonearth)
[@Avin ](https://github.com/ajsike2310)
[@Edwin K Mathew](https://github.com/EdKMathew)