Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/mousazourob/uplift
Mood changer that uses facial recognition to play a song to positively change the user's mood
https://github.com/mousazourob/uplift
keras-tensorflow matplotlib opencv pandas python raspberry-pi raspberry-pi-camera
Last synced: 28 days ago
JSON representation
Mood changer that uses facial recognition to play a song to positively change the user's mood
- Host: GitHub
- URL: https://github.com/mousazourob/uplift
- Owner: MousaZourob
- Created: 2020-12-25T21:03:27.000Z (almost 4 years ago)
- Default Branch: main
- Last Pushed: 2020-12-27T01:15:58.000Z (almost 4 years ago)
- Last Synced: 2024-10-11T16:41:17.281Z (28 days ago)
- Topics: keras-tensorflow, matplotlib, opencv, pandas, python, raspberry-pi, raspberry-pi-camera
- Language: Jupyter Notebook
- Homepage:
- Size: 84.4 MB
- Stars: 1
- Watchers: 1
- Forks: 4
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# UpLift
### Overview:
Uplift is a mood changer programmed with **Python** that plays a song to positively change the user's mood. The song choice is based on the user's facial expression, so if the user is happy, cheerful music is played to keep them motivated, and if they are sad or angry, relaxing music is played to soothe them. A combination of hardware (**Raspberry Pi**), and software (**OpenCV, Keras, Pandas**) components were used to create this project.
Other than helping individual users, a commercial application of this product is usage in restaurants and cafes to detect the most dominant emotion, and play music based on the customer’s emotions.### Components:
#### Software:
The project utilizes **OpenCV** to read each frame of the video input from the **PiCamera**, which is connected to a **Raspberry Pi**. Once this is done **OpenCV** detects where the user’s face is in the frame by pixel location. When a face is detected, a blue rectangle is drawn over the face to allow the user to see the program in action. **Keras** then takes the frames and runs them against a dataset from **Kaggle** to determine the user's facial expression. After this using **MatplotLib** visualization library and **Pandas**, a graph is displayed showing the user their emotions based on their facial expression, and music is played using the **SimpleAudio** Python package.#### Hardware:
For hardware a **PiCamera** was used as well as a **Raspberry Pi 4**, to allow for efficient and speedy processing of the video frames. These components were connected to a monitor, so that the user can see the live video feed and a graph of their emotions that was displayed using the **Matplotlib**, and hear the music playing.### Demo:
#### 1. OpenCV detects the user's face and crops an image of their face#### 2. Keras runs a model against the detected faces and determines a facial expression, the data is displayed using MatplotLib
#### 3. A song is played through the speakers attached to the Raspberry Pi to maintain or elevate the user's mood
### Libraries and Frameworks Used:
* **Keras:** https://keras.io/
* **OpenCV:** https://opencv.org/
* **Pandas:** https://pandas.pydata.org/
* **MatPlotLib:** https://matplotlib.org/
* **NumPy:** https://numpy.org/### Contributers:
* Christopher Zhang: https://github.com/ChristopherChiZhang
* Jenny Tai: https://github.com/aegerita
* Alperen Asim Kes: https://github.com/alperenkes
* Jinwoo Park: https://github.com/hilfiger1
* Mousa Zourob https://github.com/MousaZourob