https://github.com/mostafax/image-caption
End to End Deep learning model that generate image captions
https://github.com/mostafax/image-caption
cnn deep-learning image-caption image-captioning image-classifier keras lstm nuralnetwork python3 rnn text-from-image
Last synced: 5 months ago
JSON representation
End to End Deep learning model that generate image captions
- Host: GitHub
- URL: https://github.com/mostafax/image-caption
- Owner: mostafax
- Created: 2018-10-06T10:37:54.000Z (about 7 years ago)
- Default Branch: master
- Last Pushed: 2018-12-25T16:50:03.000Z (almost 7 years ago)
- Last Synced: 2024-04-28T06:07:36.421Z (over 1 year ago)
- Topics: cnn, deep-learning, image-caption, image-captioning, image-classifier, keras, lstm, nuralnetwork, python3, rnn, text-from-image
- Language: Python
- Homepage:
- Size: 12.5 MB
- Stars: 8
- Watchers: 3
- Forks: 4
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Image-caption Using End2end CNN , LSTM based Model!
The aim of the project is to generate a caption for images.
Each image has a story, Image Captioning narrates it.![]()
this model is bases on [Show and Tell: A Neural Image Caption Generator
](https://arxiv.org/pdf/1411.4555.pdf)📖 Documentation
================
## How to Run
**Install the requirements:**
```bash
pip3 install -r requirements.txt
```
**Running the Model**
```bash
python3 model.py
```## Results
The results are not bad at all! a lot of test cases gonna be so realistic, but the model still needs more training
![]()
## Paper
This project is an implementation of the [Show and Tell](https://arxiv.org/pdf/1411.4555.pdf), published 2015.## Dataset
- Dataset used is Flicker8k each image have 5 captions.
- you can request data from here [Flicker8k]
(https://forms.illinois.edu/sec/1713398).
**Sample of the data used**## Model Used
## Experiments
## Future Work
-Training, Training and more Training
-Using Resnet instead of VGG16
-Creating API for production level
-Using Word2Vec embedding.