https://github.com/nyandwi/learning-visual-attention-for-robotic-vision
Learning Visual Attention for Robotic Vision
https://github.com/nyandwi/learning-visual-attention-for-robotic-vision
Last synced: about 1 month ago
JSON representation
Learning Visual Attention for Robotic Vision
- Host: GitHub
- URL: https://github.com/nyandwi/learning-visual-attention-for-robotic-vision
- Owner: Nyandwi
- Created: 2023-03-06T14:48:50.000Z (over 2 years ago)
- Default Branch: main
- Last Pushed: 2023-05-16T13:15:58.000Z (over 2 years ago)
- Last Synced: 2025-06-27T14:07:45.969Z (4 months ago)
- Language: Jupyter Notebook
- Size: 7.66 MB
- Stars: 2
- Watchers: 2
- Forks: 1
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Learning Visual Attention for Robotic Vision
### Abstract
Visual saliency maps represent the attention-grabbing area of an image based on
the human vision nerve system. Existing deep learning architectures used to predict
saliency maps from images face an underlying challenge of generalization. In this
study, we explore the use of image transformation techniques and deeper model
architectures, such as ConvNext, a mixture of CNNs and Transformers as feature
extractors as a way of making the salency map predictions of the DeepGazeIIE
model more generalizable. This is done with aim of developing a deep learning.Learn more from the [project report](https://drive.google.com/drive/u/0/folders/1FlmtSYcd-l7DbgHO7L5uMOwwoQtNbXJE).
Contributors: Denis Musinguzi, Kevin Sebineza, Jeande Dieu Nyandwi, Muhammed Danso.
### AcknowledgmentsThis project is a part of Introduction to Deep Learning(11-785). We thank the course staffs(instructors and TAs) for supporting us over the whole semester. We also thank authors of the baseline models we used for open-sourcing their codes.