Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/lukaszbinden/learning-object-representations-by-mixing-scenes
Learning Object Representations by Mixing Scenes, MSc thesis, University of Bern, Switzerland
https://github.com/lukaszbinden/learning-object-representations-by-mixing-scenes
computer-vision disentangled-representations machine-learning representation-learning tensorflow unsupervised-learning
Last synced: 19 days ago
JSON representation
Learning Object Representations by Mixing Scenes, MSc thesis, University of Bern, Switzerland
- Host: GitHub
- URL: https://github.com/lukaszbinden/learning-object-representations-by-mixing-scenes
- Owner: lukaszbinden
- Created: 2018-08-09T14:12:13.000Z (over 6 years ago)
- Default Branch: master
- Last Pushed: 2019-10-30T21:35:30.000Z (about 5 years ago)
- Last Synced: 2024-10-26T14:35:28.513Z (2 months ago)
- Topics: computer-vision, disentangled-representations, machine-learning, representation-learning, tensorflow, unsupervised-learning
- Language: Python
- Homepage:
- Size: 8.3 MB
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Learning Object Representations by Mixing Scenes
JMCS Master Thesis, Computer Vision Group, Prof. Dr. Paolo Favaro, University of BernIn the digital age of ever increasing data amassment and accessibility, the demand for scalable machine learning models effective at refining the new oil is unprecedented. Unsupervised representation learning methods present a promising approach to exploit this invaluable yet unlabeled digital resource at scale. However, a majority of these approaches focuses on synthetic or simplified datasets of images. What if a method could learn directly from natural Internet-scale image data? In this thesis, we propose a novel approach for unsupervised learning of object representations by mixing natural image scenes. Without any human help, our method mixes visually similar images to synthesize new realistic scenes using adversarial training. In this process the model learns to represent and understand the objects prevalent in natural image data and makes them available for downstream applications. For example, it enables the transfer of objects from one scene to another. Through qualitative experiments on complex image data we show the effectiveness of our method along with its limitations. Moreover, we benchmark our approach quantitatively against state-of-the-art works on the STL-10 dataset. Our proposed method demonstrates the potential that lies in learning representations directly from natural image data and reinforces it as a promising avenue for future research.
The thesis document can be found here:
http://www.cvg.unibe.ch/media/theses/document/lukas-zbinden/2019/Master_Thesis_Lukas_Zbinden.pdf