Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/anoff/microllaborators
microllaborators ๐ฉโ๐ฆโ๐ฆ๐ฎ๐ฌ๐ฉโ๐ซ - the revolution in teaching
https://github.com/anoff/microllaborators
ar collaboration hackathon javascript microscopy zeiss
Last synced: 13 days ago
JSON representation
microllaborators ๐ฉโ๐ฆโ๐ฆ๐ฎ๐ฌ๐ฉโ๐ซ - the revolution in teaching
- Host: GitHub
- URL: https://github.com/anoff/microllaborators
- Owner: anoff
- License: mit
- Created: 2017-04-29T07:56:29.000Z (over 7 years ago)
- Default Branch: master
- Last Pushed: 2018-08-11T13:43:00.000Z (over 6 years ago)
- Last Synced: 2024-11-21T12:44:37.848Z (22 days ago)
- Topics: ar, collaboration, hackathon, javascript, microscopy, zeiss
- Language: Max
- Homepage: https://zeisshack.surge.sh
- Size: 34 MB
- Stars: 4
- Watchers: 3
- Forks: 2
- Open Issues: 0
-
Metadata Files:
- Readme: readme.md
- License: LICENSE
Awesome Lists containing this project
- awesome-hacking-lists - anoff/microllaborators - microllaborators ๐ฉโ๐ฆโ๐ฆ๐ฎ๐ฌ๐ฉโ๐ซ - the revolution in teaching (Max)
README
# microllaborators ๐ฉโ๐ฆโ๐ฆ๐ฌ๐ฉโ๐ซ
> put microscope images into AR for advanced collaboration ๐ฎ๐ฌ๐ฉโ๐ซ
Teacher/student classroom scenario for gamified collaboration on three dimensional microscopic images.
By opening the website on a device with a camera ([iOS not supported](https://webkit.org/status/#specification-webrtc) ๐) pupils and the teacher can interact visually, from remote locations and in realtime.![safari fails](./safari-fail.png)
The idea is that the teacher can ask a question regarding the current image and the students can answer it by placing markers/annotations on the image. The teacher can see all annotations and validate the students response.
> Note: This codebase was created as part of the Zeiss Hackathon 2017-04-29 in Munich. Don't blame us for the ugly code!
## โ๏ธ authors
* Tim Groรmann
* [Github](http://github.com/timgrossmann)
* [Twitter](https://twitter.com/timigrossmann)
* Andreas Offenhaeuser
* [Website](http://anoff.io)
* [Twitter](https://twitter.com/an0xff)## ๐พ๐ฆ usage
To use the augmented reality feature you need to get your hands on a markerimage of the [zeiss marker](./public/markers/zeiss.jpg). Just print it out or open it on your laptop. After you got a marker just head over to the website, see below.
Currently there are two separate views implemented for student and teachers. In a production environment there would be authentication & authorization concepts in place that automatically route the user to the correct view.### ๐จโ๐ student
[https://zeisshack.surge.sh](https://zeisshack.surge.sh)
![AR marker](./public/markers/zeiss.jpg)
### ๐ฉโ๐ซ teacher
[https://zeisshack.surge.sh/teacher](https://zeisshack.surge.sh/teacher)
### running the server
You can start a local development server with server-side autorestart and access it at [https://localhost:8083](https://localhost:8083)
```sh
yarn
npm start
```### deploying the website
If you are Tim or Andreas you can deploy the code to surge at [zeisshack.surge.sh](https://zeisshack.surge.sh)
```sh
npm run deploy
```If you're someone else just put the `public` folder on any static hosting.
All the dynamic features are **client only** and realized over [firebase](http://firebase.com).## ๐ฝ demo / workflow
This section describes how the solution would be used in a real environment.
### preparation
Teaching environment e.g. school with a teacher using a microscope that allows digital image capturing. Students use a smartphone with camera and internet/wifi.
The students open the [student website](https://zeisshackathon.surge.sh) and place the [AR marker](./public/markers/zeiss.jpg) as paper on their desk (might be handed out by teacher).
![teacher image selection](./demo/student_marker.png)
### providing the question
The teacher goes to the teacher website and uploads/selects and image from his microscope he wants to discuss. He inputs the question in the top of the website and it gets broadcasted to all students.
![teacher image selection](./demo/teacher_image_upload.png)
### gamification
The students see the new image and an updated question. Each student can place a marker, if he misplaces it he can replace it by clicking the screen again.
By moving their phone around they can look at the image from different angles with a 3d effect.![student places marker](./demo/student_3d.png)
![student AR](./demo/student_ar.gif)
### evaluation of responses
While the students place their markers the teacher gets all answers in real time in his view and can select the correct region he expects to be placed. An overview shows him how many students answered correctly. This allows him to get feedback from the whole class without them influencing each other.
![teacher results](./demo/teacher_results.png)
## โ๏ธ License
[MIT](./LICENSE)
## ๐ Used puzzle pieces / Attribution
* [AR.js](https://github.com/jeromeetienne/AR.js): helped us a lot to get off the ground
* [firebase](firebase.google.com): allowed us to do realtime collaboration without backend
* [three.js](https://github.com/jeromeetienne/AR.js/tree/master/three.js): used for 3D markers
* [material design lite](https://getmdl.io/components/index.html): fancy design
* [Zeiss](https://www.zeiss.com): thanks for hosting and guiding us during the hackathon