Results from the first Accessathons
Here, you can find some of the abstracts of the projects participants worked on during the first Accessathon in June 2017. In the meantime, we're working hard on a platform to publish detailled project documentations.
FeyeND: Assisted Living
- Team members:
- Alexander Duseti
- Anoshan Indreswaran
- Arindam Mahanta
- Sanchay Cholkar
- Ask the app to “FeyeND” something
- Point camera in any direction and click an image
- App tells you if your object is in that direction
- Guides user to the object (planned)
The idea behind FeyeND is to assist visually challenged people to locate things in their surroundings easily. It is an interactive audio based app which uses natural language processing and computer vision. On request, speak out to the app
what you are looking for. The app uses the camera to
sense the surrounding. It sends the voice command to a language
processing API to identify which object the user is looking for. It sends the image to the Microsoft computer vision API to detect the objects in the image. It checks if the object asked by the
user is on the objects found by the CV API. It speaks out the response. - Progress so far
Portable Music Player For Blind Users
(A device, not to be confused with a mobile app)
- Team members:
Technology has developed to the point where you can do almost anything with your smartphone. However, the interfaces of smart phones are not very usable for blind and visually impaired people and the user experience for a blind user is totally not enjoyable. Starting from this awareness, I was motivated to develop a music player for blind people. A small research reveals how visually impaired people struggle with current devices needing need greater independence and freedom, and how there aren't any feature rich music player for them - supporting cloud services like Spotify and such. To fill the gap, I presented the concept of a portable music player which incorporates easy navigation features which are essential in a music player by the means of tactile buttons on the device.
Pushy! : Robotic arm as an accessibility extension
- Team members:
During the introductory phase of the “Accessathon”, my team consisting of my colleague, Deep Bhatt, and I, Husam Shakeeb, were motivated by the presentation
given by Mr. Christian Bayerlein. He was explaining all the accessibility functions and tools he has at his disposal with the exception of being able to
operate buttons or carry out simple tasks that are otherwise not possible from his chair or smartphone.
We worked closely with Mr. Bayerlein and explained the simple concept of using the wireless connection capability already available in his chair to attach
a mechanical arm that can be used to push buttons and in an advanced stage to add the ability to handle objects. He was excited about the idea and he even
chose the name for the project.
The idea was to build or acquire a mechanical arm that would be used to carry out the tasks and to build an interface between the arm and the smartphone
using an Arduino microcontroller and infrared or Bluetooth directly from the phone to send commands. During the “Accessathon” using the hardware available
in the lab we were able to demonstrate the concept by building the interface with a single motor and sending signals to operate the motor from Mr. Bayerlein’s
We still hope to receive the necessary hardware in order for the project to be completed, other students may also pick up where we left off in case we