GSoC/2012/StatusReports: Difference between revisions
Line 53: | Line 53: | ||
''' This is what I learned during GSoC: ''' | ''' This is what I learned during GSoC: ''' | ||
== Yash Shah == | |||
'''Project Name:''' Multimodal Accessibility: Using Computer Vision to improve Speech Recognition in Simon | |||
'''A _Brief_ Description''' A major obstacle for command and control speech recognition systems is to differentiate commands from background noise. Many systems solve this by using physical buttons or certain key phrases to activate/deactivate the speech recognition. This project explores the use of computer vision to determine when to activate / deactivate the sound recognition using visual cues. For a media centre or robot applications, it would make a lot more sense to only activate the recognition when the user is actively looking at the screen/robot and is speaking something. This is strikingly similar to the day-to-day communication between humans! Face recognition can also be employed to provide different speech models for different people. In this way the media centre could adapt to different people in one household. | |||
[http://yashshah.com/blog/google-summer-of-code-2012-kde/ Checkout my blog post about it] | |||
''' IRC Nick: ''' yashshah | |||
''' IRC Channels: ''' #kde-accessibility | |||
'''Blog: ''' http://yashshah.com/blog/ | |||
= SoK = | = SoK = |
Revision as of 20:37, 8 June 2012
Status Report:GSoC and SoK 2012
GSoC
Puneet Goyal
Project Name: Enhancement to peer-to-peer DBus for Telepathy DBus Tubes
A _Brief_ Description Telepathy is a modular framework for real-time communications that handles voice, video, text, file transfer, and so on. It makes extensive use of the D-Bus messaging bus and a modular design. When an application connects to a peer to peer dbus tube, it must know what exactly to look for. Even When it registers for another object, the other side of the tube must know about it. So the ideas is to create a class that could ease the object to register and unregister on the DBus Tubes, and to provide you with an interface similar to the one as a DBus Server.
Status: Implementing org.freedesktop.DBus.Peer Interface and Adapter- almost complete
Screenshot:
IRC Nick: puneetgoyal
IRC Channels: #kde-telepathy
Blog: http://dcetech.com/puneet I am going to start the blog soon!
This is awesome about GSoC:
This is what I learned during GSoC:
Cyril Oblikov
Project Name: Asynchronous errors handling during file transfer
A _Brief_ Description The goal is to bring powerful, easy to use and suitable for both desktops and touch devices way to handle errors during copying or moving files. Since this project will be finished, errors will no more block copying process. The functionality for kdelibs is already done in last GSoC. You can see my blog posts about it: first, second.
Google+: munknex
Blog: http://munknex.net
Rishab Arora
Project Name: Improving Data Storage, Logs and adding DSO catalogs to KStars
A _Brief_ Description My proposal aims at improving the way the data is handled inside KStars. All user data which is currently being fed by multiple text files can be consolidated into a SQLite database which also makes the data more manageable and resolves many issues. This will also let me improve drastically on the ability to create portable logs about sky objects with a detailed interface and storage. However, in cases where reading the text files (comma/space/colon delimited) is necessary, I shall create separate provisions for processing them which is robust, tested and reliable. I shall also target the way the sky objects are stored and referenced and transfer all those objects to another SQLite database to make the data more manageable.
Status: Porting the User DB to SQLite and replacing old file usage.
Screenshot:
IRC Nick: spacetime
IRC Channels: #kde-edu #kstars #kde-in
Blog: http://blog.rishab.in/category/kde/
This is awesome about GSoC:
This is what I learned during GSoC:
Yash Shah
Project Name: Multimodal Accessibility: Using Computer Vision to improve Speech Recognition in Simon
A _Brief_ Description A major obstacle for command and control speech recognition systems is to differentiate commands from background noise. Many systems solve this by using physical buttons or certain key phrases to activate/deactivate the speech recognition. This project explores the use of computer vision to determine when to activate / deactivate the sound recognition using visual cues. For a media centre or robot applications, it would make a lot more sense to only activate the recognition when the user is actively looking at the screen/robot and is speaking something. This is strikingly similar to the day-to-day communication between humans! Face recognition can also be employed to provide different speech models for different people. In this way the media centre could adapt to different people in one household. Checkout my blog post about it
IRC Nick: yashshah
IRC Channels: #kde-accessibility
Blog: http://yashshah.com/blog/