Jump to content

GSoC/2012/StatusReports: Difference between revisions

From KDE Community Wiki
Spacetime (talk | contribs)
Yash (talk | contribs)
Line 53: Line 53:


''' This is what I learned during GSoC: '''
''' This is what I learned during GSoC: '''
== Yash Shah ==


'''Project Name:''' Multimodal Accessibility: Using Computer Vision to improve Speech Recognition in Simon
'''A _Brief_ Description'''  A major obstacle for command and control speech recognition systems is to differentiate commands from background noise. Many systems solve this by using physical buttons or certain key phrases to activate/deactivate the speech recognition. This project explores the use of computer vision to determine when to activate / deactivate the sound recognition using visual cues. For a media centre or robot applications, it would make a lot more sense to only activate the recognition when the user is actively looking at the screen/robot and is speaking something. This is strikingly similar to the day-to-day communication between humans! Face recognition can also be employed to provide different speech models for different people. In this way the media centre could adapt to different people in one household.
[http://yashshah.com/blog/google-summer-of-code-2012-kde/ Checkout my blog post about it]
''' IRC Nick: ''' yashshah
''' IRC Channels: ''' #kde-accessibility
'''Blog: ''' http://yashshah.com/blog/
= SoK =
= SoK =

Revision as of 20:37, 8 June 2012

Status Report:GSoC and SoK 2012

GSoC

Puneet Goyal

Project Name: Enhancement to peer-to-peer DBus for Telepathy DBus Tubes

A _Brief_ Description Telepathy is a modular framework for real-time communications that handles voice, video, text, file transfer, and so on. It makes extensive use of the D-Bus messaging bus and a modular design. When an application connects to a peer to peer dbus tube, it must know what exactly to look for. Even When it registers for another object, the other side of the tube must know about it. So the ideas is to create a class that could ease the object to register and unregister on the DBus Tubes, and to provide you with an interface similar to the one as a DBus Server.

Status: Implementing org.freedesktop.DBus.Peer Interface and Adapter- almost complete

Screenshot:

IRC Nick: puneetgoyal

IRC Channels: #kde-telepathy

Blog: http://dcetech.com/puneet I am going to start the blog soon!

This is awesome about GSoC:

This is what I learned during GSoC:

Cyril Oblikov

Project Name: Asynchronous errors handling during file transfer

A _Brief_ Description The goal is to bring powerful, easy to use and suitable for both desktops and touch devices way to handle errors during copying or moving files. Since this project will be finished, errors will no more block copying process. The functionality for kdelibs is already done in last GSoC. You can see my blog posts about it: first, second.

Google+: munknex

Blog: http://munknex.net


Rishab Arora

Project Name: Improving Data Storage, Logs and adding DSO catalogs to KStars

A _Brief_ Description My proposal aims at improving the way the data is handled inside KStars. All user data which is currently being fed by multiple text files can be consolidated into a SQLite database which also makes the data more manageable and resolves many issues. This will also let me improve drastically on the ability to create portable logs about sky objects with a detailed interface and storage. However, in cases where reading the text files (comma/space/colon delimited) is necessary, I shall create separate provisions for processing them which is robust, tested and reliable. I shall also target the way the sky objects are stored and referenced and transfer all those objects to another SQLite database to make the data more manageable.

Status: Porting the User DB to SQLite and replacing old file usage.

Screenshot:

IRC Nick: spacetime

IRC Channels: #kde-edu #kstars #kde-in

Blog: http://blog.rishab.in/category/kde/

This is awesome about GSoC:

This is what I learned during GSoC:

Yash Shah

Project Name: Multimodal Accessibility: Using Computer Vision to improve Speech Recognition in Simon

A _Brief_ Description A major obstacle for command and control speech recognition systems is to differentiate commands from background noise. Many systems solve this by using physical buttons or certain key phrases to activate/deactivate the speech recognition. This project explores the use of computer vision to determine when to activate / deactivate the sound recognition using visual cues. For a media centre or robot applications, it would make a lot more sense to only activate the recognition when the user is actively looking at the screen/robot and is speaking something. This is strikingly similar to the day-to-day communication between humans! Face recognition can also be employed to provide different speech models for different people. In this way the media centre could adapt to different people in one household. Checkout my blog post about it

IRC Nick: yashshah

IRC Channels: #kde-accessibility

Blog: http://yashshah.com/blog/

SoK