Hardware-Software Codesign of a Large Vocabulary Continuous Speech Recognition system.

dc.contributor.advisorDr Winser E Alexander, Committee Memberen_US
dc.contributor.advisorDr Arne A Nilsson, Committee Memberen_US
dc.contributor.advisorDr Paul D Franzon, Committee Chairen_US
dc.contributor.authorJayadev, Viveken_US
dc.date.accessioned2010-04-02T17:53:32Z
dc.date.available2010-04-02T17:53:32Z
dc.date.issued2007-04-09en_US
dc.degree.disciplineElectrical Engineeringen_US
dc.degree.levelthesisen_US
dc.degree.nameMSen_US
dc.description.abstractModern real-time applications with increasing design complexity have revolutionized the embedded design procedure. Energy budget constraints and shortening time to market have led designers to consider cooperative design of hardware and software modules for a given embedded application. In hardware-software codesign the trade offs in both the domains are carefully analyzed and the processor intensive tasks are off-loaded to the hardware to meet the performance criteria while the rest is implemented in software to provide the required features and flexibility. Speech recognition systems used in real time applications involve complex algorithms for faithful recognition. The nature of these tasks restricts the implementation to large platforms and is not feasible to meet the performance constraints for smaller embedded mobile systems and battery operated devices. This thesis proposes an idea for hardware-software codesign of a Hidden Markov Model (HMM) based large vocabulary continuous speech recognition system. The entire procedure can be divided into three phases: the initial phase deals with the spectral analysis of the speech input, the second phase deals with learning of the sound units followed by the recognition phase. Studies have shown that the recognition phase consumes more than 50% of the processor time. Keeping this in mind, we partitioned our design to perform the spectral analysis and acoustic training in software using the front end executables and the acoustic trainers provided by the CMU SPHINX. The decoder implementing the phonetic detection and viterbi algorithm was designed in hardware. In this project we simulated different speech input files in software and the relevant input vector files required for hardware analysis were tapped from the SPHINX systemen_US
dc.identifier.otheretd-12142006-154914en_US
dc.identifier.urihttp://www.lib.ncsu.edu/resolver/1840.16/156
dc.rightsI hereby certify that, if appropriate, I have obtained and attached hereto a written permission statement from the owner(s) of each third party copyrighted matter to be included in my thesis, dis sertation, or project report, allowing distribution as specified below. I certify that the version I submitted is the same as that approved by my advisory committee. I hereby grant to NC State University or its agents the non-exclusive license to archive and make accessible, under the conditions specified below, my thesis, dissertation, or project report in whole or in part in all forms of media, now or hereafter known. I retain all other ownership rights to the copyright of the thesis, dissertation or project report. I also retain the right to use in future works (such as articles or books) all or part of this thesis, dissertation, or project report.en_US
dc.subjectCMU SPHINXen_US
dc.subjectSpeech Recognitionen_US
dc.subjectCodesignen_US
dc.titleHardware-Software Codesign of a Large Vocabulary Continuous Speech Recognition system.en_US

Files

Original bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
etd.pdf
Size:
490.1 KB
Format:
Adobe Portable Document Format

Collections