TCS Innovation Labs, Thane
In TCS Innovation labs - Thane, people work on diverse research areas such as speech and natural language processing technologies, mobile-based social empowerment and performance engineering.
They seek researchers who will advance work in the core research areas, create R&I assets, and enhance the visibility of organizational capabilities. They offer an intellectually stimulating research environment to develop transformative, research-based solutions in order to address present and future business as well as technology opportunities.
## Speech Assessment Platform Speech assessment platform analyses some of primary soft skills of an individual through two speech signals i.e. read and spontaneous speech. It analyses individual speech signal to determine certain soft skill set.
Read Speech Analysis:
The above stated aspects of speech helps in determining how an individual expresses emotions and attitudes. It helps in evaluating his grammatical structure, and psychological functions in organizing speech that are easy to perceive and memorize.
Spontaneous Speech Analysis:
## Language Model for Human Resource Queries Language model for HR Queries is an end-to-end system which accepts audio speech signals (Employee queries to Human Resource department) as inputs, process them into a text format using Automatic Speech Recognition system. In next stage, text-data is processed using NLP techniques to identify appropriate query from the database and synthesize appropriate response based on context to respond back to user with audio output. Challenges observed in development of these end-to-end systems involves automatic speech recognition, which is realised by modelling the following components
In speech recognition, the acoustic model maps sound with a word. The language model provides context to distinguish between words and phrases that sound similar. In American English, the phrases “recognize speech” and “wreck a nice beach” are pronounced almost the same but mean very different things. These ambiguities can be resolved by incorporating the evidence from language model into the acoustic model.
Subsystems in Automatic Speech Recognition:
In Speech Assessment Platform, I have contributed in development of certain components in graphical user interface, pipelining various modules in centralised backend processing server, segmentating speech sentences for calculating Goodness of pronunciation, complete intonation analysis module, and testing the desired working of complete pipeline.
In this project, I have studied the effect of discounting techniques such as Witten-bell discounting and Kneser–Ney discounting in language modelling trained with 5500 Human Resource query dataset, which is categorised into 2 sets with 2 different lexicon.
NOTE: There is only limited information regarding the project implementation and no project code here due to TCS confidentiality and Intellectual property rights