Настройки

Укажите год
-

Небесная энциклопедия

Космические корабли и станции, автоматические КА и методы их проектирования, бортовые комплексы управления, системы и средства жизнеобеспечения, особенности технологии производства ракетно-космических систем

Подробнее
-

Мониторинг СМИ

Мониторинг СМИ и социальных сетей. Сканирование интернета, новостных сайтов, специализированных контентных площадок на базе мессенджеров. Гибкие настройки фильтров и первоначальных источников.

Подробнее

Форма поиска

Поддерживает ввод нескольких поисковых фраз (по одной на строку). При поиске обеспечивает поддержку морфологии русского и английского языка
Ведите корректный номера.
Ведите корректный номера.
Ведите корректный номера.
Ведите корректный номера.
Укажите год
Укажите год

Применить Всего найдено 16134. Отображено 100.
09-02-2012 дата публикации

Speech search device and speech search method

Номер: US20120036159A1
Принадлежит: Toyohashi University of Technology NUC

Provided are a speech search device, the search speed of which is very fast, the search performance of which is also excellent, and which performs fuzzy search, and a speech search method. Not only the fuzzy search is performed, but also the distance between phoneme discrimination features included in speech data is calculated to determine the similarity with respect to the speech using both a suffix array and dynamic programming, and an object to be searched for is narrowed by means of search keyword division based on a phoneme and search thresholds relative to a plurality of the divided search keywords, the object to be searched for is repeatedly searched for while increasing the search thresholds in order, and whether or not there is the keyword division is determined according to the length of the search keywords, thereby implementing speech search, the search speed of which is very fast and the search performance of which is also excellent.

Подробнее
16-02-2012 дата публикации

Teaching aid

Номер: US20120040315A1
Автор: Peter Lawrence King
Принадлежит: UNICUS INVESTMENTS Pty Ltd

The present invention relates to the field of voice and speech recognition, in one form, the invention relates to a teaching aid adapted to teach reading and spelling via a voice and/or speech recognition system adapted to assist persons having dyslexia. The invention also provides a mechanism to train a speech recognition system without the need for the used to read verbose passages of text.

Подробнее
15-03-2012 дата публикации

System and method for pronunciation modeling

Номер: US20120065975A1
Принадлежит: AT&T INTELLECTUAL PROPERTY I LP

Systems, computer-implemented methods, and tangible computer-readable media for generating a pronunciation model. The method includes identifying a generic model of speech composed of phonemes, identifying a family of interchangeable phonemic alternatives for a phoneme in the generic model of speech, labeling the family of interchangeable phonemic alternatives as referring to the same phoneme, and generating a pronunciation model which substitutes each family for each respective phoneme. In one aspect, the generic model of speech is a vocal tract length normalized acoustic model. Interchangeable phonemic alternatives can represent a same phoneme for different dialectal classes. An interchangeable phonemic alternative can include a string of phonemes.

Подробнее
16-08-2012 дата публикации

Method And Background Estimator For Voice Activity Detection

Номер: US20120209604A1
Автор: Martin Sehlstedt
Принадлежит: Individual

The present invention relates to a method and a background estimator in voice activity detector for updating a background noise estimate for an input signal. The input signal for a current frame is received and it is determined whether the current frame of the input signal comprises non-noise. Further, an additional determination is performed whether the current frame of the non-noise input comprises noise by analyzing characteristics at least related to correlation and energy level of the input signal, and background noise estimate is updated if it is determined that the current frame comprises noise.

Подробнее
16-08-2012 дата публикации

Method and apparatus for information extraction from interactions

Номер: US20120209606A1
Принадлежит: Nice Systems Ltd

Obtaining information from audio interactions associated with an organization. The information may comprise entities, relations or events. The method comprises: receiving a corpus comprising audio interactions; performing audio analysis on audio interactions of the corpus to obtain text documents; performing linguistic analysis of the text documents; matching the text documents with one or more rules to obtain one or more matches; and unifying or filtering the matches.

Подробнее
23-08-2012 дата публикации

Sound Recognition Operation Apparatus and Sound Recognition Operation Method

Номер: US20120215537A1
Автор: Yoshihiro Igarashi
Принадлежит: Individual

According to one embodiment, a sound recognition operation apparatus includes a sound detection module, a keyword detection module, an audio mute module, and a transmission module. The sound detection module is configured to detect sound. The keyword detection module is configured to detect a particular keyword using voice recognition when the sound detection module detects sound. The audio mute module is configured to transmit an operation signal for muting audio sound when the keyword detection module detects the keyword. The transmission module is configured to recognize the voice command after the keyword is detected by the keyword detection module, and transmit an operation signal corresponding to the voice command.

Подробнее
30-08-2012 дата публикации

Network apparatus and methods for user information delivery

Номер: US20120221412A1
Автор: Robert F. Gazdzinski
Принадлежит: Individual

A network apparatus useful for providing directions and other information to a user of a client device in wireless communication therewith. In one embodiment, the apparatus includes one or more wireless interfaces and a network interface for communication with a server. User speech inputs in the form of digitized representations are received by the apparatus and used by the server as the basis for retrieving information including graphical representations of location or entities that the user wishes to find.

Подробнее
04-10-2012 дата публикации

System and method for rapid customization of speech recognition models

Номер: US20120253799A1
Принадлежит: AT&T INTELLECTUAL PROPERTY I LP

Disclosed herein are systems, methods, and non-transitory computer-readable storage media for generating domain-specific speech recognition models for a domain of interest by combining and tuning existing speech recognition models when a speech recognizer does not have access to a speech recognition model for that domain of interest and when available domain-specific data is below a minimum desired threshold to create a new domain-specific speech recognition model. A system configured to practice the method identifies a speech recognition domain and combines a set of speech recognition models, each speech recognition model of the set of speech recognition models being from a respective speech recognition domain. The system receives an amount of data specific to the speech recognition domain, wherein the amount of data is less than a minimum threshold to create a new domain-specific model, and tunes the combined speech recognition model for the speech recognition domain based on the data.

Подробнее
29-11-2012 дата публикации

Number-assistant voice input system, number-assistant voice input method for voice input system and number-assistant voice correcting method for voice input system

Номер: US20120303368A1
Автор: Ting Ma
Принадлежит: Mitac International Corp

The present invention discloses a number-assistant voice input system, a number-assistant voice input method for a voice input system and a number-assistant voice correcting method for a voice input system, which apply software to drive a voice input system of an electronic device to provide a voice input logic circuit module. The voice input logic circuit module defines the pronunciation of numbers 1 to 26 as the paths to respectively input letters A to Z in the voice input system and allows users to selectively input or correct a letter by reading a number from 1 to 26 instead of a letter from A to Z.

Подробнее
03-01-2013 дата публикации

Automatic Language Model Update

Номер: US20130006640A1
Принадлежит: Google LLC

A method for generating a speech recognition model includes accessing a baseline speech recognition model, obtaining information related to recent language usage from search queries, and modifying the speech recognition model to revise probabilities of a portion of a sound occurrence based on the information. The portion of a sound may include a word. Also, a method for generating a speech recognition model, includes receiving at a search engine from a remote device an audio recording and a transcript that substantially represents at least a portion of the audio recording, synchronizing the transcript with the audio recording, extracting one or more letters from the transcript and extracting the associated pronunciation of the one or more letters from the audio recording, and generating a dictionary entry in a pronunciation dictionary.

Подробнее
07-02-2013 дата публикации

Apparatus and method for recognizing voice

Номер: US20130035938A1
Автор: Ho Young JUNG

The present invention includes a hierarchical search process. The hierarchical search process includes three steps. In a first step, a word boundary is determined using a recognition method of determining a following word dependent on a preceding word, and a word boundary detector. In a second step, word unit based recognition is performed in each area by dividing an input voice into a plurality of areas based on the determined word boundary. Finally, in a third step, a language model is applied to induce an optimal sentence recognition result with respect to a candidate word that is determined for each area. The present invention may improve the voice recognition performance, and particularly, the sentence unit based consecutive voice recognition performance.

Подробнее
21-02-2013 дата публикации

Transcript editor

Номер: US20130047059A1
Принадлежит: Avid Technology Inc

A transcript editor enables text-based editing of time-based media that includes spoken dialog. It involves an augmented transcript that includes timing metadata that associates words and phrases within the transcript with corresponding temporal locations within the time-based media where the text is spoken, and editing the augmented transcript without the need for playback of the time-based media. After editing, the augmented transcript is processed by a media editing system to automatically generate an edited version of the time-based media that only includes the segments of the time-based media that include the speech corresponding to the edited augmented transcript.

Подробнее
14-03-2013 дата публикации

Apparatus and method for generating vocal organ animation

Номер: US20130065205A1
Автор: Bong-Rae Park
Принадлежит: CLUSOFT CO Ltd

The present disclosure relates to an apparatus and method for generating a vocal organ animation very similar to a pronunciation pattern of a native speaker in order to support foreign language pronunciation education. The present disclosure checks an adjacent phonetic value in phonetic value constitution information, extracts a detail phonetic value based on the adjacent phonetic value, extracting pronunciation pattern information corresponding to the detail phonetic value and pronunciation pattern information corresponding to a transition section allocated between detail phonetic values, and performs interpolation to the extracted pronunciation pattern information, thereby generating a vocal organ animation.

Подробнее
18-04-2013 дата публикации

Recognizing device, computer-readable recording medium, recognizing method, generating device, and generating method

Номер: US20130096918A1
Автор: Shouji Harada
Принадлежит: Fujitsu Ltd

A recognizing device includes a memory and a processor coupled to the memory. The memory stores words included in a sentence and positional information indicating a position of the words in the sentence. The processor executes a process including comparing an input voice signal with reading information of a character string that connects a plurality of words stored in the memory to calculate a similarity; calculating a connection score indicating a proximity between the plurality of connected words based on positional information of the words stored in the memory; and determining a character string corresponding to the voice signal based on the similarity and the connection score.

Подробнее
23-05-2013 дата публикации

System and method for crowd-sourced data labeling

Номер: US20130132080A1
Принадлежит: AT&T INTELLECTUAL PROPERTY I LP

Disclosed herein are systems, methods, and non-transitory computer-readable storage media for crowd-sourced data labeling. The system requests a respective response from each of a set of entities. The set of entities includes crowd workers. Next, the system incrementally receives a number of responses from the set of entities until at least one of an accuracy threshold is reached and m responses are received, wherein the accuracy threshold is based on characteristics of the number of responses. Finally, the system generates an output response based on the number of responses.

Подробнее
30-05-2013 дата публикации

Method and system for generating search network for voice recognition

Номер: US20130138441A1

Disclosed is a method of generating a search network for voice recognition, the method including: generating a pronunciation transduction weighted finite state transducer by implementing a pronunciation transduction rule representing a phenomenon of pronunciation transduction between recognition units as a weighted finite state transducer; and composing the pronunciation transduction weighted finite state transducer and one or more weighted finite state transducers.

Подробнее
25-07-2013 дата публикации

Computerized information and display apparatus

Номер: US20130191750A1
Автор: Robert F. Gazdzinski
Принадлежит: West View Research LLC

Apparatus useful for obtaining and displaying information. In one embodiment, the apparatus includes a network interface, display device, and speech recognition apparatus configured to receive user speech input and enable performance of various tasks via a remote entity, such as obtaining desired information relating to directions, sports, finance, weather, or any number of other topics. The downloaded may also, in one variant, be transmitted to a personal user device, such as via a data interface.

Подробнее
08-08-2013 дата публикации

Enhanced context awareness for speech recognition

Номер: US20130204622A1
Автор: Mika Grundstrom, Wenhui Lu
Принадлежит: Nokia Oyj

A method comprising establishing a call connection ( 200 ) between at least a first and a second terminal; monitoring ( 202 ), by at least the first terminal, a conversation during the call in order to detect ( 204 ) at least one predetermined context-related keyword repeated in at least the first and the second terminal; and in response to detecting ( 210 ) at least one repeated predetermined context-related keyword, providing an indication ( 212, 214, 216 ) about the detected context-related keyword to a user of at least the first terminal, said indication enabling opening an application linked to said context-related keyword.

Подробнее
31-10-2013 дата публикации

Negative Example (Anti-Word) Based Performance Improvement For Speech Recognition

Номер: US20130289987A1
Принадлежит: Interactive Intelligence Inc

A system and method are presented for negative example based performance improvements for speech recognition. The presently disclosed embodiments address identified false positives and the identification of negative examples of keywords in an Automatic Speech Recognition (ASR) system. Various methods may be used to identify negative examples of keywords. Such methods may include, for example, human listening and learning possible negative examples from a large domain specific text source. In at least one embodiment, negative examples of keywords may be used to improve the performance of an ASR system by reducing false positives.

Подробнее
06-02-2014 дата публикации

Speech recognition models based on location indicia

Номер: US20140039888A1
Принадлежит: Google LLC

Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for performing speech recognition using models that are based on where, within a building, a speaker makes an utterance are disclosed. The methods, systems, and apparatus include actions of receiving data corresponding to an utterance, and obtaining location indicia for an area within a building where the utterance was spoken. Further actions include selecting one or more models for speech recognition based on the location indicia, wherein each of the selected one or more models is associated with a weight based on the location indicia. Additionally, the actions include generating a composite model using the selected one or more models and the respective weights of the selected one or more models. And the actions also include generating a transcription of the utterance using the composite model.

Подробнее
13-03-2014 дата публикации

Method and System for Building a Phonotactic Model for Domain Independent Speech Recognition

Номер: US20140074476A1
Автор: Giuseppe Riccardi
Принадлежит: AT&T Intellectual Property II LP

The invention concerns a method and corresponding system for building a phonotactic mode for domain independent speech recognition. The method may include recognizing phones from a user's input communication using a current phonotactic model detecting morphemes (acoustic and/or non-acoustic) from the recognized phones, and outputting die detected morphemes for processing. The method also updates the phonotactic model with the detected morphemes and stores the new model in a database for use by the system daring the next user interaction. The method may also include making task-type classification decisions based on the detected morphemes from the user's input communication.

Подробнее
06-01-2022 дата публикации

SYSTEM AND METHOD OF GENERATING FACIAL EXPRESSION OF A USER FOR VIRTUAL ENVIRONMENT

Номер: US20220005246A1
Принадлежит:

The present invention relates to a method of generating a facial expression of a user for a virtual environment. The method comprises obtaining a video and an associated speech of the user. Further, extracting in real-time at least one of one or more voice features and one or more text features based on the speech. Furthermore, identifying one or more phonemes in the speech. Thereafter, determining one or more facial features relating to the speech of the user using a pre-trained second learning model based on the one or more voice features, the one or more phonemes, the video and one or more previously generated facial features of the user. Finally, generating the facial expression of the user corresponding to the speech for an avatar representing the user in the virtual environment. 1. A method of generating a facial expression of a user for a virtual environment , the method comprises:obtaining, by a computing system, a video and an associated speech of the user;extracting in real-time, by the computing system, at least one of one or more voice features and one or more text features based on the speech of the user;identifying in real-time, by the computing system, one or more phonemes in the speech using a pre-trained first learning model based on at least one of the one or more voice features and the one or more text features;determining in real-time, by the computing system, one or more facial features relating to the speech of the user using a pre-trained second learning model based on the one or more voice features, the one or more phonemes, the video and one or more previously generated facial features of the user; andgenerating in real-time, by the computing system, the facial expression of the user corresponding to the speech for an avatar representing the user in the virtual environment based on the one or more facial features.2. The method as claimed in claim 1 , wherein obtaining the video and the associated speech comprises one of:receiving the video ...

Подробнее
06-01-2022 дата публикации

ON-DEVICE SPEECH SYNTHESIS OF TEXTUAL SEGMENTS FOR TRAINING OF ON-DEVICE SPEECH RECOGNITION MODEL

Номер: US20220005458A1
Принадлежит:

Processor(s) of a client device can: identify a textual segment stored locally at the client device; process the textual segment, using a speech synthesis model stored locally at the client device, to generate synthesized speech audio data that includes synthesized speech of the identified textual segment; process the synthesized speech, using an on-device speech recognition model that is stored locally at the client device, to generate predicted output; and generate a gradient based on comparing the predicted output to ground truth output that corresponds to the textual segment. In some implementations, the generated gradient is used, by processor(s) of the client device, to update weights of the on-device speech recognition model. In some implementations, the generated gradient is additionally or alternatively transmitted to a remote system for use in remote updating of global weights of a global speech recognition model. 1. A client device comprising:at least one microphone;at least one display;local storage storing a textual segment, an end-to-end speech recognition model, and a speech synthesis model; identify the textual segment;', 'generate synthesized speech audio data that includes synthesized speech of the identified textual segment, wherein in generating the synthesized speech audio data one or more of the processors are to process the textual segment using the speech synthesis model;', 'process, using the end-to-end speech recognition model, the synthesized speech audio data to generate a predicted textual segment;', 'generate a gradient based on comparing the predicted textual segment to the textual segment; and', 'update one or more weights of the end-to-end speech recognition model based on the generated gradient., 'one or more processors executing locally stored instructions to cause one or more of the processors to2. The client device of claim 1 , further comprising:a network interface; 'transmit, via the network interface to a remote system, the ...

Подробнее
06-01-2022 дата публикации

KEYWORD DETECTION APPARATUS, KEYWORD DETECTION METHOD, AND PROGRAM

Номер: US20220005466A1

A keyword is extracted robustly despite a voice recognition result including an error. A model storage unit stores a keyword extraction model that accepts word vector representations of a plurality of words as an input and extracts and outputs a word vector representation of a word to be extracted as a keyword. A speech detection unit detects a speech part from a voice signal. A voice recognition unit executes voice recognition on the speech part of the voice signal and outputs a confusion network which is a voice recognition result. A word vector representation generating unit generates a word vector representation including reliability of voice recognition with regard to each candidate word for each confusion set. A keyword extraction unit inputs the word vector representation of the candidate word to the keyword extraction model in descending order of the reliability and obtains the word vector representation of the keyword. 1. A keyword extraction device comprising:a model store configured to store a keyword extraction model that accepts word vector representations of a plurality of words as an input and extracts and outputs a word vector representation of a word to be extracted as a keyword;a voice recognize configured to execute voice recognition on an input voice signal and output a confusion network which is a voice recognition result;a word vector representation generator configured to generate a word vector representation including at least reliability of voice recognition with regard to each candidate word included in each confusion set included in the confusion network; anda keyword extractor configured to input the word vector representation of the candidate word included in each confusion set to the keyword extraction model in descending order of the reliability and obtain the word vector representation of the word to be extracted as the keyword.2. The keyword extraction device according to claim 1 , wherein the keyword extraction model outputs a ...

Подробнее
06-01-2022 дата публикации

MULTI-LOOK ENHANCEMENT MODELING AND APPLICATION FOR KEYWORD SPOTTING

Номер: US20220005468A1
Автор: Yu Dong, Yu Meng
Принадлежит: Tencent America LLC

A method, computer system, and computer readable medium are provided for activating speech recognition based on keyword spotting (KWS). Waveform data corresponding to one or more speakers is received. One or more direction features are extracted from the received waveform data. One or more keywords are determined from the received waveform data based on the one or more extracted features. Speech recognition is activated based on detecting the determined keyword. 1. A method of activating speech recognition , executable by a processor , comprising:receiving waveform data corresponding to one or more speakers;extracting one or more direction features from the received waveform data;determining one or more keywords from the received waveform data based on the one or more extracted features; andactivating speech recognition based on detecting the determined keywords.2. The method of claim 1 , wherein the received waveform data comprises one or more multi-channel input waveforms and one or more preset look directions.3. The method of claim 2 , further comprising mapping the multi-channel input waveforms to one or more spectrograms by a 1-D convolution layer.4. The method of claim 3 , further comprising extracting a single-channel spectral feature claim 3 , logarithm power spectrum data claim 3 , and one or more multi-channel spatial features based on the spectrograms.5. The method of claim 2 , wherein each extracted direction feature corresponds to one of the one or more preset look directions.6. The method of claim 5 , wherein a set of directions in the horizontal plane is sampled based on the direction features.7. The method of claim 6 , wherein the one or more direction feature vectors are derived from the set of directions in the horizontal plane.8. The method of claim 7 , wherein the directional feature computes the averaged cosine distance between a look direction steering vector between one or more pairs of microphones based on a phase of the look direction ...

Подробнее
05-01-2017 дата публикации

Speech recognition apparatus, speech recognition method, and electronic device

Номер: US20170004824A1
Принадлежит: SAMSUNG ELECTRONICS CO LTD

A speech recognition apparatus includes a probability calculator configured to calculate phoneme probabilities of an audio signal using an acoustic model; a candidate set extractor configured to extract a candidate set from a recognition target list; and a result returner configured to return a recognition result of the audio signal based on the calculated phoneme probabilities and the extracted candidate set.

Подробнее
05-01-2017 дата публикации

SMART HOME APPLIANCES, OPERATING METHOD OF THEREOF, AND VOICE RECOGNITION SYSTEM USING THE SMART HOME APPLIANCES

Номер: US20170004828A1
Принадлежит:

Provided is a smart home appliance. The smart home appliance includes: a voice input unit collecting a voice; a voice recognition unit recognizing a text corresponding to the voice collected through the voice input unit; a capturing unit collecting an image for detecting a user's visage or face; a memory unit mapping the text recognized by the voice recognition unit and a setting function and storing the mapped information; and a control unit determining whether to perform a voice recognition service on the basis of at least one information of image information collected by the capturing unit and voice information collected by the voice input unit. 1. A smart home appliance comprising:a voice input unit collecting a voice;a voice recognition unit recognizing a text corresponding to the voice collected through the voice input unit;a capturing unit collecting an image for detecting a user's visage;a memory unit mapping the text recognized by the voice recognition unit and a setting function and storing the mapped information; anda control unit determining whether to perform a voice recognition service on the basis of at least one information of image information collected by the capturing unit and voice information collected by the voice input unit.2. The smart home appliance according to claim 1 , wherein the control unit comprises a face detection unit recognizing that a user is in a staring state for voice input when image information on a user's visage is collected for more than a setting time through the capturing unit.3. The smart home appliance according to claim 2 , wherein the control unit determines that a voice recognition service standby state is entered when it is recognized that there is keyword information in a voice through the voice input unit and a user is in the staring state through the face detection unit.4. The smart home appliance according to claim 1 , further comprising:a filter unit removing a noise sound from the voice inputted through the ...

Подробнее
05-01-2017 дата публикации

SYSTEMS AND METHODS FOR SOURCE SIGNAL SEPARATION

Номер: US20170004844A1
Принадлежит:

A method includes receiving an input signal comprising an original domain signal and creating a first window data set and a second window data set from the signal, wherein an initiation of the second window data set is offset from an initiation of the first window data set, converting the first window data set and the second window data set to a frequency domain and storing the resulting data as data in a second domain different from the original domain, performing complex spectral phase evolution (CSPE) on the second domain data to estimate component frequencies of the first and second window data sets, using the component frequencies estimated in the CSPE, sampling a set of second-domain high resolution windows to select a mathematical representation comprising a second-domain high resolution window that fits at least one of the amplitude, phase, amplitude modulation and frequency modulation of a component of an underlying signal wherein the component comprises at least one oscillator peak, generating an output signal from the mathematical representation of the original signal as at least one of: an audio file; one or more audio signal components; and one or more speech vectors and outputting the output signal to an external system. 1. A method comprising:receiving a biomedical signal; and computing a plurality of oscillator peaks for the biomedical signal;', 'tracking one or more of the plurality of oscillator peaks to form one or more tracklets; and', 'grouping the one or more tracklets based, at least in part, on at least one characteristic harmonic of the one or more tracklets to create a first fingerprint., 'creating a fingerprint of the biomedical signal wherein said creating comprises the steps of2. The method of claim 1 , further comprising comparing the first fingerprint to a previously computed second fingerprint.3. The method of wherein the comparing further comprises comparing the first fingerprint to the previously computed second fingerprint to ...

Подробнее
07-01-2021 дата публикации

KEYPHRASE EXTRACTION BEYOND LANGUAGE MODELING

Номер: US20210004439A1
Принадлежит: Microsoft Technology Licensing, LLC

A system for extracting a key phrase from a document includes a neural key phrase extraction model (“BLING-KPE”) having a first layer to extract a word sequence from the document, a second layer to represent each word in the word sequence by ELMo embedding, position embedding, and visual features, and a third layer to concatenate the ELMo embedding, the position embedding, and the visual features to produce hybrid word embeddings. A convolutional transformer models the hybrid word embeddings to n-gram embeddings, and a feedforward layer converts the n-gram embeddings into a probability distribution over a set of n-grams and calculates a key phrase score of each n-gram. The neural key phrase extraction model is trained on annotated data based on a labeled loss function to compute cross entropy loss of the key phrase score of each n-gram as compared with a label from the annotated dataset. 1. A system for extracting a key phrase from a document , the system comprising:a processor and a memory, wherein a program executing on the processor is configured to run a neural key phrase extraction model having:a first layer to extract a word sequence from the document;a second layer to represent each word in the word sequence by word embedding and visual features;a third layer to concatenate the word embedding and the visual features to produce hybrid word embeddings;a convolutional transformer to model the hybrid word embeddings to n-gram embeddings; anda feedforward layer to convert the n-gram embeddings into a probability distribution over a set of n-grams and calculate a key phrase score of each n-gram,wherein the neural key phrase extraction model was trained on annotated data based on a labeled loss function to compute loss of the key phrase score of each n-gram as compared with a label from the annotated dataset, andwherein the key phrase score of each n-gram is used in a computer-implemented application.2. The system of claim 1 , wherein the labeled loss function is ...

Подробнее
03-01-2019 дата публикации

METHOD AND DEVICE FOR SWITCHING INPUT MODES

Номер: US20190004676A1
Автор: DENG Yuanyuan
Принадлежит:

An input mode method includes providing a first input mode of an application program which receives a first type of information from a user and detecting whether the first type of information comprises a pre-configured condition for switching the first input mode of the application program to a second input mode of the application program. The second input mode of the application program receives a second type of information from the user. The method further includes, if the first type of information comprises the pre-configured condition, switching the first input mode of the application program to the second input mode of the application program. 1. An input mode method comprising:providing a first input mode of an application program which receives a first type of information from a user;detecting whether the first type of information comprises a pre-configured condition for switching the first input mode of the application program to a second input mode of the application program, wherein the second input mode of the application program receives a second type of information from the user; andif so, switching the first input mode of the application program to the second input mode of the application program.2. The method according to claim 1 , wherein:providing the first input mode includes providing an audio input mode for the application program and displaying an input interface of the audio input mode on a display interface of a mobile terminal; andswitching to the second input mode includes switching to an image input mode for the application program and displaying an input interface of the image input mode on the display interface.3. The method according to claim 2 , wherein detecting whether the first type of information comprises a pre-configured condition includes:detecting, within a preset period of time, whether the first type of input information is not audio information.4. The method according to claim 2 , wherein:the first type of information ...

Подробнее
01-01-2015 дата публикации

APPARATUS AND METHOD FOR RECOGNIZING CONTINUOUS SPEECH

Номер: US20150006175A1

The present invention relates to an apparatus and a method for recognizing continuous speech having large vocabulary. In the present invention, large vocabulary in large vocabulary continuous speech having a lot of same kinds of vocabulary is divided to a reasonable number of clusters, then representative vocabulary for pertinent clusters is selected and first recognition is performed with the representative vocabulary, then if the representative vocabulary is recognized by use of the result of first recognition, re-recognition is performed against all words in the cluster where the recognized representative vocabulary belongs. 1. An apparatus for recognizing continuous speech , comprising:a cluster creation portion configured to create clusters from continuous speech, each of the clusters including at least one word;a representative vocabulary extraction portion configured to extract at least one representative word from each of the clusters;a continuous speech primary recognition portion configured to recognize the continuous speech primarily based on the extracted representative words and produce a recognition result; anda continuous speech final recognition portion configured to recognize the continuous speech finally based on the produced recognition result.2. The apparatus of claim 1 , wherein the cluster creation portion is configured to create a smaller number of clusters than the number of words included in the continuous speech.3. The apparatus of claim 1 , wherein the cluster creation portion comprises:a pronunciation array extraction portion configured to extract a pronunciation array from each word; anda quantization portion configured to create the clusters from the continuous speech according to vector quantization by using the extracted pronunciation array as a vector.4. The apparatus of claim 1 , wherein the representative vocabulary extraction portion is configured to extract the representative word according to a probability of appearance of words ...

Подробнее
07-01-2016 дата публикации

GENERATING COMPUTER RESPONSES TO SOCIAL CONVERSATIONAL INPUTS

Номер: US20160005395A1
Принадлежит:

Conversational interactions between humans and computer systems can be provided by a computer system that classifies an input by conversation type, and provides human authored responses for conversation types. The input classification can be performed using trained binary classifiers. Training can be performed by labeling inputs as either positive or negative examples of a conversation type. Conversational responses can be authored by the same individuals that label the inputs used in training the classifiers. In some cases, the process of training classifiers can result in a suggestion of a new conversation type, for which human authors can label inputs for a new classifier and write content for responses for that new conversation type. 1. A process for processing a conversational input , comprising:receiving input data representing a conversational input into memory;causing feature data derived from the input data to be applied to a plurality of classifiers, each classifier representing a conversation type from among a plurality of conversation types and outputting match data in response to input feature data; andselecting a conversational response according to one or more conversation types from among the plurality of conversation types according to the match data output by the classifiers.2. The process of claim 1 , wherein the features include n-grams generated from the conversational input.3. The process of claim 2 , wherein the features include data indicative of emotion related to the conversational input.4. The process of claim 2 , wherein the features include data indicative of one or more classes for subject matter of the conversational input.5. The process of claim 1 , wherein each classifier is a binary classifier such that the match data output in response to the feature data represents a probability that the feature data matches a conversational input labeled as being a positive example of the conversation type associated with the binary classifier.6. ...

Подробнее
07-01-2016 дата публикации

METHOD AND SYSTEM FOR EFFICIENT SPOKEN TERM DETECTION USING CONFUSION NETWORKS

Номер: US20160005398A1
Принадлежит:

Systems and methods for spoken term detection are provided. A method for spoken term detection, comprises receiving phone level out-of-vocabulary (OOV) keyword queries, converting the phone level OOV keyword queries to words, generating a confusion network (CN) based keyword searching (KWS) index, and using the CN based KWS index for both in-vocabulary (IV) keyword queries and the OOV keyword queries. 1. A method for spoken term detection , comprising:receiving phone level out-of-vocabulary (OOV) keyword queries;converting the phone level OOV keyword queries to words;generating a confusion network (CN) based keyword searching (KWS) index; andusing the CN based KWS index for both in-vocabulary (IV) keyword queries and the OOV keyword queries, wherein the receiving, converting, generating and using steps are performed by a computer system comprising a memory and at least one processor coupled to the memory.2. The method according to claim 1 , wherein generating the CN based KWS index comprises constructing the CN based KWS index from a plurality of confusion networks by compiling each confusion network into a weighted finite state transducer having the same topology as the confusion network.3. The method according to claim 2 , wherein each weighted finite state transducer includes input labels that are words on each arc in the corresponding confusion network.4. The method according to claim 2 , wherein each weighted finite state transducer includes output labels that encode a start time (T start) and an end time (T end) of each arc in the corresponding confusion network as T start-T end strings.5. The method according to claim 2 , wherein each weighted finite state transducer includes costs that are negative log CN posteriors for each arc in the confusion network.6. The method according to claim 2 , wherein for each weighted finite state transducer claim 2 , the method further comprises adding a new start node claim 2 , Swith zero-cost epsilon-arcs connecting Sto each ...

Подробнее
13-01-2022 дата публикации

TRANSFERRING AN AUTOMATED ASSISTANT ROUTINE BETWEEN CLIENT DEVICES DURING EXECUTION OF THE ROUTINE

Номер: US20220013121A1
Автор: Ni Yuzhao
Принадлежит:

Transferring (e.g., automatically) an automated assistant routine between client devices during execution of the automated assistant routine. The automated assistant routine can correspond to a set of actions to be performed by one or more agents and/or one or more devices. While content, corresponding to an action of the routine, is being rendered at a particular device, the user may walk away from the particular device and toward a separate device. The automated assistant routine can be automatically transferred in response, and the separate device can continue to rendering the content for the user. 1. A method implemented by one or more processors , the method comprising:receiving, at a remote server device, data transmitted from a first client device;determining, at the remote server device, that the data corresponds to a request for initialization of an automated assistant routine that corresponds to a set of automated assistant actions;in response to determining that the data corresponds to the request, generating, at the remote server device, content for an action of the set of automated assistant actions; 'transmitting the content for the action to the first client device to cause the first client device to render the content for the action;', 'in response to the data that corresponds to the request for initialization of the automated assistant routine being received from the first client devicedetermining, at the remote server device during rendering of the content for the action by the first client device, that a user has directly or indirectly indicated an interest in the automated assistant routine being continued at a second client device; and 'rendering, at the second client device, additional data that is in furtherance of the automated assistant routine.', 'in response to determining that the user has indicated the interest in the automated assistant routine being continued at the second client device2. The method of claim 1 , wherein in response to ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR INTERVIEW TRAINING WITH TIME-MATCHED FEEDBACK

Номер: US20210004768A1
Принадлежит:

The present disclosure generally relates to interview training and providing interview feedback. An exemplary method comprises: at an electronic device that is in communication with a display and one or more input devices: receiving, via the one or more input devices, media data corresponding to a user's responses to a plurality of prompts; analyzing the media data; and while displaying, on the display, a media representation of the media data, displaying a plurality of analysis representations overlaid on the media representation, wherein each of the plurality of analysis representations is associated with an analysis of content located at a given time in the media representation and is displayed in coordination with the given time in the media representation. 115-. (canceled)16. A computer-implemented method , comprising:recording data comprising speech by a user;analyzing the recorded data to calculate a metric related to the speech;determining that the calculated metric exceeds a predefined threshold; an interactive representation of the recorded data for providing a playback of the recorded data;', 'the calculated metric; and', 'an indication of whether the calculated metric exceeds the predefined threshold., 'rendering a user interface comprising17. The method of claim 16 , further comprising: rendering an overall analysis score claim 16 , wherein the overall analysis score is at least partially based on the calculated metric.18. The method of claim 16 , wherein the recorded data comprises audio data claim 16 , image data claim 16 , video data claim 16 , or any combination thereof.19. The method of claim 16 , wherein the metric related to the speech comprises a number of words per minute.20. The method of claim 16 , wherein the indication of whether the calculated metric exceeds the predefined threshold comprises an indication of whether the speech is of an appropriate talking speed.21. The method of claim 16 , wherein the indication of whether the calculated ...

Подробнее
07-01-2021 дата публикации

System and method for automated agent assistance within a cloud-based contact center

Номер: US20210004817A1
Принадлежит: Talkdesk Inc

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing.

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR SMS AND EMAIL ENABLED AUTOMATED AGENT ASSISTANCE WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004819A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:receiving a SMS text communication from a customer;performing inference processing on the SMS text to determine a customer intent;automatically analyzing the SMS text to determine a subject of the SMS text communication and key terms associated with the subject;automatically parsing a knowledgebase using the key terms for at least one responsive answer associated with the subject; andproviding the solution to an agent in a unified interface during the communication with the customer.21. The method of , further comprising:querying a customer relationship management (CRM) platform/a customer service management (CSM) platform using the key terms; anddisplaying responsive results from the CRM/CSM in the second field in the unified interface.3. The method of claim 1 , further comprising:querying a database of customer-agent transcripts using the key terms; anddisplaying responsive results from the database of customer-agent transcripts in the second field in the unified interface.4. The method of claim 1 , wherein the method is performed in real-time as the customer communication progresses with the agent.5. The method of claim 1 , further comprising concurrently displaying to the agent the text in a first field of the unified interface and the solution in a second field of the unified interface.6. The method of claim 1 , further comprising authenticating the customer claim 1 , predicting a customer gender claim 1 , predicting a customer age category claim 1 , predicting a customer accent claim 1 , and predict ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR AUTOMATED AGENT ASSISTANCE NEXT ACTIONS WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004820A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:receiving a communication from a customer;automatically analyzing the communication to determine a subject for the customer's communication;automatically parsing a knowledgebase for at least one responsive answer to the subject associated with the customer's communication;providing the solution to an agent as a clickable link within a unified interface during the communication with the customer; andreceiving selection of the clickable link to perform a subsequent action in response to the selection.2. The method of claim 1 , wherein the communication is in textual form claim 1 , the method further comprising:displaying text input by the customer in a first field of a unified interface;parsing the text input by the customer for key terms;querying the knowledgebase using the key terms; anddisplaying responsive results from the knowledgebase as the solution in a second field in the unified interface.3. The method of claim 2 , wherein the method is performed in real-time.42. The method of claim 2 , further comprising:querying a customer relationship management (CRM) platform/a customer service management (CSM) platform using the key terms; anddisplaying responsive results from the CRM/CSM in the second field in the unified interface.5. The method of claim 2 , further comprising:querying a database of customer-agent transcripts using the key terms; anddisplaying responsive results from the database of customer-agent transcripts in the second field in the unified interface.6. The method of claim 1 , further comprising ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR AUTOMATED AGENT ASSISTANCE NEXT ACTIONS WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004821A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:executing an automation infrastructure within a cloud-based contact center that includes a communication manager, speech-to-text converter, a natural language processor, and an inference processor exposed by application programming interfaces; andexecuting an agent assist functionality within the automation infrastructure that performs operations comprising:receiving a communication from a customer;automatically analyzing the communication to determine a subject for the customer's communication;automatically parsing a knowledgebase for at least one responsive answer to the subject associated with the customer's communication;providing the solution to an agent as a clickable link within a unified interface during the communication with the customer; andreceiving selection of the clickable link to perform a subsequent action in response to the selection.2. The method of claim 1 , wherein the communication is in textual form claim 1 , the method further comprising:displaying text input by the customer in a first field of a unified interface;parsing the text input by the customer for key terms;querying the knowledgebase using the key terms; anddisplaying responsive results from the knowledgebase as the solution in a second field in the unified interface.3. The method of claim 2 , wherein the method is performed in real-time.42. The method of claim 2 , further comprising:querying a customer relationship management (CRM) platform/a customer service management (CSM) platform using the key terms; anddisplaying ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR AUTOMATED SMART NOTES USING AGENT ASSIST WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004822A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:receiving a communication from a customer;displaying text associated with the communication in a first field of a unified user interface;automatically analyzing the text to determine key terms associated the communication; andautomatically populating the key terms into a notes field of a call notes user interface.2. The method of claim 1 , further comprising:querying one of a knowledgebase, a customer relationship management (CRM) platform/a customer service management (CSM) platform, and a database of customer-agent transcripts using the key terms; anddisplaying responsive results to the querying in a second field in the unified interface.3. The method of claim 2 , further comprising automatically populating query results into the notes field of the call notes user interface.4. The method of claim 3 , further comprising providing an editor in the call notes user interface to receive edits from the agent of the automatically populated key terms and query results.5. The method of claim 3 , wherein the populated key terms and query results form a call summary.6. The method of claim 1 , further comprising highlighting the key terms in the unified interface.7. The method of claim 1 , further comprising:receiving the communication as speech;converting the speech to text; andparsing the text for key terms.8. The method of claim 1 , wherein the communication is a multi-channel communication and received as one of an SMS text claim 1 , voice call claim 1 , e-mail claim 1 , chat claim 1 , interactive voice response (IVR ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR AUTOMATED SMART NOTES USING AGENT ASSIST WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004823A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:executing an automation infrastructure within a cloud-based contact center that includes a communication manager, speech-to-text converter, a natural language processor, and an inference processor exposed by application programming interfaces; andexecuting an agent assist functionality within the automation infrastructure that performs operations comprising:receiving a communication from a customer;displaying text associated with the communication in a first field of a unified user interface;automatically analyzing the text to determine key terms associated the communication; andautomatically populating the key terms into a notes field of a call notes user interface.2. The method of claim 1 , further comprising:querying one of a knowledgebase, a customer relationship management (CRM) platform/a customer service management (CSM) platform, and a database of customer-agent transcripts using the key terms; anddisplaying responsive results to the querying in a second field in the unified interface.3. The method of claim 2 , further comprising automatically populating query results into the notes field of the call notes user interface.4. The method of claim 3 , further comprising providing an editor in the call notes user interface to receive edits from the agent of the automatically populated key terms and query results.5. The method of claim 3 , wherein the populated key terms and query results form a call summary.6. The method of claim 1 , further comprising highlighting the key terms in the unified interface.7. ...

Подробнее
07-01-2021 дата публикации

System and method for automated scheduling using agent assist within a cloud-based contact center

Номер: US20210004824A1
Принадлежит: Talkdesk Inc

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing.

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR AUTOMATED SCHEDULING USING AGENT ASSIST WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004825A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:executing an automation infrastructure within a cloud-based contact center that includes a communication manager, speech-to-text converter, a natural language processor, and an inference processor exposed by application programming interfaces; andexecuting an agent assist functionality within the automation infrastructure that performs operations comprising:receiving a communication from a customer;displaying text associated with the communication in a first field of a unified user interface;automatically analyzing the text to determine key terms associated the communication; andautomatically using the key terms to schedule a subsequent interaction with the customer.2. The method of claim 1 , further comprising:querying one of a knowledgebase, a customer relationship management (CRM) platform/a customer service management (CSM) platform, and a database of customer-agent transcripts using the key terms; anddisplaying responsive results to the querying in a second field in the unified interface.3. The method of claim 2 , further comprising automatically using query results to schedule the subsequent interaction.4. The method of claim 1 , further comprising:receiving the communication as speech;converting the speech to text; andparsing the text for key terms.5. The method of claim 1 , wherein the communication is a multi-channel communication and received as one of an SMS text claim 1 , voice call claim 1 , e-mail claim 1 , chat claim 1 , interactive voice response (IVR)/intelligent virtual agent (IVA) systems ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR AN AGENT ASSIST UNIFIED INTERFACE WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004826A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:receiving a communication from a customer;presenting text associated with the communication in a unified user interface;automatically analyzing the communication to determine a subject of the customer's communication;automatically querying an information source for at least one response to the subject; andpresenting the response the unified user interface.2. The method of claim 1 , further comprising:parsing the communication for key terms; andautomatically highlighting the key terms in the unified user interface.3. The method of claim 2 , the analyzing further comprising:inferring an intent of the customer using an intent inference module; anddetermining the keywords in accordance with the intent.4. The method of claim 1 , further comprising scrolling the unified user interface to receive subsequent communication from the customer and to present subsequent responses from the information source.5. The method of claim 1 , further comprising:querying a plurality of information sources for responses to the subject; andpresenting the responses from the plurality of information sources in the unified user interface.6. The method of claim 1 , further comprising displaying an indication in the unified user interface that the communications being analyzed.7. The method of claim 1 , further comprising presenting communication controls in the unified user interface.8. The method of claim 1 , further comprising presenting information about the customer.9. The method of claim 8 , wherein the information comprises a customer ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR AN AGENT ASSIST UNIFIED INTERFACE WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004827A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:executing an automation infrastructure within a cloud-based contact center that includes a communication manager, speech-to-text converter, a natural language processor, and an inference processor exposed by application programming interfaces; andexecuting an agent assist functionality within the automation infrastructure that performs operations comprising:receiving a communication from a customer;presenting text associated with the communication in a unified user interface;automatically analyzing the communication to determine a subject of the customer's communication;automatically querying an information source for at least one response to the subject; andpresenting the response the unified user interface.2. The method of claim 1 , further comprising:parsing the communication for key terms; andautomatically highlighting the key terms in the unified user interface.3. The method of claim 2 , the analyzing further comprising:inferring an intent of the customer using an intent inference module; anddetermining the keywords in accordance with the intent.4. The method of claim 1 , further comprising scrolling the unified user interface to receive subsequent communication from the customer and to present subsequent responses from the information source.5. The method of claim 1 , further comprising:querying a plurality of information sources for responses to the subject; andpresenting the responses from the plurality of information sources in the unified user interface.6. The method of claim 1 , further comprising ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR REAL-TIME AGENT ASSIST WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004828A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:receiving a communication from a customer;automatically analyzing the communication in real-time to determine a subject of the customer's communication;automatically parsing a knowledgebase in real-time for at least one responsive answer to a question associated with the subject of the customer's communication; andproviding the at least one responsive answer in real-time to an agent during the communication with the customer.2. The method of claim 1 , wherein the communication is in textual form claim 1 , the method further comprising:displaying text input by the customer in a first field of a unified interface;parsing the text input by the customer for key terms;querying the knowledgebase using the key terms; anddisplaying responsive results from the knowledgebase as the solution in a second field in the unified interface.3. The method of claim 2 , further comprising highlighting the key terms.42. The method of claim 2 , further comprising:querying a customer relationship management (CRM) platform/a customer service management (CSM) platform using the key terms; anddisplaying responsive results from the CRM/CSM in the second field in the unified interface.5. The method of claim 2 , further comprising:querying a database of customer-agent transcripts using the key terms; anddisplaying responsive results from the database of customer-agent transcripts in the second field in the unified interface.6. The method of claim 1 , further comprising:receiving the communication as speech;converting the speech to text; ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR REAL-TIME AGENT ASSIST WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004829A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:executing an automation infrastructure within a cloud-based contact center that includes a communication manager, speech-to-text converter, a natural language processor, and an inference processor exposed by application programming interfaces; andexecuting an agent assist functionality within the automation infrastructure that performs operations comprising:receiving a communication from a customer;automatically analyzing the communication in real-time to determine a subject of the customer's communication;automatically parsing a knowledgebase in real-time for at least one responsive answer to a question associated with the subject of the customer's communication; andproviding the at least one responsive answer in real-time to an agent during the communication with the customer.2. The method of claim 1 , wherein the communication is in textual form claim 1 , the method further comprising:displaying text input by the customer in a first field of a unified interface;parsing the text input by the customer for key terms;querying the knowledgebase using the key terms; anddisplaying responsive results from the knowledgebase as the solution in a second field in the unified interface.3. The method of claim 2 , further comprising highlighting the key terms.42. The method of claim 2 , further comprising:querying a customer relationship management (CRM) platform/a customer service management (CSM) platform using the key terms; anddisplaying responsive results from the CRM/CSM in the second field in the unified interface.5. ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR AGENT ASSIST ERROR CORRECTION AND DETECTION WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004830A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:receiving a communication from a customer;routing the communication to an agent;creating a transcript of the communication;analyzing to detect predetermined groupings of words or variations of the groupings of words; andalerting a supervisor upon detecting the predetermined groupings of the words or the variations of the groupings of the words.2. The method of claim 1 , further comprising performing the analyzing to determine agent compliance.3. The method of claim 1 , further comprising determining an emotional state of the agent.4. The method of claim 1 , further comprising determining if the agent is talking too fast claim 1 , too slow or if there is a delay between words.5. The method of claim 1 , further comprising:receiving the communication as speech;converting the speech to text; anddetermining intent from the text.6. The method of claim 1 , wherein the communication is a multi-channel communication and received as one of an SMS text claim 1 , voice call claim 1 , e-mail claim 1 , chat claim 1 , interactive voice response (IVR)/intelligent virtual agent (IVA) systems claim 1 , and social media.7. The method of claim 1 , wherein the analyzing is performed in real time during the communication.8. The method of claim 1 , further comprising:populating a notes interface with a transcript of the communication; andproviding the transcript of the indication to the supervisor.9. The method of claim 8 , further comprising highlighting the predetermined groupings of the words or the variations of the groupings of ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR AGENT ASSIST ERROR CORRECTION AND DETECTION WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004831A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:executing an automation infrastructure within a cloud-based contact center that includes a communication manager, speech-to-text converter, a natural language processor, and an inference processor exposed by application programming interfaces; and receiving a communication from a customer;', 'routing the communication to an agent;', 'creating a transcript of the communication;', 'analyzing to detect predetermined groupings of words or variations of the groupings of words; and', 'alerting a supervisor upon detecting the predetermined groupings of the words or the variations of the groupings of the words., 'executing an agent assist functionality within the automation infrastructure that performs operations comprising2. The method of claim 1 , further comprising performing the analyzing to determine agent compliance.3. The method of claim 1 , further comprising determining an emotional state of the agent.4. The method of claim 1 , further comprising determining if the agent is talking too fast claim 1 , too slow or if there is a delay between words.5. The method of claim 1 , further comprising:receiving the communication as speech;converting the speech to text; anddetermining intent from the text.6. The method of claim 1 , wherein the communication is a multi-channel communication and received as one of an SMS text claim 1 , voice call claim 1 , e-mail claim 1 , chat claim 1 , interactive voice response (IVR)/intelligent virtual agent (IVA) systems claim 1 , and social media.7. The method of claim 1 , wherein the ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR ESCALATION USING AGENT ASSIST WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004832A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:executing an automation infrastructure within a cloud-based contact center that includes a communication manager, speech-to-text converter, a natural language processor, and an inference processor exposed by application programming interfaces; andexecuting an agent assist functionality within the automation infrastructure that performs operations comprising:receiving a communication from a customer;displaying text associated with the communication in a first field of a unified user interface;automatically analyzing the text to determine key terms associated the communication;automatically populating the key terms into a notes field of a call notes user interface; andescalating the communication to a supervisor for further action.2. The method of claim 1 , further comprising:querying one of a knowledgebase, a customer relationship management (CRM) platform/a customer service management (CSM) platform, and a database of customer-agent transcripts using the key terms; anddisplaying responsive results to the querying in a second field in the unified interface.3. The method of claim 2 , further comprising automatically populating query results into the notes field of the call notes user interface.4. The method of claim 3 , further comprising providing an editor in the call notes user interface to receive edits from the agent of the automatically populated key terms and query results.5. The method of claim 3 , wherein the populated key terms and query results form a call summary for use by the supervisor to understand ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR QUERYING MULTIPLE INFORMATION SOURCES USING AGENT ASSIST WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004833A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:receiving a communication from a customer;presenting text associated with the communication in a unified user interface;automatically analyzing the communication to determine a subject of the customer's communication;automatically querying an in a plurality of information sources in real time formation source for at least one response to the subject; andpresenting the at least one response from the plurality of information sources the unified user interface.2. The method of claim 1 , further comprising:parsing the communication for key terms; andautomatically highlighting the key terms in the unified user interface.3. The method of claim 2 , the analyzing further comprising:inferring an intent of the customer using an intent inference module; anddetermining the keywords in accordance with the intent.4. The method of claim 1 , further comprising scrolling the unified user interface to receive subsequent communication from the customer and to present subsequent responses from the information source.5. The method of claim 1 , further comprising ranking a plurality of responses from the plurality of information sources in accordance with a relevance of each of the plurality of responses.6. The method of claim 1 , further comprising displaying an indication in the unified user interface that the communications being analyzed.7. The method of claim 1 , further comprising presenting communication controls in the unified user interface.8. The method of claim 1 , further comprising presenting information about the ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR COMMUNICATION ANALYSIS FOR USE WITH AGENT ASSIST WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004834A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:receiving a communication from a customer;automatically analyzing the communication to determine a subject of the customer's communication;automatically querying a database of communications between other customers and other agents related to the subject of the customer's communication;determining at least one responsive answer to the subject from the database; andproviding the at least one responsive answer to an agent during the communication with the customer.2. The method of claim 1 , further comprising populating the database with information about the communications claim 1 , wherein information includes at least one of a time of the other communication claim 1 , a duration of the origination claim 1 , a topic of the other communication claim 1 , a frequency of words in the other communication.3. The method of claim 1 , wherein the communication is in textual form claim 1 , the method further comprising:displaying text input by the customer in a first field of a unified interface;parsing the text input by the customer for key terms;querying the knowledgebase using the key terms; anddisplaying responsive results from the knowledgebase as the solution in a second field in the unified interface.41. The method of claim 1 , further comprising:querying a customer relationship management (CRM) platform/a customer service management (CSM) platform using the key terms; anddisplaying responsive results from the CRM/CSM in the second field in the unified interface.5. The method of claim 1 , further comprising: ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR COMMUNICATION ANALYSIS FOR USE WITH AGENT ASSIST WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004835A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:executing an automation infrastructure within a cloud-based contact center that includes a communication manager, speech-to-text converter, a natural language processor, and an inference processor exposed by application programming interfaces; andexecuting an agent assist functionality within the automation infrastructure that performs operations comprising:receiving a communication from a customer;automatically analyzing the communication to determine a subject of the customer's communication;automatically querying a database of communications between other customers and other agents related to the subject of the customer's communication;determining at least one responsive answer to the subject from the database; andproviding the at least one responsive answer to an agent during the communication with the customer.2. The method of claim 1 , further comprising populating the database with information about the communications claim 1 , wherein information includes at least one of a time of the other communication claim 1 , a duration of the origination claim 1 , a topic of the other communication claim 1 , a frequency of words in the other communication.3. The method of claim 1 , wherein the communication is in textual form claim 1 , the method further comprising:displaying text input by the customer in a first field of a unified interface;parsing the text input by the customer for key terms;querying the knowledgebase using the key terms; anddisplaying responsive results from the knowledgebase as the solution in ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR PRE-POPULATING FORMS USING AGENT ASSIST WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210004836A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:receiving a communication from a customer;displaying text associated with the communication in a first field of a unified user interface;automatically analyzing the text to determine key terms associated the communication; andautomatically populating the key terms into a web-based form.2. The method of claim 1 , further comprising:querying one of a knowledgebase, a customer relationship management (CRM) platform/a customer service management (CSM) platform, and a database of customer-agent transcripts using the key terms; anddisplaying responsive results to the querying in a second field in the unified interface.3. The method of claim 2 , further comprising automatically populating query results into the web form.4. The method of claim 3 , further comprising providing an editor in the call notes user interface to receive edits from the agent of the automatically populated key terms and query results.5. The method of claim 3 , wherein the populated key terms and query results are responsive inputs to the web form.6. The method of claim 1 , further comprising highlighting the key terms in the unified interface.7. The method of claim 1 , further comprising:receiving the communication as speech;converting the speech to text; andparsing the text for key terms.8. The method of claim 1 , wherein the communication is a multi-channel communication and received as one of an SMS text claim 1 , voice call claim 1 , e-mail claim 1 , chat claim 1 , interactive voice response (IVR)/intelligent virtual agent (IVA) systems claim ...

Подробнее
03-01-2019 дата публикации

TOPIC SHIFT DETECTOR

Номер: US20190005123A1
Принадлежит:

Aspects detect or recognize shifts in topics in computer implemented speech recognition processes as a function of mapping keywords to non-verbal cues. An initial topic is mapped to one or more keywords extracted from a first spoken query within a user keyword ontology mapping. A query spoken subsequent in time to the first query is identified and distinguished by recognizing one or more non-verbal cues associated with the audio data input that include a time elapsed between the queries, and in some aspects a user's facial expression or motion activity. Aspects determine whether the second spoken query is directed to the initial topic or to a new topic that is different from the initial topic, as a function of mappings of the keyword(s) extracted from the first query to one or more keywords extracted from the second query and to the non-verbal cue(s) within the user ontology mapping. 1. A computer-implemented method for detecting shifts in topics in computer implemented speech recognition processes as a function of mapping keywords to non-verbal cues , the method comprising executing on a computer processor:identifying an initial topic of a first spoken query within an audio data input from a user that is mapped to at least one keyword extracted from the first spoken query within a user keyword ontology mapping;identifying a second spoken query within the audio data input that is subsequent in time to the first spoken query and is distinguished from the first query by recognizing at least one non-verbal cue associated with the audio data input, wherein the at least one non-verbal cue comprises a time elapsed between the first spoken query and the second spoken query, and a user's motion activity relative to a programmable device comprising the processor during a time between the first spoken query and the second spoken query;determining whether the second spoken query is directed to the initial topic or to a new topic that is different from the initial topic, as a ...

Подробнее
07-01-2021 дата публикации

System and Method for Audio Event Detection in Surveillance Systems

Номер: US20210005067A1
Принадлежит:

A method and system for detecting and localizing a target audio event in an audio clip is disclosed. The method and system use utilizes a hierarchical approach in which a dilated convolutional neural network to detect the presence of the target audio event anywhere in an audio clip based on high level audio features. If the target audio event is detected somewhere in the audio clip, the method and system further utilizes a robust audio vector representation that encodes the inherent state of the audio as well as a learned relationship between state of the audio and the particular target audio event that was detected in the audio clip. A bi-directional long short term memory classifier is used to model long term dependencies and determine the boundaries in time of the target audio event within the audio clip based on the audio vector representations. 1. A method for detecting and localizing a target audio event in an audio clip , the method comprising:receiving, with a processor, an audio clip;determining, with the processor, a plurality of audio features based on the audio clip;determining, with the processor, whether the target audio event is present in the audio clip using a first neural network based on the plurality of audio features;determining, with the processor, in response to determining that the target audio event is present in the audio clip, a plurality of vectors based on (i) the plurality of audio features and (ii) the target audio event, the vectors in the plurality of vectors indicating a correlation between audio features in the plurality of audio features and the target audio event; anddetermining, with the processor, a position in time of the target audio event within the audio clip using a second neural network based on the plurality of vectors.2. The method of claim 1 , the determining the plurality of audio features further comprising:segmenting the audio clip into a plurality of audio segments with a first predetermined length; anddetermining, ...

Подробнее
07-01-2021 дата публикации

Abuse Alert System by Analyzing Sound

Номер: US20210005069A1
Автор: Mangal Aryan
Принадлежит:

A method, including recognizing an indicia pattern on an audio signal recognizing its characteristics including but not limited to intensity, pitch, presence of abuse words, utilizing a portable device based sound controller system, prompting one of acceptance and disapproval of the indicia pattern, detecting a present of abuse intent and if intent is present, start taking an action e.g. recording, emailing or flashing an LED. 1. A method , comprising:recognizing an indicia pattern of abuse detection through analysing various characteristics of sound;prompting one of acceptance and disapproval of the indicia pattern determining abusal intent;start recording a sound clip from a sound input if the abuse detection indicia pattern is accepted;generate alert and send communication in any desirable format as configured; andprompting one of confirmation and refusal of the pairing.2. The method of claim 1 , further comprising detecting voice intensity by measuring any or all of amplitude claim 1 , pitch to at least one of accept the indicia pattern and disapprove the indicia pattern.3. The method of claim 1 , further comprising detecting voice intensity can be tuned based on person claim 1 , environment claim 1 , surrounding noise level and other factors that can impact sound detection.4. The method of claim 1 , further comprising detecting voice intensity duration to at least one of confirm the abuse and refuse the abuse.5. The method of claim 1 , further comprising detecting presence of abusive words by converting voice to text to at least one of confirm the abuse and refuse the abuse.6. In addition to claim 4 , ability to choose and personalize abusive or derogatory words is also claimed.7. The method of claim 1 , further comprising machine learning models that can be miniaturized in size and can be fit onto portable devices such as microcontrollers claim 1 , mobile phone claim 1 , tablets with the sole purpose of abuse detection.8. Adding to claim 7 , a feedback loop ...

Подробнее
07-01-2021 дата публикации

AUDIBLE KEYWORD DETECTION AND METHOD

Номер: US20210005181A1
Принадлежит:

The disclosure describes keyword detection in an audio processor and methods therefor including a low-power keyword detection engine (LKDE) and a high-power keyword detection engine (HKDE). In one implementation, the LKDE detects a keyword in data from a single audio source while buffering data from multiple audio sources and, upon detection of a keyword, the HKDE is awakened to verify the previously detected keyword by processing the buffered audio data from the multiple sources. 1. A digital processor for processing audio data , the processor comprising:an audio data interface;a buffer coupled to the interface and configured to buffer data received at the interface;a low-power keyword detection engine (LKDE) configured to determine likely presence of a keyword in data received at the interface while the data is buffered in the buffer; anda high-power keyword detection engine (HKDE) configured to wakeup from a low-power sleep mode if the LKDE determines likely presence of a keyword, and after awakening, verify the likely presence of the keyword detected by the LKDE by processing data in the buffer,wherein the HKDE is configured to detect keywords with higher certainty than the LKDE.2. The processor of claim 1 ,wherein the LKDE is configured to determine likely presence of a keyword with a true positive rate (TPR) above a first threshold and a false acceptance rate (FAR) below a second threshold, wherein the first and second thresholds are constrained by a maximum acceptable power consumption associated with a duty cycle with which the HKDE is awakened, andwherein the HKDE is configured to detect likely presence of a keyword with a lower FAR than the LKDE.3. The processor of claim 2 , wherein the LKDE is configured to determine likely presence of a keyword based on whether a confidence level associated with detection of the keyword satisfies a condition.4. The processor of claim 2 ,the interface is a multi-source interface and the buffer is configured to buffer data ...

Подробнее
07-01-2021 дата публикации

ORTHOGONALLY CONSTRAINED MULTI-HEAD ATTENTION FOR SPEECH TASKS

Номер: US20210005183A1
Принадлежит:

A method for operating a neural network includes receiving an input sequence at an encoder. The input sequence is encoded to produce a set of hidden representations. Attention-heads of the neural network calculate attention weights based on the hidden representations. A context vector is calculated for each attention-head based on the attention weights and the hidden representations. Each of the context vectors correspond to a portion of the input sequence. An inference is output based on the context vectors. 1. A method for operating a neural network , the method comprising:receiving an input sequence at an encoder;encoding the input sequence to produce hidden representations;calculating attention weights in attention-heads of the neural network based on the hidden representations;calculating a context vector for each attention-head based on the attention weights and the hidden representations, each context vector corresponding to a portion of the input sequence; andoutputting an inference based on the context vectors.2. The method of claim 1 , in which the input sequence comprises an acoustic feature and the inference includes an indication of whether a keyword is included in a corresponding portion of the input sequence.3. The method of claim 1 , in which each of the context vectors are orthogonal to other context vectors of other attention-heads.4. The method of claim 1 , in which the attention weight is calculated based on a score function for focus determination.5. The method of claim 1 , further comprising selectively regularizing the neural network based on orthogonality constraints of the attention weights and context vectors.6. The method of claim 5 , in which the selectively regularizing further comprises calculating regularization terms only for positive samples in the input sequence.7. An apparatus for operating a neural network claim 5 , the apparatus comprising:a memory; andat least one processor coupled to the memory, the at least one processor being ...

Подробнее
07-01-2021 дата публикации

TRAINING ACOUSTIC MODELS USING CONNECTIONIST TEMPORAL CLASSIFICATION

Номер: US20210005184A1
Принадлежит: Google LLC

Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for training acoustic models and using the trained acoustic models. A connectionist temporal classification (CTC) acoustic model is accessed, the CTC acoustic model having been trained using a context-dependent state inventory generated from approximate phonetic alignments determined by another CTC acoustic model trained without fixed alignment targets. Audio data for a portion of an utterance is received. Input data corresponding to the received audio data is provided to the accessed CTC acoustic model. Data indicating a transcription for the utterance is generated based on output that the accessed CTC acoustic model produced in response to the input data. The data indicating the transcription is provided as output of an automated speech recognition service. 1. A method comprising:receiving, at data processing hardware of a speech recognition system, audio data for a portion of an utterance spoken by a user;dividing, by the data processing hardware, the audio data for the portion of the utterance into a sequence of fixed-length frames;for each frame in the sequence of fixed-length frames, determining, by the data processing hardware, a corresponding set of log-Mel frequency cepstral coefficients;generating, by the data processing hardware, using an acoustic model configured to receive the corresponding set of log-Mel frequency cepstral coefficients generated for each frame in the sequence of fixed-length frames as input, a sequence of context-dependent states representing the portion of the utterance; andgenerating, by the data processing hardware, using a language model trained separately from the acoustic model, a streaming speech recognition result for the portion of the utterance spoken by the user, obtaining acoustic model training data that comprises training audio data and word-level transcriptions for the training audio data,', 'training a first neural network ...

Подробнее
07-01-2021 дата публикации

ANALYSIS OF A TOPIC IN A COMMUNICATION RELATIVE TO A CHARACTERISTIC OF THE COMMUNICATION

Номер: US20210005188A1
Принадлежит:

A device monitors a communication between a user associated with a user device and a service representative associated with a service representative device, and causes a natural language processing model to perform a natural language processing analysis of a user input of the communication to identify a topic associated with the communication. The device determines a first score associated with the topic, and determines a second score associated with enabling the communication, where the first score and second score indicate a service performance score of an entity. The device causes a sentiment analysis model to perform a sentiment analysis of the communication to determine a sentiment score indicating a level of satisfaction the user has relative to the topic. The device updates a transaction protocol associated with the topic based on the service performance score, and/or updates a communication processing protocol associated with the communication based on the sentiment score. 1. A method , comprising:receiving, by a device, data associated with a communication between a user and a service representative;tracking, by the device, during the communication, and using a sentiment analysis model, a level of satisfaction of the user with respect to the communication; and 'the representative rating being based on a change in a measure of user sentiment during the communication.', 'determining, by the device and based on tracking the level of satisfaction of the user, a representative rating for the service representative,'}2. The method of claim 1 , wherein determining the representative rating comprises:determining the representative rating using a classification technique to determine a categorical outcome.3. The method of claim 1 , wherein receiving the data associated with the communication comprises:receiving the data associated with the communication from a log or record of service representative performance scores.4. The method of claim 1 , further comprising: ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR TEXT-ENABLED AUTOMATED AGENT ASSISTANCE WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210005192A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method , comprising:executing an automation infrastructure within a cloud-based contact center that includes a communication manager, speech-to-text converter, a natural language processor, and an inference processor exposed by application programming interfaces; andexecuting an agent assist functionality within the automation infrastructure that performs operations comprising:receiving a text communication from a customer;performing inference processing on the text to determine a customer intent;automatically analyzing the text to determine a subject of the text communication and key terms associated with the subject;automatically parsing a knowledgebase using the key terms for at least one responsive answer associated with the subject; andproviding the solution to an agent in a unified interface during the communication with the customer.21. The method of , further comprising:querying a customer relationship management (CRM) platform/a customer service management (CSM) platform using the key terms; anddisplaying responsive results from the CRM/CSM in the second field in the unified interface.3. The method of claim 1 , further comprising:querying a database of customer-agent transcripts using the key terms; anddisplaying responsive results from the database of customer-agent transcripts in the second field in the unified interface.4. The method of claim 1 , wherein the method is performed in real-time as the customer communication progresses with the agent.5. The method of claim 1 , further comprising concurrently displaying to ...

Подробнее
07-01-2021 дата публикации

RECORDING MEDIUM RECORDING PROGRAM, INFORMATION PROCESSING APPARATUS, AND INFORMATION PROCESSING METHOD FOR TRANSCRIPTION

Номер: US20210005204A1
Автор: Sankoda Satoru
Принадлежит: FUJITSU LIMITED

A method for transcription is performed by a computer. The method includes: accepting input of a voice after causing a display unit to display a sentence including a plurality of words; acquiring first sound information being information concerning sounds corresponding to the sentence; acquiring second sound information being information concerning sounds of the voice accepted in the accepting; specifying a portion in the first sound information having a prescribed similarity to the second sound information; and correcting a character string in the sentence corresponding to the specified portion based on a character string corresponding to the second sound information. 1. A non-transitory computer-readable recording medium having stored therein a program for causing a computer to execute processing comprising:accepting input of sounds after causing a display to display a sentence including a plurality of words;acquiring first sound information being information concerning sounds corresponding to the sentence;acquiring second sound information concerning the accepted sounds;specifying a portion in the first sound information having a prescribed similarity to the second sound information; andcorrecting a character string in the sentence corresponding to the specified portion based on a character string corresponding to the second sound information.2. The recording medium according to claim 1 , whereinthe portion is specified in the specifying based on a similarity of a phoneme sequence included in the first sound information to a phoneme sequence included in the second sound information.3. The recording medium according to claim 1 , whereinthe portion is specified in the specifying based on a similarity of a waveform of a voice included in the first sound information to a waveform of a voice included in the second sound information.4. The recording medium according to claim 1 , whereinthe portion is specified in the specifying based on a phoneme sequence included in ...

Подробнее
07-01-2021 дата публикации

System and method for speech-enabled automated agent assistance within a cloud-based contact center

Номер: US20210005206A1
Принадлежит: Talkdesk Inc

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing.

Подробнее
04-01-2018 дата публикации

DEVICE INCLUDING SPEECH RECOGNITION FUNCTION AND METHOD OF RECOGNIZING SPEECH

Номер: US20180005627A1
Принадлежит:

A device including a speech recognition function which recognizes speech from a user, includes: a loudspeaker which outputs speech to a space; a microphone which collects speech in the space; a first speech recognition unit which recognizes the speech collected by the microphone; a command control unit which issues a command for controlling the device, based on the speech recognized by the first speech recognition unit; and a control unit which prohibits the command issuance unit from issuing the command, based on the speech to be output from the loudspeaker. 17-. (canceled)8. A device including a speech recognition function which recognizes user speech which is speech from a user , the device comprising:a loudspeaker which outputs speech to a space;a microphone which collects speech in the space;a speech output unit configured to output a speech signal;a first speech recognition unit configured to recognize the speech collected by the microphone;a first control unit configured to generate a command for controlling the speech output unit, based on the speech recognized by the first speech recognition unit; anda second control unit configured to permit or prohibit issuance of the command to the speech output unit, based on the speech signal,wherein the second control unit includes a second speech recognition unit configured to analyze the speech signal to recognize the speech to be output from the loudspeaker, andthe second control unit is configured to determine whether or not the speech recognized by the second speech recognition unit matches a predetermined keyword, and when the speech recognized by the second speech recognition unit matches the predetermined keyword, prohibit the issuance of the command and when the speech recognized by the second speech recognition unit does not match the predetermined keyword, permit issuance of the command.9. A method of recognizing user speech using a device including a speech recognition function , the user speech being ...

Подробнее
04-01-2018 дата публикации

PERFORMING TASKS AND RETURING AUDIO AND VISUAL ANSWERS BASED ON VOICE COMMAND

Номер: US20180005631A1
Принадлежит: KT CORPORATION

An artificial intelligence voice interactive system may provide various services to a user in response to a voice command by providing an interface between the system and a legacy system to enable providing various types of existing services in response to user speech without modifying systems for the existing services. Such system includes a central server, and the central server may perform operations of registering a plurality of service servers at the central server and storing registration information of each service server, analyzing voice command data from the user device and determining at least one task and corresponding service servers based on the analysis results, generating an instruction message based on the voice command data, the determined at least one task, and the registration information of the selected service servers, and transmitting the generated instruction message to the selected service servers, and receiving task results including audio and video data from the selected service servers and outputting the task results through at least one device associated with the user device. 1. A method for providing voice interactive services to a user through an artificial intelligence voice interactive system including a central server and a plurality of service servers each connected to the central server through a communication network , the method of the central server comprising:performing a registration operation upon generation of a predetermined event for registering the service servers and storing registration information of each service server;receiving voice command data, determining at least one task to perform based on the voice command data, and selecting at least one of the service servers based on the determined task; andgenerating an instruction message based on the voice command data, the determined at least one task, and the registration information of the selected service server, and transmitting the generated instruction message to ...

Подробнее
02-01-2020 дата публикации

Deeplearning method for voice recognition model and voice recognition device based on artificial neural network

Номер: US20200005766A1
Автор: Dami Kim
Принадлежит: LG ELECTRONICS INC

A method for training an artificial neural network-based speech recognition model is disclosed. In the method for training an artificial neural network-based speech recognition model, a user's speech is learned by using target data representing features and non-target data representing non-features as random inputs and outputs, and then the user's speech is recognized under a noise situation. A method for training an artificial neural network-based speech recognition model and speech recognition device of the present disclosure can be associated with artificial intelligence modules, drones (unmanned aerial vehicles (UAVs)), robots, augmented reality (AR) devices, virtual reality (VR) devices, devices related to 5G service, etc.

Подробнее
02-01-2020 дата публикации

METHOD AND APPARATUS FOR RECOGNIZING A VOICE

Номер: US20200005768A1
Автор: Chae Jong Hoon
Принадлежит: LG ELECTRONICS INC.

Disclosed are a speech recognition method and a speech recognition device, in which speech recognition is performed by executing an artificial intelligence (AI) algorithm and/or a machine learning algorithm provided therein. According to an embodiment of the present disclosure, the speech recognition method includes buffering an inputted spoken utterance, determining whether a preset wake-up word is present in the spoken utterance by comparing the buffered spoken utterance to the preset wake-up word, and in response to the preset wake-up word in the spoken utterance, activating a speech recognition function and isolating, from the spoken utterance, a spoken sentence as a voice command without the wake-up word, and processing the spoken sentence and outputting a processing result. 1. A speech recognition method by a speech recognition device , the speech recognition method comprising:buffering an inputted spoken utterance;comparing the buffered spoken utterance to a preset wake-up word to determine whether the preset wake-up word is present in the spoken utterance;in response to the preset wake-up word present in the spoken utterance, activating a speech recognition function and isolating, from the spoken utterance, a spoken sentence as a voice command without the wake-up word; andprocessing the spoken sentence and outputting a processing result.2. The speech recognition method of claim 1 , wherein the buffering further includes buffering by varying a buffering volume or time in accordance with an environment condition under which the spoken utterance is inputted.3. The speech recognition method of claim 1 , wherein the isolating includes isolating claim 1 , from the spoken utterance claim 1 , the spoken sentence preceding the wake-up word.4. The speech recognition method of claim 1 , wherein the isolating includes isolating claim 1 , from the spoken utterance claim 1 , a first spoken segment and a second spoken segment without the wake-up word claim 1 , wherein the ...

Подробнее
02-01-2020 дата публикации

VOICE RECOGNITION DEVICE AND VOICE RECOGNITION METHOD

Номер: US20200005775A1
Принадлежит: Mitsubishi Electric Corporation

A voice recognition device includes: a first feature vector calculating unit () for calculating a first feature vector from voice data input; an acoustic likelihood calculating unit () for calculating an acoustic likelihood of the first feature vector by using an acoustic model used for calculating an acoustic likelihood of a feature vector; a second feature vector calculating unit () for calculating a second feature vector from the voice data; a noise degree calculating unit () for calculating a noise degree of the second feature vector by using a discriminant model used for calculating a noise degree indicating whether a feature vector is noise or voice; a noise likelihood recalculating unit () for recalculating an acoustic likelihood of noise on the basis of the acoustic likelihood of the first feature vector and the noise degree of the second feature vector; and a collation unit () for performing collation with a pattern of a vocabulary word to be recognized, by using the acoustic likelihood calculated and the acoustic likelihood of noise recalculated, and outputting a recognition result of the voice data. 1. A voice recognition device comprising:a processor to execute a program; anda memory to store the program which, when executed by the processor, performs processes of,calculating a first feature vector from voice data input;calculating acoustic likelihoods of respective phonemes and an acoustic likelihood of noise of the first feature vector, by using an acoustic model used for calculating an acoustic likelihood of a feature vector;calculating a second feature vector from the voice data;calculating a noise degree of the second feature vector, by using a discriminant model used for calculating a noise degree indicating whether a feature vector is noise or voice;recalculating an acoustic likelihood of noise on a basis of a larger value between the acoustic likelihood of noise of the first feature vector, and a likelihood that is calculated by adding a maximum ...

Подробнее
02-01-2020 дата публикации

AUTOMATIC SKILL ROUTING IN CONVERSATIONAL COMPUTING FRAMEWORKS

Номер: US20200005776A1

An utterance is analyzed to identify an absence of a known invocation phrase. A skill set is constructed in response to the absence, the skill set including a first skill corresponding to the utterance and a first skill score corresponding to a likelihood that the first skill corresponds to the utterance. The first skill score is adjusted, based on the presence of the first skill in a skill history, where the skill history stores a set of history skills in an order of recency of use of each history skill in the set of history skills. The first skill score is adjusted, based on an association of the first skill with a default installed skill. An installed skill is selected, based on the adjusted first skill score, the installed skill performing an action in response to the utterance. 1. A method comprising:analyzing an utterance to identify an absence of a known invocation phrase;constructing, in response to the absence, using a processor and a memory, a skill set, the skill set including a first skill corresponding to the utterance and a first skill score corresponding to a likelihood that the first skill corresponds to the utterance;adjusting, based on the presence of the first skill in a skill history, the first skill score, wherein the skill history stores a set of history skills in an order of recency of use of each history skill in the set of history skills;adjusting, based on an association of the first skill with a default installed skill, the first skill score; andselecting, based on the adjusted first skill score, an installed skill, wherein the installed skill performs an action in response to the utterance.2. The method of claim 1 , wherein the adjusting claim 1 , based on the presence of the first skill in a skill history claim 1 , further comprises:determining that the first skill matches a history skill associated with an installed skill other than the default installed skill; andraising, responsive to the determining, the first skill score.3. The ...

Подробнее
02-01-2020 дата публикации

Method for Term-Dependent Output of Information Based on a Voice Input to a Specific Group, and System

Номер: US20200005785A1
Принадлежит:

A method for term-dependent output of information based on a voice input to a specific group includes the steps of: capturing the voice input; analyzing the captured voice input for the presence of a group-specific key term, associated with the specific group; and on detection of the group-specific key term in the analyzed voice input, outputting the information based on the voice input to the specific group. 1. A method for term-dependent output of information based on a voice input to a specific group , the method comprising the steps of:a) capturing the voice input;b) analyzing the captured voice input for a presence of a group-specific key term, associated with the specific group; andc) on detection of the group-specific key term in the analyzed voice input, outputting the information based on the voice input to the specific group.2. The method according to claim 1 , wherein the specific group is one or more of:a garden and/or forestry worker group, and the group-specific key term is a garden and/or forestry term,a hauler group, and the group-specific key term is a hauler term,a forestry and/or work management group, and the group-specific key term is a forestry and/or work management term, anda rescue group, and the group-specific key term is an emergency term and/or a rescue term.3. The method according to claim 1 , further comprising the step of:ascertaining a direction from an output location to a capture location, andwherein the information has the ascertained direction.4. The method according to claim 1 , further comprising the steps of:ascertaining a distance between a capture location and an output location; andif the ascertained distance is below a distance limit value, outputting the information in group-aspecific fashion.5. The method according to claim 4 , whereinthe distance is ascertained by a time-of-flight method.6. The method according to claim 1 , further comprising the step of:wirelessly transmitting the captured voice input and/or the ...

Подробнее
02-01-2020 дата публикации

GUIDE ROBOT AND METHOD FOR OPERATING THE SAME

Номер: US20200005787A1
Автор: MAENG Jichan, SHIN Wonho
Принадлежит: LG ELECTRONICS INC.

The present disclosure relates to a guide robot and a method of operating the same. A guide robot according to the present disclosure includes a voice receiving unit to receive a voice, a controller to determine whether the received voice includes a preset wake-up word, and a wireless communication unit to perform communication with an artificial intelligence (AI) server set to be activated by the preset wake-up word. At this time, the control unit transmits the received voice to the artificial intelligence server, receives result information from the artificial intelligence server, and outputs the received result information, when the received voice includes the preset wake-up word. And, the control unit outputs a response voice selected according to a predetermined reference when the received voice does not include the preset wake-up word. 1. A guide robot , comprising:a voice receiving unit configured to receive a voice;a control unit configured to determine whether the received voice includes a preset wake-up word; anda wireless communication unit configured to perform communication with an artificial intelligence (AI) server set to be activated by the preset wake-up word,wherein the control unit transmits the received voice to the artificial intelligence server, receives result information from the artificial intelligence server, and output the received result information, when the received voice includes the preset wake-up word, andoutputs a response voice selected according to a predetermined reference when the received voice does not include the preset wake-up word.2. The guide robot of claim 1 , wherein the control unit performs a greeting recognition operation when the received voice does not include the preset wake-up word claim 1 , and determines whether the received voice is recognized as a greeting based on a sensing signal received from at least one sensor in the greeting recognition operation.3. The guide robot of claim 2 , wherein the control unit ...

Подробнее
02-01-2020 дата публикации

INTELLIGENT VOICE RECOGNIZING METHOD, APPARATUS, AND INTELLIGENT COMPUTING DEVICE

Номер: US20200005788A1
Принадлежит: LG ELECTRONICS INC.

Disclosed are an intelligent voice recognizing method, a voice recognizing device, and an intelligent computing device. According to an embodiment of the present invention, an intelligent voice recognizing method of a voice recognizing device may obtain a microphone detection signal, recognize a user's voice from the microphone detection signal based on a pre-learned speech recognition model, output information related to a result of recognition of the user's voice, and update the speech recognition model based on the output speech recognition result information, easily updating the speech recognition model for speech recognition based on the speech recognition result information which is intuitively shown to the user. According to the present invention, one or more of the voice recognizing device, intelligent computing device, and server may be related to artificial intelligence (AI) modules, unmanned aerial vehicles (UAVs), robots, augmented reality (AR) devices, virtual reality (VR) devices, and 5G service-related devices. 1. A method of intelligently recognizing a voice by a voice recognizing device , the method comprising:obtaining a microphone detection signal;recognizing a user's voice from the microphone detection signal based on a pre-learned speech recognition model; andgenerating information related to a result of recognition of the user's voice, wherein the speech recognition model is updated based on the generated speech recognition result information.2. The method of claim 1 , wherein the speech recognition result information includes information related to whether the speech recognition succeeds.3. The method of claim 2 , wherein the information related to whether the speech recognition succeeds includes text information generated by recognizing the voice.4. The method of claim 2 , wherein the information related to whether the speech recognition succeeds includes information related to the number of words included in the text information.5. The ...

Подробнее
02-01-2020 дата публикации

METHOD AND APPARATUS FOR RECOGNIZING A VOICE

Номер: US20200005789A1
Автор: Chae Jong Hoon
Принадлежит: LG ELECTRONICS INC.

Disclosed are a speech recognition device and a speech recognition method, which perform speech recognition by executing an artificial intelligence (AI) algorithm and/or a machine learning algorithm provided therein, and which can communicate with other electronic devices and an external server in a 5G communication environment. According to an embodiment, the speech recognition method includes setting an additional wake-up word target capable of activating a speech recognition function in addition to a preset wake-up word, generating a plurality of additional wake-up word utterances formed on the basis of the additional wake-up word target being uttered under various conditions, learning a wake-up word recognition algorithm by using each of the spoken utterances of the additional wake-up word to generate an additional wake-up word recognition algorithm, and executing the additional wake-up word recognition algorithm upon receiving a select word uttered by a user to determine whether to activate the speech recognition function. 1. A speech recognition method by a speech recognition device , comprising:setting an additional wake-up word target in addition to a preset wake-up word, the additional wake-up word target being capable of activating a speech recognition function;generating a plurality of utterance data items with respect to the additional wake-up word target based on the additional wake-up word target being uttered under various conditions;generating an additional wake-up word recognition algorithm by training a wake-up word recognition algorithm by using each of the plurality of utterance data items as training input data; anddetermining whether to activate the speech recognition function by executing the additional wake-up recognition algorithm in response to a select word received and uttered by a user.2. The speech recognition method of claim 1 , the setting of the additional wake-up word target includes:receiving an additional wake-up word setting menu ...

Подробнее
02-01-2020 дата публикации

DESIRED SIGNAL SPOTTING IN NOISY, FLAWED ENVIRONMENTS

Номер: US20200005813A1
Принадлежит:

There are disclosed devices, system and methods for desired signal spotting in noisy, flawed environments by identifying a signal to be spotted, identifying a target confidence level, and then passing a pool of cabined arrays through a comparator to detect the identified signal, wherein the cabined arrays are derived from respective distinct environments. The arrays may include plural converted samples, each converted sample include a product of a conversion of a respective original sample, the conversion including filtering noise and transforming the original sample from a first form to a second form. Detecting may include measuring a confidence of the presence of the identified signal in each of plural converted samples using correlation of the identified signal to bodies of known matching samples. If the confidence for a given converted sample satisfies the target confidence level, the given sample is flagged. 1. A method of desired signal spotting in noisy , flawed environments comprising:identifying a signal to be spotted, wherein the signal is one or more spoken terms;identifying a target confidence level;passing a pool of call samples through a comparator to spot the identified signal, wherein the call samples are derived from respective distinct environments, wherein spotting the identified signal comprises measuring a confidence of a presence of the identified signal in each of the call samples using correlation of the identified signal to bodies of known matching samples that are machine created reference signals, and wherein the correlation includes the comparator performing a weighting analysis comparison of the identified signal to the bodies of known matching samples that are selected by the comparator during measuring of each of the call samples;if the confidence for a given call sample satisfies the target confidence level, flagging the given call sample as having the identified signal; andusing the flagged sample as a signal of the bodies of known ...

Подробнее
03-01-2019 дата публикации

Operating method for voice function and electronic device supporting the same

Номер: US20190005944A1
Автор: SUBHOJIT Chakladar
Принадлежит:

An electronic device is provided. The electronic device includes a memory configured to store at least a portion of a plurality of pieces of speech information used for voice recognition, and a processor operatively connected to the memory, wherein the processor selects speaker speech information from at least a portion of the plurality of pieces of speech information based on mutual similarity, and generates voice recognition information to be registered as personalized voice information based on the speaker speech information. 120-. (canceled)21. An electronic device comprising:a microphone configured to receive voice command; anda one or more processor including processing circuitry, the processor functionally connected to a memory, wherein the processor is configured to:receive a first voice command through the microphone of the electronic device,provide the first voice command to a server device to recognize the first voice command,receive a response corresponding to the first voice command from the server,execute a function according to the response from the server, andstore a voice recognition model for the first voice command in a memory of the electronic device, wherein the voice recognition model is generated by the server device when audio data corresponding to the first voice command is collected above a predetermined threshold,wherein the one or more processor is further configured to:receive a second voice command newly inputted through the microphone of the electronic device,recognize the second voice command using the voice recognition model stored in the memory, andexecute the function in response to the recognizing the second voice command using the voice recognition model.22. The electronic device of claim 21 , wherein the processor is configured to output a message providing a notification that an operation of applying the voice recognition model to a voice recognition is being performed.23. The electronic device of claim 21 , wherein claim 21 , ...

Подробнее
03-01-2019 дата публикации

METHOD AND DEVICE FOR MANAGING DIALOGUE BASED ON ARTIFICIAL INTELLIGENCE

Номер: US20190005948A1
Принадлежит:

Embodiments of the present disclosure provide a method and a device for managing a dialogue based on artificial intelligence. The method includes the followings. An optimum system action is determined from at least one candidate system action according to a current dialogue status feature, a candidate system action feature and surrounding feedback information of the at least one candidate system action and based on a decision model. Since the current dialogue status corresponding to the current dialogue status feature includes uncertain results of natural language understanding, the at least one candidate system action acquired according to the current dialogue status also includes the uncertain results of natural language understanding. 1. A method for managing a dialogue based on artificial intelligence , comprising:receiving current dialogue information;determining a user intention of the current dialogue information;determining query dimension distribution information and current single-round slot distribution information of the current dialogue information under the user intention;generating current multi-round slot distribution information according to the current single-round slot distribution information of the current dialogue information and historical multi-round slot distribution information of historical dialogue information;generating a current dialogue status according to the user intention, the query dimension distribution information and the current multi-round slot distribution information of the current dialogue information;performing a first feature extraction on the current dialogue status to obtain a current dialogue status feature;determining at least one candidate system action according to the current dialogue status and a pre-configured rule;performing a second feature extraction on the at least one candidate system action to obtain a candidate system action feature of each of the at least one candidate system; andinputting the current ...

Подробнее
03-01-2019 дата публикации

Hands free always on near field wakeword solution

Номер: US20190005953A1
Принадлежит: Amazon Technologies Inc

Apparatuses and systems for conserving power for a portable electronic device that monitors local audio for a wakeword are described herein. In a non-limiting embodiment, a portable electronic device may have two-phases. The first phase may be a first circuit that stores an audio input while determining whether human speech is present in the audio input. The second phase may be a second circuit that activates when the first circuit determines that human speech is present in the audio input. The second circuit may receive the audio input from the first circuit, store the audio input, and determine whether a wakeword is present within the audio input.

Подробнее
03-01-2019 дата публикации

WAKE-ON-VOICE METHOD, TERMINAL AND STORAGE MEDIUM

Номер: US20190005954A1
Принадлежит:

The present disclosure provides a wake-on-voice method, a terminal and a storage medium. The method includes: acquiring a wake-up voice configured to wake up a smart terminal; performing an analysis on an acoustic feature of the wake-up voice by using a preset acoustic model and a preset wake-up word recognition network of the smart terminal, so as to acquire a confidence coefficient of the acoustic feature of the wake-up voice with respect to an acoustic feature of a preset wake-up word; determining whether the confidence coefficient falls in a preset range of moderate confidence coefficients, if yes, uploading the wake-up voice to a remote server; and determining whether a linguistic feature obtained by analyzing the wake-up voice using a linguistic model matches to a linguistic feature of the preset wake-up word, if yes, receiving an instruction to wake up the smart terminal generated by the remote server. 1. A wake-on-voice method , comprising:acquiring a wake-up voice configured to wake up a smart terminal;performing an analysis on an acoustic feature of the wake-up voice by using a preset acoustic model and a preset wake-up word recognition network of the smart terminal, so as to acquire a confidence coefficient of the acoustic feature of the wake-up voice with respect to an acoustic feature of a preset wake-up word;determining whether the confidence coefficient falls in a preset range of moderate confidence coefficients, and if yes, uploading the wake-up voice to a remote server; anddetermining whether a linguistic feature obtained by analyzing the wake-up voice using a linguistic model in the remote server matches to a linguistic feature of the preset wake-up word, and if yes, receiving an instruction to wake up the smart terminal generated by the remote server.2. The wake-on-voice method according to claim 1 , wherein after it is determined that the linguistic feature obtained by analyzing the wake-up voice using the linguistic model in the remote server ...

Подробнее
03-01-2019 дата публикации

METHODS, SYSTEMS, AND MEDIA FOR VOICE-BASED CALL OPERATIONS

Номер: US20190005955A1
Принадлежит:

Methods, systems, and media for voice-based call operations are provided. In some embodiments, a method comprises: receiving, at a first user device, a communication; detecting a voice command, using the first user device, that includes a keyword; and in response to detecting the voice command, causing the communication to be transferred to a second user device that is associated with the keyword. 1. A method comprising:receiving, at a first user device, a communication;detecting a voice command, using the first user device, that includes a keyword; andin response to detecting the voice command, causing the communication to be transferred to a second user device that is associated with the keyword.2. The method of claim 1 , wherein the communication is an invitation to join a telecommunication channel.3. The method of claim 2 , wherein the telecommunication channel is a telephone channel.4. The method of claim 2 , wherein the telecommunication channel is an Internet videotelephony channel.5. The method of claim 2 , further comprising:in response to receiving the invitation to join the telecommunication channel, presenting a notification of the invitation during a period of time, wherein the period of time is terminated upon acceptance of the invitation, and wherein the voice command is detected during the period of time.6. The method of claim 2 , further comprising causing the second user device to join the telecommunication channel.7. The method of claim 1 , wherein causing the communication to be transferred to the second user device is further in response to detecting a second keyword in the voice command claim 1 , wherein the second keyword corresponds to a communication transfer operation.8. The method of claim 1 , wherein the communication is an SMS message or an MMS message.9. The method of claim 1 , further comprising:receiving an indication of the keyword; anddetermining that the keyword is to be associated with the second user device.10. A system ...

Подробнее
05-01-2017 дата публикации

ELECTRONIC DEVICE AND METHOD FOR CONTROLLING THE SAME

Номер: US20170006329A1
Принадлежит: LG ELECTRONICS INC.

A digital television can include a display; an interface to receive a control signal from a remote controller; a communication unit for internet access; a broadcast receiver; and a controller to display broadcast content received through the broadcast receiver, display a screen corresponding to a voice recognition mode in response to entering the voice recognition mode, the screen includes a symbol and text indicating the voice recognition mode, receive a voice command signal corresponding to a voice command through the remote controller in the voice recognition mode while the broadcast content is displayed, in which the voice command is input via a microphone provided in the remote controller, and the input voice command includes a search source and a search keyword, and display a web page corresponding to the search source included in the voice command, which includes a search window including the search keyword and a search result. 1. A digital television comprising:a display;an interface configured to receive a control signal for controlling an operation of the digital television from a remote controller;a communication unit configured to support internet access;a broadcast receiver configured to receive broadcast signals or broadcast associated information from an external broadcast management server via a broadcast channel; and display broadcast content received through the broadcast receiver on the display,', 'display a screen corresponding to a voice recognition mode on the display in response to entering the voice recognition mode, wherein the screen includes a symbol and text indicating the voice recognition mode,', 'receive a voice command signal corresponding to a voice command through the remote controller in the voice recognition mode while the broadcast content is displayed, wherein the voice command is input via a microphone provided in the remote controller, wherein the input voice command includes a search source and a search keyword, and', ' ...

Подробнее
07-01-2021 дата публикации

METHOD FOR CONTACT-CENTER MOBILE VIRTUAL NETWORK OPERATION

Номер: US20210006654A1
Автор: Gramacho Rui, Paiva Tiago
Принадлежит:

Methods for providing a cloud-based contact center solution configured as a mobile virtual network operator (MVNO) that can provide mobile subscription and call services and applications to contact center agents over a third-party communication system (e.g., without the contact center having its own spectrum allocation. Advanced or complex applications and services may thus be maintained/executed in the cloud-based contact center with minimal configuration to the respective mobile device that would use such applications and services. 1. A method is disclosed of configuring a mobile device for use by an agent of a cloud-based contact center to natively operate with the cloud-based contact center , the method comprising:installing in the mobile device a subscriber identification module (SIM) associated with operation with a mobile network operator comprising a cloud-based contact center;authenticating via the mobile device the subscriber identification module with the mobile network operator comprising the cloud-based contact center; andnatively routing from a network infrastructure at least a portion of a data transmission from an initiated data exchange between the mobile device and the mobile network operator comprising the cloud-based contact center, wherein the mobile network operator is configured to analyze the data transmission in one or more intelligent contact center applications.2. The method of claim 1 , further comprising:converting, at the mobile network operator comprising the cloud-based contact center, a voice portion of the routed portion of data transmission to text data; andanalyzing, at the mobile network operator comprising the cloud-based contact center, the text data to provide agent assist information in the intelligent contact center applications.3. The method of claim 2 , wherein at least one of the intelligent contact center applications is configured to:extract smart notes from the text data real-time; andtransmit the smart notes to the ...

Подробнее
07-01-2021 дата публикации

CONTACT-CENTER MOBILE VIRTUAL NETWORK OPERATOR

Номер: US20210006655A1
Автор: Gramacho Rui, Paiva Tiago
Принадлежит:

Disclosed herein are computer readable medium with instructions for a cloud-based contact center, and corresponding network devices, configured as a mobile virtual network operator (MVNO) that can provide mobile subscription and call services and applications to contact center agents over a third-party communication system. 1. A non-transitory computer readable medium comprising instructions stored thereon , wherein execution of the instructions by a processor cause the processor to:authenticate a mobile device comprising subscriber identification module, the subscriber identification module being configured with an identifier associated with a mobile network operator comprising a cloud-based contact center; andnatively route at least a portion of data transmission from an initiated data exchange between the mobile device and the mobile network operator comprising the cloud-based contact center, wherein the mobile network operator is configured to analyze the data transmission for one or more intelligent contact center applications.2. The computer readable medium of claim 1 , wherein the mobile network operator comprising the cloud-based contact center is configured to i) process a voice portion of the routed portion of data transmission and convert the voice portion to text data and ii) analyze the text data to provide agent assist information in the intelligent contact center applications.3. The computer readable medium of claim 2 , wherein the intelligent contact center applications comprise a first application configured to extract smart notes from the text data real-time and transmit the smart notes to the mobile device to be presented at a graphical user interface of the mobile device.4. The computer readable medium of claim 2 , wherein the intelligent contact center applications comprise a second application configured to search a database with the text data and transmit retrieved information from the search to the apparatus to be presented at a graphical ...

Подробнее
07-01-2021 дата публикации

SYSTEM AND METHOD FOR SPEECH-ENABLED AUTOMATED AGENT ASSISTANCE WITHIN A CLOUD-BASED CONTACT CENTER

Номер: US20210006660A1
Принадлежит:

Methods to reduce agent effort and improve customer experience quality through artificial intelligence. The Agent Assist tool provides contact centers with an innovative tool designed to reduce agent effort, improve quality and reduce costs by minimizing search and data entry tasks The Agent Assist tool is natively built and fully unified within the agent interface while keeping all data internally protected from third-party sharing. 1. A method for automated services in a geographically-distributed cloud-based contact center , the method comprising:receiving, by the contact center, a speech communication from a customer;converting, by the contact center, the speech communication to text to perform inference processing on the text to determine a customer intent;automatically analyzing, by the contact center, the text to determine a subject of the speech communication and key terms associated with the subject;automatically parsing, by the contact center, a knowledgebase using the key terms for at least one responsive answer associated with the subject; andproviding, by the contact center, the solution to an agent in a unified interface during the communication with the customer.21. The method of , further comprising:querying a customer relationship management (CRM) platform/a customer service management (CSM) platform using the key terms; anddisplaying responsive results from the CRM/CSM in the second field in the unified interface.3. The method of claim 1 , further comprising:querying a database of customer-agent transcripts using the key terms; anddisplaying responsive results from the database of customer-agent transcripts in the second field in the unified interface.4. The method of claim 1 , wherein the method is performed in real-time as the customer communication progresses with the agent.5. The method of claim 1 , further comprising concurrently displaying to the agent the text in a first field of the unified interface and the solution in a second field of ...

Подробнее
04-01-2018 дата публикации

CUSTOMER CARE DATABASE CREATION SYSTEM AND METHOD

Номер: US20180007205A1
Принадлежит:

A live service agent is associated to a first service session, where the first service session facilitates a communication exchange associated with a user inquiry between user equipment and a first service resource. The first service session remains uninterrupted by the associating of the live service agent. A verbal exchange of the communication exchange between the user equipment and the live service agent is recorded in response to the associating of the live service agent, and a number of keywords are identified. To the extent that the verbal exchange yields a satisfactory response to the user inquiry and that the recording is dissimilar to other recordings of a solution database, the recording is stored in the solution database. Other embodiments are disclosed. 1. A method , comprising:joining, by a system comprising a processing system including a processor, a live, human customer service agent to a first interactive customer service session, wherein the first interactive customer service session facilitates a communication exchange between customer equipment and a first customer service resource associated with a customer inquiry, wherein the first interactive customer service session remains uninterrupted by the joining of the live, human customer service agent;recording, by the processing system, the communication exchange responsive to the joining of the live, human customer service agent to obtain a recording, wherein the communication exchange comprises a verbal exchange between the customer equipment and the live, human customer service agent;identifying, by the processing system, a plurality of keywords within the verbal exchange;determining, by the processing system, that the verbal exchange yielded a satisfactory response to the customer inquiry;determining, by the processing system, that the recording is dissimilar to other recordings of a solution database comprising a plurality of recordings of other verbal exchanges; andstoring, by the processing ...

Подробнее
02-01-2020 дата публикации

METHOD AND APPARATUS FOR SENDING INFORMATION

Номер: US20200007656A1
Автор: Zhu Ziqiang
Принадлежит:

A method and apparatus for sending information are provided. An embodiment of the method comprises: determining, in response to acquiring a user speech audio through a target application, a user command instructed by the user speech audio, and determining whether the user command satisfies a preset trigger condition for plug-in downloading; and sending, in response to the user command satisfying the trigger condition for plug-in downloading, a request for downloading a target plug-in to a target server, the target plug-in being a plug-in in a preset plug-in set for the target application, and the plug-in being not locally installed. According to the embodiment, a terminal device may be triggered to download the plug-in based on the content instructed by the user speech audio, to implement the functional upgrading. Therefore, the self-learning capability and the self-upgrading capability of the terminal device are improved, which makes the response to the user command more accurate and more pertinent. 1. A method for sending information , comprising:determining, in response to acquiring a user speech audio through a target application, a user command instructed by the user speech audio, and determining whether the user command satisfies a preset trigger condition for plug-in downloading; andsending, in response to the user command satisfying the trigger condition for plug-in downloading, a request for downloading a target plug-in to a target server, the target plug-in being a plug-in in a preset plug-in set for the target application, and the plug-in being not locally installed.2. The method according to claim 1 , wherein the user command instructs to execute an operation claim 1 , the trigger condition for plug-in downloading includes that the operation the user command instructs to execute is inexecutable claim 1 , and the target plug-in supports the operation.3. The method according to claim 1 , wherein the trigger condition for plug-in downloading includes that ...

Подробнее
20-01-2022 дата публикации

ON-DEVICE LEARNING IN A HYBRID SPEECH PROCESSING SYSTEM

Номер: US20220020357A1
Принадлежит:

A speech interface device is configured to receive response data from a remote speech processing system for responding to user speech. This response data may be enhanced with information such as remote NLU data. The response data from the remote speech processing system may be compared to local NLU data to improve a speech processing model on the device. Thus, the device may perform supervised on-device learning based on the remote NLU data. The device may determine differences between the updated speech processing model and an original speech processing model received from the remote system and may send data indicating these differences to the remote system. The remote system may aggregate data received from a plurality of devices and may generate an improved speech processing model. 120.-. (canceled)21. A computer-implemented method , comprising:determining, by a first device, audio data representing an utterance;performing, by the first device, speech processing based on the audio data using a first model to determine output data;generating a second model configured to be used during speech processing;generating first data representing at least one difference between the first model and the second model; andsending the first data to a remote system.22. The computer-implemented method of claim 21 , further comprising:receiving, by the first device and from the remote system, second data corresponding to a third model based at least in part on the first data;determining, by the first device, second audio data representing a second utterance; andperforming, by the first device, speech processing based on the second audio data using the third model to determine second output data.23. The computer-implemented method of claim 22 , wherein the second data comprises weight data and the method further comprises claim 22 , by the first device:processing the weight data with respect to at least one of the first model or the second model to generate the third model.24. The ...

Подробнее
20-01-2022 дата публикации

Wireless audio testing

Номер: US20220020370A1
Автор: Jonathan D. Hurwitz
Принадлежит: Google LLC

A method includes outputting, by a computing device, to a remote computing device, test audio data; determining, by the computing device, whether audio data detected by an audio input device includes the test audio data; and responsive to determining that the test audio data was not detected by the audio input device, temporarily refraining, by the computing device, from outputting advisory audio data indicating the audio input device is ready to receive a spoken audio command.

Подробнее
02-01-2020 дата публикации

METHOD AND SYSTEM FOR ACCURATE AUTOMATIC CALL TRACKING AND ANALYSIS

Номер: US20200007683A1
Принадлежит:

There is disclosed a method in a data processing system for automatically and accurately determining an outcome of a phone call by using signifiers or audibles as a way to increase accuracy without altering the flow of the conversation. The disclosed method can be used in any call center, such as a high volume call center used in the financial (banking), insurance, rental (hotel or car), ticket sales, and the like. The method comprises receiving voice data of a phone call; transmitting a response communication based on the voice data, wherein the response communication includes at least one audible or signifier; identifying, by at least one processor, the at least one audible or signifier in the response communication; and automatically determining the outcome of the phone call based on the audible or signifier in the response communication. A data processing system and a non-transitory computer-readable medium for storing instructions consistent with the described method are also disclosed. 1. A method in a data processing system for automatically determining an outcome of a phone call , the method comprising:receiving voice data of a phone call;transmitting a response communication based on the voice data, wherein the response communication includes at least one audible or signifier;identifying, by at least one processor, the at least one audible or signifier in the response communication; andautomatically determining the outcome of the phone call based on the audible or signifier in the response communication.2. The method of claim 1 , wherein the response communication is generated by a call operator.3. The method of claim 1 , wherein the at least one audible or signifier includes at least one word that is not commonly used in conversation.4. The method of claim 1 , wherein the at least one audible or signifier includes a plurality of words.5. The method of claim 1 , wherein identifying the at least one audible or signifier comprises:automatically transcribing ...

Подробнее
20-01-2022 дата публикации

METHOD, DEVICE, AND PROGRAM FOR CUSTOMIZING AND ACTIVATING A PERSONAL VIRTUAL ASSISTANT SYSTEM FOR MOTOR VEHICLES

Номер: US20220020374A1
Автор: BEN ABDELAZIZ Omar
Принадлежит:

A method for customizing and activating a personal virtual assistant (PVA) in a motor vehicle. The method includes: activating a PVA management system, determining a customized mode of use of the personal virtual assistant, and activating the customized mode of use of the personal virtual assistant. A device for carrying out the method is also included. 1. A method for customizing and activating a personal virtual assistant for a motor vehicle , comprising:at least one seat suitable for accommodating a user,at least one imaging device having a field of view configured to include said user,at least one central processing unit communicating at least with said at least one imaging device, activating a PVA management system, from said central processing unit,', 'determining a customized mode of use of the personal virtual assistant, and', 'activating said customized mode of use of the personal virtual assistant., 'the method comprising at least2. Method according to claim 1 , wherein said PVA management system is activated by performing:activation of said imaging device, anddetection of at least one event indicating the presence of said user.3. Method according to claim 2 , wherein said determination of a customized mode of use of said personal virtual assistant comprises:capture of at least one image by said at least imaging device,comparison of said at least one captured image with preexisting user image data, anddetermination, by said central processing unit, of a customized mode of use of the personal virtual assistant, based on the result of said comparison.4. Method according to claim 3 , wherein said customized mode of use comprises:a default operation if the result of said comparison is an absence of recognition of the user or a recognition of a user for whom the central processing unit cannot access a language preference for said PVA system or for the language of wake words,absence of operation if the result of said comparison is recognition of a user who does ...

Подробнее
20-01-2022 дата публикации

INTERRUPT FOR NOISE-CANCELLING AUDIO DEVICES

Номер: US20220020387A1
Принадлежит:

Implementations of the subject technology provide systems and methods for determining whether to interrupt a user of an audio device that is operating in a noise-cancelling mode of operation. For example, the user may desire to be interrupted by one or more pre-designated contacts that are identified at an associated electronic device as interrupt-authorized contacts, or by a person who speaks a designated keyword to the user. 1. A device of a first user , the device comprising:secure memory storing a plurality of contacts including contacts designated at the device as interrupt-authorized contacts for a peripheral device; and provide audio content to the peripheral device, the audio content to be played by the peripheral device in a first mode of operation of the peripheral device or to be played combined with noise cancelling content by the peripheral device in a second mode of operation of the peripheral device;', 'receive, from the peripheral device, information associated with a voice input received by the peripheral device from a person other than the first user during operation of the peripheral device in the second mode of operation;', 'determine, at least in part based on the information received from the peripheral device, whether the person is one of the interrupt-authorized contacts;', 'transmit an instruction to the peripheral device to switch from the second mode of operation to the first mode of operation if it is determined that the person is one of the interrupt-authorized contacts; and', 'transmit an instruction to the peripheral device to continue operation in the second mode of operation if it is determined that the person is not one of the interrupt-authorized contact., 'one or more processors configured to2. The device of claim 1 , wherein the interrupt-authorized contacts are not authorized users of the device or the peripheral device claim 1 , and wherein determining that the person is one of the interrupt-authorized contacts does not provide ...

Подробнее
08-01-2015 дата публикации

SPEECH RECOGNITION DEVICE AND METHOD, AND SEMICONDUCTOR INTEGRATED CIRCUIT DEVICE

Номер: US20150012275A1
Автор: NONAKA Tsutomu
Принадлежит:

A semiconductor integrated circuit device for speech recognition includes a scenario setting unit that receives a command designating scenario flow information and selects prescribed speech reproduction data in a speech reproduction data storage and a prescribed conversion list, in accordance with the scenario flow information, a standard pattern extraction unit that extracts a standard pattern corresponding to at least part of individual words or sentences included in the prescribed conversion list from a speech recognition database, a speech signal synthesizer that synthesizes an output speech signal, a signal processor that generates a feature pattern representing the distribution state of the frequency component of an input speech signal, and a match detector that compares the feature pattern with the standard pattern and outputs a speech recognition result. 1. A semiconductor integrated circuit device that is used in a speech recognition device that issues a question or a message to a user based on speech reproduction data and performs speech recognition processing on speech of the user , comprising:a scenario setting unit that receives a command designating scenario flow information representing a relationship between a plurality of the speech reproduction data and a plurality of conversion lists, and, in accordance with the scenario flow information, selects prescribed speech reproduction data from among the plurality of speech reproduction data which are stored in a speech reproduction data storage, and selects a prescribed conversion list from among the plurality of conversion lists which are stored in a conversion list storage;a standard pattern extraction unit that extracts a standard pattern corresponding to at least part of individual words or sentences included in the prescribed conversion list, from a speech recognition database containing standard patterns representing a distribution state of frequency components of a plurality of phonemes that are ...

Подробнее
27-01-2022 дата публикации

SYSTEMS AND METHODS FOR PROCESSING SPEECH DIALOGUES

Номер: US20220028371A1
Автор: Han Kun, Xu Haiyang

The present disclosure is related to systems and methods for processing speech dialogue. The method includes obtaining target speech dialogue data. The method includes obtaining a text vector representation sequence, a phonetic symbol vector representation sequence, and a role vector representation sequence by performing a vector transformation on the target speech dialogue data based on a text embedding model, a phonetic symbol embedding model, and a role embedding model, respectively. The method includes determining a representation vector corresponding to the target speech dialogue data by inputting the text vector representation sequence, the phonetic symbol vector representation sequence, and the role vector representation sequence into a trained speech dialogue coding model. The method includes determining a summary of the target speech dialogue data by inputting the representation vector into a classification model. 1. A method for processing speech dialogue implemented on a computing device having at least one processor and at least one storage device , the method comprising:obtaining target speech dialogue data;obtaining a text vector representation sequence, a phonetic symbol vector representation sequence, and a role vector representation sequence by performing a vector transformation on the target speech dialogue data based on a text embedding model, a phonetic symbol embedding model, and a role embedding model, respectively;determining a representation vector corresponding to the target speech dialogue data by inputting the text vector representation sequence, the phonetic symbol vector representation sequence, and the role vector representation sequence into a trained speech dialogue coding model; anddetermining a summary of the target speech dialogue data by inputting the representation vector into a classification model.2. The method of claim 1 , further comprising:obtaining a sentence text of the summary of the target speech dialogue data; ...

Подробнее
27-01-2022 дата публикации

METHOD FOR SEMANTIC RECOGNITION, ELECTRONIC DEVICE, AND STORAGE MEDIUM

Номер: US20220028376A1

The disclosure discloses a method for semantic recognition, an electronic device, and a storage medium. The detailed solution includes: obtaining a speech recognition result of a speech to be processed, in which the speech recognition result includes a newly added recognition result fragment and a historical recognition result fragment; obtaining a semantic vector of each historical object in the historical recognition result fragment, and obtaining a semantic vector of each newly added object by inputting the semantic vector of each historical object and each newly added object in the newly added recognition result fragment into a streaming semantic coding layer; and obtaining a semantic recognition result of the speech by inputting the semantic vector of each historical object and the semantic vector of each newly added object into a streaming semantic vector fusion layer and a semantic understanding multi-task layer sequentially arranged. 1. A method for semantic recognition , comprising:obtaining a speech recognition result of a speech to be processed, wherein the speech recognition result comprises a newly added recognition result fragment and a historical recognition result fragment, and the newly added recognition result fragment is a recognition result fragment corresponding to a newly added speech fragment in the speech;obtaining a semantic vector of each historical object in the historical recognition result fragment, and obtaining a semantic vector of each newly added object by inputting the semantic vector of each historical object and each newly added object in the newly added recognition result fragment into a streaming semantic coding layer; andobtaining a semantic recognition result of the speech by inputting the semantic vector of each historical object and the semantic vector of each newly added object into a streaming semantic vector fusion layer and a semantic understanding multi-task layer sequentially arranged.2. The method of claim 1 , wherein ...

Подробнее
12-01-2017 дата публикации

METHOD FOR SYSTEM COMBINATION IN AN AUDIO ANALYTICS APPLICATION

Номер: US20170011734A1
Принадлежит:

Exemplary embodiments of the present invention provide a method of system combination in an audio analytics application including providing a plurality of language identification systems in which each of the language identification systems includes a plurality of probabilities. Each probability is associated with the system's ability to detect a particular language. The method of system combination in the audio analytics application includes receiving data at the language identification systems. The received data is different from data used to train the language identification systems. A confidence measure is determined for each of the language identification systems. The confidence measure identifies which language its system predicts for the received data and combining the language identification systems according to the confidence measures. 1. A method of system combination in an audio analytics application , comprising:providing a plurality of language identification systems, wherein each of the language identification systems includes a plurality of probabilities, wherein each probability is associated with the system's ability to detect a particular language;receiving data at the language identification systems, wherein the received data is different from data used to train the language identification systems;determining a confidence measure for each of the language identification systems, wherein the confidence measure identifies which language its system predicts for the received data; andcombining the language identification systems according to the confidence measures.2. The method of claim 1 , wherein the language identification systems have different feature extraction methods from each other.3. The method of claim 1 , wherein the language identification systems have different modeling schemes from each other.4. The method of claim 1 , wherein the language identification systems have different noise removal schemes from each other.5. The method of claim ...

Подробнее
12-01-2017 дата публикации

SPEECH RECOGNITION SYSTEM AND METHOD

Номер: US20170011735A1
Автор: Kim Dong Hyun, LEE Min Kyu
Принадлежит:

A system and a method of speech recognition which enable a spoken language to be automatically identified while recognizing speech of a person who vocalize to effectively process multilingual speech recognition without a separate process for user registration or recognized language setting such as use of a button for allowing a user to manually select a language to be vocalized and support speech recognition of each language to be automatically performed even though persons who speak different languages vocalize by using one terminal to increase convenience of the user. 1. A system of speech recognition comprising:a speech processing unit analyzing a speech signal to extract feature data; anda language identification speech recognition unit performing language identification and speech recognition by using the feature data and feeding back identified language information to the speech processing unit,wherein the speech processing unit outputs a result of the speech recognition in the language identification speech recognition unit according to the fed-back identified language information.2. The system of claim 1 , wherein the language identification speech recognition unit identifies a language for the speech signal through analysis of likelihood with respect to the feature data by referring to an acoustic model and a language model.3. The system of claim 1 , wherein the language identification speech recognition unit includesa plurality of language decoders each performing the speech recognition for the feature data in parallel and calculating a language identification score through the analysis of the likelihood every one or more speech signal frames based on the feature data by referring to the acoustic model and the language model of a corresponding language, anda language decision module deciding as the identified language a language corresponding to a selected target language decoder according to a decision rule by referring to the language identification ...

Подробнее
12-01-2017 дата публикации

Machine learning dialect identification

Номер: US20170011739A1
Автор: Fei Huang
Принадлежит: Facebook Inc

Technology is disclosed for creating and tuning classifiers for language dialects and for generating dialect-specific language modules. A computing device can receive an initial training data set as a current training data set. The selection process for the initial training data set can be achieved by receiving one or more initial content items, establishing dialect parameters of each of the initial content items, and sorting each of the initial content items into one or more dialect groups based on the established dialect parameters. The computing device can generate, based on the initial training data set, a dialect classifier configured to detect language dialects of content items to be classified. The computing device can augment the current training data set with additional training data by applying the dialect classifier to candidate content items. The computing device can then update the dialect classifier based on the augmented current training data set.

Подробнее
12-01-2017 дата публикации

DEVICE AND METHOD FOR UNDERSTANDING USER INTENT

Номер: US20170011742A1
Принадлежит: Mitsubishi Electric Corporation

A voice recognizer generates plural voice recognition results from one input speech . For each of the voice recognition results, an intent understanding processor estimates an intent to thereby output one or more candidates of intent understanding results and scores of them. A weight calculator calculates standby weights using setting information of a control target apparatus. An intent understanding corrector corrects the scores of the candidates of intent understanding result, using the standby weights, to thereby calculate their final scores, and then selects one from among the candidates of intent understanding result, as an intent understanding result , on the basis of the final scores. 110-. (canceled)11. An intent understanding device , comprising:a voice recognizer that recognizes one speech spoken in a natural language by a user, to thereby generate plural voice recognition results of highly ranked recognition scores;a morphological analyzer that converts the respective voice recognition results into morpheme strings;an intent understanding processor that estimates an intent about the speech by the user on the basis of each of the morpheme strings, to thereby output from each one of the morpheme strings, one or more candidates of intent understanding result and scores indicative of degrees of likelihood of the candidates and generate the candidates of intent understanding result in descending order of likelihoods of the plural voice recognition results;a weight calculator that calculates respective weights for the candidates of intent understanding result; andan intent understanding corrector that corrects the scores of the candidates of intent understanding result, using the weights, to thereby calculate their final scores, and then selects the candidate of intent understanding result with the final score that satisfies a preset condition first, as the intent understanding result.12. The intent understanding device of claim 11 , wherein the weight ...

Подробнее
12-01-2017 дата публикации

In-Vehicle Device, Server Device, Information System, and Content Start Method

Номер: US20170011743A1
Принадлежит:

An in-vehicle device according to the present invention includes a voice input receiving unit that receives input information with a voice, an output control unit that transmits the input information to a predetermined external device, and then outputs a plurality of received content candidates according to order of a priority degree of the content candidates, and a content execution unit that executes the content candidate upon receipt of specification of any of the output content candidates. 1. An in-vehicle device comprising:a voice input receiving unit configured to receive input information with a voice;an output control unit configured to transmit the input information to a predetermined external device, and then output a plurality of received content candidates according to order of a priority degree of the content candidates; anda content execution unit configured to execute, upon receipt of specification of any of the output content candidates, the content candidate.2. The in-vehicle device according to claim 1 , whereinthe output control unit displays a display area of the content candidate larger as the order of the priority degree is higher.3. The in-vehicle device according to claim 1 , further comprising:a biological state information acquisition unit configured to acquire biological state information indicating a state of a biological body from an external biological information detection device, whereinthe output control unit transmits the biological state information to the external device.4. The in-vehicle device according to claim 1 , further comprising:an environment information acquisition unit configured to acquire environment information, whereinthe output control unit transmits the environment information to the external device.5. A server device comprising:a keyword storage unit configured to store a word in association with a word and content related to the word;a related word search unit configured to recognize, upon receipt of information ...

Подробнее
12-01-2017 дата публикации

VIRTUAL PHOTOREALISTIC DIGITAL ACTOR SYSTEM FOR REMOTE SERVICE OF CUSTOMERS

Номер: US20170011745A1
Автор: Navaratnam Ratnakumar
Принадлежит:

A system for remote servicing of customers includes an interactive display unit at the customer location providing two-way audio/visual communication with a remote service/sales agent, wherein communication inputted by the agent is delivered to customers via a virtual Digital Actor on the display. The system also provides for remote customer service using physical mannequins with interactive capability having two-way audio visual communication ability with the remote agent, wherein communication inputted by the remote service or sales agent is delivered to customers using the physical mannequin. A web solution integrates the virtual Digital Actor system into a business website. A smart phone solution provides the remote service to customers via an App. In another embodiment, the Digital Actor is instead displayed as a 3D hologram. The Digital Actor is also used in an e-learning solution, in a movie studio suite, and as a presenter on TV, online, or other broadcasting applications. 1. A system configured to provide service to a customer from a remote service agent comprising:a video camera and a microphone configured to selectively capture video images and sounds, respectively, within a preset customer perimeter, said system configured to transmit said video images and sounds to the remote service agent;a speaker configured to emit sound within said preset customer perimeter;a sensor configured to detect a customer positioned in said preset customer perimeter, and to trigger said system to initiate said selective capture of video images and sounds therein, and said transmissions between said preset customer perimeter and the remote service agent;means for displaying a virtual digital actor to the customer; andwherein said system is configured for an input of the remote service agent to dynamically control a visual appearance of said displayed virtual digital actor on said means for displaying, and to control verbal communication emitted from said speaker, to interact ...

Подробнее