Nuance Communications, Inc.

États‑Unis d’Amérique

Retour au propriétaire

1-100 de 109 pour Nuance Communications, Inc. Trier par
Recheche Texte
Brevet
États-Unis - USPTO
Excluant les filiales
Affiner par Reset Report
Date
Nouveautés (dernières 4 semaines) 1
2024 avril (MACJ) 1
2024 février 1
2024 (AACJ) 2
2022 1
Voir plus
Classe IPC
G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité 23
G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique 21
G10L 15/00 - Reconnaissance de la parole 14
G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00 13
G10L 15/26 - Systèmes de synthèse de texte à partir de la parole 13
Voir plus
Statut
En Instance 6
Enregistré / En vigueur 103
Résultats pour  brevets
  1     2        Prochaine page

1.

System and Method for Spectral Pooling in Streaming Speech Processing

      
Numéro d'application 18162186
Statut En instance
Date de dépôt 2023-01-31
Date de la première publication 2024-04-18
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Weninger, Felix
  • Albesano, Dario
  • Zhan, Puming

Abrégé

A method, computer program product, and computing system for inserting a spectral pooling layer into a neural network of a speech processing system. An output of a hidden layer of the neural network is filtered using the spectral pooling layer with a non-integer stride. The filtered output is provided to a subsequent hidden layer of the neural network.

Classes IPC  ?

  • G10L 15/16 - Classement ou recherche de la parole utilisant des réseaux neuronaux artificiels
  • G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho

2.

INTERACTIVE VOICE RESPONSE SYSTEMS HAVING IMAGE ANALYSIS

      
Numéro d'application 17816957
Statut En instance
Date de dépôt 2022-08-02
Date de la première publication 2024-02-08
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Chawla, Akash
  • Degroot, Jenny
  • Vovk, Sergey A.

Abrégé

An interactive voice response system is provided that includes an interactive voice recognition module, an image collection module, and a data extraction module. The image collection module communicates with the voice recognition module and the user device. The extraction module communicates with the image collection module. The voice recognition module collects speech data from a user of the user device and provides an indication to the image collection module when the speech data includes complex data. The image collection module, in response to the indication, communicates with the user device in a text message. The text message includes a link that, when activated, opens a camera on the user device. The image collection module, in response to receiving an image having the complex data from the camera, communicates the image to the extraction module, which extracts the complex data from the image as textual data.

Classes IPC  ?

  • G06V 30/41 - Analyse du contenu de documents
  • G06V 30/146 - Alignement ou centrage du capteur d’image ou du champ d’image
  • G06V 30/19 - Reconnaissance utilisant des moyens électroniques
  • H04M 3/493 - Services d'information interactifs, p. ex. renseignements sur l'annuaire téléphonique
  • H04L 51/18 - Commandes ou codes exécutables
  • G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine 

3.

Automated Clinical Documentation System and Method

      
Numéro d'application 17210292
Statut En instance
Date de dépôt 2021-03-23
Date de la première publication 2022-02-17
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Gallopyn, Guido Remi Marcel
  • Sharma, Dushyant
  • Jost, Uwe Helmut
  • Owen, Donald E.
  • Naylor, Patrick
  • Nour-Eldin, Amr
  • Almendro Barreda, Daniel Paulino
  • Öz, Mehmet Mert
  • Erskine, Garret N.

Abrégé

A computer-implemented method, computer program product, and computing system for source separation is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes first audio encounter information obtained from a first encounter participant and at least second audio encounter information obtained from at least a second encounter participant. The first audio encounter information and the at least second audio encounter information are processed to eliminate audio interference between the first audio encounter information and the at least second audio encounter information. A computer-implemented method, computer program product, and computing system for source separation is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes first audio encounter information obtained from a first encounter participant and at least second audio encounter information obtained from at least a second encounter participant. The first audio encounter information and the at least second audio encounter information are processed to eliminate audio interference between the first audio encounter information and the at least second audio encounter information. A computer-implemented method, computer program product, and computing system for compartmentalizing a virtual assistant is executed on a computing device and includes obtaining encounter information via a compartmentalized virtual assistant during a user encounter, wherein the compartmentalized virtual assistant includes a core functionality module. One or more additional functionalities are added to the compartmentalized virtual assistant on an as-needed basis. A computer-implemented method, computer program product, and computing system for source separation is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes first audio encounter information obtained from a first encounter participant and at least second audio encounter information obtained from at least a second encounter participant. The first audio encounter information and the at least second audio encounter information are processed to eliminate audio interference between the first audio encounter information and the at least second audio encounter information. A computer-implemented method, computer program product, and computing system for compartmentalizing a virtual assistant is executed on a computing device and includes obtaining encounter information via a compartmentalized virtual assistant during a user encounter, wherein the compartmentalized virtual assistant includes a core functionality module. One or more additional functionalities are added to the compartmentalized virtual assistant on an as-needed basis. A computer-implemented method, computer program product, and computing system for functionality module communication is executed on a computing device and includes obtaining encounter information via a compartmentalized virtual assistant during a user encounter, wherein the compartmentalized virtual assistant includes a plurality of functionality modules. At least a portion of the encounter information may be processed via a first functionality module of the plurality of functionality modules to generate a first result. The first result may be provided to a second functionality module of the plurality of functionality modules. The first result set may be processed via the second functionality module to generate a second result. A computer-implemented method, computer program product, and computing system for source separation is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes first audio encounter information obtained from a first encounter participant and at least second audio encounter information obtained from at least a second encounter participant. The first audio encounter information and the at least second audio encounter information are processed to eliminate audio interference between the first audio encounter information and the at least second audio encounter information. A computer-implemented method, computer program product, and computing system for compartmentalizing a virtual assistant is executed on a computing device and includes obtaining encounter information via a compartmentalized virtual assistant during a user encounter, wherein the compartmentalized virtual assistant includes a core functionality module. One or more additional functionalities are added to the compartmentalized virtual assistant on an as-needed basis. A computer-implemented method, computer program product, and computing system for functionality module communication is executed on a computing device and includes obtaining encounter information via a compartmentalized virtual assistant during a user encounter, wherein the compartmentalized virtual assistant includes a plurality of functionality modules. At least a portion of the encounter information may be processed via a first functionality module of the plurality of functionality modules to generate a first result. The first result may be provided to a second functionality module of the plurality of functionality modules. The first result set may be processed via the second functionality module to generate a second result. A computer-implemented method, computer program product, and computing system for synchronizing machine vision and audio is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes machine vision encounter information and audio encounter information. The machine vision encounter information and the audio encounter information are temporally-aligned to produce a temporarily-aligned encounter recording.

Classes IPC  ?

  • G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
  • G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
  • G10L 21/0208 - Filtration du bruit
  • G06K 9/00 - Méthodes ou dispositions pour la lecture ou la reconnaissance de caractères imprimés ou écrits ou pour la reconnaissance de formes, p.ex. d'empreintes digitales

4.

Automated Clinical Documentation System and Method

      
Numéro d'application 17210253
Statut En instance
Date de dépôt 2021-03-23
Date de la première publication 2021-08-05
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Owen, Donald E.
  • Erskine, Garret N.
  • Öz, Mehmet Mert
  • Jost, Uwe Helmut
  • Almendro Barreda, Daniel Paulino
  • Sharma, Dushyant
  • Gallopyn, Guido Remi Marcel
  • Nour-Eldin, Amr
  • Naylor, Patrick A.

Abrégé

A computer-implemented method, computer program product, and computing system for rendering content is executed on a computing device and includes receiving a request to render content during a user encounter. If it is determined that the content includes sensitive content, a complete version of the content is rendered on a first device (wherein the complete version of the content includes the sensitive content) and a limited version of the content on a second device (wherein the limited version of the content excludes the sensitive content). A computer-implemented method, computer program product, and computing system for rendering content is executed on a computing device and includes receiving a request to render content during a user encounter. If it is determined that the content includes sensitive content, a complete version of the content is rendered on a first device (wherein the complete version of the content includes the sensitive content) and a limited version of the content on a second device (wherein the limited version of the content excludes the sensitive content). A modular ACD system is configured to automate clinical documentation and includes a machine vision system configured to obtain machine vision encounter information concerning a user encounter. An audio recording system is configured to obtain audio encounter information concerning the user encounter. A compute system is configured to receive the machine vision encounter information and the audio encounter information. A computer-implemented method, computer program product, and computing system for rendering content is executed on a computing device and includes receiving a request to render content during a user encounter. If it is determined that the content includes sensitive content, a complete version of the content is rendered on a first device (wherein the complete version of the content includes the sensitive content) and a limited version of the content on a second device (wherein the limited version of the content excludes the sensitive content). A modular ACD system is configured to automate clinical documentation and includes a machine vision system configured to obtain machine vision encounter information concerning a user encounter. An audio recording system is configured to obtain audio encounter information concerning the user encounter. A compute system is configured to receive the machine vision encounter information and the audio encounter information. A computer-implemented method, computer program product, and computing system for automating diarization is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. An encounter transcript is generated based, at least in part, upon the first portion of the encounter information and the at least a second portion of the encounter information. A computer-implemented method, computer program product, and computing system for rendering content is executed on a computing device and includes receiving a request to render content during a user encounter. If it is determined that the content includes sensitive content, a complete version of the content is rendered on a first device (wherein the complete version of the content includes the sensitive content) and a limited version of the content on a second device (wherein the limited version of the content excludes the sensitive content). A modular ACD system is configured to automate clinical documentation and includes a machine vision system configured to obtain machine vision encounter information concerning a user encounter. An audio recording system is configured to obtain audio encounter information concerning the user encounter. A compute system is configured to receive the machine vision encounter information and the audio encounter information. A computer-implemented method, computer program product, and computing system for automating diarization is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. An encounter transcript is generated based, at least in part, upon the first portion of the encounter information and the at least a second portion of the encounter information. A computer-implemented method, computer program product, and computing system for automating role assignment is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to associate a first portion of the encounter information with a first encounter participant. A first role is assigned to the first encounter participant.

Classes IPC  ?

  • H04N 7/18 - Systèmes de télévision en circuit fermé [CCTV], c. à d. systèmes dans lesquels le signal vidéo n'est pas diffusé
  • G06Q 30/02 - Marketing; Estimation ou détermination des prix; Collecte de fonds
  • G06F 21/62 - Protection de l’accès à des données via une plate-forme, p.ex. par clés ou règles de contrôle de l’accès

5.

Automated Clinical Documentation System and Method

      
Numéro d'application 17210233
Statut En instance
Date de dépôt 2021-03-23
Date de la première publication 2021-07-29
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Owen, Donald E.
  • Erskine, Garret N.
  • Gallopyn, Guido Remi Marcel
  • Öz, Mehmet Mert
  • Almendro Barreda, Daniel Paulino

Abrégé

A computer-implemented method, computer program product, and computing system for automated clinical documentation is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to generate an encounter transcript. At least a portion of the encounter transcript is processed to populate at least a portion of a record associated with the user encounter. A computer-implemented method, computer program product, and computing system for automated clinical documentation is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to generate an encounter transcript. At least a portion of the encounter transcript is processed to populate at least a portion of a record associated with the user encounter. A computer-implemented method, computer program product, and computing system for automating an intake process is executed on a computing device and includes prompting a user to provide encounter information via a virtual assistant during a pre-visit portion of a user encounter. Encounter information is obtained from the user in response to the prompting by the virtual assistant. A computer-implemented method, computer program product, and computing system for automated clinical documentation is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to generate an encounter transcript. At least a portion of the encounter transcript is processed to populate at least a portion of a record associated with the user encounter. A computer-implemented method, computer program product, and computing system for automating an intake process is executed on a computing device and includes prompting a user to provide encounter information via a virtual assistant during a pre-visit portion of a user encounter. Encounter information is obtained from the user in response to the prompting by the virtual assistant. A computer-implemented method, computer program product, and computing system for automating a follow-up process is executed on a computing device and includes prompting a user to provide encounter information via a virtual assistant during a post-visit portion of a user encounter. Encounter information is obtained from the user in response to the prompting by the virtual assistant. A computer-implemented method, computer program product, and computing system for automated clinical documentation is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to generate an encounter transcript. At least a portion of the encounter transcript is processed to populate at least a portion of a record associated with the user encounter. A computer-implemented method, computer program product, and computing system for automating an intake process is executed on a computing device and includes prompting a user to provide encounter information via a virtual assistant during a pre-visit portion of a user encounter. Encounter information is obtained from the user in response to the prompting by the virtual assistant. A computer-implemented method, computer program product, and computing system for automating a follow-up process is executed on a computing device and includes prompting a user to provide encounter information via a virtual assistant during a post-visit portion of a user encounter. Encounter information is obtained from the user in response to the prompting by the virtual assistant. A computer-implemented method, computer program product, and computing system for automating a monitoring process is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to determine if the encounter information is indicative of a potential situation. An inquiry is initiated concerning the potential situation.

Classes IPC  ?

  • G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
  • G16H 80/00 - TIC spécialement adaptées pour faciliter la communication entre les professionnels de la santé ou les patients, p.ex. pour le diagnostic collaboratif, la thérapie collaborative ou la surveillance collaborative de l’état de santé
  • G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
  • G16H 20/30 - TIC spécialement adaptées aux thérapies ou aux plans d’amélioration de la santé, p.ex. pour manier les prescriptions, orienter la thérapie ou surveiller l’observance par les patients concernant des thérapies ou des activités physiques, p.ex. la physiothérapie, l’acupression ou les exercices
  • A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
  • G06T 1/00 - Traitement de données d'image, d'application générale
  • G16H 30/20 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le maniement d’images médicales, p.ex. DICOM, HL7 ou PACS

6.

Automated Clinical Documentation System and Method

      
Numéro d'application 17210300
Statut En instance
Date de dépôt 2021-03-23
Date de la première publication 2021-07-29
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Owen, Donald E.
  • Erskine, Garret N.
  • Öz, Mehmet Mert
  • Almendro Barreda, Daniel Paulino

Abrégé

A computer-implemented method, computer program product, and computing system for visual diarization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. A computer-implemented method, computer program product, and computing system for visual diarization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. A computer-implemented method, computer program product, and computing system for visual compartmentalization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter portion, and associate at least a second portion of the encounter information with at least a second encounter portion. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. A computer-implemented method, computer program product, and computing system for visual diarization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. A computer-implemented method, computer program product, and computing system for visual compartmentalization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter portion, and associate at least a second portion of the encounter information with at least a second encounter portion. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. A computer-implemented method, computer program product, and computing system for reactive encounter scanning is executed on a computing device and includes obtaining encounter information of a user encounter. A request is received from a user concerning a specific condition. In response to receiving the request, the encounter information is processed to determine if the encounter information is indicative of the specific condition and to generate a result set. The result set is provided to the user. A computer-implemented method, computer program product, and computing system for visual diarization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. A computer-implemented method, computer program product, and computing system for visual compartmentalization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter portion, and associate at least a second portion of the encounter information with at least a second encounter portion. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. A computer-implemented method, computer program product, and computing system for reactive encounter scanning is executed on a computing device and includes obtaining encounter information of a user encounter. A request is received from a user concerning a specific condition. In response to receiving the request, the encounter information is processed to determine if the encounter information is indicative of the specific condition and to generate a result set. The result set is provided to the user. A computer-implemented method, computer program product, and computing system for proactive encounter scanning is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is proactively processed to determine if the encounter information is indicative of one or more conditions and to generate one or more result sets. The one or more result sets are provided to the user.

Classes IPC  ?

  • G16H 40/20 - TIC spécialement adaptées à la gestion ou à l’administration de ressources ou d’établissements de santé; TIC spécialement adaptées à la gestion ou au fonctionnement d’équipement ou de dispositifs médicaux pour la gestion ou l’administration de ressources ou d’établissements de soins de santé, p.ex. pour la gestion du personnel hospitalier ou de salles d’opération
  • G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
  • G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
  • G06Q 30/02 - Marketing; Estimation ou détermination des prix; Collecte de fonds
  • G16H 30/20 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le maniement d’images médicales, p.ex. DICOM, HL7 ou PACS
  • G06F 16/248 - Présentation des résultats de requêtes
  • G09B 19/00 - Enseignement non couvert par d'autres groupes principaux de la présente sous-classe
  • G06F 3/16 - Entrée acoustique; Sortie acoustique
  • G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine 
  • G06F 40/40 - Traitement ou traduction du langage naturel
  • G10L 25/51 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes spécialement adaptées pour un usage particulier pour comparaison ou différentiation
  • G06K 9/00 - Méthodes ou dispositions pour la lecture ou la reconnaissance de caractères imprimés ou écrits ou pour la reconnaissance de formes, p.ex. d'empreintes digitales

7.

Spectral estimation of room acoustic parameters

      
Numéro d'application 16084771
Numéro de brevet 10403300
Statut Délivré - en vigueur
Date de dépôt 2016-03-17
Date de la première publication 2019-03-14
Date d'octroi 2019-09-03
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Wolff, Tobias
  • Desiraju, Naveen Kumar

Abrégé

60, and can further estimate an additional parameter, such as Direct-to-Reverberant Ratio (DRR). The prediction filter may be adapted during a period of reverberation by minimizing a cost function. Adaptation can include using a gradient descent approach, which can operate according to a step size provided by an adaptation controller configured to determine the period of reverberation. One or more microphones can provide the signals. The reverberation parameters estimated can be applied to a reverberation suppressor, with an estimator that does not require a training phase and without relying on assumptions of the user's position relative to the microphones.

Classes IPC  ?

  • G10L 21/0224 - Traitement dans le domaine temporel
  • G10L 21/0208 - Filtration du bruit
  • G10L 21/0216 - Filtration du bruit caractérisée par le procédé d’estimation du bruit
  • H04R 3/00 - Circuits pour transducteurs
  • H04R 5/027 - Dispositions spatiales ou structurelles des microphones, p.ex. dispositifs simulant la tête humaine
  • H04R 5/04 - Circuits
  • G01H 7/00 - Mesure du temps de réverbération

8.

User dedicated automatic speech recognition

      
Numéro d'application 15876545
Numéro de brevet 10789950
Statut Délivré - en vigueur
Date de dépôt 2018-01-22
Date de la première publication 2018-06-07
Date d'octroi 2020-09-29
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Wolff, Tobias
  • Buck, Markus
  • Haulick, Tim

Abrégé

A multi-mode voice controlled user interface is described. The user interface is adapted to conduct a speech dialog with one or more possible speakers and includes a broad listening mode which accepts speech inputs from the possible speakers without spatial filtering, and a selective listening mode which limits speech inputs to a specific speaker using spatial filtering. The user interface switches listening modes in response to one or more switching cues.

Classes IPC  ?

  • G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine 
  • G10L 15/28 - Reconnaissance de la parole - Détails de structure des systèmes de reconnaissance de la parole
  • G06F 3/16 - Entrée acoustique; Sortie acoustique
  • G10L 25/51 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes spécialement adaptées pour un usage particulier pour comparaison ou différentiation
  • G10L 15/183 - Classement ou recherche de la parole utilisant une modélisation du langage naturel selon les contextes, p.ex. modèles de langage
  • G10L 21/0216 - Filtration du bruit caractérisée par le procédé d’estimation du bruit

9.

System and method for speech enhancement using a coherent to diffuse sound ratio

      
Numéro d'application 15535245
Numéro de brevet 10242690
Statut Délivré - en vigueur
Date de dépôt 2014-12-12
Date de la première publication 2017-11-16
Date d'octroi 2019-03-26
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Wolff, Tobias
  • Matheja, Timo
  • Buck, Markus

Abrégé

Embodiments of the present disclosure may include a system and method for speech enhancement using the coherent to diffuse sound ratio. Embodiments may include receiving an audio signal at one or more microphones and controlling one or more adaptive filters of a beamformer using a coherent to diffuse ratio (“CDR”).

Classes IPC  ?

  • G10L 21/0216 - Filtration du bruit caractérisée par le procédé d’estimation du bruit
  • G10L 21/0208 - Filtration du bruit
  • H04B 1/62 - TRANSMISSION - Détails des systèmes de transmission non caractérisés par le milieu utilisé pour la transmission pour produire une prédistorsion du signal à l'émission et une correction correspondante à la réception, p.ex. pour améliorer le rapport signal/bruit
  • H04R 3/00 - Circuits pour transducteurs

10.

System and method for generating a self-steering beamformer

      
Numéro d'application 15535264
Numéro de brevet 10924846
Statut Délivré - en vigueur
Date de dépôt 2014-12-12
Date de la première publication 2017-11-09
Date d'octroi 2021-02-16
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Wolff, Tobias
  • Buck, Markus

Abrégé

A system and method for generating a self-steering beamformer is provided. Embodiments may include receiving, at one or more microphones, a first audio signal and adapting one or more blocking filters based upon, at least in part, the first audio signal. Embodiments may also include generating, using the one or more blocking filters, one or more noise reference signals. Embodiments may further include providing the one or more noise reference signals to an adaptive interference canceller to reduce a beamformer output power level.

Classes IPC  ?

  • H04R 1/40 - Dispositions pour obtenir la fréquence désirée ou les caractéristiques directionnelles pour obtenir la caractéristique directionnelle désirée uniquement en combinant plusieurs transducteurs identiques
  • G10L 21/0208 - Filtration du bruit
  • G10L 21/0216 - Filtration du bruit caractérisée par le procédé d’estimation du bruit
  • H04R 1/24 - Combinaisons structurelles de transducteurs séparés ou de parties du même transducteur et sensibles respectivement à plusieurs bandes de fréquences
  • G10L 21/0272 - Séparation du signal de voix

11.

Text message generation for emergency services as a backup to voice communications

      
Numéro d'application 15134733
Numéro de brevet 09930502
Statut Délivré - en vigueur
Date de dépôt 2016-04-21
Date de la première publication 2016-08-11
Date d'octroi 2018-03-27
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Basore, David L.
  • Lawser, John Jutten

Abrégé

A mobile device may detect when a calling party dials an emergency service to request emergency assistance. Following input of the dialed digits, the device may automatically generate a text message in addition to initiating a voice call, both of which may be transmitted over a wireless data network. The wireless network nay correlate the two calls as originating from the same emergency situation and may attempt to deliver the two calls to a Public Services Answering Position (PSAP) at an appropriate emergency center. If the PSAP does not receive a voice call, the PSAP may communicate with the device via test messaging.

Classes IPC  ?

  • H04W 4/14 - Services d'envoi de messages courts, p.ex. SMS ou données peu structurées de services supplémentaires [USSD]
  • H04W 4/22 - Mise en œuvre de liaisons de secours
  • H04W 4/12 - Messagerie; Boîtes aux lettres; Annonces

12.

Voice commerce

      
Numéro d'application 14855334
Numéro de brevet 09626703
Statut Délivré - en vigueur
Date de dépôt 2015-09-15
Date de la première publication 2016-03-17
Date d'octroi 2017-04-18
Propriétaire
  • VB ASSETS, LLC (USA)
  • NUANCE COMMUNICATIONS, INC. (USA)
  • VB ASSETS, LLC (USA)
Inventeur(s) Kennewick, Sr., Michael R.

Abrégé

In certain implementations, a system for facilitating voice commerce is provided. A user input comprising a natural language utterance related to a product or service to be purchased may be received. A first product or service that is to be purchased may be determined based on the utterance. First payment information that is to be used to purchase the first product or service may be obtained. First shipping information that is to be used to deliver the first product or service may be obtained. A purchase transaction for the first product or service may completed based on the first payment information and the first shipping information without further user input, after the receipt of utterance, that identifies a product or service type or a product or service, seller information, payment information, shipping information, or other information related to purchasing the first product or service.

Classes IPC  ?

  • G06Q 30/06 - Transactions d’achat, de vente ou de crédit-bail
  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel
  • G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine 

13.

Task switching in dialogue processing

      
Numéro d'application 14478121
Numéro de brevet 09607102
Statut Délivré - en vigueur
Date de dépôt 2014-09-05
Date de la première publication 2016-03-10
Date d'octroi 2017-03-28
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Lavallee, Jean-Francois
  • Goussard, Jacques-Olivier
  • Beaufort, Richard

Abrégé

Disclosed methods and systems are directed to task switching in dialog processing. The methods and systems may include activating a primary task, receiving, one or more ambiguous natural language commands, and identifying a first candidate task for each of the one or more ambiguous natural language commands. The methods and system may also include identifying, for each of the one or more ambiguous natural language commands and based on one or more rules, a second candidate task of the plurality of tasks corresponding to the ambiguous natural language command, determining whether to modify at least one of the one or more rules-based task switching rules based on whether a quality metric satisfies a threshold quantity, and when the second quality metric satisfies the threshold quantity, changing the task switching rule for the corresponding candidate task from a rules-based model to the optimized statistical based task switching model.

Classes IPC  ?

  • G06F 17/40 - Acquisition et consignation de données
  • G06F 17/30 - Recherche documentaire; Structures de bases de données à cet effet
  • G10L 15/00 - Reconnaissance de la parole
  • G06F 3/16 - Entrée acoustique; Sortie acoustique

14.

System and method for delivering targeted advertisements and/or providing natural language processing based on advertisements

      
Numéro d'application 14836606
Numéro de brevet 09406078
Statut Délivré - en vigueur
Date de dépôt 2015-08-26
Date de la première publication 2015-12-17
Date d'octroi 2016-08-02
Propriétaire
  • VB ASSETS, LLC (USA)
  • NUANCE COMMUNICATIONS, INC. (USA)
  • VB ASSETS, LLC (USA)
Inventeur(s)
  • Freeman, Tom
  • Kennewick, Mike

Abrégé

The system and method described herein may use various natural language models to deliver targeted advertisements and/or provide natural language processing based on advertisements. In one implementation, an advertisement associated with a product or service may be provided for presentation to a user. A natural language utterance of the user may be received. The natural language utterance may be interpreted based on the advertisement and, responsive to the existence of a pronoun in the natural language utterance, a determination of whether the pronoun refers to one or more of the product or service or a provider of the product or service may be effectuated.

Classes IPC  ?

  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
  • G06Q 30/02 - Marketing; Estimation ou détermination des prix; Collecte de fonds
  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique

15.

System and method for adapting automatic speech recognition pronunciation by acoustic model restructuring

      
Numéro d'application 14698183
Numéro de brevet 09305547
Statut Délivré - en vigueur
Date de dépôt 2015-04-28
Date de la première publication 2015-08-27
Date d'octroi 2016-04-05
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Ljolje, Andrej
  • Conkie, Alistair D.
  • Syrdal, Ann K.

Abrégé

Disclosed herein are systems, computer-implemented methods, and computer-readable storage media for recognizing speech by adapting automatic speech recognition pronunciation by acoustic model restructuring. The method identifies an acoustic model and a matching pronouncing dictionary trained on typical native speech in a target dialect. The method collects speech from a new speaker resulting in collected speech and transcribes the collected speech to generate a lattice of plausible phonemes. Then the method creates a custom speech model for representing each phoneme used in the pronouncing dictionary by a weighted sum of acoustic models for all the plausible phonemes, wherein the pronouncing dictionary does not change, but the model of the acoustic space for each phoneme in the dictionary becomes a weighted sum of the acoustic models of phonemes of the typical native speech. Finally the method includes recognizing via a processor additional speech from the target speaker using the custom speech model.

Classes IPC  ?

  • G10L 15/04 - Segmentation; Détection des limites de mots
  • G10L 15/187 - Contexte phonémique, p.ex. règles de prononciation, contraintes phonotactiques ou n-grammes de phonèmes
  • G10L 15/07 - Adaptation au locuteur
  • G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
  • G10L 15/14 - Classement ou recherche de la parole utilisant des modèles statistiques, p.ex. des modèles de Markov cachés [HMM]

16.

Online maximum-likelihood mean and variance normalization for speech recognition

      
Numéro d'application 14640912
Numéro de brevet 09280979
Statut Délivré - en vigueur
Date de dépôt 2015-03-06
Date de la première publication 2015-08-06
Date d'octroi 2016-03-08
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s) Willett, Daniel

Abrégé

A feature transform for speech recognition is described. An input speech utterance is processed to produce a sequence of representative speech vectors. A time-synchronous speech recognition pass is performed using a decoding search to determine a recognition output corresponding to the speech input. The decoding search includes, for each speech vector after some first threshold number of speech vectors, estimating a feature transform based on the preceding speech vectors in the utterance and partial decoding results of the decoding search. The current speech vector is then adjusted based on the current feature transform, and the adjusted speech vector is used in a current frame of the decoding search.

Classes IPC  ?

  • G10L 19/02 - Techniques d'analyse ou de synthèse de la parole ou des signaux audio pour la réduction de la redondance, p.ex. dans les vocodeurs; Codage ou décodage de la parole ou des signaux audio utilisant les modèles source-filtre ou l’analyse psychoacoustique utilisant l'analyse spectrale, p.ex. vocodeurs à transformée ou vocodeurs à sous-bandes
  • G10L 15/02 - Extraction de caractéristiques pour la reconnaissance de la parole; Sélection d'unités de reconnaissance 
  • G10L 15/08 - Classement ou recherche de la parole
  • G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
  • G10L 19/00 - Techniques d'analyse ou de synthèse de la parole ou des signaux audio pour la réduction de la redondance, p.ex. dans les vocodeurs; Codage ou décodage de la parole ou des signaux audio utilisant les modèles source-filtre ou l’analyse psychoacoustique
  • G10L 15/34 - Adaptation d’un reconnaisseur unique pour traitement en parallèle, p.ex. par utilisation de processeurs multiples ou informatique en nuage

17.

Techniques for evaluation, building and/or retraining of a classification model

      
Numéro d'application 14686099
Numéro de brevet 09311609
Statut Délivré - en vigueur
Date de dépôt 2015-04-14
Date de la première publication 2015-08-06
Date d'octroi 2016-04-12
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s) Marcheret, Etienne

Abrégé

Techniques for evaluation and/or retraining of a classification model built using labeled training data. In some aspects, a classification model having a first set of weights is retrained by using unlabeled input to reweight the labeled training data to have a second set of weights, and by retraining the classification model using the labeled training data weighted according to the second set of weights. In some aspects, a classification model is evaluated by building a similarity model that represents similarities between unlabeled input and the labeled training data and using the similarity model to evaluate the labeled training data to identify a subset of the plurality of items of labeled training data that is more similar to the unlabeled input than a remainder of the labeled training data.

Classes IPC  ?

  • G06N 99/00 - Matière non prévue dans les autres groupes de la présente sous-classe
  • G06N 7/00 - Agencements informatiques fondés sur des modèles mathématiques spécifiques

18.

Multiple web-based content category searching in mobile search application

      
Numéro d'application 14570404
Numéro de brevet 09619572
Statut Délivré - en vigueur
Date de dépôt 2014-12-15
Date de la première publication 2015-04-09
Date d'octroi 2017-04-11
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Phillips, Michael S.
  • Nguyen, John N.

Abrégé

In embodiments of the present invention improved capabilities are described for multiple web-based content category searching for web content on a mobile communication facility comprising capturing speech presented by a user using a resident capture facility on the mobile communication facility; transmitting at least a portion of the captured speech as data through a wireless communication facility to a speech recognition facility; generating speech-to-text results for the captured speech utilizing the speech recognition facility; and transmitting the text results and a plurality of formatting rules specifying how search text may be used to form a query for a search capability on the mobile communications facility, wherein each formatting rule is associated with a category of content to be searched.

Classes IPC  ?

  • G06F 17/30 - Recherche documentaire; Structures de bases de données à cet effet
  • G10L 15/30 - Reconnaissance distribuée, p.ex. dans les systèmes client-serveur, pour les applications en téléphonie mobile ou réseaux
  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
  • G10L 25/48 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes spécialement adaptées pour un usage particulier

19.

Dealing with switch latency in speech recognition

      
Numéro d'application 14537418
Numéro de brevet 09495956
Statut Délivré - en vigueur
Date de dépôt 2014-11-10
Date de la première publication 2015-03-12
Date d'octroi 2016-11-15
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Meisel, William S.
  • Phillips, Michael S.
  • Nguyen, John N.

Abrégé

In embodiments of the present disclosure, capabilities are described for interacting with a mobile communication facility, which may include receiving a switch activation from a user to initiate a speech recognition recording session, recording the speech recognition recording session using a mobile communication facility resident capture facility, recognizing a portion of the voice command as an indication that user speech for recognition will begin following the end of the portion of the voice command, recognizing the recorded speech using a speech recognition facility to produce an external output, and using the selected output to perform a function on the mobile communication facility. The speech recognition recording session may include a voice command from the user followed by the speech to be recognized from the user.

Classes IPC  ?

  • G10L 15/08 - Classement ou recherche de la parole
  • G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine 
  • G06F 3/16 - Entrée acoustique; Sortie acoustique
  • G10L 17/22 - Procédures interactives; Interfaces homme-machine

20.

Method and system for dictionary noise removal

      
Numéro d'application 14010903
Numéro de brevet 09336195
Statut Délivré - en vigueur
Date de dépôt 2013-08-27
Date de la première publication 2015-03-05
Date d'octroi 2016-05-10
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s) Barrett, Neil D.

Abrégé

A method and system of removing noise from a dictionary using a weighted graph is presented. The method can include mapping, by a noise reducing agent executing on a processor, a plurality of dictionaries to a plurality of vertices of a graphical representation, wherein the plurality of vertices is connected by weighted edges representing noise. The plurality of dictionaries may further comprise a plurality of entries, wherein each entry further comprises a plurality of tokens. The method can include selecting a subset of the weighted edges, constructing an acyclic graphical representation from the selected subset of weighted edges, and determining an ordering based on the acyclic graphical representation. The selected subset of weighted edges may approximate a solution to the Maximum Acyclic Subgraph problem. The method can include removing noise from the plurality of dictionaries according to the determined ordering.

Classes IPC  ?

  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G06F 17/30 - Recherche documentaire; Structures de bases de données à cet effet
  • G06F 19/00 - Équipement ou méthodes de traitement de données ou de calcul numérique, spécialement adaptés à des applications spécifiques (spécialement adaptés à des fonctions spécifiques G06F 17/00;systèmes ou méthodes de traitement de données spécialement adaptés à des fins administratives, commerciales, financières, de gestion, de surveillance ou de prévision G06Q;informatique médicale G16H)

21.

System and method for delivering targeted advertisements and/or providing natural language processing based on advertisements

      
Numéro d'application 14537598
Numéro de brevet 09269097
Statut Délivré - en vigueur
Date de dépôt 2014-11-10
Date de la première publication 2015-03-05
Date d'octroi 2016-02-23
Propriétaire
  • VB ASSETS, LLC (USA)
  • NUANCE COMMUNICATIONS, INC. (USA)
  • VB ASSETS, LLC (USA)
Inventeur(s)
  • Freeman, Tom
  • Kennewick, Mike

Abrégé

The system and method described herein may use various natural language models to deliver targeted advertisements and/or provide natural language processing based on advertisements. In one implementation, an advertisement associated with a product or service may be provided for presentation to a user. A natural language utterance of the user may be received. The natural language utterance may be interpreted based on the advertisement and, responsive to the existence of a pronoun in the natural language utterance, a determination of whether the pronoun refers to one or more of the product or service or a provider of the product or service may be effectuated.

Classes IPC  ?

  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel
  • G06Q 30/02 - Marketing; Estimation ou détermination des prix; Collecte de fonds
  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique

22.

Text message generation for emergency services as a backup to voice communications

      
Numéro d'application 14478048
Numéro de brevet 09351142
Statut Délivré - en vigueur
Date de dépôt 2014-09-05
Date de la première publication 2014-12-18
Date d'octroi 2016-05-24
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Basore, David L.
  • Lawser, John Jutten

Abrégé

A mobile device may detect when a calling party dials an emergency service to request emergency assistance. Following input of the dialed digits, the device may automatically generate a text message in addition to initiating a voice call, both of which may be transmitted over a wireless data network. The wireless network may correlate the two calls as originating from the same emergency situation and may attempt to deliver the two calls to a Public Services Answering Position (PSAP) at an appropriate emergency center. If the PSAP does not receive a voice call, the PSAP may communicate with the device via test messaging.

Classes IPC  ?

  • H04W 4/22 - Mise en œuvre de liaisons de secours
  • H04W 4/12 - Messagerie; Boîtes aux lettres; Annonces

23.

System and method for providing network coordinated conversational services

      
Numéro d'application 14448216
Numéro de brevet 09761241
Statut Délivré - en vigueur
Date de dépôt 2014-07-31
Date de la première publication 2014-11-20
Date d'octroi 2017-09-12
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Maes, Stephane H.
  • Gopalakrishnan, Ponani S.

Abrégé

A system and method for providing automatic and coordinated sharing of conversational resources, e.g., functions and arguments, between network-connected servers and devices and their corresponding applications. In one aspect, a system for providing automatic and coordinated sharing of conversational resources includes a network having a first and second network device, the first and second network device each comprising a set of conversational resources, a dialog manager for managing a conversation and executing calls requesting a conversational service, and a communication stack for communicating messages over the network using conversational protocols, wherein the conversational protocols establish coordinated network communication between the dialog managers of the first and second network device to automatically share the set of conversational resources of the first and second network device, when necessary, to perform their respective requested conversational service.

Classes IPC  ?

  • G10L 15/00 - Reconnaissance de la parole
  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • H04L 29/06 - Commande de la communication; Traitement de la communication caractérisés par un protocole
  • H04L 12/24 - Dispositions pour la maintenance ou la gestion

24.

Machine translation using global lexical selection and sentence reconstruction

      
Numéro d'application 14336297
Numéro de brevet 09323745
Statut Délivré - en vigueur
Date de dépôt 2014-07-21
Date de la première publication 2014-11-06
Date d'octroi 2016-04-26
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Bangalore, Srinivas
  • Haffner, Patrick
  • Kanthak, Stephan

Abrégé

Disclosed are systems, methods, and computer-readable media for performing translations from a source language to a target language. The method comprises receiving a source phrase, generating a target bag of words based on a global lexical selection of words that loosely couples the source words/phrases and target words/phrases, and reconstructing a target phrase or sentence by considering all permutations of words with a conditional probability greater than a threshold.

Classes IPC  ?

  • G06F 17/28 - Traitement ou traduction du langage naturel
  • G11B 27/10 - Indexation; Adressage; Minutage ou synchronisation; Mesure de l'avancement d'une bande
  • G11B 27/28 - Indexation; Adressage; Minutage ou synchronisation; Mesure de l'avancement d'une bande en utilisant une information détectable sur le support d'enregistrement en utilisant des signaux d'information enregistrés par le même procédé que pour l'enregistrement principal
  • H04N 5/445 - Circuits de réception pour visualisation d'information additionnelle
  • H04N 5/45 - Image dans l'image
  • H04N 5/765 - Circuits d'interface entre un appareil d'enregistrement et un autre appareil
  • H04N 21/232 - Opération de récupération de contenu au sein d'un serveur, p.ex. lecture de flux vidéo du réseau de disques
  • H04N 21/233 - Traitement de flux audio élémentaires
  • H04N 21/235 - Traitement de données additionnelles, p.ex. brouillage de données additionnelles ou traitement de descripteurs de contenu
  • H04N 21/258 - Gestion de données liées aux clients ou aux utilisateurs finaux, p.ex. gestion des capacités des clients, préférences ou données démographiques des utilisateurs, traitement des multiples préférences des utilisateurs finaux pour générer des données co
  • H04N 21/482 - Interface pour utilisateurs finaux pour la sélection de programmes
  • H04N 21/81 - Composants mono média du contenu
  • H04N 21/84 - Génération ou traitement de données de description, p.ex. descripteurs de contenu
  • H04N 21/845 - Structuration du contenu, p.ex. décomposition du contenu en segments temporels
  • H04N 21/8547 - Création de contenu impliquant des marquages temporels pour synchroniser le contenu
  • H04N 21/2662 - Contrôle de la complexité du flux vidéo, p.ex. en mettant à l'échelle la résolution ou le débit binaire du flux vidéo en fonction des capacités du client
  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole

25.

Method for determining a set of filter coefficients for an acoustic echo compensator

      
Numéro d'application 14314106
Numéro de brevet 09264805
Statut Délivré - en vigueur
Date de dépôt 2014-06-25
Date de la première publication 2014-10-16
Date d'octroi 2016-02-16
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Buck, Markus
  • Schmidt, Gerhard Uwe
  • Wolff, Tobias

Abrégé

Methods and apparatus for beamforming and performing echo compensation for the beamformed signal with an echo canceller including calculating a set of filter coefficients as an estimate for a new steering direction without a complete adaptation of the echo canceller.

Classes IPC  ?

  • H04R 3/00 - Circuits pour transducteurs
  • G10L 21/0208 - Filtration du bruit
  • H04M 9/08 - Systèmes téléphoniques à haut-parleur à double sens comportant des moyens pour conditionner le signal, p.ex. pour supprimer les échos dans l'une ou les deux directions du trafic 
  • H04R 1/40 - Dispositions pour obtenir la fréquence désirée ou les caractéristiques directionnelles pour obtenir la caractéristique directionnelle désirée uniquement en combinant plusieurs transducteurs identiques

26.

System and method for handling missing speech data

      
Numéro d'application 14299745
Numéro de brevet 09305546
Statut Délivré - en vigueur
Date de dépôt 2014-06-09
Date de la première publication 2014-09-25
Date d'octroi 2016-04-05
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Ljolje, Andrej
  • Conkie, Alistair D.

Abrégé

Disclosed herein are systems, computer-implemented methods, and tangible computer-readable media for handling missing speech data. The computer-implemented method includes receiving speech with a missing segment, generating a plurality of hypotheses for the missing segment, identifying a best hypothesis for the missing segment, and recognizing the received speech by inserting the identified best hypothesis for the missing segment. In another method embodiment, the final step is replaced with synthesizing the received speech by inserting the identified best hypothesis for the missing segment. In one aspect, the method further includes identifying a duration for the missing segment and generating the plurality of hypotheses of the identified duration for the missing segment. The step of identifying the best hypothesis for the missing segment can be based on speech context, a pronouncing lexicon, and/or a language model. Each hypothesis can have an identical acoustic score.

Classes IPC  ?

  • G10L 15/00 - Reconnaissance de la parole
  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress

27.

Biometric authorization for real time access control

      
Numéro d'application 13787774
Numéro de brevet 09348988
Statut Délivré - en vigueur
Date de dépôt 2013-03-06
Date de la première publication 2014-09-11
Date d'octroi 2016-05-24
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Dykstra-Erickson, Elizabeth Ann
  • Daniel, Susan Dawnstarr
  • Mauro, David Andrew

Abrégé

A method of providing biometric authorization comprising enabling a user to log into an account, and determining whether there is a hold on the account. When there is a hold on the account, informing the user of the hold, and enabling the user to respond to a transaction that caused the hold. The method, in one embodiment further comprising prompting the user to enter a biometric authentication, in conjunction with the response, and processing the unblock request in real-time upon receiving and validating the biometric authentication.

Classes IPC  ?

  • G06F 7/04 - Contrôle d'égalité, c. à d. pour valeurs égales ou non
  • G06F 21/32 - Authentification de l’utilisateur par données biométriques, p.ex. empreintes digitales, balayages de l’iris ou empreintes vocales

28.

Speaker localization

      
Numéro d'application 14178309
Numéro de brevet 09622003
Statut Délivré - en vigueur
Date de dépôt 2014-02-12
Date de la première publication 2014-09-04
Date d'octroi 2017-04-11
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Schmidt, Gerhard Uwe
  • Wolff, Tobias
  • Buck, Markus
  • Valbuena, Olga Gonzalez
  • Wirsching, Gunther

Abrégé

Methods and apparatus for determining phase shift information between the first and second microphone signals for a sound signal, and determining an angle of incidence of the sound in relation to the first and second positions of the first and second microphones from the phase shift information of a band-limited test signal received by the first and second microphones for a frequency range of interest.

Classes IPC  ?

  • H04R 3/00 - Circuits pour transducteurs
  • H04R 29/00 - Dispositifs de contrôle; Dispositifs de tests
  • G10L 21/0272 - Séparation du signal de voix
  • G10L 21/0216 - Filtration du bruit caractérisée par le procédé d’estimation du bruit

29.

Machine translation using global lexical selection and sentence reconstruction

      
Numéro d'application 11686681
Numéro de brevet 08788258
Statut Délivré - en vigueur
Date de dépôt 2007-03-15
Date de la première publication 2014-07-22
Date d'octroi 2014-07-22
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Bangalore, Srinivas
  • Haffner, Patrick
  • Kanthak, Stephan

Abrégé

Disclosed are systems, methods, and computer-readable media for performing translations from a source language to a target language. The method comprises receiving a source phrase, generating a target bag of words based on a global lexical selection of words that loosely couples the source words/phrases and target words/phrases, and reconstructing a target phrase or sentence by considering all permutations of words with a conditional probability greater than a threshold.

Classes IPC  ?

  • G06F 17/28 - Traitement ou traduction du langage naturel

30.

Beamforming pre-processing for speaker localization

      
Numéro d'application 14176351
Numéro de brevet 09414159
Statut Délivré - en vigueur
Date de dépôt 2014-02-10
Date de la première publication 2014-06-05
Date d'octroi 2016-08-09
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Wolff, Tobias
  • Buck, Markus
  • Schmidt, Gerhard Uwe

Abrégé

Methods and apparatus to beamform a first plurality of microphone signals using at least one beamforming weight to obtain a first beamformed signal, beamform a second plurality of microphone signals using the at least one beamforming weight to obtain a second beamformed signal, and adjust the at least one beamforming weight so that the power density of at least one perturbation component present in the first or the second plurality of microphone signals is reduced.

Classes IPC  ?

31.

Text message generation for emergency services as a backup to voice communications

      
Numéro d'application 13689396
Numéro de brevet 08874070
Statut Délivré - en vigueur
Date de dépôt 2012-11-29
Date de la première publication 2014-05-29
Date d'octroi 2014-10-28
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Basore, David L.
  • Lawser, John Jutten

Abrégé

A mobile device may detect when a calling party dials an emergency service to request emergency assistance. Following input of the dialed digits, the device may automatically generate a text message in addition to initiating a voice call, both of which may be transmitted over a wireless data network. The wireless network may correlate the two calls as originating from the same emergency situation and may attempt to deliver the two calls to a Public Services Answering Position (PSAP) at an appropriate emergency center. If the PSAP does not receive a voice call, the PSAP may communicate with the device via test messaging.

Classes IPC  ?

  • H04M 11/04 - Systèmes de communication téléphonique spécialement adaptés pour être combinés avec d'autres systèmes électriques avec systèmes d’alarme, p.ex. systèmes d’alarme d'incendie, de police ou systèmes antivol  
  • H04W 4/22 - Mise en œuvre de liaisons de secours

32.

Accuracy improvement of spoken queries transcription using co-occurrence information

      
Numéro d'application 14156788
Numéro de brevet 09330661
Statut Délivré - en vigueur
Date de dépôt 2014-01-16
Date de la première publication 2014-05-15
Date d'octroi 2016-05-03
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Mamou, Jonathan
  • Sethy, Abhinav
  • Ramabhadran, Bhuvana
  • Hoory, Ron
  • Vozila, Paul Joseph
  • Bodenstab, Nathan

Abrégé

Techniques disclosed herein include systems and methods for voice-enabled searching. Techniques include a co-occurrence based approach to improve accuracy of the 1-best hypothesis for non-phrase voice queries, as well as for phrased voice queries. A co-occurrence model is used in addition to a statistical natural language model and acoustic model to recognize spoken queries, such as spoken queries for searching a search engine. Given an utterance and an associated list of automated speech recognition n-best hypotheses, the system rescores the different hypotheses using co-occurrence information. For each hypothesis, the system estimates a frequency of co-occurrence within web documents. Combined scores from a speech recognizer and a co-occurrence engine can be combined to select a best hypothesis with a lower word error rate.

Classes IPC  ?

  • G10L 15/00 - Reconnaissance de la parole
  • G10L 15/16 - Classement ou recherche de la parole utilisant des réseaux neuronaux artificiels
  • G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
  • G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
  • G10L 15/04 - Segmentation; Détection des limites de mots
  • G10L 15/14 - Classement ou recherche de la parole utilisant des modèles statistiques, p.ex. des modèles de Markov cachés [HMM]
  • G10L 15/28 - Reconnaissance de la parole - Détails de structure des systèmes de reconnaissance de la parole
  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel
  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • G10L 25/00 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes
  • G10L 15/08 - Classement ou recherche de la parole
  • G06F 7/00 - Procédés ou dispositions pour le traitement de données en agissant sur l'ordre ou le contenu des données maniées
  • G06F 17/30 - Recherche documentaire; Structures de bases de données à cet effet

33.

System and method for delivering targeted advertisements and tracking advertisement interactions in voice recognition contexts

      
Numéro d'application 14016757
Numéro de brevet 08886536
Statut Délivré - en vigueur
Date de dépôt 2013-09-03
Date de la première publication 2014-01-09
Date d'octroi 2014-11-11
Propriétaire
  • VB ASSETS, LLC (USA)
  • NUANCE COMMUNICATIONS, INC. (USA)
  • VB ASSETS, LLC (USA)
Inventeur(s)
  • Freeman, Tom
  • Kennwick, Mike

Abrégé

The system and method described herein may use various natural language models to deliver targeted advertisements and track advertisement interactions in voice recognition contexts. In particular, in response to an input device receiving an utterance, a conversational language processor may select and deliver one or more advertisements targeted to a user that spoke the utterance based on cognitive models associated with the user, various users having similar characteristics to the user, an environment in which the user spoke the utterance, or other criteria. Further, subsequent interaction with the targeted advertisements may be tracked to build and refine the cognitive models and thereby enhance the information used to deliver targeted advertisements in response to subsequent utterances.

Classes IPC  ?

  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • G06Q 30/00 - Commerce
  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
  • G06Q 30/02 - Marketing; Estimation ou détermination des prix; Collecte de fonds

34.

System and method for a cooperative conversational voice user interface

      
Numéro d'application 13987645
Numéro de brevet 09015049
Statut Délivré - en vigueur
Date de dépôt 2013-08-19
Date de la première publication 2013-12-19
Date d'octroi 2015-04-21
Propriétaire
  • VB ASSETS, LLC (USA)
  • NUANCE COMMUNICATIONS, INC. (USA)
  • VB ASSETS, LLC (USA)
Inventeur(s)
  • Baldwin, Larry
  • Freeman, Tom
  • Tjalve, Michael
  • Ebersold, Blane
  • Weider, Chris

Abrégé

A cooperative conversational voice user interface is provided. The cooperative conversational voice user interface may build upon short-term and long-term shared knowledge to generate one or more explicit and/or implicit hypotheses about an intent of a user utterance. The hypotheses may be ranked based on varying degrees of certainty, and an adaptive response may be generated for the user. Responses may be worded based on the degrees of certainty and to frame an appropriate domain for a subsequent utterance. In one implementation, misrecognitions may be tolerated, and conversational course may be corrected based on subsequent utterances and/or responses.

Classes IPC  ?

  • G10L 15/00 - Reconnaissance de la parole
  • G10L 25/00 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G06F 3/16 - Entrée acoustique; Sortie acoustique
  • G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine 
  • G10L 15/183 - Classement ou recherche de la parole utilisant une modélisation du langage naturel selon les contextes, p.ex. modèles de langage
  • G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur

35.

Automatic updating of confidence scoring functionality for speech recognition systems with respect to a receiver operating characteristic curve

      
Numéro d'application 13977174
Numéro de brevet 09330665
Statut Délivré - en vigueur
Date de dépôt 2011-01-07
Date de la première publication 2013-10-17
Date d'octroi 2016-05-03
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Morales, Nicolas
  • Connolly, Dermot
  • Halberstadt, Andrew

Abrégé

Automatically adjusting confidence scoring functionality is described for a speech recognition engine. Operation of the speech recognition system is revised so as to change an associated receiver operating characteristic (ROC) curve describing performance of the speech recognition system with respect to rates of false acceptance (FA) versus correct acceptance (CA). Then a confidence scoring functionality related to recognition reliability for a given input utterance is automatically adjusted such that where the ROC curve is better for a given operating point after revising the operation of the speech recognition system, the adjusting reflects a double gain constraint to maintain FA and CA rates at least as good as before revising operation of the speech recognition system.

Classes IPC  ?

  • G10L 15/01 - Estimation ou évaluation des systèmes de reconnaissance de la parole
  • G10L 15/065 - Adaptation
  • G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine 
  • G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur

36.

Integrating multimedia and voicemail

      
Numéro d'application 13868278
Numéro de brevet 09313624
Statut Délivré - en vigueur
Date de dépôt 2013-04-23
Date de la première publication 2013-09-05
Date d'octroi 2016-04-12
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Shaw, Venson M.
  • Silverman, Alexander E.

Abrégé

Integrated multimedia voicemail systems and methods allow the creation of voicemail with associated multimedia content. A user can compose a voicemail and select or create multimedia content to be associated with the voicemail. A user can associate files, webpage addresses, applications, and user-created content with a voicemail. A user may operate an interface on a user device to select content and instruct a voicemail system to associate such content with a voicemail. The voicemail with integrated multimedia content may be an originating voicemail or a voicemail in response to another voicemail.

Classes IPC  ?

  • H04M 11/00 - Systèmes de communication téléphonique spécialement adaptés pour être combinés avec d'autres systèmes électriques
  • H04W 4/12 - Messagerie; Boîtes aux lettres; Annonces
  • H04M 3/53 - Dispositions centralisées pour enregistrer des messages entrants
  • H04L 12/58 - Systèmes de commutation de messages
  • H04M 1/725 - Téléphones sans fil

37.

Message translations

      
Numéro d'application 13755903
Numéro de brevet 08688433
Statut Délivré - en vigueur
Date de dépôt 2013-01-31
Date de la première publication 2013-06-06
Date d'octroi 2014-04-01
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Davis, Joel A.
  • Kent, Jr., Larry G.
  • Daniell, W. Todd
  • Daigle, Brian K.

Abrégé

Systems for translating text messages in an instant messaging system comprise a translation engine for translating text messages into a preferred language of a recipient of the text messages. The systems are preferably configured to send and receive the text messages and to determine whether the text messages that are received in a source language are in the preferred language of the recipients so that the text messages are displayed in the preferred language of the recipients of the text messages. Other systems and methods are also provided.

Classes IPC  ?

  • G06F 17/28 - Traitement ou traduction du langage naturel

38.

System and method for structuring speech recognized text into a pre-selected document format

      
Numéro d'application 13718568
Numéro de brevet 09396166
Statut Délivré - en vigueur
Date de dépôt 2012-12-18
Date de la première publication 2013-05-02
Date d'octroi 2016-07-19
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Rosen, Lee
  • Roe, Ed
  • Poust, Wade

Abrégé

A system for creating a structured report using a template having at least one predetermined heading and formatting data associated with each heading. The steps include recording a voice file, creating a speech recognized text file corresponding to the voice file, identifying the location of each heading in the text file, and the text corresponding thereto, populating the template with the identified text corresponding to each heading, and formatting the populated template to create the structured report.

Classes IPC  ?

  • G06F 17/21 - Traitement de texte
  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
  • G06F 17/24 - Edition, p.ex. insertion/suppression

39.

Automated sentence planning in a task classification system

      
Numéro d'application 13470913
Numéro de brevet 08620669
Statut Délivré - en vigueur
Date de dépôt 2012-05-14
Date de la première publication 2013-02-14
Date d'octroi 2013-12-31
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Walker, Marilyn A.
  • Rambow, Owen Christopher
  • Rogati, Monica

Abrégé

Disclosed is a task classification system that interacts with a user. The task classification system may include a recognizer that may recognize symbols in the user's input communication, and a natural language understanding unit that may determine whether the user's input communication can be understood. If the user's input communication can be understood, the natural language understanding unit may generate understanding data. The system may also include a communicative goal generator that may generate communicative goals based on the symbols recognized by the recognizer and understanding data from the natural language understanding unit. The generated communicative goals may be related to information needed to be obtained from the user. The system may further include a sentence planning unit that may automatically plan one or more sentences based on the communicative goals generated by the communicative goal generator with at least one of the sentences plans being output to the user.

Classes IPC  ?

  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G10L 15/00 - Reconnaissance de la parole
  • G10L 25/00 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes

40.

Acoustic localization of a speaker

      
Numéro d'application 13478941
Numéro de brevet 09338549
Statut Délivré - en vigueur
Date de dépôt 2012-05-23
Date de la première publication 2012-11-22
Date d'octroi 2016-05-10
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Haulick, Tim
  • Schmidt, Gerhard Uwe
  • Buck, Markus
  • Wolff, Tobias

Abrégé

A system locates a speaker in a room containing a loudspeaker and a microphone array. The loudspeaker transmits a sound that is partly reflected by a speaker. The microphone array detects the reflected sound and converts the sound into a microphone array, the speaker's distance from the microphone array, or both, based on the characteristics of the microphone signals.

Classes IPC  ?

  • G01S 3/80 - Radiogoniomètres pour déterminer la direction d'où proviennent des ondes infrasonores, sonores, ultrasonores ou électromagnétiques ou des émissions de particules sans caractéristiques de direction utilisant des ondes ultrasonores, sonores ou infrasonores
  • H04R 3/00 - Circuits pour transducteurs
  • H04R 1/40 - Dispositions pour obtenir la fréquence désirée ou les caractéristiques directionnelles pour obtenir la caractéristique directionnelle désirée uniquement en combinant plusieurs transducteurs identiques
  • G01S 3/808 - Systèmes pour déterminer une direction ou une déviation par rapport à une direction prédéterminée utilisant des transducteurs espacés et mesurant la différence de phase ou de temps entre les signaux provenant de ces transducteurs, c. à d. systèmes à différence de parcours
  • G01S 5/30 - Détermination de distances absolues à partir de plusieurs points espacés d'emplacement connu
  • G01S 15/00 - Systèmes utilisant la réflexion ou la reradiation d'ondes acoustiques, p.ex. systèmes sonar
  • G01S 7/52 - DÉTERMINATION DE LA DIRECTION PAR RADIO; RADIO-NAVIGATION; DÉTERMINATION DE LA DISTANCE OU DE LA VITESSE EN UTILISANT DES ONDES RADIO; LOCALISATION OU DÉTECTION DE LA PRÉSENCE EN UTILISANT LA RÉFLEXION OU LA RERADIATION D'ONDES RADIO; DISPOSITIONS ANALOGUES UTILISANT D'AUTRES ONDES - Détails des systèmes correspondant aux groupes , , de systèmes selon le groupe
  • G01S 15/42 - Mesure simultanée de la distance et d'autres coordonnées
  • G01S 15/87 - Combinaisons de systèmes sonar
  • H04S 7/00 - Dispositions pour l'indication; Dispositions pour la commande, p.ex. pour la commande de l'équilibrage
  • G01S 13/00 - Systèmes utilisant la réflexion ou la reradiation d'ondes radio, p.ex. systèmes radar; Systèmes analogues utilisant la réflexion ou la reradiation d'ondes dont la nature ou la longueur d'onde sont sans importance ou non spécifiées
  • G01S 7/292 - Récepteurs avec extraction de signaux d'échos recherchés
  • H04R 29/00 - Dispositifs de contrôle; Dispositifs de tests
  • H04M 1/60 - COMMUNICATIONS TÉLÉPHONIQUES Équipement de sous-station, p.ex. pour utilisation par l'abonné comprenant des amplificateurs de parole
  • G01S 13/42 - Mesure simultanée de la distance et d'autres coordonnées
  • G01S 5/02 - Localisation par coordination de plusieurs déterminations de direction ou de ligne de position; Localisation par coordination de plusieurs déterminations de distance utilisant les ondes radioélectriques
  • G01S 15/06 - Systèmes déterminant les données relatives à la position d'une cible
  • G01S 3/802 - Systèmes pour déterminer une direction ou une déviation par rapport à une direction prédéterminée
  • G01S 7/523 - DÉTERMINATION DE LA DIRECTION PAR RADIO; RADIO-NAVIGATION; DÉTERMINATION DE LA DISTANCE OU DE LA VITESSE EN UTILISANT DES ONDES RADIO; LOCALISATION OU DÉTECTION DE LA PRÉSENCE EN UTILISANT LA RÉFLEXION OU LA RERADIATION D'ONDES RADIO; DISPOSITIONS ANALOGUES UTILISANT D'AUTRES ONDES - Détails des systèmes correspondant aux groupes , , de systèmes selon le groupe - Détails de systèmes à impulsions
  • H04B 7/08 - Systèmes de diversité; Systèmes à plusieurs antennes, c. à d. émission ou réception utilisant plusieurs antennes utilisant plusieurs antennes indépendantes espacées à la station de réception
  • G01S 3/04 - Radiogoniomètres pour déterminer la direction d'où proviennent des ondes infrasonores, sonores, ultrasonores ou électromagnétiques ou des émissions de particules sans caractéristiques de direction utilisant des ondes radio - Détails
  • G01S 5/18 - Localisation par coordination de plusieurs déterminations de direction ou de ligne de position; Localisation par coordination de plusieurs déterminations de distance utilisant des ondes ultrasonores, sonores ou infrasonores
  • G10K 11/34 - Procédés ou dispositifs pour transmettre, conduire ou diriger le son pour focaliser ou pour diriger le son, p.ex. balayage par commande électrique de systèmes de transducteurs, p.ex. en dirigeant un faisceau acoustique
  • G01S 3/00 - Radiogoniomètres pour déterminer la direction d'où proviennent des ondes infrasonores, sonores, ultrasonores ou électromagnétiques ou des émissions de particules sans caractéristiques de direction
  • G10K 15/08 - Dispositions pour produire une réverbération sonore ou un écho

41.

Method and system for automatic transcription prioritization

      
Numéro d'application 13354142
Numéro de brevet 08407050
Statut Délivré - en vigueur
Date de dépôt 2012-01-19
Date de la première publication 2012-06-28
Date d'octroi 2013-03-26
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Kobal, Jeffrey S.
  • Dhanakshirur, Girish

Abrégé

A visual toolkit for prioritizing speech transcription is provided. The toolkit can include a logger (102) for capturing information from a speech recognition system, a processor (104) for determining an accuracy rating of the information, and a visual display (106) for categorizing the information and prioritizing a transcription of the information based on the accuracy rating. The prioritizing identifies spoken utterances having a transcription priority in view of the recognized result. The visual display can include a transcription category (156) having a modifiable textbox entry with a text entry initially corresponding to a text of the recognized result, and an accept button (157) for validating a transcription of the recognized result. The categories can be automatically ranked by the accuracy rating in an ordered priority for increasing an efficiency of transcription.

Classes IPC  ?

  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole

42.

System and method for delivering targeted advertisements and tracking advertisement interactions in voice recognition contexts

      
Numéro d'application 13371870
Numéro de brevet 08527274
Statut Délivré - en vigueur
Date de dépôt 2012-02-13
Date de la première publication 2012-06-14
Date d'octroi 2013-09-03
Propriétaire
  • VB ASSETS, LLC (USA)
  • NUANCE COMMUNICATIONS, INC. (USA)
  • VB ASSETS, LLC (USA)
Inventeur(s)
  • Freeman, Tom
  • Kennewick, Mike

Abrégé

The system and method described herein may use various natural language models to deliver targeted advertisements and track advertisement interactions in voice recognition contexts. In particular, in response to an input device receiving an utterance, a conversational language processor may select and deliver one or more advertisements targeted to a user that spoke the utterance based on cognitive models associated with the user, various users having similar characteristics to the user, an environment in which the user spoke the utterance, or other criteria. Further, subsequent interaction with the targeted advertisements may be tracked to build and refine the cognitive models and thereby enhance the information used to deliver targeted advertisements in response to subsequent utterances.

Classes IPC  ?

  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel
  • G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00

43.

System and method for isolating and processing common dialog cues

      
Numéro d'application 11246604
Numéro de brevet 08185400
Statut Délivré - en vigueur
Date de dépôt 2005-10-07
Date de la première publication 2012-05-22
Date d'octroi 2012-05-22
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Goffin, Vincent J.
  • Parthasarathy, Sarangarajan

Abrégé

A method, system and machine-readable medium are provided. Speech input is received at a speech recognition component and recognized output is produced. A common dialog cue from the received speech input or input from a second source is recognized. An action is performed corresponding to the recognized common dialog cue. The performed action includes sending a communication from the speech recognition component to the speech generation component while bypassing a dialog component.

Classes IPC  ?

  • G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00
  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • G10L 13/08 - Analyse de texte ou génération de paramètres pour la synthèse de la parole à partir de texte, p.ex. conversion graphème-phonème, génération de prosodie ou détermination de l'intonation ou de l'accent tonique
  • G10L 17/00 - Identification ou vérification du locuteur
  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
  • G10L 15/00 - Reconnaissance de la parole
  • G10L 15/28 - Reconnaissance de la parole - Détails de structure des systèmes de reconnaissance de la parole
  • G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress

44.

Text entry with word prediction, completion, or correction supplemented by search of shared corpus

      
Numéro d'application 12943856
Numéro de brevet 09626429
Statut Délivré - en vigueur
Date de dépôt 2010-11-10
Date de la première publication 2012-05-10
Date d'octroi 2017-04-18
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s) Unruh, Erland

Abrégé

Searching a shared corpus is used to supplement word prediction, completion, and/or correction of text entry. A user input device at a client device receives user entry of text input comprising a string of symbols. The client device wirelessly transmits instructions to a remote site to conduct a search of a corpus using the string as a contiguous search term. From the remote site, the client device receives results of the search, including multiple sets of one or more words, each set occurring in the corpus immediately after the search term. The client device uses the received sets in word prediction, completion, and/or correction.

Classes IPC  ?

  • G06F 17/30 - Recherche documentaire; Structures de bases de données à cet effet
  • G06F 3/023 - Dispositions pour convertir sous une forme codée des éléments d'information discrets, p.ex. dispositions pour interpréter des codes générés par le clavier comme codes alphanumériques, comme codes d'opérande ou comme codes d'instruction
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique

45.

Multi-state barge-in models for spoken dialog systems

      
Numéro d'application 13279443
Numéro de brevet 08612234
Statut Délivré - en vigueur
Date de dépôt 2011-10-24
Date de la première publication 2012-04-26
Date d'octroi 2013-12-17
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s) Ljolje, Andrej

Abrégé

A method is disclosed for applying a multi-state barge-in acoustic model in a spoken dialogue system. The method includes receiving an audio speech input from the user during the presentation of a prompt, accumulating the audio speech input from the user, applying a non-speech component having at least two one-state Hidden Markov Models (HMMs) to the audio speech input from the user, applying a speech component having at least five three-state HMMs to the audio speech input from the user, in which each of the five three-state HMMs represents a different phonetic category, determining whether the audio speech input is a barge-in-speech input from the user, and if the audio speech input is determined to be the barge-in-speech input from the user, terminating the presentation of the prompt.

Classes IPC  ?

46.

Voicemail system and method for providing voicemail to text message conversion

      
Numéro d'application 11954267
Numéro de brevet 08139726
Statut Délivré - en vigueur
Date de dépôt 2007-12-12
Date de la première publication 2012-03-20
Date d'octroi 2012-03-20
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s) Zetterberg, Carl Peter

Abrégé

A method and system for allowing a calling party to send a voicemail message as a text message. A calling party leaves a voicemail message and that message is converted from voice to a text message. If the calling party wishes to confirm the conversion, the text message is then converted to a voicemail message. The converted voicemail message is presented to the calling party so that the calling party can review and edit the message. The calling party can review and edit any portion of the converted voicemail message. The edits of the voicemail message are applied and the voicemail message is converted to a new text message. If the calling party wishes to further review and edit the text message, it is converted to a new voicemail; otherwise the text message is sent to the called party.

Classes IPC  ?

  • H04M 11/00 - Systèmes de communication téléphonique spécialement adaptés pour être combinés avec d'autres systèmes électriques

47.

System and method for a cooperative conversational voice user interface

      
Numéro d'application 13251712
Numéro de brevet 08515765
Statut Délivré - en vigueur
Date de dépôt 2011-10-03
Date de la première publication 2012-01-26
Date d'octroi 2013-08-20
Propriétaire
  • VB ASSETS, LLC (USA)
  • NUANCE COMMUNICATIONS, INC. (USA)
  • VB ASSETS, LLC (USA)
Inventeur(s)
  • Baldwin, Larry
  • Freeman, Tom
  • Tjalve, Michael
  • Ebersold, Blane
  • Weider, Chris

Abrégé

A cooperative conversational voice user interface is provided. The cooperative conversational voice user interface may build upon short-term and long-term shared knowledge to generate one or more explicit and/or implicit hypotheses about an intent of a user utterance. The hypotheses may be ranked based on varying degrees of certainty, and an adaptive response may be generated for the user. Responses may be worded based on the degrees of certainty and to frame an appropriate domain for a subsequent utterance. In one implementation, misrecognitions may be tolerated, and conversational course may be corrected based on subsequent utterances and/or responses.

Classes IPC  ?

  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00
  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité

48.

Multi-pass echo residue detection with speech application intelligence

      
Numéro d'application 13236968
Numéro de brevet 08244529
Statut Délivré - en vigueur
Date de dépôt 2011-09-20
Date de la première publication 2012-01-12
Date d'octroi 2012-08-14
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s) Wong, Ngai Chiu

Abrégé

A method is provided for multi-pass echo residue detection. The method includes detecting audio data, and determining whether the audio data is recognized as speech. Additionally, the method categorizes the audio data recognized as speech as including an acceptable level of residual echo, and categorizes categorizing unrecognizable audio data as including an unacceptable level of residual echo. Furthermore, the method determines whether the unrecognizable audio data contains a user input, and also determines whether a duration of the user input is at least a predetermined duration, and when the user input is at least the predetermined duration, the method extracts the predetermined duration of the user input from a total duration of the user input.

Classes IPC  ?

  • G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
  • G10L 11/02 - Détection de présence ou d'absence de signaux de parole

49.

Method and system for using input signal quality in speech recognition

      
Numéro d'application 13205775
Numéro de brevet 08190430
Statut Délivré - en vigueur
Date de dépôt 2011-08-09
Date de la première publication 2012-01-05
Date d'octroi 2012-05-29
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Doyle, John
  • Pickering, John Brian

Abrégé

A method and system for using input signal quality in an automatic speech recognition system. The method includes measuring the quality of an input signal into a speech recognition system and varying a rejection threshold of the speech recognition system at runtime in dependence on the measurement of the input signal quality. If the measurement of the input signal quality is low, the rejection threshold is reduced and, if the measurement of the input signal quality is high, the rejection threshold is increased. The measurement of the input signal quality may be based on one or more of the measurements of signal-to-noise ratio, loudness, including clipping, and speech signal duration.

Classes IPC  ?

50.

Method and system for identifying and correcting accent-induced speech recognition difficulties

      
Numéro d'application 13228879
Numéro de brevet 08285546
Statut Délivré - en vigueur
Date de dépôt 2011-09-09
Date de la première publication 2011-12-29
Date d'octroi 2012-10-09
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s) Reich, David E.

Abrégé

A system for use in speech recognition includes an acoustic module accessing a plurality of distinct-language acoustic models, each based upon a different language; a lexicon module accessing at least one lexicon model; and a speech recognition output module. The speech recognition output module generates a first speech recognition output using a first model combination that combines one of the plurality of distinct-language acoustic models with the at least one lexicon model. In response to a threshold determination, the speech recognition output module generates a second speech recognition output using a second model combination that combines a different one of the plurality of distinct-language acoustic models with the at least one distinct-language lexicon model.

Classes IPC  ?

  • G10L 15/14 - Classement ou recherche de la parole utilisant des modèles statistiques, p.ex. des modèles de Markov cachés [HMM]
  • G10L 15/00 - Reconnaissance de la parole
  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel
  • G10L 15/10 - Classement ou recherche de la parole utilisant des mesures de distance ou de distorsion entre la parole inconnue et les gabarits de référence
  • G10L 15/28 - Reconnaissance de la parole - Détails de structure des systèmes de reconnaissance de la parole
  • G10L 17/00 - Identification ou vérification du locuteur

51.

Automated sentence planning in a task classification system

      
Numéro d'application 13230254
Numéro de brevet 08180647
Statut Délivré - en vigueur
Date de dépôt 2011-09-12
Date de la première publication 2011-12-29
Date d'octroi 2012-05-15
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Walker, Marilyn A.
  • Rambow, Owen Christopher
  • Rogati, Monica

Abrégé

The invention relates to a task classification system (900) that interacts with a user. The task classification system (900) may include a recognizer (920) that may recognize symbols in the user's input communication, and a natural language understanding unit (900) that may determine whether the user's input communication can be understood. If the user's input communication can be understood, the natural language understanding unit (930) may generate understanding data. The system may also include a communicative goal generator that may generate communicative goals based on the symbols recognized by the recognizer (920) and understanding data from the natural language understanding unit (930). The generated communicative goals may be related to information needed to be obtained from the user. The system may further include a sentence planning unit (120) that may automatically plan one or more sentences based on the communicative goals generated by the communicative goal generator with at least one of the sentences plans being output to the user.

Classes IPC  ?

  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00
  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité

52.

Method for automated sentence planning in a task classification system

      
Numéro d'application 13110628
Numéro de brevet 08209186
Statut Délivré - en vigueur
Date de dépôt 2011-05-18
Date de la première publication 2011-09-08
Date d'octroi 2012-06-26
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Walker, Marilyn A.
  • Rambow, Owen Christopher
  • Rogati, Monica

Abrégé

The invention relates to a method for sentence planning (120) in a task classification system that interacts with a user. The method may include recognizing symbols in the user's input communication and determining whether the user's input communication can be understood. If the user's communication can be understood, understanding data may be generated (220). The method may further include generating communicative goals (3010) based on the recognized symbols and understanding data. The generated communicative goals (3010) may be related to information needed to be obtained form the user. The method may also include automatically planning one or more sentences (3020) based on the generated communicative goals and outputting at least one of the sentence plans to the user (3080).

Classes IPC  ?

  • G10L 21/06 - Transformation de la parole en une représentation non audible, p.ex. visualisation de la parole ou traitement de la parole pour les aides tactiles
  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique

53.

Integrating multimedia and voicemail

      
Numéro d'application 12606503
Numéro de brevet 08447261
Statut Délivré - en vigueur
Date de dépôt 2009-10-27
Date de la première publication 2011-04-28
Date d'octroi 2013-05-21
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Shaw, Venson M.
  • Silverman, Alexander E.

Abrégé

Integrated multimedia voicemail systems and methods allow the creation of voicemail with associated multimedia content. A user can compose a voicemail and select or create multimedia content to be associated with the voicemail. A user can associate files, webpage addresses, applications, and user-created content with a voicemail. A user may operate an interface on a user device to select content and instruct a voicemail system to associate such content with a voicemail. The voicemail with integrated multimedia content may be an originating voicemail or a voicemail in response to another voicemail.

Classes IPC  ?

  • H04M 11/00 - Systèmes de communication téléphonique spécialement adaptés pour être combinés avec d'autres systèmes électriques
  • H04M 11/10 - Systèmes de communication téléphonique spécialement adaptés pour être combinés avec d'autres systèmes électriques avec systèmes d'enregistrement et de reproduction de dictée
  • H04M 1/00 - COMMUNICATIONS TÉLÉPHONIQUES Équipement de sous-station, p.ex. pour utilisation par l'abonné

54.

System and method for improving robustness of speech recognition using vocal tract length normalization codebooks

      
Numéro d'application 12869039
Numéro de brevet 08160875
Statut Délivré - en vigueur
Date de dépôt 2010-08-26
Date de la première publication 2010-12-23
Date d'octroi 2012-04-17
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s) Gilbert, Mazin

Abrégé

Disclosed are systems, methods, and computer readable media for performing speech recognition. The method embodiment comprises selecting a codebook from a plurality of codebooks with a minimal acoustic distance to a received speech sample, the plurality of codebooks generated by a process of (a) computing a vocal tract length for a each of a plurality of speakers, (b) for each of the plurality of speakers, clustering speech vectors, and (c) creating a codebook for each speaker, the codebook containing entries for the respective speaker's vocal tract length, speech vectors, and an optional vector weight for each speech vector, (2) applying the respective vocal tract length associated with the selected codebook to normalize the received speech sample for use in speech recognition, and (3) recognizing the received speech sample based on the respective vocal tract length associated with the selected codebook.

Classes IPC  ?

  • G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur

55.

System and method for selecting and presenting advertisements based on natural language processing of voice-based input

      
Numéro d'application 12847564
Numéro de brevet 08145489
Statut Délivré - en vigueur
Date de dépôt 2010-07-30
Date de la première publication 2010-11-25
Date d'octroi 2012-03-27
Propriétaire
  • VB ASSETS, LLC (USA)
  • NUANCE COMMUNICATIONS, INC. (USA)
  • VB ASSETS, LLC (USA)
Inventeur(s)
  • Freeman, Tom
  • Kennewick, Mike

Abrégé

A system and method for selecting and presenting advertisements based on natural language processing of voice-based inputs is provided. A user utterance may be received at an input device, and a conversational, natural language processor may identify a request from the utterance. At least one advertisement may be selected and presented to the user based on the identified request. The advertisement may be presented as a natural language response, thereby creating a conversational feel to the presentation of advertisements. The request and the user's subsequent interaction with the advertisement may be tracked to build user statistical profiles, thus enhancing subsequent selection and presentation of advertisements.

Classes IPC  ?

  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel
  • G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G06Q 30/00 - Commerce

56.

Automatic setting of reminders in telephony using speech recognition

      
Numéro d'application 12465731
Numéro de brevet 08145274
Statut Délivré - en vigueur
Date de dépôt 2009-05-14
Date de la première publication 2010-11-18
Date d'octroi 2012-03-27
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Gandhi, Salil P.
  • Kottawar, Saidas T.
  • Macias, Mike V.
  • Mahajan, Sandip D.

Abrégé

Systems and methods for automatically setting reminders. A method for automatically setting reminders includes receiving utterances, determining whether the utterances match a stored phrase, and in response to determining that there is a match, automatically setting a reminder in a mobile communication device. Various filters can be applied to determine whether or not to set a reminder. Examples of suitable filters include location, date/time, callee's phone number, etc.

Classes IPC  ?

  • H04B 1/38 - TRANSMISSION - Détails des systèmes de transmission non caractérisés par le milieu utilisé pour la transmission Émetteurs-récepteurs, c. à d. dispositifs dans lesquels l'émetteur et le récepteur forment un ensemble structural et dans lesquels au moins une partie est utilisée pour des fonctions d'émission et de réception

57.

Automated sentence planning in a task classification system

      
Numéro d'application 12789883
Numéro de brevet 08185401
Statut Délivré - en vigueur
Date de dépôt 2010-05-28
Date de la première publication 2010-09-23
Date d'octroi 2012-05-22
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Walker, Marilyn A.
  • Rambow, Owen Christopher
  • Rogati, Monica

Abrégé

The invention relates to a system that interacts with a user in an automated dialog system (100). The system may include a communicative goal generator (210) that generates communicative goals based on a first communication received from the user. The generated communicative goals (210) may be related to information needed to be obtained from the user. The system may further include a sentence planning unit (220) that automatically plans one or more sentences based on the communicative goals generated by the communicative goal generator (210). At least one of the planned sentences may be then output to the user (230).

Classes IPC  ?

  • G10L 21/06 - Transformation de la parole en une représentation non audible, p.ex. visualisation de la parole ou traitement de la parole pour les aides tactiles
  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique

58.

Method for determining a set of filter coefficients for an acoustic echo compensator

      
Numéro d'application 12708172
Numéro de brevet 08787560
Statut Délivré - en vigueur
Date de dépôt 2010-02-18
Date de la première publication 2010-08-26
Date d'octroi 2014-07-22
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Buck, Markus
  • Schmidt, Gerhard
  • Wolff, Tobias

Abrégé

The invention provides a method for determining a set of filter coefficients for an acoustic echo compensator in a beamformer arrangement. The acoustic echo compensator compensates for echoes within the beamformed signal. A plurality of sets of filter coefficients for the acoustic echo compensator is provided. Each set of filter coefficients corresponds to one of a predetermined number of steering directions of the beamformer arrangement. The predetermined number of steering directions is equal to or greater than the number of microphones in the microphone array. For a current steering direction, a current set of filter coefficients for the acoustic echo compensator is determined based on the provided sets of filter coefficients.

Classes IPC  ?

  • H04M 9/08 - Systèmes téléphoniques à haut-parleur à double sens comportant des moyens pour conditionner le signal, p.ex. pour supprimer les échos dans l'une ou les deux directions du trafic 
  • G01S 15/00 - Systèmes utilisant la réflexion ou la reradiation d'ondes acoustiques, p.ex. systèmes sonar
  • H04R 3/00 - Circuits pour transducteurs

59.

Speech recognition of a list entry

      
Numéro d'application 12706245
Numéro de brevet 08532990
Statut Délivré - en vigueur
Date de dépôt 2010-02-16
Date de la première publication 2010-08-19
Date d'octroi 2013-09-10
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Hillebrecht, Christian
  • Schwarz, Markus

Abrégé

The present invention relates to a method of generating a candidate list from a list of entries in accordance with a string of subword units corresponding to a speech input in a speech recognition system, the list of entries including plural list entries each comprising at least one fragment having one or more subword units. For each list entry, the fragments of the list entry are compared with the string of subword units. A matching score for each of the compared fragments based on the comparison is determined. The matching score for a fragment is further based on a comparison of at least one other fragment of the same list entry with the string of subword units. A total score for each list entry is determined based on the matching scores for the compared fragments of the respective list entry. A candidate list with the best matching entries from the list of entries based on the total scores of the list entries is generated.

Classes IPC  ?

60.

System and method for enhancing speech recognition accuracy

      
Numéro d'application 12339802
Numéro de brevet 08160879
Statut Délivré - en vigueur
Date de dépôt 2008-12-19
Date de la première publication 2010-06-24
Date d'octroi 2012-04-17
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s) Czahor, Michael

Abrégé

Systems, computer-implemented methods, and computer-readable media for enhancing speech recognition accuracy. The method includes dividing a system dialog turn into segments based on timing of probable user responses, generating a weighted grammar for each segment, exclusively activating the weighted grammar generated for a current segment of the dialog turn during the current segment of the dialog turn, and recognizing user speech received during the current segment using the activated weighted grammar generated for the current segment. The method can further include assigning probability to the weighted grammar based on historical user responses and activating each weighted grammar is based on the assigned probability. Weighted grammars can be generated based on a user profile. A weighted grammar can be generated for two or more segments. Exclusively activating each weighted grammar can include a transition period blending the previously activated grammar and the grammar to be activated.

Classes IPC  ?

  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel

61.

User intention based on N-best list of recognition hypotheses for utterances in a dialog

      
Numéro d'application 12325786
Numéro de brevet 08140328
Statut Délivré - en vigueur
Date de dépôt 2008-12-01
Date de la première publication 2010-06-03
Date d'octroi 2012-03-20
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s) Williams, Jason

Abrégé

Disclosed herein are systems, computer-implemented methods, and tangible computer-readable media for using alternate recognition hypotheses to improve whole-dialog understanding accuracy. The method includes receiving an utterance as part of a user dialog, generating an N-best list of recognition hypotheses for the user dialog turn, selecting an underlying user intention based on a belief distribution across the generated N-best list and at least one contextually similar N-best list, and responding to the user based on the selected underlying user intention. Selecting an intention can further be based on confidence scores associated with recognition hypotheses in the generated N-best lists, and also on the probability of a user's action given their underlying intention. A belief or cumulative confidence score can be assigned to each inferred user intention.

Classes IPC  ?

  • G10L 15/14 - Classement ou recherche de la parole utilisant des modèles statistiques, p.ex. des modèles de Markov cachés [HMM]

62.

Method and device for locating a sound source

      
Numéro d'application 12547681
Numéro de brevet 08194500
Statut Délivré - en vigueur
Date de dépôt 2009-08-26
Date de la première publication 2010-03-04
Date d'octroi 2012-06-05
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Wolff, Tobias
  • Buck, Markus
  • Schmidt, Gerhard
  • Valbuena, Olga González
  • Wirsching, Günther

Abrégé

A method of locating a sound source based on sound received at an array of microphones comprises the steps of determining a correlation function of signals provided by microphones of the array and establishing a direction in which the sound source is located based on at least one eigenvector of a matrix having matrix elements which are determined based on the correlation function. The correlation function has first and second frequency components associated with a first and second frequency band, respectively. The first frequency component is determined based on signals from microphones having a first distance, and the second frequency component is determined based on signals from microphones having a second distance different from the first distance.

Classes IPC  ?

  • G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
  • G01S 3/80 - Radiogoniomètres pour déterminer la direction d'où proviennent des ondes infrasonores, sonores, ultrasonores ou électromagnétiques ou des émissions de particules sans caractéristiques de direction utilisant des ondes ultrasonores, sonores ou infrasonores
  • H04R 3/00 - Circuits pour transducteurs

63.

Method and apparatus for providing voice control for accessing teleconference services

      
Numéro d'application 12553700
Numéro de brevet 08184792
Statut Délivré - en vigueur
Date de dépôt 2009-09-03
Date de la première publication 2009-12-31
Date d'octroi 2012-05-22
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Croak, Marian
  • Eslambolchi, Hossein

Abrégé

A method and apparatus for providing access to teleconference services using voice recognition technology to receive information on packet networks such as Voice over Internet Protocol (VoIP) and Service over Internet Protocol (SoIP) networks are disclosed. In one embodiment, the service provider enables a caller to enter access information for accessing a conference service using at least one natural language response.

Classes IPC  ?

  • H04M 3/42 - Systèmes fournissant des fonctions ou des services particuliers aux abonnés

64.

Method and system for training a text-to-speech synthesis system using a specific domain speech database

      
Numéro d'application 12540441
Numéro de brevet 08135591
Statut Délivré - en vigueur
Date de dépôt 2009-08-13
Date de la première publication 2009-12-03
Date d'octroi 2012-03-13
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s) Schroeter, Horst Juergen

Abrégé

A method and system are disclosed that train a text-to-speech synthesis system for use in speech synthesis. The method includes generating a speech database of audio files comprising domain-specific voices having various prosodies, and training a text-to-speech synthesis system using the speech database by selecting audio segments having a prosody based on at least one dialog state. The system includes a processor, a speech database of audio files, and modules for implementing the method.

Classes IPC  ?

  • G10L 13/00 - Synthèse de la parole; Systèmes de synthèse de la parole à partir de texte

65.

Low latency real-time vocal tract length normalization

      
Numéro d'application 12490634
Numéro de brevet 08909527
Statut Délivré - en vigueur
Date de dépôt 2009-06-24
Date de la première publication 2009-10-15
Date d'octroi 2014-12-09
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Goffin, Vincent
  • Ljolje, Andrej
  • Saraclar, Murat

Abrégé

A method and system for training an automatic speech recognition system are provided. The method includes separating training data into speaker specific segments, and for each speaker specific segment, performing the following acts: generating spectral data, selecting a first warping factor and warping the spectral data, and comparing the warped spectral data with a speech model. The method also includes iteratively performing the steps of selecting another warping factor and generating another warped spectral data, comparing the other warped spectral data with the speech model, and if the other warping factor produces a closer match to the speech model, saving the other warping factor as the best warping factor for the speaker specific segment. The system includes modules configured to control a processor in the system to perform the steps of the method.

Classes IPC  ?

  • G10L 15/12 - Classement ou recherche de la parole utilisant des techniques de programmation dynamique, p.ex. normalisation temporelle par comparaison dynamique [DTW]
  • G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
  • G10L 15/02 - Extraction de caractéristiques pour la reconnaissance de la parole; Sélection d'unités de reconnaissance 

66.

System for distinguishing desired audio signals from noise

      
Numéro d'application 12269837
Numéro de brevet 08131544
Statut Délivré - en vigueur
Date de dépôt 2008-11-12
Date de la première publication 2009-09-10
Date d'octroi 2012-03-06
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Herbig, Tobias
  • Gaupp, Oliver
  • Gerl, Franz

Abrégé

A system distinguishes a primary audio source and background noise to improve the quality of an audio signal. A speech signal from a microphone may be improved by identifying and dampening background noise to enhance speech. Stochastic models may be used to model speech and to model background noise. The models may determine which portions of the signal are speech and which portions are noise. The distinction may be used to improve the signal's quality, and for speaker identification or verification.

Classes IPC  ?

  • G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress

67.

Voice response system

      
Numéro d'application 12253849
Numéro de brevet 08145494
Statut Délivré - en vigueur
Date de dépôt 2008-10-17
Date de la première publication 2009-06-04
Date d'octroi 2012-03-27
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Horioka, Masaru
  • Atake, Yoshinori
  • Tahara, Yoshinori

Abrégé

A voice response system attempts to respond to spoken user input and to provide computer-generated responses. If the system decides it cannot provide valid responses, the current state of user session is determined and forwarded to a human operator for further action. The system maintains a recorded history of the session in the form of a dialog history log. The dialog history and information as to the reliability of past speech recognition efforts is employed in making the current state determination. The system includes formatting rules for controlling the display of information presented to the human operator.

Classes IPC  ?

  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité

68.

System and method for conducting a search using a wireless mobile device

      
Numéro d'application 12350848
Numéro de brevet 08285273
Statut Délivré - en vigueur
Date de dépôt 2009-01-08
Date de la première publication 2009-05-07
Date d'octroi 2012-10-09
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s) Roth, Daniel L.

Abrégé

A method and system are provided by which a wireless mobile device takes a vocally entered query and transmits it in a text message format over a wireless network to a search engine; receives search results based on the query from the search engine over the wireless network; and displays the search results.

Classes IPC  ?

  • H04W 4/00 - Services spécialement adaptés aux réseaux de télécommunications sans fil; Leurs installations

69.

Voice conversion method and system

      
Numéro d'application 12240148
Numéro de brevet 08234110
Statut Délivré - en vigueur
Date de dépôt 2008-09-29
Date de la première publication 2009-04-02
Date d'octroi 2012-07-31
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Meng, Fan Ping
  • Qin, Yong
  • Shi, Qin
  • Shuang, Zhi Wei

Abrégé

A method, system and computer program product for voice conversion. The method includes performing speech analysis on the speech of a source speaker to achieve speech information; performing spectral conversion based on said speech information, to at least achieve a first spectrum similar to the speech of a target speaker; performing unit selection on the speech of said target speaker at least using said first spectrum as a target; replacing at least part of said first spectrum with the spectrum of the selected target speaker's speech unit; and performing speech reconstruction at least based on the replaced spectrum.

Classes IPC  ?

  • G10L 19/06 - Détermination ou codage des caractéristiques spectrales, p.ex. des coefficients de prédiction à court terme

70.

Creation and use of application-generic class-based statistical language models for automatic speech recognition

      
Numéro d'application 11845015
Numéro de brevet 08135578
Statut Délivré - en vigueur
Date de dépôt 2007-08-24
Date de la première publication 2009-02-26
Date d'octroi 2012-03-13
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s) Hébert, Matthieu

Abrégé

A method of creating an application-generic class-based SLM includes, for each of a plurality of speech applications, parsing a corpus of utterance transcriptions to produce a first output set, in which expressions identified in the corpus are replaced with corresponding grammar tags from a grammar that is specific to the application. The method further includes, for each of the plurality of speech applications, replacing each of the grammar tags in the first output set with a class identifier of an application-generic class, to produce a second output set. The method further includes processing the resulting second output sets with a statistical language model (SLM) trainer to generate an application-generic class-based SLM.

Classes IPC  ?

  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique

71.

Using speech recognition results based on an unstructured language model in a mobile communication facility application

      
Numéro d'application 12184375
Numéro de brevet 08886540
Statut Délivré - en vigueur
Date de dépôt 2008-08-01
Date de la première publication 2009-01-29
Date d'octroi 2014-11-11
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Cerra, Joseph P.
  • Nguyen, John N.
  • Phillips, Michael S.
  • Shu, Han
  • Mischke, Alexandra Beth

Abrégé

A method and system for entering information into a software application resident on a mobile communication facility is provided. The method and system may include recording speech presented by a user using a mobile communication facility resident capture facility, transmitting the recording through a wireless communication facility to a speech recognition facility, transmitting information relating to the software application to the speech recognition facility, generating results utilizing the speech recognition facility using an unstructured language model based at least in part on the information relating to the software application and the recording, transmitting the results to the mobile communications facility, loading the results into the software application and simultaneously displaying the results as a set of words and as a set of application results based on those words.

Classes IPC  ?

  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • G10L 15/30 - Reconnaissance distribuée, p.ex. dans les systèmes client-serveur, pour les applications en téléphonie mobile ou réseaux

72.

System and method of performing user-specific automatic speech recognition

      
Numéro d'application 12207175
Numéro de brevet 08145481
Statut Délivré - en vigueur
Date de dépôt 2008-09-09
Date de la première publication 2009-01-01
Date d'octroi 2012-03-27
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Gajic, Bojana
  • Narayanan, Shrikanth Sambasivan
  • Parthasarathy, Sarangarajan
  • Rose, Richard Cameron
  • Rosenberg, Aaron Edward

Abrégé

Speech recognition models are dynamically re-configurable based on user information, application information, background information such as background noise and transducer information such as transducer response characteristics to provide users with alternate input modes to keyboard text entry. Word recognition lattices are generated for each data field of an application and dynamically concatenated into a single word recognition lattice. A language model is applied to the concatenated word recognition lattice to determine the relationships between the word recognition lattices and repeated until the generated word recognition lattices are acceptable or differ from a predetermined value only by a threshold amount. These techniques of dynamic re-configurable speech recognition provide for deployment of speech recognition on small devices such as mobile phones and personal digital assistants as well environments such as office, home or vehicle while maintaining the accuracy of the speech recognition.

Classes IPC  ?

  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité

73.

Method and device for providing speech-to-text encoding and telephony service

      
Numéro d'application 12200292
Numéro de brevet 08265931
Statut Délivré - en vigueur
Date de dépôt 2008-08-28
Date de la première publication 2008-12-25
Date d'octroi 2012-09-11
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Caldwell, Charles David
  • Harlow, John Bruce
  • Sayko, Robert J.
  • Shaye, Norman

Abrégé

A machine-readable medium and a network device are provided for speech-to-text translation. Speech packets are received at a broadband telephony interface and stored in a buffer. The speech packets are processed and textual representations thereof are displayed as words on a display device. Speech processing is activated and deactivated in response to a command from a subscriber.

Classes IPC  ?

  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • H04M 1/64 - Dispositions automatiques pour répondre aux appels; Dispositions automatiques pour enregistrer des messages pour abonnés absents; Dispositions pour enregistrer des conversations

74.

Method and system for speech based document history tracking

      
Numéro d'application 12096068
Numéro de brevet 08140338
Statut Délivré - en vigueur
Date de dépôt 2006-11-10
Date de la première publication 2008-12-18
Date d'octroi 2012-03-20
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Grobauer, Gerhard
  • Papai, Miklos

Abrégé

A method and a system of history tracking corrections in a speech based document are disclosed. The speech based document comprises one or more sections of text recognized or transcribed from sections of speech, wherein the sections of speech are dictated by a user and processed by a speech recognizer in a speech recognition system into corresponding sections of text of the speech based document. The method comprises associating of at least one speech attribute (14) to each section of text in the speech based document, said speech attribute (14) comprising information related to said section of text, respectively; presenting said speech based document on a presenting unit (8); detecting an action being performed within any of said sections of text; and updating information of said speech attributes (14) related to the kind of action detected on one of said sections of text for updating said speech based document, whereby said updated information of said speech attributes (14) is used for history tracking corrections of said speech based document.

Classes IPC  ?

  • G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00
  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
  • G06F 3/00 - Dispositions d'entrée pour le transfert de données destinées à être traitées sous une forme maniable par le calculateur; Dispositions de sortie pour le transfert de données de l'unité de traitement à l'unité de sortie, p.ex. dispositions d'interface

75.

Speech recognition system with huge vocabulary

      
Numéro d'application 12096046
Numéro de brevet 08140336
Statut Délivré - en vigueur
Date de dépôt 2006-12-06
Date de la première publication 2008-11-27
Date d'octroi 2012-03-20
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s) Saffer, Zsolt

Abrégé

The invention deals with speech recognition, such as a system for recognizing words in continuous speech. A speech recognition system is disclosed which is capable of recognizing a huge number of words, and in principle even an unlimited number of words. The speech recognition system comprises a word recognizer for deriving a best path through a word graph, and wherein words are assigned to the speech based on the best path. The word score being obtained from applying a phonemic language model to each word of the word graph. Moreover, the invention deals with an apparatus and a method for identifying words from a sound block and to computer readable code for implementing the method.

Classes IPC  ?

  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel
  • G10L 15/04 - Segmentation; Détection des limites de mots
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique

76.

Acoustic localization of a speaker

      
Numéro d'application 12104836
Numéro de brevet 08204248
Statut Délivré - en vigueur
Date de dépôt 2008-04-17
Date de la première publication 2008-11-20
Date d'octroi 2012-06-19
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Haulick, Tim
  • Schmidt, Gerhard Uwe
  • Buck, Markus
  • Wolff, Tobias

Abrégé

A system locates a speaker in a room containing a loudspeaker and a microphone array. The loudspeaker transmits a sound that is partly reflected by a speaker. The microphone array detects the reflected sound and converts the sound into a microphone signal. A processor determines the speaker's direction relative to the microphone array, the speaker's distance from the microphone array, or both, based on the characteristics of the microphone signals.

Classes IPC  ?

77.

Categorization of information using natural language processing and predefined templates

      
Numéro d'application 12121527
Numéro de brevet 08185553
Statut Délivré - en vigueur
Date de dépôt 2008-05-15
Date de la première publication 2008-10-16
Date d'octroi 2012-05-22
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Carus, Alwin B.
  • Ogrinc, Harry J.

Abrégé

A computer implemented method for generating a report that includes latent information, comprising receiving an input data stream that includes latent information, performing one of normalization, validation, and extraction of the input data stream, processing the input data stream to identify latent information within the data stream that is required for generation of a particular report, wherein said processing of the input data stream to identify latent information comprises of identifying a relevant portion of the input data stream, bounding the relevant portion of the input data stream, classifying and normalizing the bounded data, activating a relevant report template based on said identified latent information, populating said template with template-specified data, and processing the template-specified data to generate a report.

Classes IPC  ?

  • G06F 17/30 - Recherche documentaire; Structures de bases de données à cet effet

78.

Method for dialog management

      
Numéro d'application 12140805
Numéro de brevet 08600747
Statut Délivré - en vigueur
Date de dépôt 2008-06-17
Date de la première publication 2008-10-09
Date d'octroi 2013-12-03
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Abella, Alicia
  • Gorin, Allen Louis

Abrégé

A spoken dialog system and method having a dialog management module are disclosed. The dialog management module includes a plurality of dialog motivators for handling various operations during a spoken dialog. The dialog motivators comprise an error handling, disambiguation, assumption, confirmation, missing information, and continuation. The spoken dialog system uses the assumption dialog motivator in either a-priori or a-posteriori modes. A-priori assumption is based on predefined requirements for the call flow and a-posteriori assumption can work with the confirmation dialog motivator to assume the content of received user input and confirm received user input.

Classes IPC  ?

  • G10L 15/00 - Reconnaissance de la parole
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G06F 3/048 - Techniques d’interaction fondées sur les interfaces utilisateur graphiques [GUI]
  • G06F 9/46 - Dispositions pour la multiprogrammation
  • G06F 9/44 - Dispositions pour exécuter des programmes spécifiques
  • G06F 17/20 - Manipulation de données en langage naturel
  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
  • H04M 1/64 - Dispositions automatiques pour répondre aux appels; Dispositions automatiques pour enregistrer des messages pour abonnés absents; Dispositions pour enregistrer des conversations
  • H04M 11/00 - Systèmes de communication téléphonique spécialement adaptés pour être combinés avec d'autres systèmes électriques

79.

Natural error handling in speech recognition

      
Numéro d'application 12135452
Numéro de brevet 08355920
Statut Délivré - en vigueur
Date de dépôt 2008-06-09
Date de la première publication 2008-10-02
Date d'octroi 2013-01-15
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Gopinath, Ramesh A.
  • Maison, Benoit
  • Wu, Brian C.

Abrégé

A user interface, and associated techniques, that permit a fast and efficient way of correcting speech recognition errors, or of diminishing their impact. The user may correct mistakes in a natural way, essentially by repeating the information that was incorrectly recognized previously. Such a mechanism closely approximates what human-to-human dialogue would be in similar circumstances. Such a system fully takes advantage of all the information provided by the user, and on its own estimates the quality of the recognition in order to determine the correct sequence of words in the fewest number of steps.

Classes IPC  ?

  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00

80.

Method and apparatus for data capture using a voice activated workstation

      
Numéro d'application 12089033
Numéro de brevet 08165876
Statut Délivré - en vigueur
Date de dépôt 2006-09-04
Date de la première publication 2008-09-25
Date d'octroi 2012-04-24
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Emam, Ossama
  • Gamal, Khaled

Abrégé

A method and apparatus for capturing data in a workstation, wherein a large number of data associated with a sample which is viewed, by a user, through an optical device, such as a microscope, is to be entered in a computer related file. The optical device can be moved to a data-sampling position utilizing voice commands. A pointer can then be moved to an appropriate place in the file to receive the data relating to the data-sampling position. Data can be then entered in the appropriate position utilizing a voice command. The steps of moving the pointer and entering the data can then be repeated until all data is provided with respect to the data-sampling positions.

Classes IPC  ?

81.

Invoking tapered prompts in a multimodal application

      
Numéro d'application 11678920
Numéro de brevet 08150698
Statut Délivré - en vigueur
Date de dépôt 2007-02-26
Date de la première publication 2008-08-28
Date d'octroi 2012-04-03
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Ativanichayaphong, Soonthorn
  • Cross, Jr., Charles W.
  • Mccobb, Gerald M.

Abrégé

Methods, apparatus, and computer program products are described for invoking tapered prompts in a multimodal application implemented with a multimodal browser and a multimodal application operating on a multimodal device supporting multiple modes of user interaction with the multimodal application, the modes of user interaction including a voice mode and one or more non-voice modes. Embodiments include identifying, by a multimodal browser, a prompt element in a multimodal application; identifying, by the multimodal browser, one or more attributes associated with the prompt element; and playing a speech prompt according to the one or more attributes associated with the prompt element.

Classes IPC  ?

  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité

82.

System and method for selecting and presenting advertisements based on natural language processing of voice-based input

      
Numéro d'application 11671526
Numéro de brevet 07818176
Statut Délivré - en vigueur
Date de dépôt 2007-02-06
Date de la première publication 2008-08-07
Date d'octroi 2010-10-19
Propriétaire
  • VB ASSETS, LLC (USA)
  • NUANCE COMMUNICATIONS, INC. (USA)
  • VB ASSETS, LLC (USA)
Inventeur(s)
  • Freeman, Tom
  • Kennewick, Mike

Abrégé

A system and method for selecting and presenting advertisements based on natural language processing of voice-based inputs is provided. A user utterance may be received at an input device, and a conversational, natural language processor may identify a request from the utterance. At least one advertisement may be selected and presented to the user based on the identified request. The advertisement may be presented as a natural language response, thereby creating a conversational feel to the presentation of advertisements. The request and the user's subsequent interaction with the advertisement may be tracked to build user statistical profiles, thus enhancing subsequent selection and presentation of advertisements.

Classes IPC  ?

  • G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00
  • G10L 15/18 - Classement ou recherche de la parole utilisant une modélisation du langage naturel
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G06Q 30/00 - Commerce

83.

Method and an apparatus to disambiguate requests

      
Numéro d'application 11701811
Numéro de brevet 08175248
Statut Délivré - en vigueur
Date de dépôt 2007-02-02
Date de la première publication 2008-08-07
Date d'octroi 2012-05-08
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Agarwal, Rajeev
  • Ardman, David
  • Master, Muneeb
  • Mauro, David Andrew
  • Raman, Vijay R.
  • Ulug, Amy E.
  • Valli, Zulfikar

Abrégé

A method and an apparatus to disambiguate requests are presented. In one embodiment, the method includes receiving a request for information from a user. Then data is retrieved from a back-end database in response to the request. Based on a predetermined configuration of a disambiguation system and the data retrieved, the ambiguity within the request is dynamically resolved.

Classes IPC  ?

  • H04M 3/42 - Systèmes fournissant des fonctions ou des services particuliers aux abonnés

84.

Method and apparatus for recognizing and reacting to user personality in accordance with speech recognition system

      
Numéro d'application 12055952
Numéro de brevet 08719035
Statut Délivré - en vigueur
Date de dépôt 2008-03-26
Date de la première publication 2008-07-24
Date d'octroi 2014-05-06
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Stewart, Osamuyimen Thompson
  • Dai, Liwei

Abrégé

Techniques are disclosed for recognizing user personality in accordance with a speech recognition system. For example, a technique for recognizing a personality trait associated with a user interacting with a speech recognition system includes the following steps/operations. One or more decoded spoken utterances of the user are obtained. The one or more decoded spoken utterances are generated by the speech recognition system. The one or more decoded spoken utterances are analyzed to determine one or more linguistic attributes (morphological and syntactic filters) that are associated with the one or more decoded spoken utterances. The personality trait associated with the user is then determined based on the analyzing step/operation.

Classes IPC  ?

  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G10L 15/00 - Reconnaissance de la parole
  • G10L 25/00 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes
  • G09B 3/00 - Matériel d'enseignement à commande manuelle ou mécanique procédant par questions et réponses
  • G09B 7/00 - Dispositifs ou appareils d'enseignement à commande électrique procédant par questions et réponses
  • G09B 19/00 - Enseignement non couvert par d'autres groupes principaux de la présente sous-classe
  • G09B 19/04 - Elocution
  • G09B 17/04 - Enseignement de la lecture pour l'accroissement de la cadence de lecture; Contrôle de la cadence de lecture
  • G09B 1/00 - Matériel à but éducatif à commande manuelle ou mécanique utilisant des éléments formant ou comportant des symboles, des signes, des images ou similaires, qui sont agencés ou adaptés pour être disposés selon un ou plusieurs schémas particuliers

85.

Software program and method for providing promotions on a phone prior to call connection

      
Numéro d'application 11636334
Numéro de brevet 08160552
Statut Délivré - en vigueur
Date de dépôt 2006-12-08
Date de la première publication 2008-06-12
Date d'octroi 2012-04-17
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s) Stone, Kevin M.

Abrégé

The present invention includes a method and software application for providing a promotion to a user on a phone. The software application resides on a user's phone and “listens” for phone numbers dialed by a user. In response to the user dialing a phone number, the software determines whether a promotion or an offer for a promotion should be provided to the user. In response to determining to play or offer to play a promotion to the user, the software application on the phone effectively “intercepts” the call and plays to the user either a promotion or an offer to hear about a promotion prior to placing an outbound voice call. The software application may retrieve the promotion from local memory or may connect with a remote server to download an applicable promotion.

Classes IPC  ?

  • H04M 3/42 - Systèmes fournissant des fonctions ou des services particuliers aux abonnés

86.

Web integrated interactive voice response

      
Numéro d'application 11961005
Numéro de brevet 08204184
Statut Délivré - en vigueur
Date de dépôt 2007-12-20
Date de la première publication 2008-05-08
Date d'octroi 2012-06-19
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Gao, Xiaofeng
  • Scott, David
  • Zellner, Sam

Abrégé

One embodiment of a representative system for web integrated interactive voice response includes an interactive voice response system adapted to provide a plurality of voice menus to a user over a telephone and a graphical user interface system adapted to provide a plurality of menus in a graphical format to the user over a network connection. Information provided in the voice menus corresponds to information provided in the menus in the graphical format and is responsive to commands received by the graphical user interface system from the user. Other systems and methods are also provided.

Classes IPC  ?

  • H04M 11/06 - Transmission simultanée téléphonique et de données, p.ex. transmission télégraphique sur les mêmes conducteurs
  • G06F 3/00 - Dispositions d'entrée pour le transfert de données destinées à être traitées sous une forme maniable par le calculateur; Dispositions de sortie pour le transfert de données de l'unité de traitement à l'unité de sortie, p.ex. dispositions d'interface
  • G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00

87.

Methods for voice activated dialing

      
Numéro d'application 11959822
Numéro de brevet 08150001
Statut Délivré - en vigueur
Date de dépôt 2007-12-19
Date de la première publication 2008-05-01
Date d'octroi 2012-04-03
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Bishop, Michael
  • Koch, Robert

Abrégé

Methods for routing a call based on voice activated dialing (VAD). A VAD device module may respond to a VAD instruction, or to a call received with a VAD instruction with a corresponding call destination number obtained from a personal VAD directory. If the personal VAD directory fails to include the call destination number, the VAD device module may route the call or initiate a call through a gateway to a VAD network module. The VAD network module may obtain call destination information from the VAD instruction, and may use the call destination information obtain the call destination number. The VAD network module may obtain additional information from the call or other source, and use the additional information to obtain the call destination number. The call then is routed to the call destination number. The call destination number may be added to the personal VAD directory.

Classes IPC  ?

  • H04M 1/64 - Dispositions automatiques pour répondre aux appels; Dispositions automatiques pour enregistrer des messages pour abonnés absents; Dispositions pour enregistrer des conversations

88.

System and method for a cooperative conversational voice user interface

      
Numéro d'application 11580926
Numéro de brevet 08073681
Statut Délivré - en vigueur
Date de dépôt 2006-10-16
Date de la première publication 2008-04-17
Date d'octroi 2011-12-06
Propriétaire
  • VB ASSETS, LLC (USA)
  • NUANCE COMMUNICATIONS, INC. (USA)
  • VB ASSETS, LLC (USA)
Inventeur(s)
  • Baldwin, Larry
  • Freeman, Tom
  • Tjalve, Michael
  • Ebersold, Blane
  • Weider, Chris

Abrégé

A cooperative conversational voice user interface is provided. The cooperative conversational voice user interface may build upon short-term and long-term shared knowledge to generate one or more explicit and/or implicit hypotheses about an intent of a user utterance. The hypotheses may be ranked based on varying degrees of certainty, and an adaptive response may be generated for the user. Responses may be worded based on the degrees of certainty and to frame an appropriate domain for a subsequent utterance. In one implementation, misrecognitions may be tolerated, and conversational course may be corrected based on subsequent utterances and/or responses.

Classes IPC  ?

  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique
  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00

89.

Establishing a preferred mode of interaction between a user and a multimodal application

      
Numéro d'application 11530599
Numéro de brevet 08145493
Statut Délivré - en vigueur
Date de dépôt 2006-09-11
Date de la première publication 2008-03-13
Date d'octroi 2012-03-27
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Cross, Jr., Charles W.
  • Pike, Hilary A.

Abrégé

Establishing a preferred mode of interaction between a user and a multimodal application, including evaluating, by a multimodal application operating on a multimodal device supporting multiple modes of interaction including a voice mode and one or more non-voice modes, user modal preference, and dynamically configuring multimodal content of the multimodal application in dependence upon the evaluation of user modal preference.

Classes IPC  ?

  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité

90.

Method and apparatus for recognizing a user personality trait based on a number of compound words used by the user

      
Numéro d'application 11436295
Numéro de brevet 08150692
Statut Délivré - en vigueur
Date de dépôt 2006-05-18
Date de la première publication 2007-11-22
Date d'octroi 2012-04-03
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Stewart, Osamuyimen Thompson
  • Dai, Liwei

Abrégé

Techniques for recognizing a personality trait associated with a user. Input from the user is analyzed to determine a number of words, including a number of compound words. The personality trait associated with the user is determined based, at least in part, on the number of compound words exceeding a threshold.

Classes IPC  ?

  • G10L 15/08 - Classement ou recherche de la parole

91.

Mass-scale, user-independent, device-independent voice messaging system

      
Numéro d'application 11673746
Numéro de brevet 08903053
Statut Délivré - en vigueur
Date de dépôt 2007-02-12
Date de la première publication 2007-06-07
Date d'octroi 2014-12-02
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s) Doulton, Daniel Michael

Abrégé

A mass-scale, user-independent, device-independent, voice messaging system that converts unstructured voice messages into text for display on a screen is disclosed. The system comprises (i) computer implemented sub-systems and also (ii) a network connection to human operators providing transcription and quality control; the system being adapted to optimize the effectiveness of the human operators by further comprising 3 core sub-systems, namely (i) a pre-processing front end that determines an appropriate conversion strategy; (ii) one or more conversion resources; and (iii) a quality control sub-system.

Classes IPC  ?

  • H04M 1/64 - Dispositions automatiques pour répondre aux appels; Dispositions automatiques pour enregistrer des messages pour abonnés absents; Dispositions pour enregistrer des conversations
  • H04M 3/51 - Dispositions centralisées de réponse aux appels demandant l'intervention d'un opérateur
  • H04M 3/493 - Services d'information interactifs, p. ex. renseignements sur l'annuaire téléphonique
  • H04M 3/533 - Systèmes de messagerie vocale
  • G10L 15/26 - Systèmes de synthèse de texte à partir de la parole

92.

System and method for conducting a search using a wireless mobile device

      
Numéro d'application 11263601
Numéro de brevet 07477909
Statut Délivré - en vigueur
Date de dépôt 2005-10-31
Date de la première publication 2007-05-03
Date d'octroi 2009-01-13
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s) Roth, Daniel Lawrence

Abrégé

A method and system are provided by which a wireless mobile device takes a vocally entered query and transmits it in a text message format over a wireless network to a search engine; receives search results based on the query from the search engine over the wireless network; and displays the search results.

Classes IPC  ?

  • H04N 7/173 - Systèmes à secret analogiques; Systèmes à abonnement analogiques à deux voies, p.ex. l'abonné envoyant un signal de sélection du programme
  • G06F 17/30 - Recherche documentaire; Structures de bases de données à cet effet
  • H04M 3/00 - Centraux automatiques ou semi-automatiques
  • H04Q 7/20 -

93.

Method, system and apparatus for data reuse

      
Numéro d'application 11545414
Numéro de brevet 08370734
Statut Délivré - en vigueur
Date de dépôt 2006-10-10
Date de la première publication 2007-02-15
Date d'octroi 2013-02-05
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Boone, Keith W.
  • Chaparala, Sunitha
  • Fordyce, Cameron
  • Gervais, Sean
  • Manoukian, Roubik
  • Ogrinc, Harry J.
  • Titemore, Robert G.
  • Hopkins, Jeffrey G.

Abrégé

A system and method may be disclosed for facilitating the creation or modification of a document by providing a mechanism for locating relevant data from external sources and organizing and incorporating some or all of said data into the document. In the method for reusing data, there may be a set of documents that may be queried, where each document may be divided into a plurality of sections. A plurality of section text groups may be formed based on the set of documents, where each section text group may be associated with a respective section from the plurality of sections and each section group includes a plurality of items. Each item may be associated with a respective section from each document of the set of documents. A selected item within a selected section text group may be focused. The selected item may be extracted to a current document. The current document may be exported to a host application.

Classes IPC  ?

  • G06F 17/00 - TRAITEMENT ÉLECTRIQUE DE DONNÉES NUMÉRIQUES Équipement ou méthodes de traitement de données ou de calcul numérique, spécialement adaptés à des fonctions spécifiques

94.

Methods and apparatus for automatically extending the voice vocabulary of mobile communications devices

      
Numéro d'application 11347666
Numéro de brevet 08160884
Statut Délivré - en vigueur
Date de dépôt 2006-02-03
Date de la première publication 2006-08-03
Date d'octroi 2012-04-17
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Roth, Daniel L.
  • Cohen, Jordan
  • Behrakis, Elias P.

Abrégé

The invention is a method of improving the performance of a speech recognizer. The method generally involves: providing a lexicon for the speech recognizer; monitoring a user's interaction with a network; accessing a plurality of words associated with the monitored interaction; and including the plurality of words in the lexicon.

Classes IPC  ?

  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité

95.

Speech signal processing with combined noise reduction and echo compensation

      
Numéro d'application 11218687
Numéro de brevet 07747001
Statut Délivré - en vigueur
Date de dépôt 2005-09-02
Date de la première publication 2006-07-13
Date d'octroi 2010-06-29
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Kellermann, Walter
  • Herbordt, Wolfgang

Abrégé

A speech signal processing system combines acoustic noise reduction and echo cancellation to enhance acoustic performance. The speech signal processing system may be used in vehicles or other environments where noise-suppressed communication is desirable. The system includes an adaptive beamforming signal processing unit, an adaptive echo compensating unit to reduce acoustic echoes, and an adaptation unit to combine noise reduction and adaptive echo compensating.

Classes IPC  ?

  • H04M 9/08 - Systèmes téléphoniques à haut-parleur à double sens comportant des moyens pour conditionner le signal, p.ex. pour supprimer les échos dans l'une ou les deux directions du trafic 

96.

System and method of providing an automated data-collection in spoken dialog systems

      
Numéro d'application 11029798
Numéro de brevet 08185399
Statut Délivré - en vigueur
Date de dépôt 2005-01-05
Date de la première publication 2006-07-06
Date d'octroi 2012-05-22
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Di Fabbrizio, Giuseppe
  • Hakkani-Tur, Dilek Z.
  • Rahim, Mazin G.
  • Renger, Bernard S.
  • Tur, Gokhan

Abrégé

The invention relates to a system and method for gathering data for use in a spoken dialog system. An aspect of the invention is generally referred to as an automated hidden human that performs data collection automatically at the beginning of a conversation with a user in a spoken dialog system. The method comprises presenting an initial prompt to a user, recognizing a received user utterance using an automatic speech recognition engine and classifying the recognized user utterance using a spoken language understanding module. If the recognized user utterance is not understood or classifiable to a predetermined acceptance threshold, then the method re-prompts the user. If the recognized user utterance is not classifiable to a predetermined rejection threshold, then the method transfers the user to a human as this may imply a task-specific utterance. The received and classified user utterance is then used for training the spoken dialog system.

Classes IPC  ?

  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • G10L 19/00 - Techniques d'analyse ou de synthèse de la parole ou des signaux audio pour la réduction de la redondance, p.ex. dans les vocodeurs; Codage ou décodage de la parole ou des signaux audio utilisant les modèles source-filtre ou l’analyse psychoacoustique
  • G06F 17/27 - Analyse automatique, p.ex. analyse grammaticale, correction orthographique

97.

System and method for providing network coordinated conversational services

      
Numéro d'application 11303768
Numéro de brevet 07519536
Statut Délivré - en vigueur
Date de dépôt 2005-12-16
Date de la première publication 2006-05-25
Date d'octroi 2009-04-14
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Maes, Stephane H.
  • Gopalakrishnan, Ponani

Abrégé

A system and method for providing automatic and coordinated sharing of conversational resources, e.g., functions and arguments, between network-connected servers and devices and their corresponding applications. In one aspect, a system for providing automatic and coordinated sharing of conversational resources includes a network having a first and second network device, the first and second network device each comprising a set of conversational resources, a dialog manager for managing a conversation and executing calls requesting a conversational service, and a communication stack for communicating messages over the network using conversational protocols, wherein the conversational protocols establish coordinated network communication between the dialog managers of the first and second network device to automatically share the set of conversational resources of the first and second network device, when necessary, to perform their respective requested conversational service.

Classes IPC  ?

  • G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
  • G10L 11/00 - Détermination ou détection des caractéristiques de la parole ou des signaux audio qui ne se limitent pas à un seul des groupes ; G10L 15/00-G10L 21/00
  • G06F 15/16 - Associations de plusieurs calculateurs numériques comportant chacun au moins une unité arithmétique, une unité programme et un registre, p.ex. pour le traitement simultané de plusieurs programmes

98.

Method and system of generating a speech signal with overlayed random frequency signal

      
Numéro d'application 10957222
Numéro de brevet 07558389
Statut Délivré - en vigueur
Date de dépôt 2004-10-01
Date de la première publication 2006-04-06
Date d'octroi 2009-07-07
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s) Desimone, Joseph

Abrégé

A method and apparatus utilizing prosody modification of a speech signal output by a text-to-speech (TTS) system to substantially prevent an interactive voice response (IVR) system from understanding the speech signal without significantly degrading the speech signal with respect to human understanding. The present invention involves modifying the prosody of the speech output signal by using the prosody of the user's response to a prompt. In addition, a randomly generated overlay frequency is used to modify the speech signal to further prevent an IVR system from recognizing the TTS output. The randomly generated frequency may be periodically changed using an overlay timer that changes the random frequency signal at a predetermined intervals.

Classes IPC  ?

  • H04L 9/00 - Dispositions pour les communications secrètes ou protégées; Protocoles réseaux de sécurité
  • H04N 7/167 - Systèmes rendant le signal de télévision inintelligible et ensuite intelligible
  • G10L 19/00 - Techniques d'analyse ou de synthèse de la parole ou des signaux audio pour la réduction de la redondance, p.ex. dans les vocodeurs; Codage ou décodage de la parole ou des signaux audio utilisant les modèles source-filtre ou l’analyse psychoacoustique

99.

Combined speech recognition and sound recording

      
Numéro d'application 11005568
Numéro de brevet 07505911
Statut Délivré - en vigueur
Date de dépôt 2004-12-05
Date de la première publication 2005-07-21
Date d'octroi 2009-03-17
Propriétaire NUANCE COMMUNICATIONS, INC. (USA)
Inventeur(s)
  • Roth, Daniel L.
  • Cohen, Jordan R.
  • Johnston, David F.
  • Porter, Edward W.

Abrégé

A handheld device with both large-vocabulary speech recognition and audio recoding allows users to switch between at least two of the following three modes: (1) recording audio without corresponding speech recognition; (2) recording with speech recognition; and (3) speech recognition without audio recording. A handheld device with both large-vocabulary speech recognition and audio recoding enables a user to select a portion of previously recorded sound and have speech recognition performed upon it. A system enables a user to search for a text label associated with portions of unrecognized recorded sound by uttering the label's words. A large-vocabulary system allows users to switch between playing back recorded audio and speech recognition with a single input, with successive audio playbacks automatically starting slightly before the end of prior playback. And a cell phone that allows both large-vocabulary speech recognition and audio recording and playback.

Classes IPC  ?

  • G01L 21/06 - Indicateurs de vide ayant une chambre de compression dans laquelle le gaz dont on doit mesurer la pression est comprimé dans lesquels la chambre est fermée par un liquide; Indicateurs de vide du type MacLeod actionnés en faisant tourner ou en renversant le dispositif de mesure

100.

Electronic device and user interface and input method therefor

      
Numéro d'application 10719576
Numéro de brevet 08136050
Statut Délivré - en vigueur
Date de dépôt 2003-11-21
Date de la première publication 2005-05-26
Date d'octroi 2012-03-13
Propriétaire Nuance Communications, Inc. (USA)
Inventeur(s)
  • Sacher, Heiko K.
  • Romera, Maria E.
  • Nagel, Jens

Abrégé

A portable electronic device (100,400) and user interface (425) are operated using a method including initiating entry of a content string; determining the most probable completion alternative or a content prediction using a personalized and learning database (430); displaying the most probable completion alternative or next content prediction; determining whether a user has accepted the most probable completion alternative or next content prediction; and adding the most probable completion alternative or next content prediction to the content string upon user acceptance.

Classes IPC  ?

  • G06F 3/048 - Techniques d’interaction fondées sur les interfaces utilisateur graphiques [GUI]
  1     2        Prochaine page