There is provided a method that includes (a) obtaining a first voice vector that was derived from a signal of a voice that was sampled at a first sampling frequency, (b) obtaining a second voice vector that was derived from a signal of a voice that was sampled at a second sampling frequency, (c) mapping the second voice vector into a mapped voice vector in accordance with a machine learning model, and (d) comparing the first voice vector to the mapped voice vector to yield a score that indicates a probability that the first voice vector and the second voice vector originated from a same person.
G10L 17/04 - Entraînement, enrôlement ou construction de modèle
G10L 17/20 - Transformations de motifs ou opérations ayant pout but d’augmenter la robustesse du système, p.ex. contre le bruit du canal ou les différentes conditions de fonctionnement
G10L 17/00 - Identification ou vérification du locuteur
2.
AUTOMATIC CANONICALIZATION IN A SEMANTIC TAGGER AND SPEECH-TO-TEXT PIPELINE
A method of correcting an automatic speech recognition (ASR) output of an ASR module, includes: providing a corrector model configured to receive the ASR output; pre-training and training the corrector model to map the ASR output to desired formatting of a natural language understanding (NLU) dataset; and fine-tuning the corrector model. The mapping of the ASR output utilizes deep neural network (DNN). Out-of-domain data simulating the ASR output is utilized as the ASR output for the pre-training of the corrector model; in-domain data is utilized for the training of the corrector model; and project-specific data is utilized for fine-tuning the corrector model. The data simulating the ASR output is generated by a simulated ASR runtime process including: feeding raw text into a tokenizer to generate spelled-out text; and feeding the spelled-out text into a formatter to generate formatted text as the data simulating the ASR output.
A method, computer program product, and computing system for obtaining one or more speech signals from a first device, thus defining one or more first device speech signals. One or more speech signals may be obtained from a second device, thus defining one or more second device speech signals. An acoustic relative transfer function may be selected from a plurality of acoustic relative transfer functions based upon, at least in part, the one or more first device speech signals and the one or more second device speech signals. The one or more second device speech signals may be augmented, at run-time, based upon, at least in part, the acoustic relative transfer function
A method, computer program product, and computing system for obtaining one or more speech signals from a first device, thus defining one or more first device speech signals. One or more speech signals may be obtained from a second device, thus defining one or more second device speech signals. An acoustic relative transfer function may be selected from a plurality of acoustic relative transfer functions based upon, at least in part, the one or more first device speech signals and the one or more second device speech signals. The one or more second device speech signals may be augmented, at run-time, based upon, at least in part, the acoustic relative transfer function
A method, computer program product, and computing system for obtaining one or more speech signals from a first device, thus defining one or more first device speech signals. One or more speech signals may be obtained from a second device, thus defining one or more second device speech signals. A noise component model may be selected from a plurality of noise component models based upon, at least in part, the one or more first device speech signals and the one or more second device speech signals. The one or more second device speech signals may be augmented, at run-time, based upon, at least in part, the noise component model.
A method, computer program product, and computing system for obtaining one or more speech signals from a first device, thus defining one or more first device speech signals. One or more speech signals may be obtained from a second device, thus defining one or more second device speech signals. One or more acoustic relative transfer functions mapping reverberation from the one or more first device speech signals to the one or more second device speech signals may be generated. One or more augmented second device speech signals may be generated based upon, at least in part, the one or more acoustic relative transfer functions and first device training data
A method, computer program product, and computer system for placing, by a computing device, a user into a first queue on a first communication channel to converse with a second user. A trigger may be identified for the first communication channel. The user may be sent a self-service option based upon, at least in part, identifying the trigger for the first communication channel, wherein the self-service option is sent on an alternate communication channel while the user is in the first queue.
A method of processing speech includes: providing a first set of audio data having audio features in a first bandwidth; down-sampling the first set of audio data to a second bandwidth lower than the first bandwidth; producing, by a high frequency reconstruction network (HFRN), an estimate of audio features in the first bandwidth for the first set of audio data, based on at least the down-sampled audio data; inputting, into the HFRN, a second set of audio data having audio features in the second bandwidth; producing, by the HFRN, based on a second set of audio data having audio features in the second bandwidth, an estimate of audio features in the first bandwidth for the second set of audio data; and training a speech processing system (SPS) using the estimates of audio features in the first bandwidth for the first and second sets of audio data.
G10L 21/0388 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho utilisant des techniques d’étalement de bande - Détails du traitement à cet effet
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
G10L 13/00 - Synthèse de la parole; Systèmes de synthèse de la parole à partir de texte
9.
SYSTEM AND METHOD FOR GENERATING SYNTHETIC COHORTS USING GENERATIVE MODELING
A method, computer program product, and computing system for generating a generative model representative of a plurality of natural biometric profiles. A plurality of random samples are generated from the generative model. A plurality of synthetic biometric profiles are generated based upon, at least in part, the plurality of random samples.
A method, computer program product, and computing system for receiving a notification that a patient has arrived to a telehealth session before the telehealth session begins. The notification is received via a computing device. In response to receiving the notification that the patient has arrived to the telehealth session before the telehealth session begins, information associated with the patient is automatically pulled by a virtual assistant. The patient is prompted by the virtual assistant to complete a task before the telehealth session begins. A question is received from the patient before the telehealth session begins. Patient data may be obtained from one or more sources. The obtained patient data is processed to determine if the patient data is indicative of a possible medical condition and the medical condition is provided to a medical professional. An answer to the question is provided. The answer is personalized to the patient.
G16H 80/00 - TIC spécialement adaptées pour faciliter la communication entre les professionnels de la santé ou les patients, p.ex. pour le diagnostic collaboratif, la thérapie collaborative ou la surveillance collaborative de l’état de santé
G16H 50/20 - TIC spécialement adaptées au diagnostic médical, à la simulation médicale ou à l’extraction de données médicales; TIC spécialement adaptées à la détection, au suivi ou à la modélisation d’épidémies ou de pandémies pour le diagnostic assisté par ordinateur, p.ex. basé sur des systèmes experts médicaux
11.
MULTI-ENCODER END-TO-END AUTOMATIC SPEECH RECOGNITION (ASR) FOR JOINT MODELING OF MULTIPLE INPUT DEVICES
An end-to-end automatic speech recognition (ASR) system includes: first encoder configured for close-talk input captured by a close-talk input mechanism; second encoder configured for far-talk input captured by far-talk input mechanism; and encoder selection layer configured to select at least one of first and second encoders for use in producing ASR output. The selection is made based on at least one of short-time Fourier transform (STFT), Mel-frequency Cepstral Coefficient (MFCC) and filter bank derived from at least one of the close-talk input and far-talk input. If signals from both the close-talk input mechanism and far-talk input mechanism are present for a speech segment, the encoder selection layer dynamically selects between the close-talk encoder and far-talk encoder to select the encoder that better recognizes the speech segment. An encoder-decoder model is used to produce ASR output.
G10L 15/34 - Adaptation d’un reconnaisseur unique pour traitement en parallèle, p.ex. par utilisation de processeurs multiples ou informatique en nuage
G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine
A method, computer program product, and computing system for: monitoring a conversation between a patient and a medical entity; identifying a portion of the conversation associated with the patient, thus identifying a current patient conversation portion; and processing the current patient conversation portion to identify a condition associated with the patient, thus identifying a patient condition.
A61B 1/00 - Instruments pour procéder à l'examen médical de l'intérieur des cavités ou des conduits du corps par inspection visuelle ou photographique, p.ex. endoscopes; Dispositions pour l'éclairage dans ces instruments
A computer-implemented method, computer program product and computing system for enabling a user to initiate a problem-reporting procedure in response to an inaccurate result generated by an application when processing confidential data; processing the confidential data to generate at least one instantiation of non-confidential data that is related to the confidential data; and providing a preferred instantiation of the non-confidential data for troubleshooting the application.
G06Q 10/06 - Ressources, gestion de tâches, des ressources humaines ou de projets; Planification d’entreprise ou d’organisation; Modélisation d’entreprise ou d’organisation
H04L 9/32 - Dispositions pour les communications secrètes ou protégées; Protocoles réseaux de sécurité comprenant des moyens pour vérifier l'identité ou l'autorisation d'un utilisateur du système
14.
SYSTEM AND METHOD FOR SELF-ATTENTION-BASED COMBINING OF MULTICHANNEL SIGNALS FOR SPEECH PROCESSING
A method, computer program product, and computing system for receiving a plurality of signals from a plurality of microphones, thus defining a plurality of channels. A weighted multichannel representation of the plurality of channels may be generated. A plurality of weights for each channel of the plurality of channels may be generated based upon, at least in part, the weighted multichannel representation of the plurality of channels. A single channel representation of the plurality of channels may be generated based upon, at least in part, the weighted multichannel representation of the plurality of channels and the plurality of weights generated for each channel of the plurality of channels.
G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
H04R 1/40 - Dispositions pour obtenir la fréquence désirée ou les caractéristiques directionnelles pour obtenir la caractéristique directionnelle désirée uniquement en combinant plusieurs transducteurs identiques
A method, computer program product, and computer system for processing one or more portions of an input sequence to generate one or more candidate output sequences, thus defining a plurality of prediction scores for the candidate output sequences. One or more specialized entities may be identified from the candidate output sequences. A first scoring methodology may be applied on the candidate output sequences based upon the portions of the input sequence, thus defining a first set of prediction scores for the one or more candidate output sequences. A second scoring methodology may be applied on the specialized entities from the candidate output sequences based upon the portions of the input sequence, thus defining a second set of prediction scores for the specialized entities. The plurality of predictions scores for the specialized entities may be at least partially modified based upon the first set and the second set of prediction scores.
A method, computer program product, and computing system for: monitoring a conversation between a plurality of participants of a telehealth session; identifying an addressable issue within the conversation; and initiating an action to mitigate the addressable issue. A first participant of the plurality of participants may be a medical professional. A second participant of the plurality of participants may be a patient. The addressable issue may be a potential language barrier with one of the participants of the telehealth session. Initiating an action to mitigate the addressable issue may include: translating audio received by one of the participants of the telehealth session from a first language to a second language.
G06F 40/35 - Représentation du discours ou du dialogue
G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 80/00 - TIC spécialement adaptées pour faciliter la communication entre les professionnels de la santé ou les patients, p.ex. pour le diagnostic collaboratif, la thérapie collaborative ou la surveillance collaborative de l’état de santé
A method, computer program product, and computing system for: monitoring a conversation between a patient and a medical professional; identifying one or more medication recommendations for the patient based, at least in part, upon the conversation; and providing the one or more medication recommendations to the medical professional.
A computer-implemented method, computer program product, and computing system for monitoring a plurality of encounter participants is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to: associate at least a first portion of the encounter information with at least one known encounter participant, and associate at least a second portion of the encounter information with at least one unknown encounter participant.
G06F 21/00 - Dispositions de sécurité pour protéger les calculateurs, leurs composants, les programmes ou les données contre une activité non autorisée
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
19.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A computer-implemented method, computer program product, and computing system for visual diarization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information.
G06F 21/00 - Dispositions de sécurité pour protéger les calculateurs, leurs composants, les programmes ou les données contre une activité non autorisée
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
20.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A computer-implemented method, computer program product, and computing system for automated clinical documentation is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to generate an encounter transcript. At least a portion of the encounter transcript is processed to populate at least a portion of a record associated with the user encounter.
G16H 10/00 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients
G16H 10/20 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des essais ou des questionnaires cliniques électroniques
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
21.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A computer-implemented method, computer program product, and computing system for rendering content is executed on a computing device and includes receiving a request to render content during a user encounter. If it is determined that the content includes sensitive content, a complete version of the content is rendered on a first device (wherein the complete version of the content includes the sensitive content) and a limited version of the content on a second device (wherein the limited version of the content excludes the sensitive content).
G06F 21/00 - Dispositions de sécurité pour protéger les calculateurs, leurs composants, les programmes ou les données contre une activité non autorisée
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
22.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A computer-implemented method, computer program product, and computing system for source separation is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes first audio encounter information obtained from a first encounter participant and at least second audio encounter information obtained from at least a second encounter participant. The first audio encounter information and the at least second audio encounter information are processed to eliminate audio interference between the first audio encounter information and the at least second audio encounter information.
G06F 21/00 - Dispositions de sécurité pour protéger les calculateurs, leurs composants, les programmes ou les données contre une activité non autorisée
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
23.
SYSTEM AND METHOD FOR DATA AUGMENTATION AND SPEECH PROCESSING IN DYNAMIC ACOUSTIC ENVIRONMENTS
A method, computer program product, and computing system for defining a model representative of a plurality of acoustic variations to a speech signal, thus defining a plurality of time-varying spectral modifications. The plurality of time-varying spectral modifications may be applied to a reference signal using a filtering operation, thus generating a time-varying spectrally-augmented signal.
A method, computer program product, and computing system for receiving one or more inputs indicative of at least one of: a relative location of a speaker and a microphone array, and a relative orientation of the speaker and the microphone array. One or more reference signals may be received. A speech processing system may be trained using the one or more inputs and the one or more reference signals.
G06N 99/00 - Matière non prévue dans les autres groupes de la présente sous-classe
G10K 11/00 - Procédés ou dispositifs pour transmettre, conduire ou diriger le son en général; Procédés ou dispositifs de protection contre le bruit ou les autres ondes acoustiques ou pour amortir ceux-ci, en général
G10L 13/033 - Procédés d'élaboration de parole synthétique; Synthétiseurs de parole Édition de voix, p.ex. transformation de la voix du synthétiseur
G10L 13/047 - Architecture des synthétiseurs de parole
G10L 13/10 - Règles de prosodie dérivées du texte; Intonation ou accent tonique
25.
SYSTEM AND METHOD FOR DATA AUGMENTATION AND SPEECH PROCESSING IN DYNAMIC ACOUSTIC ENVIRONMENTS
A method, computer program product, and computing system for defining model representative of a plurality of acoustic variations to a speech signal, thus defining a plurality of time-varying spectral modifications. The plurality of time-varying spectral modifications may be applied to a plurality of feature coefficients of a target domain of a reference signal, thus generating a plurality of time-varying spectrally-augmented feature coefficients of the reference signal.
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a telehealth medical encounter; gathering information during the telehealth medical encounter, thus generating gathered encounter information; and rendering an informational window concerning the telehealth medical encounter for review by the patient and/or the medical entity, wherein the informational window is configured to provide supplemental information based, at least in part, upon the gathered encounter information.
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information, wherein the gathered encounter information includes image-based content of the patient; generating image-based content information via artificial intelligence, wherein the image-based content information is based at least in part upon the image-based content and/or the gathered encounter information and is configured to provide guidance to the medical entity concerning the image-based content; and providing the image-based content information to the medical entity.
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 30/20 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le maniement d’images médicales, p.ex. DICOM, HL7 ou PACS
G16H 30/40 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le traitement d’images médicales, p.ex. l’édition
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information, wherein the gathered encounter information includes audio-based content of the patient; generating audio-based content information via artificial intelligence, wherein the audio-based content information is based at least in part upon the audio-based content and/or the gathered encounter information and is configured to provide guidance to the medical entity concerning the audio-based content; and providing the audio-based content information to the medical entity.
G06F 3/0482 - Interaction avec des listes d’éléments sélectionnables, p.ex. des menus
G16H 10/20 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des essais ou des questionnaires cliniques électroniques
G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information, wherein the gathered encounter information includes video-based content of the patient; generating video-based content information via artificial intelligence, wherein the video-based content information is based at least in part upon the video-based content and/or the gathered encounter information and is configured to provide guidance to the medical entity concerning the video-based content; and providing the video-based content information to the medical entity.
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 30/20 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le maniement d’images médicales, p.ex. DICOM, HL7 ou PACS
G16H 30/40 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le traitement d’images médicales, p.ex. l’édition
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information; and providing supplemental information to the medical entity based, at least in part, upon the gathered encounter information.
G06F 3/0482 - Interaction avec des listes d’éléments sélectionnables, p.ex. des menus
G16H 10/20 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des essais ou des questionnaires cliniques électroniques
G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information; generating medical encounter topical information via artificial intelligence, wherein the medical encounter topical information is based at least in part upon the gathered encounter information and is configured to provide guidance to the medical entity concerning one or more topics to be discussed during the medical encounter; and providing the medical encounter topical information to the medical entity.
A method, computer program product, and computing system for selecting a reference audio acquisition device from a plurality of audio acquisition devices of an audio recording system. Audio encounter information of the reference microphone may be encoded, thus defining encoded reference audio encounter information. A plurality of acoustic relative transfer functions between the reference microphone and the plurality of audio acquisition devices of the audio recording system may be generated. The encoded reference audio encounter information and a representation of the plurality of acoustic relative transfer functions may be transmitted.
G10L 19/00 - Techniques d'analyse ou de synthèse de la parole ou des signaux audio pour la réduction de la redondance, p.ex. dans les vocodeurs; Codage ou décodage de la parole ou des signaux audio utilisant les modèles source-filtre ou l’analyse psychoacoustique
33.
MULTI-CHANNEL SPEECH COMPRESSION SYSTEM AND METHOD
A method, computer program product, and computing system for generating a plurality of acoustic relative transfer functions between a plurality of audio acquisition devices of an audio recording system based upon, at least in part, one or more of a predefined speech processing application and a predefined acoustic environment. An acoustic relative transfer function codebook may be generated using the plurality of acoustic relative transfer functions. One or more channels from the plurality of audio acquisition devices of the audio recording system may be encoded using the acoustic relative transfer function codebook.
G10L 15/02 - Extraction de caractéristiques pour la reconnaissance de la parole; Sélection d'unités de reconnaissance
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 25/51 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes spécialement adaptées pour un usage particulier pour comparaison ou différentiation
34.
FIRST AND SECOND EMBEDDING OF ACOUSTIC RELATIVE TRANSFER FUNCTIONS
A method, computer program product, and computing system for generating a plurality of acoustic relative transfer functions for a plurality of audio acquisition devices of an audio recording system deployed in an acoustic environment. The plurality of acoustic relative transfer functions may be encoded into a first embedding of acoustic relative transfer functions and at least a second embedding of acoustic relative transfer functions. Information may be extracted from at least the first embedding of acoustic relative transfer functions.
G10L 19/008 - Codage ou décodage du signal audio multi-canal utilisant la corrélation inter-canaux pour réduire la redondance, p.ex. stéréo combinée, codage d’intensité ou matriçage
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information; generating medical encounter workflow information via artificial intelligence, wherein the medical encounter workflow information is based at least in part upon the gathered encounter information and is configured to provide guidance to the medical entity concerning a desired workflow for the medical encounter; and providing the medical encounter workflow information to the medical entity.
G06Q 10/06 - Ressources, gestion de tâches, des ressources humaines ou de projets; Planification d’entreprise ou d’organisation; Modélisation d’entreprise ou d’organisation
G16H 10/20 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des essais ou des questionnaires cliniques électroniques
G16H 10/40 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données relatives aux analyses de laboratoire, p.ex. pour des analyses d’échantillon de patient
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information; and processing the gathered encounter information via artificial intelligence to determine whether a third-party encounter should be initiated.
A method, computer program product, and computing system for receiving audio-based content from a user who is reviewing an image on a display screen; receiving gaze information that defines a gaze location of the user; and temporally aligning the audio-based content and the gaze information to form location-based content.
G06F 3/00 - Dispositions d'entrée pour le transfert de données destinées à être traitées sous une forme maniable par le calculateur; Dispositions de sortie pour le transfert de données de l'unité de traitement à l'unité de sortie, p.ex. dispositions d'interface
38.
COMPARING ACOUSTIC RELATIVE TRANSFER FUNCTIONS FROM AT LEAST A PAIR OF TIME FRAMES
A method, computer program product, and computing system for generating a plurality of acoustic relative transfer functions associated with a plurality of audio acquisition devices of an audio recording system deployed in an acoustic environment. At least a pair of the plurality of acoustic relative transfer functions from time frames may be compared. A change in the acoustic environment may be detected based upon, at least in part, the comparison of the plurality of acoustic relative transfer functions from at least the pair of time frames.
G10L 25/51 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes spécialement adaptées pour un usage particulier pour comparaison ou différentiation
39.
MULTI-CHANNEL SPEECH COMPRESSION SYSTEM AND METHOD
A method, computer program product, and computing system for generating a plurality of acoustic relative transfer functions associated with a plurality of audio acquisition devices of an audio recording system deployed in an acoustic environment. Acoustic relative transfer functions of at least a pair of audio acquisition devices of the plurality of audio acquisition devices may be compared. Location information associated with an acoustic source within the acoustic environment may be determined based upon, at least in part, the comparison of the acoustic relative transfer functions of the at least a pair of audio acquisition devices of the plurality of audio acquisition devices.
H04S 3/00 - Systèmes utilisant plus de deux canaux, p.ex. systèmes quadriphoniques
H04S 5/00 - Systèmes pseudo-stéréophoniques, p.ex. dans lesquels les signaux d'un canal supplémentaire sont dérivés du signal monophonique par déphasage, retardement ou réverbération
40.
MULTI-CHANNEL SPEECH COMPRESSION SYSTEM AND METHOD
A method, computer program product, and computing system for encoding audio encounter information of a reference audio acquisition device of a plurality of audio acquisition devices of an audio recording system, thus defining encoded reference audio encounter information. Location information may be estimated, via a machine vision system, for an acoustic source within an acoustic environment. One or more acoustic relative transfer functions may be selected from a plurality of acoustic relative transfer functions for the plurality of audio acquisition devices of the audio recording system based upon, at least in part, the location information. The encoded reference audio encounter information and a representation of the selected one or more acoustic relative transfer function may be transmitted.
G10L 15/10 - Classement ou recherche de la parole utilisant des mesures de distance ou de distorsion entre la parole inconnue et les gabarits de référence
G10L 15/16 - Classement ou recherche de la parole utilisant des réseaux neuronaux artificiels
H04R 1/40 - Dispositions pour obtenir la fréquence désirée ou les caractéristiques directionnelles pour obtenir la caractéristique directionnelle désirée uniquement en combinant plusieurs transducteurs identiques
41.
MULTI-CHANNEL SPEECH COMPRESSION SYSTEM AND METHOD
A method, computer program product, and computing system for obtaining machine vision encounter information using one or more machine vision systems. Audio encounter information may be obtained using a plurality of audio acquisition devices of an audio recording system. The audio encounter information may be encoded using an audio codec. The encoding of the audio encounter information by the audio codec may be adapted based upon, at least in part, the machine vision encounter information.
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
G10L 15/02 - Extraction de caractéristiques pour la reconnaissance de la parole; Sélection d'unités de reconnaissance
G10L 15/10 - Classement ou recherche de la parole utilisant des mesures de distance ou de distorsion entre la parole inconnue et les gabarits de référence
A computer-implemented method, computer program product and computing system for defining a test truth set from a master truth set; processing the test truth set using an automated analysis process to generate an automated result set; determining a process efficacy for the automated analysis process based, at least in part, upon the test truth set and the automated result set; and rendering the process efficacy of the automated analysis process.
G06F 19/00 - Équipement ou méthodes de traitement de données ou de calcul numérique, spécialement adaptés à des applications spécifiques (spécialement adaptés à des fonctions spécifiques G06F 17/00;systèmes ou méthodes de traitement de données spécialement adaptés à des fins administratives, commerciales, financières, de gestion, de surveillance ou de prévision G06Q;informatique médicale G16H)
G06K 9/62 - Méthodes ou dispositions pour la reconnaissance utilisant des moyens électroniques
A computer-implemented method, computer program product and computing system for defining a test truth set from a master truth set; processing the test truth set using an automated analysis process to generate an automated result set; determining a process efficacy for the automated analysis process based, at least in part, upon the test truth set and the automated result set; and rendering the process efficacy of the automated analysis process.
A method, computer program product, and computing system for monitoring a plurality of conversations within a monitored space to generate a conversation data set; processing the conversation data set using machine learning to: define a system-directed command for an ACI system, and associate one or more conversational contexts with the system-directed command; detecting the occurrence of a specific conversational context within the monitored space, wherein the specific conversational context is included in the one or more conversational contexts associated with the system-directed command; and executing, in whole or in part, functionality associated with the system-directed command in response to detecting the occurrence of the specific conversational context without requiring the utterance of the system-directed command and/or a wake-up word / phrase.
A computer-implemented method, computer program product and computing system for defining a test truth set from a master truth set; processing the test truth set using an automated analysis process to generate an automated result set; determining a process efficacy for the automated analysis process based, at least in part, upon the test truth set and the automated result set; and rendering the process efficacy of the automated analysis process.
G06F 19/00 - Équipement ou méthodes de traitement de données ou de calcul numérique, spécialement adaptés à des applications spécifiques (spécialement adaptés à des fonctions spécifiques G06F 17/00;systèmes ou méthodes de traitement de données spécialement adaptés à des fins administratives, commerciales, financières, de gestion, de surveillance ou de prévision G06Q;informatique médicale G16H)
G06K 9/62 - Méthodes ou dispositions pour la reconnaissance utilisant des moyens électroniques
A computer-implemented method, computer program product and computing system for receiving a result set for content processed by an automated analysis process; receiving human feedback concerning the result set; and providing feedback information to the developer of the automated analysis process based, at least in part, upon the result set and the human feedback.
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
47.
AMBIENT COOPERATIVE INTELLIGENCE SYSTEM AND METHOD
A method, computer program product, and computing system for initiating a session within an ACI platform; receiving an authentication request from a requester; and authenticating that the requester has the authority to access the ACI platform.
A method, computer program product, and computing system for detecting the issuance of a verbal command by a requester to a virtual assistant; authenticating that the requester has the authority to issue the verbal command to the virtual assistant; if the requester is authenticated, allowing the effectuation of the verbal command to the virtual assistant; and if the requester is not authenticated, preventing the effectuation of the verbal command to the virtual assistant.
A method, computer program product, and computing system for receiving input information concerning a conversation between a caller and a recipient; processing the input information to assess a fraud-threat-level; defining a targeted response based, at least in part, upon the fraud-threat-level assessed, wherein the targeted response is intended to refine the assessed fraud-threat-level; and effectuating the targeted response.
A method, computer program product, and computer system for predicting responses to at least one conversational phrase. At least one conversational phrase may be received. A first probability for a subset of candidate responses of a plurality of candidate responses may be determined based upon, at least in part, context associated with the at least one conversational phrase, the at least one conversational phrase, and each context associated with the plurality of candidate responses. A second probability for the subset of candidate responses may be determined based upon, at least in part, the subset of candidate responses, the at least one conversational phrase, and the context associated with the at least one conversational phrase. At least one candidate response for the at least one conversational phrase may be determined based upon, at least in part, the first probability and the second probability.
G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
A method, computer program product, and computing system for performing an assessment of initial input information, concerning a communication from a caller, to define an initial fraud-threat-level; if the initial fraud-threat-level is below a defined threat threshold, providing the communication to a recipient so that a conversation may occur between the recipient and the caller; performing an assessment of subsequent input information, concerning the conversation, to define a subsequent fraud-threat-level; and effectuating a targeted response based, at least in part, upon the subsequent fraud-threat-level, wherein the targeted response is intended to refine the subsequent fraud-threat-level.
G06F 21/32 - Authentification de l’utilisateur par données biométriques, p.ex. empreintes digitales, balayages de l’iris ou empreintes vocales
G06Q 50/26 - Services gouvernementaux ou services publics
G10L 17/06 - Techniques de prise de décision; Stratégies d’alignement de motifs
G10L 17/26 - Reconnaissance de caractéristiques spéciales de voix, p.ex. pour utilisation dans les détecteurs de mensonge; Reconnaissance des voix d’animaux
H04M 3/22 - Dispositions de supervision, de contrôle ou de test
H04M 3/42 - Systèmes fournissant des fonctions ou des services particuliers aux abonnés
52.
SYSTEM AND METHOD FOR DATA AUGMENTATION FOR MULTI-MICROPHONE SIGNAL PROCESSING
A method, computer program product, and computing system for receiving a speech signal from each microphone of a plurality of microphones, thus defining a plurality of signals. One or more noise signals associated with microphone self-noise may be received. One or more self-noise-based augmentations may be performed on the plurality of signals based upon, at least in part, the one or more noise signals associated with microphone self-noise, thus defining one or more self-noise-based augmented signals.
A method, computer program product, and computing system for receiving a signal from each microphone of a plurality of microphones, thus defining a plurality of signals. Harmonic distortion associated with at least one microphone may be determined. One or more harmonic distortion-based augmentations may be performed on the plurality of signals based upon, at least in part, the harmonic distortion associated with the at least one microphone, thus defining one or more harmonic distortion-based augmented signals.
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
54.
SYSTEM AND METHOD FOR MULTI-MICROPHONE AUTOMATED CLINICAL DOCUMENTATION
A method, computer program product, and computing system for receiving information associated with an acoustic environment. A plurality of filters may be predefined to produce a plurality of beams based upon, at least in part, the information associated with the acoustic environment. The plurality of filters may be predefined to produce a plurality of nulls based upon, at least in part, the information associated with the acoustic environment. Audio encounter information may be obtained, via one or more microphone arrays, using the plurality of beams and the plurality of nulls produced by the plurality of predefined filters.
A method, computer program product, and computing system for receiving audio encounter information from a microphone array. Speech activity within one or more portions of the audio encounter information may be identified based upon, at least in part, a correlation among the audio encounter information received from the microphone array. Location information for the one or more portions of the audio encounter information may be determined based upon, at least in part, the correlation among the signals received by each microphone of the microphone array. The one or more portions of the audio encounter information may be labeled with the speech activity and the location information.
A method, computer program product, and computing system for receiving information associated with an acoustic environment. Acoustic metadata associated with audio encounter information received by a first microphone system may be received. One or more speaker representations may be defined based upon, at least in part, the acoustic metadata associated with the audio encounter information and the information associated with the acoustic environment. One or more portions of the audio encounter information may be labeled with the one or more speaker representations and a speaker location within the acoustic environment.
A method, computer program product, and computing system for receiving a plurality of predefined beams associated with a microphone array. A plurality of predefined nulls associated with the microphone array may be received. One or more predefined beams from the plurality of predefined beams or one or more predefined nulls from the plurality of predefined nulls may be selected. A microphone array may obtain audio encounter information, via the microphone array, using at least one of the one or more selected beams and the one or more selected nulls.
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
58.
SYSTEM AND METHOD FOR MULTI-MICROPHONE AUTOMATED CLINICAL DOCUMENTATION
A method, computer program product, and computing system for receiving audio encounter information from a first microphone system, thus defining a first audio stream. Audio encounter information may be received from a second microphone system, thus defining a second audio stream. Speech activity may be detected in one or more portions of the first audio stream, thus defining one or more speech portions of the first audio stream. Speech activity may be detected in one or more portions of the second audio stream, thus defining one or more speech portions of the second audio stream. The first audio stream and the second audio stream may be aligned based upon, at least in part, the one or more speech portions of the first audio stream and the one or more speech portions of the second audio stream.
A method, computer program product, and computing system for receiving a signal from each microphone of a plurality of microphones, thus defining a plurality of signals. One or more inter-microphone gain-based augmentations may be performed on the plurality of signals, thus defining one or more inter-microphone gain-augmented signals. Performing the one or more inter-microphone gain-based augmentations on the plurality of signals may include applying a gain level from a plurality of gain levels to the signal from each microphone. Applying the gain level from the plurality of gain levels to the signal from each microphone may include applying a gain level, from a predefined range of gain levels, to the signal from each microphone. Applying the gain level from the plurality of gain levels to the signal from each microphone may include applying a random gain level, from the predefined range of gain levels, to the signal from each microphone.
A method, computer program product, and computing system for receiving a signal from each microphone of a plurality of microphones, thus defining a plurality of signals. One or more microphone frequency responses associated with at least one microphone may be received. One or more microphone frequency response-based augmentations may be performed on the plurality of signals based upon, at least in part, the one or more microphone frequency responses, thus defining one or more microphone frequency response-based augmented signals.
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
61.
SYSTEM AND METHOD FOR DATA AUGMENTATION OF FEATURE-BASED VOICE DATA
A method, computer program product, and computing system for receiving feature-based voice data associated with a first acoustic domain. One or more gain-based augmentations may be performed on at least a portion of the feature-based voice data, thus defining gain-augmented feature-based voice data.
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
A method, computer program product, and computing system for receiving feature-based voice data associated with a first acoustic domain. One or more rate-based augmentations may be performed on at least a portion of the feature-based voice data, thus defining rate-based augmented feature-based voice data.
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
A method, computer program product, and computing system for receiving feature-based voice data associated with a first acoustic domain. One or more audio feature-based augmentations may be performed on at least a portion of the feature-based voice data. Performing the one or more audio feature-based augmentations may include adding one or more audio features to the at least a portion of the feature-based voice data and/or removing one or more audio features from the at least a portion of the feature-based voice data.
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
A A method, computer program product, and computing system for generating a three-dimensional model of at least a portion of a three-dimensional space incorporating an ACI system via a video recording subsystem of an ACI calibration platform; and generating one or more audio calibration signals for receipt by an audio recording system included within the ACI system via an audio generation subsystem of the ACI calibration platform.
A method, computer program product, and computing system for obtaining calibration information for a three-dimensional space incorporating an ACI system; and processing the calibration information to calibrate the ACI system.
A method, computer program product, and computing system for extracting acoustic metadata from a signal. The signal may be converted from the time domain to the feature domain, thus defining feature-based voice data associated with the signal. The feature-based voice data associated with the signal may be processed based upon, at least in part, the acoustic metadata.
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
A method, computer program product, and computing system for receiving feature based voice data associated with a first acoustic domain. One or more reverberation-based augmentations may be performed on at least a portion of the feature-based voice data, thus defining reverberation-augmented feature-based voice data. Performing the one or more reverberation-based augmentations to the at least a portion of the feature-based voice data may include performing the one or more reverberation-based augmentations to the at least a portion of the feature-based voice data based upon, at least in part, the target acoustic domain.
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
A method, computer program product, and computing system for receiving feature-based voice data. One or more data augmentation characteristics may be received. One or more augmentations of the feature-based voice data may be generated, via a machine learning model, based upon, at least in part, the feature-based voice data and the one or more data augmentation characteristics
G06F 15/18 - dans lesquels un programme est modifié en fonction de l'expérience acquise par le calculateur lui-même au cours d'un cycle complet; Machines capables de s'instruire (systèmes de commande adaptatifs G05B 13/00;intelligence artificielle G06N)
69.
SYSTEM AND METHOD FOR REVIEW OF AUTOMATED CLINICAL DOCUMENTATION
A method, computer program product, and computing system for obtaining, by a computing device, encounter information of a patient encounter, wherein the encounter information may include audio encounter information and video encounter information obtained from at least a first encounter participant. A report of the patient encounter may be generated based upon, at least in part, the encounter information. A relative importance of a word in the report may be determined. A portion of the video encounter information that corresponds to the word in the report may be determined. The portion of the video encounter information that corresponds to the word in the report may be stored at a first location, wherein the video encounter information may be stored at a second location remote from the first location.
G10L 19/00 - Techniques d'analyse ou de synthèse de la parole ou des signaux audio pour la réduction de la redondance, p.ex. dans les vocodeurs; Codage ou décodage de la parole ou des signaux audio utilisant les modèles source-filtre ou l’analyse psychoacoustique
70.
VEHICLE AVATAR DEVICES FOR INTERACTIVE VIRTUAL ASSISTANT
A system and method for providing avatar device (115, 125, 135, 145) status indicators for voice assistants in multi-zone vehicles. The method comprises: receiving at least one signal from a plurality of microphones (114, 124, 134, 144), wherein each microphone (114, 124, 134, 144) is associated with one of a plurality of spatial zones (110, 120, 130, 140), and one of a plurality of avatar devices (115, 125, 135, 145); wherein the at least one signal further comprises a speech signal component from a speaker; wherein the speech signal component is a voice command or question; sending zone information associated with the speaker and with one of the plurality of spatial zones (110, 120, 130, 140) to an avatar (115, 125, 135, 145); activating one the plurality of avatar devices (115, 125, 135, 145) in a respective one of the plurality of spatial zones (110, 120, 130, 140) associated with the speaker.
A method, computer program product, and computing system includes generating a graph data structure including a plurality of data points. A query for the graph data structure may be received via a user interface. At least one data point from the plurality of data points may be identified, via the user interface, in the graph data structure based upon, at least in part, the query. A selection of a data point from the identified at least one data point may be received via the user interface. The selected data point may be provided to one or more electronic data sources.
A method, computer program product, and computing system for obtaining encounter information during a patient encounter; processing the encounter information to detect the execution of a physical event during the patient encounter, thus defining a detected physical event; and deriving information for the detected physical event
G16H 10/20 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des essais ou des questionnaires cliniques électroniques
G16H 10/40 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données relatives aux analyses de laboratoire, p.ex. pour des analyses d’échantillon de patient
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 10/65 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients stockées sur des supports d’enregistrement portables, p.ex. des cartes à puce, des étiquettes d’identification radio-fréquence [RFID] ou des CD
73.
MULTI-CHANNEL MICROPHONE SIGNAL GAIN EQUALIZATION BASED ON EVALUATION OF CROSS TALK COMPONENTS
Gain mismatch and related problems can be solved by a system and method that applies an automatic microphone signal gain equalization without any direct absolute reference or calibration phase. The system and method performs the steps of receiving, by a computing device, a speech signal from a speaking person via a plurality of microphones, determining a speech signal component in the time- frequency domain for each microphone of the plurality of microphones, calculating an instantaneous cross-talk coupling matrix based on the speech signal components across the microphones, estimating gain factors based on calculated cross-talk couplings and a given expected cross-talk attenuation, limiting the gain factors to appropriate maximum and minimum values, and applying the gain factors to the speech signal used in the control path to control further speech enhancement algorithms or used in the signal path for direct influence on the speech enhanced audio output signal.
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
H04B 3/32 - Réduction de la diaphonie, p.ex. par compensation
H03G 3/00 - Commande de gain dans les amplificateurs ou les changeurs de fréquence
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
H04J 1/12 - Dispositions pour réduire la diaphonie entre canaux
H04J 3/10 - Dispositions pour reduire la diaphonie entre canaux
G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
74.
MULTI-MICROPHONE SPEECH DIALOG SYSTEM FOR MULTIPLE SPATIAL ZONES
There is provided a speech dialog system that includes a first microphone, a second microphone, a processor and a memory. The first microphone captures first audio from a first spatial zone, and produces a first audio signal. The second microphone captures second audio from a second spatial zone, and produces a second audio signal. The processor receives the first audio signal and the second audio signal, and the memory contains instructions that control the processor to perform operations of a speech enhancement module, an automatic speech recognition module, and a speech dialog module that performs a zone-dedicated speech dialog.
G10L 17/06 - Techniques de prise de décision; Stratégies d’alignement de motifs
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine
Disclosed are systems and methods aware of ongoing conversations and configured to intelligently schedule a speech prompt to an intended addressee. A method for intelligently scheduling a speech prompt in a speech dialog system includes monitoring an acoustic environment to detect an intended addressee's availability for a speech prompt having a measure of urgency corresponding therewith. Based on the intended addressee's availability, the method predicts a time that is convenient to present the speech prompt to the intended addressee, and schedules the speech prompt based on the predicted time and the measure of urgency. A measure of rudeness can be estimated using a cost function that includes cost for presence of an utterance, cost for presence of a conversation, and cost for involvement of the intended addressee in the conversation. Scheduling the speech prompt can include trading off the measure of urgency and the measure of rudeness.
A method, computer program product, and computer system for identifying, by a computing device, a plurality of sources, wherein a first source of the plurality of sources is a source of interest and wherein a second source of the plurality of sources is an interference source. The first source and the second source may be monitored simultaneously by implementing a spatial pre-filter for acoustic source localization.
G01S 5/22 - Position de source déterminée par coordination de plusieurs lignes de position définies par des mesures de différence de parcours
G10L 19/008 - Codage ou décodage du signal audio multi-canal utilisant la corrélation inter-canaux pour réduire la redondance, p.ex. stéréo combinée, codage d’intensité ou matriçage
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
A method, computer program product, and computer system for identifying, by a computing device, a plurality of sources. One or more feature values of a plurality of features may be assigned to a first source of the plurality of sources. One or more feature values of the plurality of features may be assigned to a second source of the plurality of sources. A first score for the first source and a second score for the second source may be determined based upon, at least in part, the one or more feature values assigned to the first source and the second source. One of the first source and the second source may be selected for spatial processing based upon, at least in part, the first score for the first source and the second score for the second source.
G01S 3/00 - Radiogoniomètres pour déterminer la direction d'où proviennent des ondes infrasonores, sonores, ultrasonores ou électromagnétiques ou des émissions de particules sans caractéristiques de direction
H04R 1/00 - HAUT-PARLEURS, MICROPHONES, TÊTES DE LECTURE POUR TOURNE-DISQUES OU TRANSDUCTEURS ACOUSTIQUES ÉLECTROMÉCANIQUES ANALOGUES; APPAREILS POUR SOURDS; SYSTÈMES D'ANNONCE EN PUBLIC - Détails des transducteurs
H04R 1/32 - Dispositions pour obtenir la fréquence désirée ou les caractéristiques directionnelles pour obtenir la caractéristique directionnelle désirée uniquement
H04R 1/40 - Dispositions pour obtenir la fréquence désirée ou les caractéristiques directionnelles pour obtenir la caractéristique directionnelle désirée uniquement en combinant plusieurs transducteurs identiques
A method, computer program product, and computer system for identifying, by a computing device, a model for predicting conversational phrases for a communication between at least a first user and a second user. The model may be trained based upon, at least in part, an attribute associated with the second user. At least one conversational phrase may be predicted for the communication between the first user and the second user. The at least one conversational phrase may be provided to the second user as an optional phrase to be sent to the first user.
Provided are a call deflection and response system and method, wherein a voice call from a caller device is received, a skill group is determined to resolve an issue associated with the call, and a callback or a text response to the issue is provided to the caller device, providing a context-based personalized response. A caller leaves a detailed voicemail explaining an issue needing resolution, which is electronically transcribed and then run through a classifier to determine concepts and intents associated with the call. Based on the concepts and intents, responsibility for the call and associated files are transferred to a particular skill group on a response system for resolution. A response entity from the appropriate skill group determines and provides an issue response via callback or text message to the caller device, e.g., to the caller's mobile phone.
H04M 3/523 - Dispositions centralisées de réponse aux appels demandant l'intervention d'un opérateur avec répartition ou mise en file d'attente des appels
H04M 3/493 - Services d'information interactifs, p. ex. renseignements sur l'annuaire téléphonique
A system, method and computer-readable storage device are disclosed for managing a mute and unmute feature on a device which is used to communicate data in a communication conference. The method includes detecting, when the device is set to mute, whether the user is speaking and whether the speech is meant for the conference. Background noises are distinguished from the speech of the user. If the user is speaking and the device is set to mute, the device will automatically switch to and unmute setting such that people in the indication conference can hear the user speak. Facial recognition, and gaze detection or other data can also be used to determine when to automatically mute or unmute the device and can aid in inferring an intent of the user to speak to the conference participants.
G06K 9/00 - Méthodes ou dispositions pour la lecture ou la reconnaissance de caractères imprimés ou écrits ou pour la reconnaissance de formes, p.ex. d'empreintes digitales
A system and method for detecting multi-tone sirens despite environmental noises that may be present obtains a microphone input signal, applies, in real time, a time-frequency analysis to the microphone input signal to determine a time-frequency representation, provides at least one multi-tone model that has a plurality of tone duration patterns, performs multi-tone siren detection on the time-frequency representation, the detection based on the at least one multi-tone model and factoring of doppler shifts, and generates a detection result that can be used in systems for automated vehicles.
A system for inputting and processing handwritten, multi -character text may comprise a handwriting recognition subsystem, a word completion subsystem, and an audio feedback system. The handwriting recognition system may be configured to capture a series of handwritten characters formed by a user and to convert the handwritten characters into a set of candidate partial text strings. The word completion subsystem may be configured to identify if a candidate partial text string constitutes a word segment and if so, generate one or both of (i) at least one clarifying word and (ii) at least one clarifying phrase that includes the clarifying word. The word segment may be an arbitrary string and not correspond to a valid complete word in a language associated with the system. The audio feedback subsystem may be configured to produce an audio representation of the word segment(s), the clarifying word(s), and the clarifying phrase(s).
G06F 40/274 - Conversion de symboles en mots; Anticipation des mots à partir des lettres déjà entrées
G10L 13/02 - Procédés d'élaboration de parole synthétique; Synthétiseurs de parole
G06F 3/0488 - Techniques d’interaction fondées sur les interfaces utilisateur graphiques [GUI] utilisant des caractéristiques spécifiques fournies par le périphérique d’entrée, p.ex. des fonctions commandées par la rotation d’une souris à deux capteurs, ou par la nature du périphérique d’entrée, p.ex. des gestes en fonction de la pression exer utilisant un écran tactile ou une tablette numérique, p.ex. entrée de commandes par des tracés gestuels
83.
SYSTEM AND METHOD FOR ACOUSTIC SPEAKER LOCALIZATION
A method, computer program product, and computing system for acoustic speech localization, comprising receiving, via a plurality of microphones, a plurality of audio signals. Modulation properties of the plurality of audio signals may be analyzed. Speech sounds may be localized from the plurality of audio signals based upon, at least in part, the modulation properties of the plurality of audio signals.
G10L 19/008 - Codage ou décodage du signal audio multi-canal utilisant la corrélation inter-canaux pour réduire la redondance, p.ex. stéréo combinée, codage d’intensité ou matriçage
G10L 19/06 - Détermination ou codage des caractéristiques spectrales, p.ex. des coefficients de prédiction à court terme
G10L 21/028 - Séparation du signal de voix utilisant les propriétés des sources sonores
G10L 21/0308 - Séparation du signal de voix caractérisée par le type de mesure du paramètre, p.ex. techniques de corrélation, techniques de passage par zéro ou techniques prédictives
H04R 1/00 - HAUT-PARLEURS, MICROPHONES, TÊTES DE LECTURE POUR TOURNE-DISQUES OU TRANSDUCTEURS ACOUSTIQUES ÉLECTROMÉCANIQUES ANALOGUES; APPAREILS POUR SOURDS; SYSTÈMES D'ANNONCE EN PUBLIC - Détails des transducteurs
H04R 5/027 - Dispositions spatiales ou structurelles des microphones, p.ex. dispositifs simulant la tête humaine
Provided are a system and method of mixing a second audio stream with a first audio stream in an audio output device. The system is configured to execute the method, comprising buffering and outputting the first audio stream via the audio output device as unmodified output, determining at least one insertion spot within the first audio stream, modifying the first audio stream at an insertion spot to avoid content loss, outputting the second audio stream at the insertion spot, and resuming unmodified output of the first audio stream at or near a completion of the second audio stream. Modifying the first audio stream can include pausing and/or warping the first audio stream at the insertion spot. The audio output device can be a vehicle head unit or a wireless device, such as a mobile phone.
H04H 20/62 - Dispositions spécialement adaptées à des applications spécifiques, p.ex. aux informations sur le trafic ou aux récepteurs mobiles à la radiodiffusion locale, p.ex. la radiodiffusion en interne aux systèmes de transport, p.ex. dans les véhicules
A method, computer program product, and computing system for automatically generating a dialogue graph is executed on a computing device and includes receiving a plurality of conversation data. A plurality of utterance pairs from the plurality of conversation data may be clustered into a plurality of utterance pair clusters. A dialogue graph may be generated with a plurality of nodes representative of the plurality of utterance pair clusters.
A method, computer program product, and computer system for transforming, by a computing device, a speech signal into a speech signal representation. A regression deep neural network may be trained with a cost function to minimize a mean squared error between actual values of the speech signal representation and estimated values of the speech signal representation, wherein the cost function may include one or more discriminative terms. Bandwidth of the speech signal may be extended by extending the speech signal representation of the speech signal using the regression deep neural
A system for processing user requests provides for recommending products or services for an agent to offer a customer or potential customer. The system classifies a user request by an intent, and presents documentation to assist the agent in handling the request. The system further parses the user request to detect life events experienced by the user that may raise the prospect of the user's interest in other products or services. Based on the detected life events, a number of offers are presented to an agent for recommendation to the user.
A method, computer program product, and computing system for obtaining, by a computing device, encounter information of a patient encounter, wherein the encounter information may include audio encounter information obtained from at least a first encounter participant. The audio encounter information obtained from at least the first encounter participant may be processed. A user interface may be generated displaying a plurality of layers associated with the audio encounter information obtained from at least the first encounter participant.
G06F 19/00 - Équipement ou méthodes de traitement de données ou de calcul numérique, spécialement adaptés à des applications spécifiques (spécialement adaptés à des fonctions spécifiques G06F 17/00;systèmes ou méthodes de traitement de données spécialement adaptés à des fins administratives, commerciales, financières, de gestion, de surveillance ou de prévision G06Q;informatique médicale G16H)
89.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A method, computer program product, and computing system for: receiving an initial portion of an encounter record; processing the initial portion of the encounter record to generate initial content for a medical report; receiving one or more additional portions of the encounter record; and processing the one or more additional portions of the encounter record to modify the medical report.
A method, computer program product, and computing system for obtaining, by a computing device, encounter information of a patient encounter, wherein the encounter information may include audio encounter information obtained from at least a first encounter participant. The audio encounter information obtained from at least the first encounter participant may be processed. A user interface may be generated displaying a plurality of layers associated with the audio encounter information obtained from at least the first encounter participant, wherein at least one of the plurality of layers is one of exposed to the user interface and not exposed to the user interface based upon, at least in part, a confidence level.
A method, computer program product, and computing system for obtaining, by a computing device, encounter information of a patient encounter, wherein the encounter information may include audio encounter information obtained from at least a first encounter participant. The audio encounter information obtained from at least the first encounter participant may be processed. A user interface may be generated displaying a plurality of layers associated with the audio encounter information obtained from at least the first encounter participant. A user input may be received from a peripheral device to navigate through each of the plurality of layers associated with the audio encounter information displayed on the user interface.
G06F 3/00 - Dispositions d'entrée pour le transfert de données destinées à être traitées sous une forme maniable par le calculateur; Dispositions de sortie pour le transfert de données de l'unité de traitement à l'unité de sortie, p.ex. dispositions d'interface
92.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A method, computer program product, and computing system for initially aligning two or more audio signals to address coarse temporal misalignment between the two or more audio signals. The two or more audio signals are detected by two or more audio detection systems within a monitored space. The two or more audio signals are subsequently realigned to address ongoing temporal signal drift between the two or more audio signals. A method, computer program product, and computing system for determining a time delay between a first audio signal received on a first audio detection system and a second audio signal received on a second audio detection system. The first and second audio detection systems are located within a monitored space. The first audio detection system is located with respect to the second audio detection system within the monitored space.
G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
G10L 25/00 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes
A method, computer program product, and computer system for mapping, by a computing device, an automatic speech recognition output of a conversation to a concept marker and a verbalized version of a value associated with the concept marker based upon, at least in part, the automatic speech recognition output of the conversation. The concept marker and the verbalized version of the value associated with the concept marker may be replaced with a formatted version. A plurality of user selectable format configurations of the formatted version may be provided as a textual output in a user interface.
A method, computer program product, and computing system for maintaining an automated clinical documentation mesh network that includes a plurality of end point devices; receiving content within the automated clinical documentation mesh network; processing the content to identify a routing rule for the content; and routing the content to one or more endpoint devices based, at least in part, upon the routing rule.
G06F 19/00 - Équipement ou méthodes de traitement de données ou de calcul numérique, spécialement adaptés à des applications spécifiques (spécialement adaptés à des fonctions spécifiques G06F 17/00;systèmes ou méthodes de traitement de données spécialement adaptés à des fins administratives, commerciales, financières, de gestion, de surveillance ou de prévision G06Q;informatique médicale G16H)
G06Q 10/06 - Ressources, gestion de tâches, des ressources humaines ou de projets; Planification d’entreprise ou d’organisation; Modélisation d’entreprise ou d’organisation
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 30/20 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le maniement d’images médicales, p.ex. DICOM, HL7 ou PACS
G16H 40/20 - TIC spécialement adaptées à la gestion ou à l’administration de ressources ou d’établissements de santé; TIC spécialement adaptées à la gestion ou au fonctionnement d’équipement ou de dispositifs médicaux pour la gestion ou l’administration de ressources ou d’établissements de soins de santé, p.ex. pour la gestion du personnel hospitalier ou de salles d’opération
G16H 40/60 - TIC spécialement adaptées à la gestion ou à l’administration de ressources ou d’établissements de santé; TIC spécialement adaptées à la gestion ou au fonctionnement d’équipement ou de dispositifs médicaux pour le fonctionnement d’équipement ou de dispositifs médicaux
95.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A method, computer program product, and computing system for obtaining encounter information of a patient encounter, wherein the encounter information includes machine vision encounter information; and processing the encounter information to generate an encounter transcript. A method, computer program product, and computing system for obtaining encounter information of a patient encounter, wherein the encounter information includes machine vision encounter information; processing the machine vision encounter information to identify one or more humanoid shapes; and steering one or more audio recording beams toward the one or more humanoid shapes to capture audio encounter information.
A method, computer program product, and computing system for obtaining, by a computing device, encounter information of a patient encounter, wherein the encounter information may include audio encounter information obtained from at least a first encounter participant. The audio encounter information obtained from at least the first encounter participant may be processed. An alert may be generated to obtain additional encounter information of the patient encounter.
G06F 3/00 - Dispositions d'entrée pour le transfert de données destinées à être traitées sous une forme maniable par le calculateur; Dispositions de sortie pour le transfert de données de l'unité de traitement à l'unité de sortie, p.ex. dispositions d'interface
97.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A method, computer program product, and computing system for obtaining encounter information of a patient encounter; processing the encounter information to generate an encounter transcript; processing at least a portion of the encounter transcript to populate at least a portion of a medical record associated with the patient encounter; providing the medical record to a medical professional; and enabling the medical professional to influence the medical record. A method, computer program product, and computing system for obtaining encounter information of a patient encounter; processing the encounter information to generate an encounter transcript; and processing the encounter transcript to locate one or more procedural events within the encounter transcript.
A method, computer program product, and computing system for generating a conversation print for voice-based content based, at least in part, upon a text-based transcript and speech-pattern indicia, and utilizing the conversation print to identify various third-parties (including fraudsters).
G06F 21/00 - Dispositions de sécurité pour protéger les calculateurs, leurs composants, les programmes ou les données contre une activité non autorisée
A method, computer program product, and computing system for receiving a request for information, concerning a geographically-proximate entity, on a consumer electronic device included within a vehicle. A location of the vehicle is determined; and the geographically-proximate entity is identified based, at least in part, upon the location of the vehicle.
A method, computer program product, and computer system for receiving, by a computing device, a first signal emitted from one or more sources. A second signal may be received emitted from the one or more sources. A first confidence level that the wake-up-word is included in the first signal may be determined. A second confidence level that the wake-up-word is included in the second signal may be determined. It may be identified that the wake-up-word originated from a first source of the one or more sources based upon, at least in part, the first and second confidence levels. The first source may be enabled to participate in a dialogue phase