A method, computer program product, and computing system for inserting a spectral pooling layer into a neural network of a speech processing system. An output of a hidden layer of the neural network is filtered using the spectral pooling layer with a non-integer stride. The filtered output is provided to a subsequent hidden layer of the neural network.
Example solutions for voice biometrics for anonymous identification and personalization capture an audio signal containing voice signal from a speaker. A plurality of unlabeled voiceprints are stored that are each associated with an anonymous label. The speaker's voice signal is recognized as matching one of the unlabeled voiceprints, enabling identification of the associated anonymous label. Historical information associated with the identified anonymous label is used to generate an alert specific to the speaker. Example practical applications include leveraging a customer relations management (CRM) interaction record to provide a personalized experience to the speaker and providing a warning to a user that the speaker is on a watchlist. These and other practical applications are possible, even though the speaker's identity may be unknown, and the speaker has not enrolled in a voice biometric system. Solutions for generating the unlabeled voiceprints are also disclosed.
G10L 17/00 - Identification ou vérification du locuteur
G10L 17/02 - Opérations de prétraitement, p.ex. sélection de segment; Représentation ou modélisation de motifs, p.ex. fondée sur l’analyse linéaire discriminante [LDA] ou les composantes principales; Sélection ou extraction des caractéristiques
G10L 17/04 - Entraînement, enrôlement ou construction de modèle
G10L 17/26 - Reconnaissance de caractéristiques spéciales de voix, p.ex. pour utilisation dans les détecteurs de mensonge; Reconnaissance des voix d’animaux
G06F 21/32 - Authentification de l’utilisateur par données biométriques, p.ex. empreintes digitales, balayages de l’iris ou empreintes vocales
3.
SYSTEM AND METHOD FOR WATERMARKING TRAINING DATA FOR MACHINE LEARNING MODELS
A method, computer program product, and computing system for identifying a target output token associated with an output of a machine learning model. A portion of training data corresponding to the target output token is modified with a watermark feature, thus defining watermarked training data.
G10L 15/16 - Classement ou recherche de la parole utilisant des réseaux neuronaux artificiels
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
G10L 19/018 - Mise en place d’un filigrane audio, c. à d. insertion de données inaudibles dans le signal audio
4.
SYSTEM AND METHOD FOR WATERMARKING DATA FOR TRACING ACCESS
A method, computer program product, and computing system for receiving, from a requesting party, a request to access data from a storage device. Identity information associated with the requesting party is determined. A bespoke identity-based watermark is generated for the requesting party. The bespoke identity-based watermark is encoded into the data. The watermarked data is provided to the requesting party.
A method, computer program product, and computing system for processing audio information associated with a speech processing system and encoding a watermark in a non-disruptive portion of the audio information.
An interactive voice response system is provided that includes an interactive voice recognition module, an image collection module, and a data extraction module. The image collection module communicates with the voice recognition module and the user device. The extraction module communicates with the image collection module. The voice recognition module collects speech data from a user of the user device and provides an indication to the image collection module when the speech data includes complex data. The image collection module, in response to the indication, communicates with the user device in a text message. The text message includes a link that, when activated, opens a camera on the user device. The image collection module, in response to receiving an image having the complex data from the camera, communicates the image to the extraction module, which extracts the complex data from the image as textual data.
A method, computer program product, and computing system for generating an obscured speech signal from an input speech signal and an obscured transcription from a transcription of the input speech signal. A speaker embedding may be extracted from the input speech signal. A speaker embedding delta may be generated based upon, at least in part, the extracted speaker embedding and a synthetic speaker embedding. A synthetic speech signal may be generated from the obscured speech signal using the synthetic speaker embedding. A residual signal may be generated based upon, at least in part, the obscured speech signal and the speaker embedding delta. A speech processing system may be trained using the obscured transcription, the synthetic speech signal, the speaker embedding delta, and the residual signal.
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
G10L 13/00 - Synthèse de la parole; Systèmes de synthèse de la parole à partir de texte
G10L 15/16 - Classement ou recherche de la parole utilisant des réseaux neuronaux artificiels
8.
SYSTEM AND METHOD FOR SECURE TRANSCRIPTION GENERATION
A method, computer program product, and computing system for receiving an input speech signal. A transcription of the input speech signal may be generated via an automated speech recognition (ASR) system. One or more splitting points between one or more sensitive content portions and one or more non-sensitive content portions from the transcription may be identified. The input speech signal maybe split into the one or more sensitive content portions and the one or more non-sensitive content portions based upon, at least in part, the one or more splitting points, thus defining one or more sensitive content signals and one or more non-sensitive content signals.
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
A method, computer program product, and computing system for receiving an input speech signal. A transcription of the input speech signal may be received. One or more sensitive content portions may be identified from the transcription of the input speech signal. The one or more sensitive content portions from the transcription of the input speech signal may be obscured, thus defining an obscured transcription of the input speech signal. An obscured speech signal may be generated based upon, at least in part, the input speech signal, the transcription of the input speech signal, and the obscured transcription of the input speech signal.
G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
A contextual end-to-end automatic speech recognition (ASR) system includes: an audio encoder configured to process input audio signal to produce as output encoded audio signal; a bias encoder configured to produce as output at least one bias entry corresponding to a word to bias for recognition by the ASR system; a transcription token probability prediction network configured to produce as output a probability of a selected transcription token, based at least in part on the output of the bias encoder and the output of the audio encoder; a first attention mechanism configured to receive the at least one bias entry and determine whether the at least one bias entry is suitable to be transcribed at a specific moment of an ongoing transcription; and a second attention mechanism configured to produce prefix penalties for restricting the first attention mechanism to only entries fitting a current transcription context.
There is provided a method that includes (a) obtaining a first voice vector that was derived from a signal of a voice that was sampled at a first sampling frequency, (b) obtaining a second voice vector that was derived from a signal of a voice that was sampled at a second sampling frequency, (c) mapping the second voice vector into a mapped voice vector in accordance with a machine learning model, and (d) comparing the first voice vector to the mapped voice vector to yield a score that indicates a probability that the first voice vector and the second voice vector originated from a same person.
G10L 17/04 - Entraînement, enrôlement ou construction de modèle
G10L 17/20 - Transformations de motifs ou opérations ayant pout but d’augmenter la robustesse du système, p.ex. contre le bruit du canal ou les différentes conditions de fonctionnement
G10L 17/00 - Identification ou vérification du locuteur
12.
AUTOMATIC CANONICALIZATION IN A SEMANTIC TAGGER AND SPEECH-TO-TEXT PIPELINE
A method of correcting an automatic speech recognition (ASR) output of an ASR module, includes: providing a corrector model configured to receive the ASR output; pre-training and training the corrector model to map the ASR output to desired formatting of a natural language understanding (NLU) dataset; and fine-tuning the corrector model. The mapping of the ASR output utilizes deep neural network (DNN). Out-of-domain data simulating the ASR output is utilized as the ASR output for the pre-training of the corrector model; in-domain data is utilized for the training of the corrector model; and project-specific data is utilized for fine-tuning the corrector model. The data simulating the ASR output is generated by a simulated ASR runtime process including: feeding raw text into a tokenizer to generate spelled-out text; and feeding the spelled-out text into a formatter to generate formatted text as the data simulating the ASR output.
A method, computer program product, and computing system for obtaining one or more speech signals from a first device, thus defining one or more first device speech signals. One or more speech signals may be obtained from a second device, thus defining one or more second device speech signals. An acoustic relative transfer function may be selected from a plurality of acoustic relative transfer functions based upon, at least in part, the one or more first device speech signals and the one or more second device speech signals. The one or more second device speech signals may be augmented, at run-time, based upon, at least in part, the acoustic relative transfer function
A method, computer program product, and computing system for obtaining one or more speech signals from a first device, thus defining one or more first device speech signals. One or more speech signals may be obtained from a second device, thus defining one or more second device speech signals. An acoustic relative transfer function may be selected from a plurality of acoustic relative transfer functions based upon, at least in part, the one or more first device speech signals and the one or more second device speech signals. The one or more second device speech signals may be augmented, at run-time, based upon, at least in part, the acoustic relative transfer function
A method, computer program product, and computing system for obtaining one or more speech signals from a first device, thus defining one or more first device speech signals. One or more speech signals may be obtained from a second device, thus defining one or more second device speech signals. A noise component model may be selected from a plurality of noise component models based upon, at least in part, the one or more first device speech signals and the one or more second device speech signals. The one or more second device speech signals may be augmented, at run-time, based upon, at least in part, the noise component model.
A method, computer program product, and computing system for obtaining one or more speech signals from a first device, thus defining one or more first device speech signals. One or more speech signals may be obtained from a second device, thus defining one or more second device speech signals. One or more acoustic relative transfer functions mapping reverberation from the one or more first device speech signals to the one or more second device speech signals may be generated. One or more augmented second device speech signals may be generated based upon, at least in part, the one or more acoustic relative transfer functions and first device training data
A method, computer program product, and computer system for placing, by a computing device, a user into a first queue on a first communication channel to converse with a second user. A trigger may be identified for the first communication channel. The user may be sent a self-service option based upon, at least in part, identifying the trigger for the first communication channel, wherein the self-service option is sent on an alternate communication channel while the user is in the first queue.
A method of processing speech includes: providing a first set of audio data having audio features in a first bandwidth; down-sampling the first set of audio data to a second bandwidth lower than the first bandwidth; producing, by a high frequency reconstruction network (HFRN), an estimate of audio features in the first bandwidth for the first set of audio data, based on at least the down-sampled audio data; inputting, into the HFRN, a second set of audio data having audio features in the second bandwidth; producing, by the HFRN, based on a second set of audio data having audio features in the second bandwidth, an estimate of audio features in the first bandwidth for the second set of audio data; and training a speech processing system (SPS) using the estimates of audio features in the first bandwidth for the first and second sets of audio data.
G10L 21/0388 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho utilisant des techniques d’étalement de bande - Détails du traitement à cet effet
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
G10L 13/00 - Synthèse de la parole; Systèmes de synthèse de la parole à partir de texte
19.
SYSTEM AND METHOD FOR GENERATING SYNTHETIC COHORTS USING GENERATIVE MODELING
A method, computer program product, and computing system for generating a generative model representative of a plurality of natural biometric profiles. A plurality of random samples are generated from the generative model. A plurality of synthetic biometric profiles are generated based upon, at least in part, the plurality of random samples.
A method, computer program product, and computing system for receiving a notification that a patient has arrived to a telehealth session before the telehealth session begins. The notification is received via a computing device. In response to receiving the notification that the patient has arrived to the telehealth session before the telehealth session begins, information associated with the patient is automatically pulled by a virtual assistant. The patient is prompted by the virtual assistant to complete a task before the telehealth session begins. A question is received from the patient before the telehealth session begins. Patient data may be obtained from one or more sources. The obtained patient data is processed to determine if the patient data is indicative of a possible medical condition and the medical condition is provided to a medical professional. An answer to the question is provided. The answer is personalized to the patient.
G16H 80/00 - TIC spécialement adaptées pour faciliter la communication entre les professionnels de la santé ou les patients, p.ex. pour le diagnostic collaboratif, la thérapie collaborative ou la surveillance collaborative de l’état de santé
G16H 50/20 - TIC spécialement adaptées au diagnostic médical, à la simulation médicale ou à l’extraction de données médicales; TIC spécialement adaptées à la détection, au suivi ou à la modélisation d’épidémies ou de pandémies pour le diagnostic assisté par ordinateur, p.ex. basé sur des systèmes experts médicaux
21.
MULTI-ENCODER END-TO-END AUTOMATIC SPEECH RECOGNITION (ASR) FOR JOINT MODELING OF MULTIPLE INPUT DEVICES
An end-to-end automatic speech recognition (ASR) system includes: first encoder configured for close-talk input captured by a close-talk input mechanism; second encoder configured for far-talk input captured by far-talk input mechanism; and encoder selection layer configured to select at least one of first and second encoders for use in producing ASR output. The selection is made based on at least one of short-time Fourier transform (STFT), Mel-frequency Cepstral Coefficient (MFCC) and filter bank derived from at least one of the close-talk input and far-talk input. If signals from both the close-talk input mechanism and far-talk input mechanism are present for a speech segment, the encoder selection layer dynamically selects between the close-talk encoder and far-talk encoder to select the encoder that better recognizes the speech segment. An encoder-decoder model is used to produce ASR output.
G10L 15/34 - Adaptation d’un reconnaisseur unique pour traitement en parallèle, p.ex. par utilisation de processeurs multiples ou informatique en nuage
G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine
A method, computer program product, and computing system for: monitoring a conversation between a patient and a medical entity; identifying a portion of the conversation associated with the patient, thus identifying a current patient conversation portion; and processing the current patient conversation portion to identify a condition associated with the patient, thus identifying a patient condition.
A61B 1/00 - Instruments pour procéder à l'examen médical de l'intérieur des cavités ou des conduits du corps par inspection visuelle ou photographique, p.ex. endoscopes; Dispositions pour l'éclairage dans ces instruments
A computer-implemented method, computer program product and computing system for enabling a user to initiate a problem-reporting procedure in response to an inaccurate result generated by an application when processing confidential data; processing the confidential data to generate at least one instantiation of non-confidential data that is related to the confidential data; and providing a preferred instantiation of the non-confidential data for troubleshooting the application.
G06Q 10/06 - Ressources, gestion de tâches, des ressources humaines ou de projets; Planification d’entreprise ou d’organisation; Modélisation d’entreprise ou d’organisation
H04L 9/32 - Dispositions pour les communications secrètes ou protégées; Protocoles réseaux de sécurité comprenant des moyens pour vérifier l'identité ou l'autorisation d'un utilisateur du système
24.
SYSTEM AND METHOD FOR SELF-ATTENTION-BASED COMBINING OF MULTICHANNEL SIGNALS FOR SPEECH PROCESSING
A method, computer program product, and computing system for receiving a plurality of signals from a plurality of microphones, thus defining a plurality of channels. A weighted multichannel representation of the plurality of channels may be generated. A plurality of weights for each channel of the plurality of channels may be generated based upon, at least in part, the weighted multichannel representation of the plurality of channels. A single channel representation of the plurality of channels may be generated based upon, at least in part, the weighted multichannel representation of the plurality of channels and the plurality of weights generated for each channel of the plurality of channels.
G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
H04R 1/40 - Dispositions pour obtenir la fréquence désirée ou les caractéristiques directionnelles pour obtenir la caractéristique directionnelle désirée uniquement en combinant plusieurs transducteurs identiques
A method, computer program product, and computer system for processing one or more portions of an input sequence to generate one or more candidate output sequences, thus defining a plurality of prediction scores for the candidate output sequences. One or more specialized entities may be identified from the candidate output sequences. A first scoring methodology may be applied on the candidate output sequences based upon the portions of the input sequence, thus defining a first set of prediction scores for the one or more candidate output sequences. A second scoring methodology may be applied on the specialized entities from the candidate output sequences based upon the portions of the input sequence, thus defining a second set of prediction scores for the specialized entities. The plurality of predictions scores for the specialized entities may be at least partially modified based upon the first set and the second set of prediction scores.
A method, computer program product, and computing system for: monitoring a conversation between a plurality of participants of a telehealth session; identifying an addressable issue within the conversation; and initiating an action to mitigate the addressable issue. A first participant of the plurality of participants may be a medical professional. A second participant of the plurality of participants may be a patient. The addressable issue may be a potential language barrier with one of the participants of the telehealth session. Initiating an action to mitigate the addressable issue may include: translating audio received by one of the participants of the telehealth session from a first language to a second language.
G06F 40/35 - Représentation du discours ou du dialogue
G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 80/00 - TIC spécialement adaptées pour faciliter la communication entre les professionnels de la santé ou les patients, p.ex. pour le diagnostic collaboratif, la thérapie collaborative ou la surveillance collaborative de l’état de santé
A method, computer program product, and computing system for: monitoring a conversation between a patient and a medical professional; identifying one or more medication recommendations for the patient based, at least in part, upon the conversation; and providing the one or more medication recommendations to the medical professional.
A computer-implemented method, computer program product, and computing system for monitoring a plurality of encounter participants is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to: associate at least a first portion of the encounter information with at least one known encounter participant, and associate at least a second portion of the encounter information with at least one unknown encounter participant.
G06F 21/00 - Dispositions de sécurité pour protéger les calculateurs, leurs composants, les programmes ou les données contre une activité non autorisée
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
29.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A computer-implemented method, computer program product, and computing system for visual diarization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information.
G06F 21/00 - Dispositions de sécurité pour protéger les calculateurs, leurs composants, les programmes ou les données contre une activité non autorisée
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
30.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A computer-implemented method, computer program product, and computing system for automated clinical documentation is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to generate an encounter transcript. At least a portion of the encounter transcript is processed to populate at least a portion of a record associated with the user encounter.
G16H 10/00 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients
G16H 10/20 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des essais ou des questionnaires cliniques électroniques
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
31.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A computer-implemented method, computer program product, and computing system for rendering content is executed on a computing device and includes receiving a request to render content during a user encounter. If it is determined that the content includes sensitive content, a complete version of the content is rendered on a first device (wherein the complete version of the content includes the sensitive content) and a limited version of the content on a second device (wherein the limited version of the content excludes the sensitive content).
G06F 21/00 - Dispositions de sécurité pour protéger les calculateurs, leurs composants, les programmes ou les données contre une activité non autorisée
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
32.
AUTOMATED CLINICAL DOCUMENTATION SYSTEM AND METHOD
A computer-implemented method, computer program product, and computing system for source separation is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes first audio encounter information obtained from a first encounter participant and at least second audio encounter information obtained from at least a second encounter participant. The first audio encounter information and the at least second audio encounter information are processed to eliminate audio interference between the first audio encounter information and the at least second audio encounter information.
G06F 21/00 - Dispositions de sécurité pour protéger les calculateurs, leurs composants, les programmes ou les données contre une activité non autorisée
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
33.
SYSTEM AND METHOD FOR DATA AUGMENTATION AND SPEECH PROCESSING IN DYNAMIC ACOUSTIC ENVIRONMENTS
A method, computer program product, and computing system for defining a model representative of a plurality of acoustic variations to a speech signal, thus defining a plurality of time-varying spectral modifications. The plurality of time-varying spectral modifications may be applied to a reference signal using a filtering operation, thus generating a time-varying spectrally-augmented signal.
A method, computer program product, and computing system for receiving one or more inputs indicative of at least one of: a relative location of a speaker and a microphone array, and a relative orientation of the speaker and the microphone array. One or more reference signals may be received. A speech processing system may be trained using the one or more inputs and the one or more reference signals.
G06N 99/00 - Matière non prévue dans les autres groupes de la présente sous-classe
G10K 11/00 - Procédés ou dispositifs pour transmettre, conduire ou diriger le son en général; Procédés ou dispositifs de protection contre le bruit ou les autres ondes acoustiques ou pour amortir ceux-ci, en général
G10L 13/033 - Procédés d'élaboration de parole synthétique; Synthétiseurs de parole Édition de voix, p.ex. transformation de la voix du synthétiseur
G10L 13/047 - Architecture des synthétiseurs de parole
G10L 13/10 - Règles de prosodie dérivées du texte; Intonation ou accent tonique
35.
SYSTEM AND METHOD FOR DATA AUGMENTATION AND SPEECH PROCESSING IN DYNAMIC ACOUSTIC ENVIRONMENTS
A method, computer program product, and computing system for defining model representative of a plurality of acoustic variations to a speech signal, thus defining a plurality of time-varying spectral modifications. The plurality of time-varying spectral modifications may be applied to a plurality of feature coefficients of a target domain of a reference signal, thus generating a plurality of time-varying spectrally-augmented feature coefficients of the reference signal.
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a telehealth medical encounter; gathering information during the telehealth medical encounter, thus generating gathered encounter information; and rendering an informational window concerning the telehealth medical encounter for review by the patient and/or the medical entity, wherein the informational window is configured to provide supplemental information based, at least in part, upon the gathered encounter information.
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information, wherein the gathered encounter information includes image-based content of the patient; generating image-based content information via artificial intelligence, wherein the image-based content information is based at least in part upon the image-based content and/or the gathered encounter information and is configured to provide guidance to the medical entity concerning the image-based content; and providing the image-based content information to the medical entity.
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 30/20 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le maniement d’images médicales, p.ex. DICOM, HL7 ou PACS
G16H 30/40 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le traitement d’images médicales, p.ex. l’édition
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information, wherein the gathered encounter information includes audio-based content of the patient; generating audio-based content information via artificial intelligence, wherein the audio-based content information is based at least in part upon the audio-based content and/or the gathered encounter information and is configured to provide guidance to the medical entity concerning the audio-based content; and providing the audio-based content information to the medical entity.
G06F 3/0482 - Interaction avec des listes d’éléments sélectionnables, p.ex. des menus
G16H 10/20 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des essais ou des questionnaires cliniques électroniques
G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information, wherein the gathered encounter information includes video-based content of the patient; generating video-based content information via artificial intelligence, wherein the video-based content information is based at least in part upon the video-based content and/or the gathered encounter information and is configured to provide guidance to the medical entity concerning the video-based content; and providing the video-based content information to the medical entity.
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 30/20 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le maniement d’images médicales, p.ex. DICOM, HL7 ou PACS
G16H 30/40 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le traitement d’images médicales, p.ex. l’édition
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information; and providing supplemental information to the medical entity based, at least in part, upon the gathered encounter information.
G06F 3/0482 - Interaction avec des listes d’éléments sélectionnables, p.ex. des menus
G16H 10/20 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des essais ou des questionnaires cliniques électroniques
G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information; generating medical encounter topical information via artificial intelligence, wherein the medical encounter topical information is based at least in part upon the gathered encounter information and is configured to provide guidance to the medical entity concerning one or more topics to be discussed during the medical encounter; and providing the medical encounter topical information to the medical entity.
A method, computer program product, and computing system for selecting a reference audio acquisition device from a plurality of audio acquisition devices of an audio recording system. Audio encounter information of the reference microphone may be encoded, thus defining encoded reference audio encounter information. A plurality of acoustic relative transfer functions between the reference microphone and the plurality of audio acquisition devices of the audio recording system may be generated. The encoded reference audio encounter information and a representation of the plurality of acoustic relative transfer functions may be transmitted.
G10L 19/00 - Techniques d'analyse ou de synthèse de la parole ou des signaux audio pour la réduction de la redondance, p.ex. dans les vocodeurs; Codage ou décodage de la parole ou des signaux audio utilisant les modèles source-filtre ou l’analyse psychoacoustique
43.
MULTI-CHANNEL SPEECH COMPRESSION SYSTEM AND METHOD
A method, computer program product, and computing system for generating a plurality of acoustic relative transfer functions between a plurality of audio acquisition devices of an audio recording system based upon, at least in part, one or more of a predefined speech processing application and a predefined acoustic environment. An acoustic relative transfer function codebook may be generated using the plurality of acoustic relative transfer functions. One or more channels from the plurality of audio acquisition devices of the audio recording system may be encoded using the acoustic relative transfer function codebook.
G10L 15/02 - Extraction de caractéristiques pour la reconnaissance de la parole; Sélection d'unités de reconnaissance
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 25/51 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes spécialement adaptées pour un usage particulier pour comparaison ou différentiation
44.
FIRST AND SECOND EMBEDDING OF ACOUSTIC RELATIVE TRANSFER FUNCTIONS
A method, computer program product, and computing system for generating a plurality of acoustic relative transfer functions for a plurality of audio acquisition devices of an audio recording system deployed in an acoustic environment. The plurality of acoustic relative transfer functions may be encoded into a first embedding of acoustic relative transfer functions and at least a second embedding of acoustic relative transfer functions. Information may be extracted from at least the first embedding of acoustic relative transfer functions.
G10L 19/008 - Codage ou décodage du signal audio multi-canal utilisant la corrélation inter-canaux pour réduire la redondance, p.ex. stéréo combinée, codage d’intensité ou matriçage
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information; generating medical encounter workflow information via artificial intelligence, wherein the medical encounter workflow information is based at least in part upon the gathered encounter information and is configured to provide guidance to the medical entity concerning a desired workflow for the medical encounter; and providing the medical encounter workflow information to the medical entity.
G06Q 10/06 - Ressources, gestion de tâches, des ressources humaines ou de projets; Planification d’entreprise ou d’organisation; Modélisation d’entreprise ou d’organisation
G16H 10/20 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des essais ou des questionnaires cliniques électroniques
G16H 10/40 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données relatives aux analyses de laboratoire, p.ex. pour des analyses d’échantillon de patient
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
A method, computer program product, and computing system for: monitoring a meeting between a patient and a medical entity during a medical encounter; gathering information during the medical encounter, thus generating gathered encounter information; and processing the gathered encounter information via artificial intelligence to determine whether a third-party encounter should be initiated.
A method, computer program product, and computing system for receiving audio-based content from a user who is reviewing an image on a display screen; receiving gaze information that defines a gaze location of the user; and temporally aligning the audio-based content and the gaze information to form location-based content.
G06F 3/00 - Dispositions d'entrée pour le transfert de données destinées à être traitées sous une forme maniable par le calculateur; Dispositions de sortie pour le transfert de données de l'unité de traitement à l'unité de sortie, p.ex. dispositions d'interface
48.
COMPARING ACOUSTIC RELATIVE TRANSFER FUNCTIONS FROM AT LEAST A PAIR OF TIME FRAMES
A method, computer program product, and computing system for generating a plurality of acoustic relative transfer functions associated with a plurality of audio acquisition devices of an audio recording system deployed in an acoustic environment. At least a pair of the plurality of acoustic relative transfer functions from time frames may be compared. A change in the acoustic environment may be detected based upon, at least in part, the comparison of the plurality of acoustic relative transfer functions from at least the pair of time frames.
G10L 25/51 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes spécialement adaptées pour un usage particulier pour comparaison ou différentiation
49.
MULTI-CHANNEL SPEECH COMPRESSION SYSTEM AND METHOD
A method, computer program product, and computing system for generating a plurality of acoustic relative transfer functions associated with a plurality of audio acquisition devices of an audio recording system deployed in an acoustic environment. Acoustic relative transfer functions of at least a pair of audio acquisition devices of the plurality of audio acquisition devices may be compared. Location information associated with an acoustic source within the acoustic environment may be determined based upon, at least in part, the comparison of the acoustic relative transfer functions of the at least a pair of audio acquisition devices of the plurality of audio acquisition devices.
H04S 3/00 - Systèmes utilisant plus de deux canaux, p.ex. systèmes quadriphoniques
H04S 5/00 - Systèmes pseudo-stéréophoniques, p.ex. dans lesquels les signaux d'un canal supplémentaire sont dérivés du signal monophonique par déphasage, retardement ou réverbération
50.
MULTI-CHANNEL SPEECH COMPRESSION SYSTEM AND METHOD
A method, computer program product, and computing system for encoding audio encounter information of a reference audio acquisition device of a plurality of audio acquisition devices of an audio recording system, thus defining encoded reference audio encounter information. Location information may be estimated, via a machine vision system, for an acoustic source within an acoustic environment. One or more acoustic relative transfer functions may be selected from a plurality of acoustic relative transfer functions for the plurality of audio acquisition devices of the audio recording system based upon, at least in part, the location information. The encoded reference audio encounter information and a representation of the selected one or more acoustic relative transfer function may be transmitted.
G10L 15/10 - Classement ou recherche de la parole utilisant des mesures de distance ou de distorsion entre la parole inconnue et les gabarits de référence
G10L 15/16 - Classement ou recherche de la parole utilisant des réseaux neuronaux artificiels
H04R 1/40 - Dispositions pour obtenir la fréquence désirée ou les caractéristiques directionnelles pour obtenir la caractéristique directionnelle désirée uniquement en combinant plusieurs transducteurs identiques
51.
MULTI-CHANNEL SPEECH COMPRESSION SYSTEM AND METHOD
A method, computer program product, and computing system for obtaining machine vision encounter information using one or more machine vision systems. Audio encounter information may be obtained using a plurality of audio acquisition devices of an audio recording system. The audio encounter information may be encoded using an audio codec. The encoding of the audio encounter information by the audio codec may be adapted based upon, at least in part, the machine vision encounter information.
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
G10L 15/02 - Extraction de caractéristiques pour la reconnaissance de la parole; Sélection d'unités de reconnaissance
G10L 15/10 - Classement ou recherche de la parole utilisant des mesures de distance ou de distorsion entre la parole inconnue et les gabarits de référence
A computer-implemented method, computer program product and computing system for defining a test truth set from a master truth set; processing the test truth set using an automated analysis process to generate an automated result set; determining a process efficacy for the automated analysis process based, at least in part, upon the test truth set and the automated result set; and rendering the process efficacy of the automated analysis process.
G06F 19/00 - Équipement ou méthodes de traitement de données ou de calcul numérique, spécialement adaptés à des applications spécifiques (spécialement adaptés à des fonctions spécifiques G06F 17/00;systèmes ou méthodes de traitement de données spécialement adaptés à des fins administratives, commerciales, financières, de gestion, de surveillance ou de prévision G06Q;informatique médicale G16H)
G06K 9/62 - Méthodes ou dispositions pour la reconnaissance utilisant des moyens électroniques
A computer-implemented method, computer program product and computing system for defining a test truth set from a master truth set; processing the test truth set using an automated analysis process to generate an automated result set; determining a process efficacy for the automated analysis process based, at least in part, upon the test truth set and the automated result set; and rendering the process efficacy of the automated analysis process.
A method, computer program product, and computing system for monitoring a plurality of conversations within a monitored space to generate a conversation data set; processing the conversation data set using machine learning to: define a system-directed command for an ACI system, and associate one or more conversational contexts with the system-directed command; detecting the occurrence of a specific conversational context within the monitored space, wherein the specific conversational context is included in the one or more conversational contexts associated with the system-directed command; and executing, in whole or in part, functionality associated with the system-directed command in response to detecting the occurrence of the specific conversational context without requiring the utterance of the system-directed command and/or a wake-up word / phrase.
A computer-implemented method, computer program product and computing system for defining a test truth set from a master truth set; processing the test truth set using an automated analysis process to generate an automated result set; determining a process efficacy for the automated analysis process based, at least in part, upon the test truth set and the automated result set; and rendering the process efficacy of the automated analysis process.
G06F 19/00 - Équipement ou méthodes de traitement de données ou de calcul numérique, spécialement adaptés à des applications spécifiques (spécialement adaptés à des fonctions spécifiques G06F 17/00;systèmes ou méthodes de traitement de données spécialement adaptés à des fins administratives, commerciales, financières, de gestion, de surveillance ou de prévision G06Q;informatique médicale G16H)
G06K 9/62 - Méthodes ou dispositions pour la reconnaissance utilisant des moyens électroniques
A computer-implemented method, computer program product and computing system for receiving a result set for content processed by an automated analysis process; receiving human feedback concerning the result set; and providing feedback information to the developer of the automated analysis process based, at least in part, upon the result set and the human feedback.
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
57.
AMBIENT COOPERATIVE INTELLIGENCE SYSTEM AND METHOD
A method, computer program product, and computing system for initiating a session within an ACI platform; receiving an authentication request from a requester; and authenticating that the requester has the authority to access the ACI platform.
A method, computer program product, and computing system for detecting the issuance of a verbal command by a requester to a virtual assistant; authenticating that the requester has the authority to issue the verbal command to the virtual assistant; if the requester is authenticated, allowing the effectuation of the verbal command to the virtual assistant; and if the requester is not authenticated, preventing the effectuation of the verbal command to the virtual assistant.
A method, computer program product, and computing system for receiving input information concerning a conversation between a caller and a recipient; processing the input information to assess a fraud-threat-level; defining a targeted response based, at least in part, upon the fraud-threat-level assessed, wherein the targeted response is intended to refine the assessed fraud-threat-level; and effectuating the targeted response.
A method, computer program product, and computer system for predicting responses to at least one conversational phrase. At least one conversational phrase may be received. A first probability for a subset of candidate responses of a plurality of candidate responses may be determined based upon, at least in part, context associated with the at least one conversational phrase, the at least one conversational phrase, and each context associated with the plurality of candidate responses. A second probability for the subset of candidate responses may be determined based upon, at least in part, the subset of candidate responses, the at least one conversational phrase, and the context associated with the at least one conversational phrase. At least one candidate response for the at least one conversational phrase may be determined based upon, at least in part, the first probability and the second probability.
G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
61.
Automated Clinical Documentation System and Method
A computer-implemented method, computer program product, and computing system for source separation is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes first audio encounter information obtained from a first encounter participant and at least second audio encounter information obtained from at least a second encounter participant. The first audio encounter information and the at least second audio encounter information are processed to eliminate audio interference between the first audio encounter information and the at least second audio encounter information.
A computer-implemented method, computer program product, and computing system for source separation is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes first audio encounter information obtained from a first encounter participant and at least second audio encounter information obtained from at least a second encounter participant. The first audio encounter information and the at least second audio encounter information are processed to eliminate audio interference between the first audio encounter information and the at least second audio encounter information.
A computer-implemented method, computer program product, and computing system for compartmentalizing a virtual assistant is executed on a computing device and includes obtaining encounter information via a compartmentalized virtual assistant during a user encounter, wherein the compartmentalized virtual assistant includes a core functionality module. One or more additional functionalities are added to the compartmentalized virtual assistant on an as-needed basis.
A computer-implemented method, computer program product, and computing system for source separation is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes first audio encounter information obtained from a first encounter participant and at least second audio encounter information obtained from at least a second encounter participant. The first audio encounter information and the at least second audio encounter information are processed to eliminate audio interference between the first audio encounter information and the at least second audio encounter information.
A computer-implemented method, computer program product, and computing system for compartmentalizing a virtual assistant is executed on a computing device and includes obtaining encounter information via a compartmentalized virtual assistant during a user encounter, wherein the compartmentalized virtual assistant includes a core functionality module. One or more additional functionalities are added to the compartmentalized virtual assistant on an as-needed basis.
A computer-implemented method, computer program product, and computing system for functionality module communication is executed on a computing device and includes obtaining encounter information via a compartmentalized virtual assistant during a user encounter, wherein the compartmentalized virtual assistant includes a plurality of functionality modules. At least a portion of the encounter information may be processed via a first functionality module of the plurality of functionality modules to generate a first result. The first result may be provided to a second functionality module of the plurality of functionality modules. The first result set may be processed via the second functionality module to generate a second result.
A computer-implemented method, computer program product, and computing system for source separation is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes first audio encounter information obtained from a first encounter participant and at least second audio encounter information obtained from at least a second encounter participant. The first audio encounter information and the at least second audio encounter information are processed to eliminate audio interference between the first audio encounter information and the at least second audio encounter information.
A computer-implemented method, computer program product, and computing system for compartmentalizing a virtual assistant is executed on a computing device and includes obtaining encounter information via a compartmentalized virtual assistant during a user encounter, wherein the compartmentalized virtual assistant includes a core functionality module. One or more additional functionalities are added to the compartmentalized virtual assistant on an as-needed basis.
A computer-implemented method, computer program product, and computing system for functionality module communication is executed on a computing device and includes obtaining encounter information via a compartmentalized virtual assistant during a user encounter, wherein the compartmentalized virtual assistant includes a plurality of functionality modules. At least a portion of the encounter information may be processed via a first functionality module of the plurality of functionality modules to generate a first result. The first result may be provided to a second functionality module of the plurality of functionality modules. The first result set may be processed via the second functionality module to generate a second result.
A computer-implemented method, computer program product, and computing system for synchronizing machine vision and audio is executed on a computing device and includes obtaining encounter information of a user encounter, wherein the encounter information includes machine vision encounter information and audio encounter information. The machine vision encounter information and the audio encounter information are temporally-aligned to produce a temporarily-aligned encounter recording.
G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G06K 9/00 - Méthodes ou dispositions pour la lecture ou la reconnaissance de caractères imprimés ou écrits ou pour la reconnaissance de formes, p.ex. d'empreintes digitales
A method, computer program product, and computing system for performing an assessment of initial input information, concerning a communication from a caller, to define an initial fraud-threat-level; if the initial fraud-threat-level is below a defined threat threshold, providing the communication to a recipient so that a conversation may occur between the recipient and the caller; performing an assessment of subsequent input information, concerning the conversation, to define a subsequent fraud-threat-level; and effectuating a targeted response based, at least in part, upon the subsequent fraud-threat-level, wherein the targeted response is intended to refine the subsequent fraud-threat-level.
G06F 21/32 - Authentification de l’utilisateur par données biométriques, p.ex. empreintes digitales, balayages de l’iris ou empreintes vocales
G06Q 50/26 - Services gouvernementaux ou services publics
G10L 17/06 - Techniques de prise de décision; Stratégies d’alignement de motifs
G10L 17/26 - Reconnaissance de caractéristiques spéciales de voix, p.ex. pour utilisation dans les détecteurs de mensonge; Reconnaissance des voix d’animaux
H04M 3/22 - Dispositions de supervision, de contrôle ou de test
H04M 3/42 - Systèmes fournissant des fonctions ou des services particuliers aux abonnés
63.
SYSTEM AND METHOD FOR DATA AUGMENTATION FOR MULTI-MICROPHONE SIGNAL PROCESSING
A method, computer program product, and computing system for receiving a speech signal from each microphone of a plurality of microphones, thus defining a plurality of signals. One or more noise signals associated with microphone self-noise may be received. One or more self-noise-based augmentations may be performed on the plurality of signals based upon, at least in part, the one or more noise signals associated with microphone self-noise, thus defining one or more self-noise-based augmented signals.
A method, computer program product, and computing system for receiving a signal from each microphone of a plurality of microphones, thus defining a plurality of signals. Harmonic distortion associated with at least one microphone may be determined. One or more harmonic distortion-based augmentations may be performed on the plurality of signals based upon, at least in part, the harmonic distortion associated with the at least one microphone, thus defining one or more harmonic distortion-based augmented signals.
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
65.
SYSTEM AND METHOD FOR MULTI-MICROPHONE AUTOMATED CLINICAL DOCUMENTATION
A method, computer program product, and computing system for receiving information associated with an acoustic environment. A plurality of filters may be predefined to produce a plurality of beams based upon, at least in part, the information associated with the acoustic environment. The plurality of filters may be predefined to produce a plurality of nulls based upon, at least in part, the information associated with the acoustic environment. Audio encounter information may be obtained, via one or more microphone arrays, using the plurality of beams and the plurality of nulls produced by the plurality of predefined filters.
A method, computer program product, and computing system for receiving audio encounter information from a microphone array. Speech activity within one or more portions of the audio encounter information may be identified based upon, at least in part, a correlation among the audio encounter information received from the microphone array. Location information for the one or more portions of the audio encounter information may be determined based upon, at least in part, the correlation among the signals received by each microphone of the microphone array. The one or more portions of the audio encounter information may be labeled with the speech activity and the location information.
A method, computer program product, and computing system for receiving information associated with an acoustic environment. Acoustic metadata associated with audio encounter information received by a first microphone system may be received. One or more speaker representations may be defined based upon, at least in part, the acoustic metadata associated with the audio encounter information and the information associated with the acoustic environment. One or more portions of the audio encounter information may be labeled with the one or more speaker representations and a speaker location within the acoustic environment.
A method, computer program product, and computing system for receiving a plurality of predefined beams associated with a microphone array. A plurality of predefined nulls associated with the microphone array may be received. One or more predefined beams from the plurality of predefined beams or one or more predefined nulls from the plurality of predefined nulls may be selected. A microphone array may obtain audio encounter information, via the microphone array, using at least one of the one or more selected beams and the one or more selected nulls.
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
69.
SYSTEM AND METHOD FOR MULTI-MICROPHONE AUTOMATED CLINICAL DOCUMENTATION
A method, computer program product, and computing system for receiving audio encounter information from a first microphone system, thus defining a first audio stream. Audio encounter information may be received from a second microphone system, thus defining a second audio stream. Speech activity may be detected in one or more portions of the first audio stream, thus defining one or more speech portions of the first audio stream. Speech activity may be detected in one or more portions of the second audio stream, thus defining one or more speech portions of the second audio stream. The first audio stream and the second audio stream may be aligned based upon, at least in part, the one or more speech portions of the first audio stream and the one or more speech portions of the second audio stream.
A method, computer program product, and computing system for receiving a signal from each microphone of a plurality of microphones, thus defining a plurality of signals. One or more inter-microphone gain-based augmentations may be performed on the plurality of signals, thus defining one or more inter-microphone gain-augmented signals. Performing the one or more inter-microphone gain-based augmentations on the plurality of signals may include applying a gain level from a plurality of gain levels to the signal from each microphone. Applying the gain level from the plurality of gain levels to the signal from each microphone may include applying a gain level, from a predefined range of gain levels, to the signal from each microphone. Applying the gain level from the plurality of gain levels to the signal from each microphone may include applying a random gain level, from the predefined range of gain levels, to the signal from each microphone.
A method, computer program product, and computing system for receiving a signal from each microphone of a plurality of microphones, thus defining a plurality of signals. One or more microphone frequency responses associated with at least one microphone may be received. One or more microphone frequency response-based augmentations may be performed on the plurality of signals based upon, at least in part, the one or more microphone frequency responses, thus defining one or more microphone frequency response-based augmented signals.
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 15/26 - Systèmes de synthèse de texte à partir de la parole
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
72.
SYSTEM AND METHOD FOR DATA AUGMENTATION OF FEATURE-BASED VOICE DATA
A method, computer program product, and computing system for receiving feature-based voice data associated with a first acoustic domain. One or more gain-based augmentations may be performed on at least a portion of the feature-based voice data, thus defining gain-augmented feature-based voice data.
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
A method, computer program product, and computing system for receiving feature-based voice data associated with a first acoustic domain. One or more rate-based augmentations may be performed on at least a portion of the feature-based voice data, thus defining rate-based augmented feature-based voice data.
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
A method, computer program product, and computing system for receiving feature-based voice data associated with a first acoustic domain. One or more audio feature-based augmentations may be performed on at least a portion of the feature-based voice data. Performing the one or more audio feature-based augmentations may include adding one or more audio features to the at least a portion of the feature-based voice data and/or removing one or more audio features from the at least a portion of the feature-based voice data.
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
A A method, computer program product, and computing system for generating a three-dimensional model of at least a portion of a three-dimensional space incorporating an ACI system via a video recording subsystem of an ACI calibration platform; and generating one or more audio calibration signals for receipt by an audio recording system included within the ACI system via an audio generation subsystem of the ACI calibration platform.
A method, computer program product, and computing system for obtaining calibration information for a three-dimensional space incorporating an ACI system; and processing the calibration information to calibrate the ACI system.
A method, computer program product, and computing system for extracting acoustic metadata from a signal. The signal may be converted from the time domain to the feature domain, thus defining feature-based voice data associated with the signal. The feature-based voice data associated with the signal may be processed based upon, at least in part, the acoustic metadata.
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
A method, computer program product, and computing system for receiving feature based voice data associated with a first acoustic domain. One or more reverberation-based augmentations may be performed on at least a portion of the feature-based voice data, thus defining reverberation-augmented feature-based voice data. Performing the one or more reverberation-based augmentations to the at least a portion of the feature-based voice data may include performing the one or more reverberation-based augmentations to the at least a portion of the feature-based voice data based upon, at least in part, the target acoustic domain.
G10L 15/06 - Création de gabarits de référence; Entraînement des systèmes de reconnaissance de la parole, p.ex. adaptation aux caractéristiques de la voix du locuteur
A method, computer program product, and computing system for receiving feature-based voice data. One or more data augmentation characteristics may be received. One or more augmentations of the feature-based voice data may be generated, via a machine learning model, based upon, at least in part, the feature-based voice data and the one or more data augmentation characteristics
G06F 15/18 - dans lesquels un programme est modifié en fonction de l'expérience acquise par le calculateur lui-même au cours d'un cycle complet; Machines capables de s'instruire (systèmes de commande adaptatifs G05B 13/00;intelligence artificielle G06N)
80.
Automated Clinical Documentation System and Method
A computer-implemented method, computer program product, and computing system for rendering content is executed on a computing device and includes receiving a request to render content during a user encounter. If it is determined that the content includes sensitive content, a complete version of the content is rendered on a first device (wherein the complete version of the content includes the sensitive content) and a limited version of the content on a second device (wherein the limited version of the content excludes the sensitive content).
A computer-implemented method, computer program product, and computing system for rendering content is executed on a computing device and includes receiving a request to render content during a user encounter. If it is determined that the content includes sensitive content, a complete version of the content is rendered on a first device (wherein the complete version of the content includes the sensitive content) and a limited version of the content on a second device (wherein the limited version of the content excludes the sensitive content).
A modular ACD system is configured to automate clinical documentation and includes a machine vision system configured to obtain machine vision encounter information concerning a user encounter. An audio recording system is configured to obtain audio encounter information concerning the user encounter. A compute system is configured to receive the machine vision encounter information and the audio encounter information.
A computer-implemented method, computer program product, and computing system for rendering content is executed on a computing device and includes receiving a request to render content during a user encounter. If it is determined that the content includes sensitive content, a complete version of the content is rendered on a first device (wherein the complete version of the content includes the sensitive content) and a limited version of the content on a second device (wherein the limited version of the content excludes the sensitive content).
A modular ACD system is configured to automate clinical documentation and includes a machine vision system configured to obtain machine vision encounter information concerning a user encounter. An audio recording system is configured to obtain audio encounter information concerning the user encounter. A compute system is configured to receive the machine vision encounter information and the audio encounter information.
A computer-implemented method, computer program product, and computing system for automating diarization is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. An encounter transcript is generated based, at least in part, upon the first portion of the encounter information and the at least a second portion of the encounter information.
A computer-implemented method, computer program product, and computing system for rendering content is executed on a computing device and includes receiving a request to render content during a user encounter. If it is determined that the content includes sensitive content, a complete version of the content is rendered on a first device (wherein the complete version of the content includes the sensitive content) and a limited version of the content on a second device (wherein the limited version of the content excludes the sensitive content).
A modular ACD system is configured to automate clinical documentation and includes a machine vision system configured to obtain machine vision encounter information concerning a user encounter. An audio recording system is configured to obtain audio encounter information concerning the user encounter. A compute system is configured to receive the machine vision encounter information and the audio encounter information.
A computer-implemented method, computer program product, and computing system for automating diarization is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. An encounter transcript is generated based, at least in part, upon the first portion of the encounter information and the at least a second portion of the encounter information.
A computer-implemented method, computer program product, and computing system for automating role assignment is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to associate a first portion of the encounter information with a first encounter participant. A first role is assigned to the first encounter participant.
A computer-implemented method, computer program product, and computing system for automated clinical documentation is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to generate an encounter transcript. At least a portion of the encounter transcript is processed to populate at least a portion of a record associated with the user encounter.
A computer-implemented method, computer program product, and computing system for automated clinical documentation is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to generate an encounter transcript. At least a portion of the encounter transcript is processed to populate at least a portion of a record associated with the user encounter.
A computer-implemented method, computer program product, and computing system for automating an intake process is executed on a computing device and includes prompting a user to provide encounter information via a virtual assistant during a pre-visit portion of a user encounter. Encounter information is obtained from the user in response to the prompting by the virtual assistant.
A computer-implemented method, computer program product, and computing system for automated clinical documentation is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to generate an encounter transcript. At least a portion of the encounter transcript is processed to populate at least a portion of a record associated with the user encounter.
A computer-implemented method, computer program product, and computing system for automating an intake process is executed on a computing device and includes prompting a user to provide encounter information via a virtual assistant during a pre-visit portion of a user encounter. Encounter information is obtained from the user in response to the prompting by the virtual assistant.
A computer-implemented method, computer program product, and computing system for automating a follow-up process is executed on a computing device and includes prompting a user to provide encounter information via a virtual assistant during a post-visit portion of a user encounter. Encounter information is obtained from the user in response to the prompting by the virtual assistant.
A computer-implemented method, computer program product, and computing system for automated clinical documentation is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to generate an encounter transcript. At least a portion of the encounter transcript is processed to populate at least a portion of a record associated with the user encounter.
A computer-implemented method, computer program product, and computing system for automating an intake process is executed on a computing device and includes prompting a user to provide encounter information via a virtual assistant during a pre-visit portion of a user encounter. Encounter information is obtained from the user in response to the prompting by the virtual assistant.
A computer-implemented method, computer program product, and computing system for automating a follow-up process is executed on a computing device and includes prompting a user to provide encounter information via a virtual assistant during a post-visit portion of a user encounter. Encounter information is obtained from the user in response to the prompting by the virtual assistant.
A computer-implemented method, computer program product, and computing system for automating a monitoring process is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is processed to determine if the encounter information is indicative of a potential situation. An inquiry is initiated concerning the potential situation.
G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
G16H 80/00 - TIC spécialement adaptées pour faciliter la communication entre les professionnels de la santé ou les patients, p.ex. pour le diagnostic collaboratif, la thérapie collaborative ou la surveillance collaborative de l’état de santé
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 20/30 - TIC spécialement adaptées aux thérapies ou aux plans d’amélioration de la santé, p.ex. pour manier les prescriptions, orienter la thérapie ou surveiller l’observance par les patients concernant des thérapies ou des activités physiques, p.ex. la physiothérapie, l’acupression ou les exercices
A61B 5/00 - Mesure servant à établir un diagnostic ; Identification des individus
G06T 1/00 - Traitement de données d'image, d'application générale
G16H 30/20 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le maniement d’images médicales, p.ex. DICOM, HL7 ou PACS
82.
Automated Clinical Documentation System and Method
A computer-implemented method, computer program product, and computing system for visual diarization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information.
A computer-implemented method, computer program product, and computing system for visual diarization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information.
A computer-implemented method, computer program product, and computing system for visual compartmentalization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter portion, and associate at least a second portion of the encounter information with at least a second encounter portion. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information.
A computer-implemented method, computer program product, and computing system for visual diarization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information.
A computer-implemented method, computer program product, and computing system for visual compartmentalization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter portion, and associate at least a second portion of the encounter information with at least a second encounter portion. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information.
A computer-implemented method, computer program product, and computing system for reactive encounter scanning is executed on a computing device and includes obtaining encounter information of a user encounter. A request is received from a user concerning a specific condition. In response to receiving the request, the encounter information is processed to determine if the encounter information is indicative of the specific condition and to generate a result set. The result set is provided to the user.
A computer-implemented method, computer program product, and computing system for visual diarization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter participant, and associate at least a second portion of the encounter information with at least a second encounter participant. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information.
A computer-implemented method, computer program product, and computing system for visual compartmentalization of a user encounter is executed on a computing device and includes obtaining encounter information of the user encounter. The encounter information is processed to: associate a first portion of the encounter information with a first encounter portion, and associate at least a second portion of the encounter information with at least a second encounter portion. A visual representation of the encounter information is rendered. A first visual representation of the first portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information. At least a second visual representation of the at least a second portion of the encounter information is rendered that is temporally-aligned with the visual representation of the encounter information.
A computer-implemented method, computer program product, and computing system for reactive encounter scanning is executed on a computing device and includes obtaining encounter information of a user encounter. A request is received from a user concerning a specific condition. In response to receiving the request, the encounter information is processed to determine if the encounter information is indicative of the specific condition and to generate a result set. The result set is provided to the user.
A computer-implemented method, computer program product, and computing system for proactive encounter scanning is executed on a computing device and includes obtaining encounter information of a user encounter. The encounter information is proactively processed to determine if the encounter information is indicative of one or more conditions and to generate one or more result sets. The one or more result sets are provided to the user.
G16H 40/20 - TIC spécialement adaptées à la gestion ou à l’administration de ressources ou d’établissements de santé; TIC spécialement adaptées à la gestion ou au fonctionnement d’équipement ou de dispositifs médicaux pour la gestion ou l’administration de ressources ou d’établissements de soins de santé, p.ex. pour la gestion du personnel hospitalier ou de salles d’opération
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 15/00 - TIC spécialement adaptées aux rapports médicaux, p.ex. leur création ou leur transmission
G06Q 30/02 - Marketing; Estimation ou détermination des prix; Collecte de fonds
G16H 30/20 - TIC spécialement adaptées au maniement ou au traitement d’images médicales pour le maniement d’images médicales, p.ex. DICOM, HL7 ou PACS
G06F 16/248 - Présentation des résultats de requêtes
G09B 19/00 - Enseignement non couvert par d'autres groupes principaux de la présente sous-classe
G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine
G06F 40/40 - Traitement ou traduction du langage naturel
G10L 25/51 - Techniques d'analyses de la parole ou de la voix qui ne se limitent pas à un seul des groupes spécialement adaptées pour un usage particulier pour comparaison ou différentiation
G06K 9/00 - Méthodes ou dispositions pour la lecture ou la reconnaissance de caractères imprimés ou écrits ou pour la reconnaissance de formes, p.ex. d'empreintes digitales
83.
SYSTEM AND METHOD FOR REVIEW OF AUTOMATED CLINICAL DOCUMENTATION
A method, computer program product, and computing system for obtaining, by a computing device, encounter information of a patient encounter, wherein the encounter information may include audio encounter information and video encounter information obtained from at least a first encounter participant. A report of the patient encounter may be generated based upon, at least in part, the encounter information. A relative importance of a word in the report may be determined. A portion of the video encounter information that corresponds to the word in the report may be determined. The portion of the video encounter information that corresponds to the word in the report may be stored at a first location, wherein the video encounter information may be stored at a second location remote from the first location.
G10L 19/00 - Techniques d'analyse ou de synthèse de la parole ou des signaux audio pour la réduction de la redondance, p.ex. dans les vocodeurs; Codage ou décodage de la parole ou des signaux audio utilisant les modèles source-filtre ou l’analyse psychoacoustique
84.
VEHICLE AVATAR DEVICES FOR INTERACTIVE VIRTUAL ASSISTANT
A system and method for providing avatar device (115, 125, 135, 145) status indicators for voice assistants in multi-zone vehicles. The method comprises: receiving at least one signal from a plurality of microphones (114, 124, 134, 144), wherein each microphone (114, 124, 134, 144) is associated with one of a plurality of spatial zones (110, 120, 130, 140), and one of a plurality of avatar devices (115, 125, 135, 145); wherein the at least one signal further comprises a speech signal component from a speaker; wherein the speech signal component is a voice command or question; sending zone information associated with the speaker and with one of the plurality of spatial zones (110, 120, 130, 140) to an avatar (115, 125, 135, 145); activating one the plurality of avatar devices (115, 125, 135, 145) in a respective one of the plurality of spatial zones (110, 120, 130, 140) associated with the speaker.
A method, computer program product, and computing system includes generating a graph data structure including a plurality of data points. A query for the graph data structure may be received via a user interface. At least one data point from the plurality of data points may be identified, via the user interface, in the graph data structure based upon, at least in part, the query. A selection of a data point from the identified at least one data point may be received via the user interface. The selected data point may be provided to one or more electronic data sources.
A method, computer program product, and computing system for obtaining encounter information during a patient encounter; processing the encounter information to detect the execution of a physical event during the patient encounter, thus defining a detected physical event; and deriving information for the detected physical event
G16H 10/20 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des essais ou des questionnaires cliniques électroniques
G16H 10/40 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données relatives aux analyses de laboratoire, p.ex. pour des analyses d’échantillon de patient
G16H 10/60 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients
G16H 10/65 - TIC spécialement adaptées au maniement ou au traitement des données médicales ou de soins de santé relatives aux patients pour des données spécifiques de patients, p.ex. pour des dossiers électroniques de patients stockées sur des supports d’enregistrement portables, p.ex. des cartes à puce, des étiquettes d’identification radio-fréquence [RFID] ou des CD
87.
MULTI-CHANNEL MICROPHONE SIGNAL GAIN EQUALIZATION BASED ON EVALUATION OF CROSS TALK COMPONENTS
Gain mismatch and related problems can be solved by a system and method that applies an automatic microphone signal gain equalization without any direct absolute reference or calibration phase. The system and method performs the steps of receiving, by a computing device, a speech signal from a speaking person via a plurality of microphones, determining a speech signal component in the time- frequency domain for each microphone of the plurality of microphones, calculating an instantaneous cross-talk coupling matrix based on the speech signal components across the microphones, estimating gain factors based on calculated cross-talk couplings and a given expected cross-talk attenuation, limiting the gain factors to appropriate maximum and minimum values, and applying the gain factors to the speech signal used in the control path to control further speech enhancement algorithms or used in the signal path for direct influence on the speech enhanced audio output signal.
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
H04B 3/32 - Réduction de la diaphonie, p.ex. par compensation
H03G 3/00 - Commande de gain dans les amplificateurs ou les changeurs de fréquence
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
H04J 1/12 - Dispositions pour réduire la diaphonie entre canaux
H04J 3/10 - Dispositions pour reduire la diaphonie entre canaux
G10L 21/00 - Traitement du signal de parole ou de voix pour produire un autre signal audible ou non audible, p.ex. visuel ou tactile, afin de modifier sa qualité ou son intelligibilité
88.
MULTI-MICROPHONE SPEECH DIALOG SYSTEM FOR MULTIPLE SPATIAL ZONES
There is provided a speech dialog system that includes a first microphone, a second microphone, a processor and a memory. The first microphone captures first audio from a first spatial zone, and produces a first audio signal. The second microphone captures second audio from a second spatial zone, and produces a second audio signal. The processor receives the first audio signal and the second audio signal, and the memory contains instructions that control the processor to perform operations of a speech enhancement module, an automatic speech recognition module, and a speech dialog module that performs a zone-dedicated speech dialog.
G10L 17/06 - Techniques de prise de décision; Stratégies d’alignement de motifs
G10L 15/20 - Techniques de reconnaissance de la parole spécialement adaptées de par leur robustesse contre les perturbations environnantes, p.ex. en milieu bruyant ou reconnaissance de la parole émise dans une situation de stress
G10L 15/22 - Procédures utilisées pendant le processus de reconnaissance de la parole, p.ex. dialogue homme-machine
Disclosed are systems and methods aware of ongoing conversations and configured to intelligently schedule a speech prompt to an intended addressee. A method for intelligently scheduling a speech prompt in a speech dialog system includes monitoring an acoustic environment to detect an intended addressee's availability for a speech prompt having a measure of urgency corresponding therewith. Based on the intended addressee's availability, the method predicts a time that is convenient to present the speech prompt to the intended addressee, and schedules the speech prompt based on the predicted time and the measure of urgency. A measure of rudeness can be estimated using a cost function that includes cost for presence of an utterance, cost for presence of a conversation, and cost for involvement of the intended addressee in the conversation. Scheduling the speech prompt can include trading off the measure of urgency and the measure of rudeness.
A method, computer program product, and computer system for identifying, by a computing device, a plurality of sources, wherein a first source of the plurality of sources is a source of interest and wherein a second source of the plurality of sources is an interference source. The first source and the second source may be monitored simultaneously by implementing a spatial pre-filter for acoustic source localization.
G01S 5/22 - Position de source déterminée par coordination de plusieurs lignes de position définies par des mesures de différence de parcours
G10L 19/008 - Codage ou décodage du signal audio multi-canal utilisant la corrélation inter-canaux pour réduire la redondance, p.ex. stéréo combinée, codage d’intensité ou matriçage
G10L 21/02 - Amélioration de l'intelligibilité de la parole, p.ex. réduction de bruit ou annulation d'écho
A method, computer program product, and computer system for identifying, by a computing device, a plurality of sources. One or more feature values of a plurality of features may be assigned to a first source of the plurality of sources. One or more feature values of the plurality of features may be assigned to a second source of the plurality of sources. A first score for the first source and a second score for the second source may be determined based upon, at least in part, the one or more feature values assigned to the first source and the second source. One of the first source and the second source may be selected for spatial processing based upon, at least in part, the first score for the first source and the second score for the second source.
G01S 3/00 - Radiogoniomètres pour déterminer la direction d'où proviennent des ondes infrasonores, sonores, ultrasonores ou électromagnétiques ou des émissions de particules sans caractéristiques de direction
H04R 1/00 - HAUT-PARLEURS, MICROPHONES, TÊTES DE LECTURE POUR TOURNE-DISQUES OU TRANSDUCTEURS ACOUSTIQUES ÉLECTROMÉCANIQUES ANALOGUES; APPAREILS POUR SOURDS; SYSTÈMES D'ANNONCE EN PUBLIC - Détails des transducteurs
H04R 1/32 - Dispositions pour obtenir la fréquence désirée ou les caractéristiques directionnelles pour obtenir la caractéristique directionnelle désirée uniquement
H04R 1/40 - Dispositions pour obtenir la fréquence désirée ou les caractéristiques directionnelles pour obtenir la caractéristique directionnelle désirée uniquement en combinant plusieurs transducteurs identiques
A method, computer program product, and computer system for identifying, by a computing device, a model for predicting conversational phrases for a communication between at least a first user and a second user. The model may be trained based upon, at least in part, an attribute associated with the second user. At least one conversational phrase may be predicted for the communication between the first user and the second user. The at least one conversational phrase may be provided to the second user as an optional phrase to be sent to the first user.
Provided are a call deflection and response system and method, wherein a voice call from a caller device is received, a skill group is determined to resolve an issue associated with the call, and a callback or a text response to the issue is provided to the caller device, providing a context-based personalized response. A caller leaves a detailed voicemail explaining an issue needing resolution, which is electronically transcribed and then run through a classifier to determine concepts and intents associated with the call. Based on the concepts and intents, responsibility for the call and associated files are transferred to a particular skill group on a response system for resolution. A response entity from the appropriate skill group determines and provides an issue response via callback or text message to the caller device, e.g., to the caller's mobile phone.
H04M 3/523 - Dispositions centralisées de réponse aux appels demandant l'intervention d'un opérateur avec répartition ou mise en file d'attente des appels
H04M 3/493 - Services d'information interactifs, p. ex. renseignements sur l'annuaire téléphonique
A system, method and computer-readable storage device are disclosed for managing a mute and unmute feature on a device which is used to communicate data in a communication conference. The method includes detecting, when the device is set to mute, whether the user is speaking and whether the speech is meant for the conference. Background noises are distinguished from the speech of the user. If the user is speaking and the device is set to mute, the device will automatically switch to and unmute setting such that people in the indication conference can hear the user speak. Facial recognition, and gaze detection or other data can also be used to determine when to automatically mute or unmute the device and can aid in inferring an intent of the user to speak to the conference participants.
G06K 9/00 - Méthodes ou dispositions pour la lecture ou la reconnaissance de caractères imprimés ou écrits ou pour la reconnaissance de formes, p.ex. d'empreintes digitales
A system and method for detecting multi-tone sirens despite environmental noises that may be present obtains a microphone input signal, applies, in real time, a time-frequency analysis to the microphone input signal to determine a time-frequency representation, provides at least one multi-tone model that has a plurality of tone duration patterns, performs multi-tone siren detection on the time-frequency representation, the detection based on the at least one multi-tone model and factoring of doppler shifts, and generates a detection result that can be used in systems for automated vehicles.
A system for inputting and processing handwritten, multi -character text may comprise a handwriting recognition subsystem, a word completion subsystem, and an audio feedback system. The handwriting recognition system may be configured to capture a series of handwritten characters formed by a user and to convert the handwritten characters into a set of candidate partial text strings. The word completion subsystem may be configured to identify if a candidate partial text string constitutes a word segment and if so, generate one or both of (i) at least one clarifying word and (ii) at least one clarifying phrase that includes the clarifying word. The word segment may be an arbitrary string and not correspond to a valid complete word in a language associated with the system. The audio feedback subsystem may be configured to produce an audio representation of the word segment(s), the clarifying word(s), and the clarifying phrase(s).
G06F 40/274 - Conversion de symboles en mots; Anticipation des mots à partir des lettres déjà entrées
G10L 13/02 - Procédés d'élaboration de parole synthétique; Synthétiseurs de parole
G06F 3/0488 - Techniques d’interaction fondées sur les interfaces utilisateur graphiques [GUI] utilisant des caractéristiques spécifiques fournies par le périphérique d’entrée, p.ex. des fonctions commandées par la rotation d’une souris à deux capteurs, ou par la nature du périphérique d’entrée, p.ex. des gestes en fonction de la pression exer utilisant un écran tactile ou une tablette numérique, p.ex. entrée de commandes par des tracés gestuels
97.
SYSTEM AND METHOD FOR ACOUSTIC SPEAKER LOCALIZATION
A method, computer program product, and computing system for acoustic speech localization, comprising receiving, via a plurality of microphones, a plurality of audio signals. Modulation properties of the plurality of audio signals may be analyzed. Speech sounds may be localized from the plurality of audio signals based upon, at least in part, the modulation properties of the plurality of audio signals.
G10L 19/008 - Codage ou décodage du signal audio multi-canal utilisant la corrélation inter-canaux pour réduire la redondance, p.ex. stéréo combinée, codage d’intensité ou matriçage
G10L 19/06 - Détermination ou codage des caractéristiques spectrales, p.ex. des coefficients de prédiction à court terme
G10L 21/028 - Séparation du signal de voix utilisant les propriétés des sources sonores
G10L 21/0308 - Séparation du signal de voix caractérisée par le type de mesure du paramètre, p.ex. techniques de corrélation, techniques de passage par zéro ou techniques prédictives
H04R 1/00 - HAUT-PARLEURS, MICROPHONES, TÊTES DE LECTURE POUR TOURNE-DISQUES OU TRANSDUCTEURS ACOUSTIQUES ÉLECTROMÉCANIQUES ANALOGUES; APPAREILS POUR SOURDS; SYSTÈMES D'ANNONCE EN PUBLIC - Détails des transducteurs
H04R 5/027 - Dispositions spatiales ou structurelles des microphones, p.ex. dispositifs simulant la tête humaine
Provided are a system and method of mixing a second audio stream with a first audio stream in an audio output device. The system is configured to execute the method, comprising buffering and outputting the first audio stream via the audio output device as unmodified output, determining at least one insertion spot within the first audio stream, modifying the first audio stream at an insertion spot to avoid content loss, outputting the second audio stream at the insertion spot, and resuming unmodified output of the first audio stream at or near a completion of the second audio stream. Modifying the first audio stream can include pausing and/or warping the first audio stream at the insertion spot. The audio output device can be a vehicle head unit or a wireless device, such as a mobile phone.
H04H 20/62 - Dispositions spécialement adaptées à des applications spécifiques, p.ex. aux informations sur le trafic ou aux récepteurs mobiles à la radiodiffusion locale, p.ex. la radiodiffusion en interne aux systèmes de transport, p.ex. dans les véhicules
A method, computer program product, and computing system for automatically generating a dialogue graph is executed on a computing device and includes receiving a plurality of conversation data. A plurality of utterance pairs from the plurality of conversation data may be clustered into a plurality of utterance pair clusters. A dialogue graph may be generated with a plurality of nodes representative of the plurality of utterance pair clusters.
A method, computer program product, and computer system for transforming, by a computing device, a speech signal into a speech signal representation. A regression deep neural network may be trained with a cost function to minimize a mean squared error between actual values of the speech signal representation and estimated values of the speech signal representation, wherein the cost function may include one or more discriminative terms. Bandwidth of the speech signal may be extended by extending the speech signal representation of the speech signal using the regression deep neural