A method, system and product includes displaying to a user, via a mobile device, a map view depicting locations of at least a portion of a plurality of people relative to a location of the user, wherein the user and the plurality of people are located in an environment, the user having the mobile device used for obtaining user input, the user having a hearable device used for providing audio output to the user; receiving, via the mobile device, an activation selection of a target person from the map view; capturing a noisy audio signal from the environment; processing the noisy audio signal by applying speech separation on the target person, whereby generating an enhanced audio signal; and outputting to the user via the hearable device the enhanced audio signal.
G10L 17/02 - Preprocessing operations, e.g. segment selection; Pattern representation or modelling, e.g. based on linear discriminant analysis [LDA] or principal components; Feature selection or extraction
G10L 17/04 - Training, enrolment or model building
G10L 17/06 - Decision making techniques; Pattern matching strategies
2.
PROCESSING AND UTILIZING AUDIO SIGNALS BASED ON SPEECH SEPARATION
A method, system and product includes capturing a noisy audio signal from an environment of a user, a plurality of people is located in the environment, the user having a mobile device used for obtaining user input, the user having at least one hearable device used for providing audio output to the user, the method comprising; processing the noisy audio signal to generate a first separate audio signal that represents a first voice, and a second separate audio signal that represents a second voice, said processing is performed based on first and second acoustic fingerprints that correspond to the first and second voices, respectively; combining the first and second separate audio signals to obtain an enhanced audio signal; and outputting to the user, via the at least one hearable device, the enhanced audio signal.
A method, system and product includes obtaining a first activation selection of a first person in an environment of a user, the environment comprising the first person and a second person, the user having a mobile device and a hearable device; capturing a first signal from the environment; applying speech separation on the first signal to extract a first audio signal associated with the first person; outputting to the user via the hearable device, a first output signal comprising the first audio signal; obtaining a second activation selection of the second person; capturing a second signal from the environment; applying speech separation on the second signal to extract a second audio signal associated with the second person; and outputting to the user, via the hearable device, a second output signal comprising the second audio signal.
A method, system and product includes capturing a noisy audio signal from an environment of a user in which a plurality of people participate in a conversation, the plurality of people comprising a person, the noisy audio signal includes audio emitted by a non-human object and audio emitted by the person; applying speech separation on the noisy audio signal to obtain a separate audio signal that represents a sound-of-interest, the separate audio signal is based on the audio emitted by the non-human object; generating an enhanced audio signal based on the separate audio signal, wherein said generating comprises ensuring that the separate audio signal is present in the enhanced audio signal; and outputting the enhanced audio signal to the user via at least one hearable device.
G10L 25/30 - Speech or voice analysis techniques not restricted to a single one of groups characterised by the analysis technique using neural networks
G10L 25/54 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination for retrieval
A method, system and product comprising: capturing a noisy audio signal from an environment of a user, a plurality of people is located in the environment, the user having a mobile device used for obtaining user input, the user having a hearable device used for providing audio output to the user; processing the noisy audio signal to generate first and second separate audio signals that represents first and second voices, said processing is performed based on first and second acoustic fingerprints that correspond to the first and second voices, respectively; combining the first and second separate audio signals to obtain an enhanced audio signal; and outputting to the user, via the hearable device, the enhanced audio signal.
G10L 21/02 - Speech enhancement, e.g. noise reduction or echo cancellation
G10L 17/02 - Preprocessing operations, e.g. segment selection; Pattern representation or modelling, e.g. based on linear discriminant analysis [LDA] or principal components; Feature selection or extraction
Embodiments are provided for a system comprising a camera, a microphone, and at least one processor programmed to execute a method, which may include: identifying at least one individual speaker in a first environment of a user; applying a voice classification model to classify at least a portion of an audio signal into one of a plurality of voice classifications based on at least one voice characteristic, the voice classifications denoting an emotional state of the at least one individual speaker; applying a context classification model to classify the first environment of the user into one of a plurality of contexts; associating, in at least one database, the at least one individual speaker with the voice classification, and the context classification of the first environment; and providing, to the user, at least one of an audible, visible, or tactile indication of the association.
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
G10L 17/06 - Decision making techniques; Pattern matching strategies
G10L 25/63 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination for estimating an emotional state
G06V 10/764 - Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
G06V 40/16 - Human faces, e.g. facial parts, sketches or expressions
A reading device may include a light source configured to illuminate an object; a trigger button configured to activate the light source, the trigger being operable by a finger of a hand of a user; a camera configured to capture images from an environment of the user; an audio output device configured to output audio signals; and at least one processor. The at least one processor may be programmed to: in response to operation of the trigger, project light from the light source to illuminate an area of the object; capture at least one image of the illuminated area of the object, wherein the at least one image includes a representation of written material; analyze the at least one image to recognize text; transform the recognized text into at least one audio signal; and output the at least one audio signal using the audio output device.
G10L 13/00 - Speech synthesis; Text to speech systems
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06V 30/142 - Image acquisition using hand-held instruments; Constructional details of the instruments
A system for generating processed audio signals may include an earphone comprising a microphone configured to capture sounds from an environment of the earphone, a first communication component, and a first processor; and an earphone case comprising an earphone compartment configured to receive the earphone, a second communication component, and a second processor. The first processor may be programmed to receive a captured audio signal representative of the sounds captured by the microphone; and transmit the captured audio signal via the first communication component. The second processor may be programmed to receive the captured audio signal via the second communication component; generate a processed audio signal based on analysis of the captured audio signal; and transmit the processed audio signal via the second communication component to the first communication component.
G06F 3/0481 - Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
A system for selectively amplifying audio signals may include a microphone configured to capture sounds from an environment of a user. The system may also include a processor programmed to: receive audio signals representative of the sounds captured by the microphone; cause selective conditioning of at least one audio signal received by the microphone from a region associated with the recognized individual; and cause transmission of the at least one conditioned audio signal to a hearing interface device configured to provide sound to an ear of the user.
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
G06V 10/80 - Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
A method for adjusting an audio transmission when a user of the system is being spoken to by another person includes receiving audio signals representative of sounds from an environment of the user captured by at least one microphone; determining at least from the received audio signals that the another person is speaking to user; and subject to the user being spoken to by the another person, adjusting the audio transmission to the user and signaling to the user that the user is being spoken to.
System and methods for processing audio signals are disclosed. In one implementation, a system may include a wearable device. The wearable device may include a microphone having an audio sensor configured to capture the audio signals from an environment of the user; and a processor. The processor may be programmed to receive the audio signals captured by the microphone; analyze the audio signals to generate a transcription; generate a summary of the transcription; cause the summary to be displayed to the user; receive a confirmation input from the user indicating that the displayed summary is correct; and cause the displayed summary to be stored.
G10L 13/02 - Methods for producing synthetic speech; Speech synthesisers
G06F 40/103 - Formatting, i.e. changing of presentation of documents
G10L 15/25 - Speech recognition using non-acoustical features using position of the lips, movement of the lips or face analysis
G06F 3/0481 - Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
A hearing aid and related systems and methods. In one implementation, a hearing aid system may comprise a wearable camera configured to capture images from an environment of a user, a microphone configured to capture sounds from the environment of the user, and a processor. The processor may be programmed to receive images captured by the camera; receive audio signals representative of sounds captured by the microphone; operate in a first mode to cause a first selective conditioning of a first audio signal; determine, based on analysis of at least one of the images or the audio signals, to switch to a second mode to cause a second selective conditioning of the first audio signal; and cause transmission of the first audio signal selectively conditioned in the second mode to a hearing interface device configured to provide sound to an ear of the user.
09 - Scientific and electric apparatus and instruments
10 - Medical apparatus and instruments
35 - Advertising and business services
Goods & Services
Computer hardware; computer peripherals; wearable computer peripherals in the nature of wearable cameras worn on the head; computer hardware for remotely accessing and transmitting data; computer peripherals for remotely accessing and transmitting data; computer hardware in the nature of computers in the nature of wearable cameras that are configured to be attached to frames of eyeglasses worn on the head; computer peripherals in the nature of wearable cameras worn on the head that can be attached to glasses; computer hardware for capturing audio, data and video; computer peripherals for capturing audio, data and video; downloadable computer software applications for setting up, configuring, and controlling wearable computer hardware and wearable computer peripherals; sunglasses, spectacles, eyeglasses frames, sunglasses frames, spectacles frames; battery charger; hearing aids that are used as ancillary equipment for computers and audio devices that are not for medical purposes, namely, headphones; in-the-canal headphones; completely-in-canal headphones; in-the-ear headphones; behind-the-ear headphones; on-the-ear headphones; receiver-in-canal headphones; open fit headphones; power cable; audio processing and transmitting apparatus, namely, receivers and speakers; noise cancelling headphones; noise reducing headphones; sound separating and manipulating headphones; recorded computer programs for monitoring social interactions; recorded computer programs for managing business tasks; recorded computer programs for controlling external devices based on detected context of the user; wearable computing device, namely, smartwatches; handheld text to speech reading devices, namely, electronic book readers; handheld devices for assessing and reporting a users' performance of reading activities, namely, digital book readers; handheld computers for generating questions based on a text; handheld computers for answering questions related to a text hearing aids; in-the-canal hearing aids; completely-in-canal hearing aids; in-the-ear hearing aids; behind-the- ear hearing aids; on-the-ear hearing aids; receiver-in-canal hearing aids; open fit hearing aids Message transcription; business administration; providing office functions
14.
WEARABLE SYSTEMS AND METHODS FOR SELECTIVELY READING TEXT
Systems and methods are disclosed for selectively reading text. A system may comprise an image capture device, an audio capture device, and a processor. The processor may be configured to receive images captured by the image capture device and audio signals captured by the audio capture device. The processor may analyze the image to identify text represented in the image; identify, based on the image, a structural element of the text; identify a request to read a first portion of the text associated with the structural element, the request being identified by at least one of analyzing the audio signals to detect a spoken request or detecting a gesture in the plurality of images; and present the first portion of text to the user of the wearable device.
09 - Scientific and electric apparatus and instruments
10 - Medical apparatus and instruments
35 - Advertising and business services
Goods & Services
Computer hardware; computer peripherals; wearable computer peripherals; computer hardware for remotely accessing and transmitting data; computer peripherals for remotely accessing and transmitting data; computer hardware worn on the head; computer peripherals worn on the head; computer hardware for capturing audio, data and video; computer peripherals for capturing audio, data and video; computer software applications for setting up, configuring, and controlling wearable computer hardware and wearable computer peripherals; sunglasses, spectacles, eyeglasses frames, sunglasses frames, spectacles frames; chargers; hearing aids that are used as ancillary equipment for computers and audio devices that are not for medical purposes, namely, headphones; in-the-canal headphones; completely-in-canal headphones; in-the-ear headphones; behind-the-ear headphones; on-the-ear headphones; receiver-in-canal headphones; open fit headphones power cables; audio processing and transmitting apparatus; noise cancelling apparatus; noise reducing apparatus; sound separating and manipulating apparatus; recorded computer programs for monitoring social interactions; recorded computer programs for managing tasks; recorded computer programs for controlling external devices based on detected context of the user; wearable computing devices; handheld text to speech reading devices; handheld devices for assessing and reporting a user's performance of reading activities; handheld devices for generating questions based on a text; handheld devices for answering questions related to a text. hearing aids; in-the-canal hearing aids; completely-in-canal hearing aids; in-the-ear hearing aids; behind-the- ear hearing aids; on-the-ear hearing aids; receiver-in-canal hearing aids; open fit hearing aids. Transcription; business administration; office functions.
16.
RESPONDING TO A USER QUERY BASED ON CAPTURED IMAGES AND AUDIO
A method for responding to a user query based on captured images and audio. An audio signal captured by at least one microphone is analyzed to determine at least one word. At least one image captured by at least one image sensor is analyzed to determine at least one identifier of at least one of a person, an object, a location, or an event represented in the image. The at least one word and the at least one identifier are stored in a database. A question is received from the user and is analyzed to determine at least one term. The database is searched to determine a correlation between the at least one term and the at least one word or between the at least one term and the at least one identifier. A response to the question is generated based on the correlation and is provided to the user.
A hearing aid system for generating processed audio signals may include at least one processor. The processor may be configured to receive an audio signal representative of the sounds captured by the at least one microphone; generate a processed audio signal based on analysis of the audio signal; determine a quality degradation of at least one aspect of the processed audio signal; and provide an auditory cue to the user indicating the quality degradation.
Systems and methods are disclosed for locating an object for a user. A system may comprise an image capture device, an audio capture device, and a processor. The processor may be configured to receive images captured by the image capture device and audio signals received by the audio capture device. The processor may analyze the audio signals to identify a descriptor word describing the object and retrieve a visual characteristic of the object based on the descriptor word. The processor may then determine a location of the object in the images based on the visual characteristic, determine a location of a hand of the user in the images, and determine a direction between the hand and the object. The processor may then determine feedback indicative of the direction and provide the feedback to the user.
A hearing interface device for generating processed audio signals is disclosed. In one implementation, the hearing interface device may include a housing configured to be at least partially inserted into an ear of a user, a microphone, a camera, and a processor included in the housing. The processor may be configured to receive a captured audio signal representative of sounds captured by the microphone; receive an image captured by the camera; generate a processed audio signal based on analysis of at least one of the captured audio signal or the image; and cause at least a portion of the processed audio signal to be presented to the ear of the user.
A hearing aid and related systems and methods are disclosed. In one implementation, a hearing aid system may include a wearable camera; a microphone; and a processor. The processor may be programmed to receive images captured by the camera; receive audio signals representative of sounds received by the at least one microphone; determine a look direction of the user based on analysis of the images; determine an amplitude of a first audio signal associated with an individual or object in a region associated with the look direction of the user; determine an amplitude of a second audio signal from a region other than the look direction of the user; adjust the second amplitude in accordance with the first amplitude; and cause transmission of the second audio signal at the adjusted amplitude to a hearing interface device configured to provide sound to an ear of the user.
A hearing aid and related systems and methods are disclosed. In one implementation, a system may comprise a microphone, a wearable camera, and a processor. The processor may be configured to receive a composite audio signal representative of sounds captured by the microphone, the composite audio signal including a representation of an audio source and additional audio source in the environment of the user; obtain an indication of a direction of arrival associated with the audio source, the direction of arrival representing a position of the audio source relative to the user; provide the composite audio signal, information associated with the plurality of images, and the indication of the direction of arrival to a trained model; and extract, based on an output from the trained model, an isolated audio signal from the composite audio signal, the isolated audio signal representing sounds emanating from the audio source.
A hearing aid and related systems and methods are disclosed. In one implementation, a system may comprise a microphone and a processor. The processor may be configured to receive an original audio signal representative of sounds captured by the microphone; determine that the original audio signal includes a voice of the user; process the original audio signal according to a first processing scheme to generate a first processed audio signal; transmit the first processed audio signal to a hearing interface device after a first time delay associated with the first processing scheme; determine that the original audio signal includes an additional sound; process the original audio signal according to a second processing scheme to generate a second processed audio signal; and transmit the second processed audio signal to the hearing interface device after a second time delay associated with the second processing scheme.
A hearing aid and related systems and methods is disclosed. In one implementation, a system may comprise a microphone and a processor. The processor may be configured to receive an audio signal representative of sounds captured by the microphone; process a first sample period of the audio signal using a first engine; determine, based on the processing of the first sample period using the first engine, that the audio signal is not to be transmitted to a hearing interface device; process a second sample period of the audio signal using a second engine; determine, based on the processing of the second sample period using the second engine, that the audio signal is to be transmitted to the hearing interface device; and transmit at least a part of the first portion of the audio signal to the hearing interface device at an increased rate.
A wearable apparatus and methods for operating a wearable apparatus. In one implementation, a system for automatically tracking and guiding one or more individuals in an environment includes at least one tracking subsystem including one or more cameras. The tracking subsystem includes a camera unit configured to be worn by a user, and the at least one tracking subsystem includes at least one processor programmed to: receive a plurality of images from the one or more cameras; identify at least one individual represented by the plurality of images; determine at least one characteristic of the at least one individual; and generate and send an alert based on the at least one characteristic.
Systems and methods for modifying audio signals based on context may include at least one microphone configured to capture sounds from an environment of a user; and at least one processor. The processor may be programmed to receive an audio signal representative of sounds captured by the at least one microphone; and determine a context associated with the captured sounds based on the audio signal. Subject to the context being included in a set of stored contexts, the processor may be programmed to determine at least one first speaker whose speech is to be amplified; identify at least one first portion of the audio signal associated with the determined at least one first speaker; amplify the at least one first portion of the audio signal; and transmit to a hearing interface device the amplified at least one first portion of the audio signal.
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
G06K 9/72 - Methods or arrangements for recognition using electronic means using context analysis based on the provisionally recognised identity of a number of successive patterns, e.g. a word
G10L 17/02 - Preprocessing operations, e.g. segment selection; Pattern representation or modelling, e.g. based on linear discriminant analysis [LDA] or principal components; Feature selection or extraction
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
G10L 17/06 - Decision making techniques; Pattern matching strategies
26.
Systems and methods for selectively attenuating a voice
Systems and methods for amplifying and/or attenuating audio signals are disclosed. In one implementation, a system for selectively amplifying audio signals may include at least one microphone for capturing sounds from an environment of the user and a processor. The processor may be programmed to receive an audio signal representative of sounds captured by the at least one microphone; determine whether the audio signal comprises speech by a user of the system; subject to the audio signal comprising speech by the user, modify the audio signal by attenuating a first part of the audio signal comprising the speech by the user; subject to the audio signal comprising audio other than speech by the user, modify the audio signal by amplifying a second part of the audio signal comprising audio other than the speech by the user; and transmit the modified audio signal to a hearing interface device.
System and methods for processing audio signals are disclosed. In one implementation, a system may include a microphone configured to capture sounds from an environment of a user; and at least one processor. The processor may be programmed to receive at least one audio signal representative of the sounds captured by the microphone; transcribe at least a portion of the at least one audio signal into text; generate metadata based on the transcribed text; after receiving the at least one audio signal, receive a request for information associated with a topic; select an information source from a plurality of information sources based on the received request for information; search the selected information source for a word or phrase based on the request; and output the word or phrase for entry into a record associated with the topic.
System and methods for processing audio signals are disclosed. In one implementation, a system may include a wearable apparatus including an image sensor to capture images from an environment of a user; an audio sensor to capture an audio signal from the environment of the user; and at least one processor. The processor may be programmed to receive the audio signal captured by the audio sensor; identify at least one segment including speech in the audio signal; receive an image including a representation of a code; analyze the code to determine whether the code is associated with the user and/or the wearable apparatus; and after determining that the code is associated with the user and/or the wearable apparatus, transmit at least one segment of the audio signal, at least one image of the plurality of images, and/or other information to a computing platform.
G06K 19/06 - Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code
G06F 21/32 - User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
G06V 40/16 - Human faces, e.g. facial parts, sketches or expressions
29.
Selectively conditioning audio signals based on an audioprint of an object
A hearing aid system for selectively conditioning audio signals associated with a recognized object may include at least one processor. The processor may be programmed to receive audio signals acquired by a wearable microphone; analyze the received audio signals to obtain an isolated audio stream associated with a sound-emanating object in the environment of the user; determine an audioprint from the isolated audio stream; and use the audioprint to retrieve from a database information relating to the particular sound-emanating object. Based on the retrieved information, the processor may cause selective conditioning of at least one audio signal received by the wearable microphone from a region associated with the at least one sound-emanating object; and cause transmission of the at least one conditioned audio signal to a hearing interface device configured to provide sounds to an ear of the user.
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
G10L 17/06 - Decision making techniques; Pattern matching strategies
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
G06K 9/46 - Extraction of features or characteristics of the image
A hearing aid system for selective modification of background noises may include at least one processor. The processor may be programmed to receive a plurality of images from an environment of a user captured by a wearable camera during a time period and receive an audio signal representative of sounds acquired by a wearable microphone during the time period. The processor may determine that at least one of the sounds was generated by a sound-emanating object in the environment of the user, but outside of a field of view of the wearable camera and retrieve from a database information associated with the sound-emanating object. Based on the retrieved information, the processor may cause selective conditioning of audio signals acquired by the wearable microphone during the time period and cause transmission of the conditioned audio signals to a hearing interface device configured to provide sounds to an ear of the user.
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
System and methods for processing audio signals are disclosed. In one implementation, a system may include a microphone configured to capture sounds from an environment of a user; and at least one processor. The processor may be programmed to receive at least one audio signal representative of the sounds captured by the microphone; analyze the at least one audio signal to distinguish a plurality of voices in the at least one audio signal; transcribe at least a portion of speech associated with at least one voice in the plurality of voices; and cause at least a part of the transcribed portion to be displayed to the user via a display device.
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
System and methods for processing audio signals are disclosed. In one implementation, a system may include at least one microphone configured to capture sounds from an environment of a user; and at least one processor. The processor may be programmed to receive at least one audio signal representative of at least part of the sounds captured by the microphone; identify at least one word in the at least one audio signal; and in response to identifying the at least one word, cause feedback to be provided the user.
G10L 15/22 - Procedures used during a speech recognition process, e.g. man-machine dialog
G06F 1/16 - Constructional details or arrangements
G10L 17/26 - Recognition of special voice characteristics, e.g. for use in lie detectors; Recognition of animal voices
G08B 25/10 - Alarm systems in which the location of the alarm condition is signalled to a central station, e.g. fire or police telegraphic systems characterised by the transmission medium using wireless transmission systems
A wearable device may include an image sensor configured to capture a plurality of images from an environment, a microphone configured to capture sounds from the environment, and at least one processor. The at least one processor may be programmed to receive audio signals representative of the sounds captured by the at least one microphone, and receive a first image including a representation of a first individual from among the plurality of images captured by the image sensor. The at least one processor may also be programmed to obtain a first audio segment from the audio signals using the first image. The first audio segment may include a first portion of the audio signals in which the first individual is speaking. The at least one processor may also be programmed to receive a second image including a representation of a second individual from among the plurality of images captured by the image sensor, and obtain a second audio segment from the audio signals using the second image. The second audio segment may include a second portion of the audio signals in which the second individual is speaking. The at least one processor may also be programmed to receive a third image including a representation of the first individual from among the plurality of images captured by the image sensor, and using the third image, obtain a third audio segment from the audio signals. The audio segment may include a third portion of the audio signals in which the first individual is speaking. The at least one processor may also associate the first and third audio segments with the first individual and associate the second audio segment with the second individual.
G06F 16/683 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
System and methods for selectively amplifying audio signals are disclosed. In one implementation, a method includes receiving at least one image of a plurality of images captured by a wearable camera; receiving a first audio signal representative of the sounds captured by a microphone; determining a looking direction of a user; and processing the first audio signal by amplifying audio coming from the looking direction of the user and attenuating audio coming from at least one other direction; receiving a second audio signal representative of the sounds captured by a hearing interface device; transmitting the second audio signal to a speaker associated with the hearing interface device; transmitting an additional audio signal to the speaker, wherein the transmission of the additional audio signal at least partially overlaps the transmission of the second audio signal; and transmitting the processed first audio signal to the speaker.
H04R 1/32 - Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only
G10L 21/00 - Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
H04R 3/02 - Circuits for transducers for preventing acoustic reaction
A wearable apparatus may include an image sensor configured to capture a first image from an environment of a user of the wearable apparatus and at least one processor. The processor may be programmed to receive, from an external device, a second image and an identifying detail associated with the second image; store the second image and the identifying detail in association with the second image; and recognize a person depicted in the first image based on the second image and the identifying detail associated with the second image.
G06K 7/14 - Methods or arrangements for sensing record carriers by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
G06F 1/16 - Constructional details or arrangements
G16H 40/63 - ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation
37.
Processing audio and video in a hearing aid system
System and methods for processing audio signals are disclosed. In one implementation, a system may include at least one microphone configured to capture sounds from an environment of a user; a wearable camera configured to capture a plurality of images from the environment of the user; and at least one processor. The processor may be configured to receive at least one image of the plurality of images; receive at least one audio signal representative of the sounds captured by the at least one microphone; identify an item of information based on at least one of the at least one image or the at least one audio signal; determine at least a beginning of an informational time period for providing the item of information to the user; and transmit an informational audio signal representative of the item of information to a hearing interface device during the informational time period.
G10L 13/08 - Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
A hearing aid system for individual identification of a hearing aid system may include a wearable camera, a microphone, and at least one processor. The processor may be programmed to receive a plurality of images captured by the wearable camera; receive audio signals representative of sounds captured by the microphone; and identify a first audio signal, from among the received audio signals, representative of a voice of a first individual. The processor may transcribe and store, in a memory, text corresponding to speech associated with the voice of the first individual and determine whether the first individual is a recognized individual. If the first individual is a recognized individual, the processor may associate an identifier of the first recognized individual with the stored text corresponding to the speech associated with the voice of the first individual.
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
G06V 10/80 - Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
39.
SELECTIVE INPUT FOR A HEARING AID BASED ON IMAGE DATA
Disclosed is a hearing aid system for selectively conditioning audio signals. The system may comprise a processor programmed to: receive a plurality of images captured by a wearable camera, wherein the plurality of images depict objects in an environment of a user; receive audio signals acquired by a wearable microphone, wherein the audio signals are representative of sounds emanating from the objects; analyze the plurality of images to identify at least one sound-emanating object in the environment of the user; retrieve, from a database, information about the at least one identified sound-emanating object; causing, based on the retrieved information, selective conditioning of at least one audio signal received by the wearable microphone from a region associated with the at least one sound-emanating object; and cause transmission of the at least one conditioned audio signal to a hearing interface device configured to provide sounds to an ear of the user.
A system may include a wearable camera configured to capture images and a microphone configured to capture sounds. The system may also include a processor programmed to receive audio signals from the microphone and detect, based on analysis of the audio signals, a first audio signal associated with a first time period. The first audio signal may be representative of a voice of a single individual. The processor may also be programmed to detect, based on analysis of the audio signals, a second audio signal associated with a second time period. The second time period may be different from the first time period, and the second audio signal may be representative of overlapping voices of two or more individuals. The processor may further be programmed to selectively condition the first and second audio signals, and cause transmission of the conditioned first audio signal to a hearing interface device.
G06V 10/80 - Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
Disclosed is a system for identifying sound-emanating objects in an environment of a user. The system may comprise at least one memory device and at least one processor programmed to: receive a plurality of images captured by a wearable camera; analyze the received at least one of the plurality of images to determine one or more visual characteristics associated with the at least one sound-emanating object; identify within a database in view of the one or more visual characteristics, the at least one sound-emanating object and determine a degree of certainty of identification; receive audio signals acquired by a wearable microphone; analyze the received audio signals to determine a voiceprint of the at least one sound emanating object; identify the at least one sound-emanating object based on the determined voiceprint: and initiate at least one action based on an identity of the at least one sound-emanating object.
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
G10L 17/04 - Training, enrolment or model building
A system may include a wearable camera configured to capture images and a microphone configured to capture sounds. The system may also include a processor programmed to receive the images; identify a representation of one or more individuals in the images; receive from the microphone a first audio signal associated with a voice; determine, based on analysis of the images, that the first audio signal is not associated with a voice of any of the one or more individuals; receive from the microphone a second audio signal associated with a voice; determine, based on analysis of the images, that the second audio signal is associated with a voice of one of the one or more individuals; and cause a first amplification of the first audio signal and a second amplification of the second audio signal. The first amplification may differ from the second amplification in one aspect.
G10L 17/00 - Speaker identification or verification
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
G10L 17/04 - Training, enrolment or model building
G10L 17/06 - Decision making techniques; Pattern matching strategies
G06V 10/80 - Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
System and methods for processing audio signals are disclosed. In one implementation, a system may comprise a wearable camera configured to capture images from an environment of a user; a microphone; and a processor. The processor may be configured to receive an audio signal representative of sounds captured by the microphone during a time period; and receive the images captured by the wearable camera. The processor may process the audio signal in a first mode based on audio data accumulated in a buffer prior to the time period; detect a change in the active speaker from the first individual to a second individual; and cease processing in the first mode and process the audio signal in a second mode that differs from the first mode.
G10L 25/57 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination for processing of video signals
G10L 21/0356 - Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude for synchronising with other signals, e.g. video signals
G10L 17/00 - Speaker identification or verification
System and methods for processing audio signals are disclosed. In one implementation, a system may comprise a wearable camera configured to capture images from an environment of a user; a microphone configured to capture sounds from the environment of the user; and a processor. The processor may be configured to receive at least one image of the plurality of images, the at least one image comprising a plurality of image portions associated with corresponding image portion timestamps; receive at least one audio signal representative of the sounds captured by the at least one microphone; identify an audio timestamp associated with a portion of the audio signal; identify an image portion from among the plurality of image portions, the image portion having an image portion timestamp associated with the audio timestamp; and analyze the image portion to identify a voice originating from an object represented in the image.
A hearing aid and related systems and methods. In one implementation, a hearing aid system may comprise a wearable camera configured to capture images from an environment of a user, a microphone configured to capture sounds from the environment of the user, and a processor. The processor may be programmed to receive images captured by the camera; receive audio signals representative of sounds captured by the microphone; operate in a first mode to cause a first selective conditioning of a first audio signal; determine, based on analysis of at least one of the images or the audio signals, to switch to a second mode to cause a second selective conditioning of the first audio signal; and cause transmission of the first audio signal selectively conditioned in the second mode to a hearing interface device configured to provide sound to an ear of the user.
G06F 3/0484 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
G10L 21/0216 - Noise filtering characterised by the method used for estimating noise
A hearing aid and related systems and methods. In one implementation, a hearing aid system may comprise a wearable camera configured to capture images from an environment of a user, a microphone configured to capture sounds from the environment of the user, and a processor. The processor may be programmed to receive images captured by the camera; receive audio signals representative of sounds captured by the microphone; operate in a first mode to cause a first selective conditioning of a first audio signal; determine, based on analysis of at least one of the images or the audio signals, to switch to a second mode to cause a second selective conditioning of the first audio signal; and cause transmission of the first audio signal selectively conditioned in the second mode to a hearing interface device configured to provide sound to an ear of the user.
Systems and methods are disclosed for selectively reading text. A system may comprise an image capture device, an audio capture device, and a processor. The processor may be configured to receive images captured by the image capture device and audio signals captured by the audio capture device. The processor may analyze the image to identify text represented in the image; identify, based on the image, a structural element of the text; identify a request to read a first portion of the text associated with the structural element, the request being identified by at least one of analyzing the audio signals to detect a spoken request or detecting a gesture in the plurality of images; and present the first portion of text to the user of the wearable device.
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G09B 21/00 - Teaching, or communicating with, the blind, deaf or mute
A system may include a camera configured to capture images from an environment of a user and a microphone configured to capture sounds from an environment of the user. The system may also include a processor programmed to: receive the images; identify a representation of a first individual and a representation of a second individual in the images; receive, from the microphone, a first audio signal associated with a voice of the first individual and a second audio signal associated with a voice of the second individual; detect an amplification criteria indicative of a voice amplification priority between the first individual and the second individual; selectively amplify the first audio signal relative to the second audio signal when the amplification criteria indicates that the first individual has voice amplification priority over the second individual; and cause transmission of the selectively amplified first audio signal to a hearing interface device.
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
A system for selectively amplifying audio signals may include a wearable camera configured to capture a plurality of images from an environment of a user and a microphone configured to capture sounds from an environment of the user. The system may also include a processor programmed to: receive the plurality of images captured by the camera; identify a representation of at least one recognized individual in at least one of the plurality of images; receive audio signals representative of the sounds captured by the microphone; cause selective conditioning of at least one audio signal received by the microphone from a region associated with the at least one recognized individual; and cause transmission of the at least one conditioned audio signal to a hearing interface device configured to provide sound to an ear of the user.
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
G06V 10/80 - Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
A system may include a wearable camera configured to capture a plurality of images from an environment of a user and a microphone configured to capture sounds from an environment of the user. The system may also include a processor programmed to receive the images; identify a representation of one individual in one of the images; identify a lip movement associated with a mouth of the individual, based on analysis of the images; receive audio signals representative of the sounds; identify, based on analysis of the sounds, a first audio signal associated with a first voice and a second audio signal associated with a second voice; cause selective conditioning of the first audio signal based on a determination that the first audio signal is associated with the identified lip movement; and cause transmission of the selectively conditioned first audio signal to a hearing interface device.
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
Systems and methods are disclosed for locating an object for a user. A system may comprise an image capture device, an audio capture device, and a processor. The processor may be configured to receive images captured by the image capture device and audio signals received by the audio capture device. The processor may analyze the audio signals to identify a descriptor word describing the object and retrieve a visual characteristic of the object based on the descriptor word. The processor may then determine a location of the object in the images based on the visual characteristic, determine a location of a hand of the user in the images, and determine a direction between the hand and the object. The processor may then determine feedback indicative of the direction and provide the feedback to the user.
A wearable apparatus may automatically monitor consumption by a user of the wearable apparatus by analyzing images captured from an environment of the user. The wearable apparatus may include at least one image capture device configured to capture a plurality of images from an environment of the user of the wearable apparatus. The wearable apparatus may also include at least one processing device configured to: analyze the plurality of images to detect a consumable product represented in at least one of the plurality of images; based on the detection of the consumable product represented in at least one of the plurality of images, analyze one or more of the plurality of images to determine a type indicator associated with the detected consumable product; analyze the one or more of the plurality of images to estimate an amount of the consumable product consumed by the user; determine a feedback based on the type indicator the detected consumable product and the estimated amount of the consumable product consumed by the user; and cause the feedback to be outputted to the user.
G06T 7/246 - Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
G16H 20/60 - ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to nutrition control, e.g. diets
A hearing aid and related systems and methods are disclosed. In one implementation, a hearing aid system (2300) may include a wearable camera (2301); a microphone (2302); and a processor (2303). The processor (2303) may be programmed to receive images captured by the camera (2301); receive audio signals representative of sounds received by the at least one microphone (2302); determine a look direction (2030) of the user based on analysis of the images; determine an amplitude of a first audio signal associated with an individual or object in a region associated with the look direction of the user; determine an amplitude of a second audio signal from a region other than the look direction of the user; adjust the second amplitude in accordance with the first amplitude; and cause transmission of the second audio signal at the adjusted amplitude to a hearing interface device configured to provide sound to an ear of the user (100).
A system may include a wearable camera configured to capture images and a microphone configured to capture sounds, and a processor programmed to receive the images captured by the camera and audio signals representative of sounds received by the microphone. The processor may also be programmed to determine a look direction for a user based upon detection of a representation of a body part of the user in at least one of the captured images and a pointing direction of the body part relative to an optical axis associated with the wearable camera. The processor may further be programmed to cause selective conditioning of an audio signal received by the microphone from a region associated with the look direction of the user and cause transmission of the conditioned audio signal to an interface device.
A wearable device may include a housing, a sensor in the housing configured to generate an output, and a transmitter in the housing. The wearable device may also include a processor programmed to alternatively operate in a normal radiation mode and a low radiation mode. The transmitter may be permitted to function at a normal capacity when operating in the normal radiation mode and may be caused to function at a reduced capacity when operating in the low radiation mode. During operation at the normal capacity, the transmitter may transmit at a higher radiation intensity than during operation at the reduced capacity. The processor may also be programmed to detect, based on the output generated by the sensor, whether the housing is currently worn by the user, and cause the transmitter to operate in the low radiation mode after detecting that the housing is being worn by the user.
G06F 21/32 - User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
G06F 1/16 - Constructional details or arrangements
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
H04L 9/32 - Arrangements for secret or secure communications; Network security protocols including means for verifying the identity or authority of a user of the system
57.
Systems and methods for directing audio output of a wearable apparatus
The present disclosure relates to systems and methods for directing the audio output of a wearable device having a plurality of speakers. In one implementation, the system may include an image sensor configured to capture one or more images from an environment of the user of the wearable apparatus, a plurality of speakers, and at least one processing device. The at least one processing device may be configured to analyze the one or more images to determine at least one indicator of head orientation of the user of the wearable apparatus, select at least one of the plurality of speakers based on the at least one indicator of head orientation, and output the audio to the user of the wearable apparatus via the selected at least one of the plurality of speakers.
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
H04R 1/40 - Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
H04R 1/34 - Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by using a single transducer with sound reflecting, diffracting, directing or guiding means
Devices and a method are provided for providing feedback to a user. In one implementation, the method comprises obtaining a plurality of images from an image sensor. The image sensor is configured to be positioned for movement with the user's head. The method further comprises monitoring the images, and determining whether relative motion occurs between a first portion of a scene captured in the plurality of images and other portions of the scene captured in the plurality of images. If the first portion of the scene moves less than at least one other portion of the scene, the method comprises obtaining contextual information from the first portion of the scene. The method further comprises providing the feedback to the user based on at least part of the contextual information.
G09B 21/00 - Teaching, or communicating with, the blind, deaf or mute
G06F 1/16 - Constructional details or arrangements
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
H04M 1/72403 - User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
G06V 30/224 - Character recognition characterised by the type of writing of printed characters having additional code marks or containing code marks
A61F 9/08 - Devices or methods enabling eye-patients to replace direct visual perception by another kind of perception
H04M 1/72463 - User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions to restrict the functionality of the device
Systems and methods are disclosed for using a wearable apparatus in social events. In one implementation, a system may comprise an image sensor, an audio sensor, and a processor. The processor may be configured to receive images captured by the image sensor and receive an audio signal representative of sound captured by the audio sensor. The processor may determine, based on the images or the audio signal, whether an individual is a recognized individual of the user. When the individual is not recognized, the processor may identify the individual based on an external resource. The processor may further identify a content source associated with the individual, identify a content item associated with the individual, and provide the content item to a computing device associated with the user.
G16H 80/00 - ICT specially adapted for facilitating communication between medical practitioners or patients, e.g. for collaborative diagnosis, therapy or health monitoring
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
A hearing aid and related systems and methods. In one implementation, a hearing aid system may selectively amplify sounds emanating from a detected look direction of a user of the hearing aid system. The system may include a wearable camera configured to capture a plurality of images from an environment of the user; at least one microphone configured to capture sounds from an environment of the user; and at least one processor programmed to receive the plurality of images captured by the camera, receive audio signals representative of sounds received by the at least one microphone from the environment of the user, determine a look direction for the user based on analysis of at least one of the plurality of images, cause selective conditioning of at least one audio signal received by the at least one microphone from a region associated with the look direction of the user, and cause transmission of the at least one conditioned audio signal to a hearing interface device configured to provide sound to an ear of the user.
G10L 17/26 - Recognition of special voice characteristics, e.g. for use in lie detectors; Recognition of animal voices
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
G10L 25/78 - Detection of presence or absence of voice signals
A wearable apparatus and methods may analyze images. In one implementation, the wearable apparatus may comprise a wearable image sensor and at least one processor. The at least one processor may be programmed to: receive, from the wearable image sensor, a facial image of an individual with whom a user of the wearable apparatus is interacting; receive sound data captured during the interacting; process at least a portion of the sound data to determine a spoken name of the individual; convert the spoken name to text; store, in memory, text associated with the spoken name in a manner associating the text with the facial image; after a subsequent encounter with the individual, receive, from the wearable image sensor, a subsequent facial image of the individual; perform a look-up of an identity of the individual based on the subsequent facial image; receive, from the memory the text of the spoken name of the individual; and cause a display in text of the name of the individual on a device paired with the wearable apparatus.
A wearable apparatus and methods may analyze images. In one implementation, a wearable apparatus for capturing and processing images may comprise a wearable image sensor configured to capture a plurality of images from an environment of a user of the wearable apparatus and at least one programming device. The at least one processing device may be programmed to: perform a first analysis of the plurality of images to detect at least two persons; perform a second analysis of the plurality of images to determine association information related to the at least two detected persons; and update a social representation based on the determined association information.
G06V 10/147 - Optical characteristics of the device performing the acquisition or on the illumination arrangements - Details of sensors, e.g. sensor lenses
G06V 20/20 - Scenes; Scene-specific elements in augmented reality scenes
G06V 20/30 - Scenes; Scene-specific elements in albums, collections or shared content, e.g. social network photos or video
G06V 40/20 - Movements or behaviour, e.g. gesture recognition
G06V 40/16 - Human faces, e.g. facial parts, sketches or expressions
63.
Retrieving and displaying key words from prior conversations
A wearable apparatus is provided. The wearable apparatus may include: a wearable image sensor configured to capture a plurality of images from an environment of a user; and at least one processor programmed to: receive, from the wearable image sensor, a facial image of an individual with whom a user interacted in a first interaction during a time window; receive sound data captured in a vicinity of the image sensor during a part of the time window; process the sound data to identify a key word; store an association between the key word and the facial image; receive another facial image of the individual during a second interaction; determine that the individual is the individual in the second interaction; access the memory to locate the key word from the first interaction; and cause a display of the key word on a display visible to the user.
G06V 10/147 - Optical characteristics of the device performing the acquisition or on the illumination arrangements - Details of sensors, e.g. sensor lenses
G06V 20/20 - Scenes; Scene-specific elements in augmented reality scenes
G06V 20/30 - Scenes; Scene-specific elements in albums, collections or shared content, e.g. social network photos or video
G06V 40/20 - Movements or behaviour, e.g. gesture recognition
G06V 40/16 - Human faces, e.g. facial parts, sketches or expressions
64.
Remotely identifying a location of a wearable apparatus
A wearable apparatus and methods may analyze images. In one implementation, a wearable apparatus for capturing and processing images may comprise a wearable image sensor configured to capture a plurality of images from an environment of a user of the wearable apparatus and at least one processing device. The at least one processing device may be programmed to: analyze the plurality of images to identify a plurality of people; analyze the plurality of images to determine an affinity level between the user and each of the plurality of people; obtain an image representation of each of the plurality of people; and generate, based on the affinity levels, a visualization comprising the image representations.
A wearable apparatus for providing information to a user of the wearable apparatus may include at least one image sensor configured to capture a plurality of images from an environment of the user of the wearable apparatus, at least one communication device, and at least one processing device. The processing device may be programmed to analyze at least one of the plurality of images to detect an object in the environment of the user of the wearable apparatus, determine a measurement of an estimated physical distance from the user to the object, and transmit, based on the measurement and using the at least one communication device, information related to the detected object.
G01S 5/00 - Position-fixing by co-ordinating two or more direction or position-line determinations; Position-fixing by co-ordinating two or more distance determinations
G06F 1/16 - Constructional details or arrangements
G06F 16/51 - Indexing; Data structures therefor; Storage structures
G06F 16/583 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
Devices and a method are provided for providing feedback to a user. In one implementation, the method comprises obtaining a plurality of images from an image sensor. The image sensor is configured to be positioned for movement with the user's head. The method further comprises monitoring the images, and determining whether relative motion occurs between a first portion of a scene captured in the plurality of images and other portions of the scene captured in the plurality of images. If the first portion of the scene moves less than at least one other portion of the scene, the method comprises obtaining contextual information from the first portion of the scene. The method further comprises providing the feedback to the user based on at least part of the contextual information.
A wearable apparatus and method are provided for capturing image data. In one implementation, a wearable apparatus for capturing image data is provided. The wearable apparatus includes a plurality of image sensors for capturing image data of an environment of a user. Each of the image sensors is associated with a different field of view. The wearable apparatus also includes a processing device programmed to process image data captured by at least two of the image sensors to identify an object in the environment. The processing device is also programmed to identify a first image sensor, which has a first optical axis closer to the object than a second optical axis of a second image sensor. After identifying the first image sensor, the processing device is also programmed to process image data from the first image sensor using a first processing scheme, and process image data from the second image sensor using a second processing scheme.
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06Q 30/02 - Marketing; Price estimation or determination; Fundraising
H04N 5/232 - Devices for controlling television cameras, e.g. remote control
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06F 3/038 - Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
H04N 5/235 - Circuitry for compensating for variation in the brightness of the object
G01S 3/786 - Systems for determining direction or deviation from predetermined direction using adjustment of orientation of directivity characteristics of a detector or detector system to give a desired condition of signal derived from that detector or detector system the desired condition being maintained automatically
Apparatuses and methods are provided for storing information related to objects associated with a hand of a user via a wearable camera system. In one implementation, a wearable apparatus for storing the information is provided comprising a wearable image sensor configured to capture a plurality of images from the environment of the user, and at least one processing device programmed to process the images. The processing device may detect the hand of the user, and an object associated with the user's hand. The processing device may proceed to store information related to the object. Consistent with disclosed embodiments, the stored information may be used for various purposes, such as warning the user of dangers, catering advertising to the user, and helping the user find objects when they are lost.
G01S 3/786 - Systems for determining direction or deviation from predetermined direction using adjustment of orientation of directivity characteristics of a detector or detector system to give a desired condition of signal derived from that detector or detector system the desired condition being maintained automatically
H04N 5/232 - Devices for controlling television cameras, e.g. remote control
G06Q 30/02 - Marketing; Price estimation or determination; Fundraising
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06F 3/038 - Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
H04N 5/235 - Circuitry for compensating for variation in the brightness of the object
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
The present disclosure relates to systems and methods for monitoring posture of a user of a wearable apparatus. In one implementation, the system may include an image sensor configured to capture one or more images from an environment of the user of the wearable apparatus and at least one processing device. The at least one processing device may be configured to analyze the one or more images to determine at least one indicator of the posture of the user, determine feedback for the user based on the at least one indicator of the posture of the user, and cause the feedback to be output to the user.
G06K 9/78 - Combination of image acquisition and recognition functions
G09B 5/04 - Electrically-operated educational appliances with audible presentation of the material to be studied
G09B 5/02 - Electrically-operated educational appliances with visual presentation of the material to be studied, e.g. using film strip
G08B 7/06 - Signalling systems according to more than one of groups ; Personal calling systems according to more than one of groups using electric transmission
A wearable apparatus may automatically monitor consumption by a user of the wearable apparatus by analyzing images captured from an environment of the user. The wearable apparatus may include at least one image capture device configured to capture a plurality of images from an environment of the user of the wearable apparatus. The wearable apparatus may also include at least one processing device configured to: analyze the plurality of images to detect a consumable product represented in at least one of the plurality of images; based on the detection of the consumable product represented in at least one of the plurality of images, analyze one or more of the plurality of images to determine a type indicator associated with the detected consumable product; analyze the one or more of the plurality of images to estimate an amount of the consumable product consumed by the user; determine a feedback based on the type indicator of the detected consumable product and the estimated amount of the consumable product consumed by the user; and cause the feedback to be outputted to the user.
G16H 20/60 - ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to nutrition control, e.g. diets
G16H 30/40 - ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
09 - Scientific and electric apparatus and instruments
35 - Advertising and business services
Goods & Services
(1) Computer hardware; computer peripherals; wearable computer peripherals; computer hardware for remotely accessing and transmitting data; computer peripherals for remotely accessing and transmitting data; computer hardware worn on the head; computer peripherals worn on the head; computer hardware for capturing data and video; computer peripherals for capturing data and video; computer software applications for setting up, configuring, and controlling wearable computer hardware and wearable computer peripherals; sunglasses, spectacles, eyeglasses frames, sunglasses frames, spectacles frames; charger; power cable; audio processing and transmitting apparatus; noise cancelling apparatus; recorded computer programs for monitoring social interactions; recorded computer programs for managing tasks; recorded computer programs for controlling external devices based on detected context of the user wearable computing device; (1) Transcription; business administration; office functions
72.
WEARABLE CAMERA SYSTEMS AND METHODS FOR AUTHENTICATING IDENTITY
Systems and methods may authenticate an identity of a wearer of a wearable device and manage radiation exposure of a wearer of a wearable device. In one implementation, a wearable device may include a housing configured to be worn by the wearer and at least one sensor in the housing. The at least one sensor may be configured to generate an output indicative of at least one aspect of an environment of the wearer. The at least one processor may be programmed to: alternatively operate in an unrestricted operation mode and a restricted operation mode; detect, based on the output generated by the at least one sensor, whether the wearer of the housing is authenticated with the wearable device; and operate in the unrestricted operation mode after the at least one processor detects that the wearer of the housing is authenticated with the wearable device.
The present disclosure relates to a user-augmented wearable camera system with variable image processing based on content. In one implementation, a wearable apparatus includes a wearable image sensor configured to capture a plurality of images from an environment of a user of the wearable apparatus. The wearable apparatus may include at least one processing device. The at least one processing device may be programmed to analyze at least one image to identify a visual context; determine, based on at least the visual context, feedback information for a user; provide the feedback information to the user; receive an input from the user, wherein the input reflects a determination by the user that the feedback information was insufficient or incorrect; and transmit, based on the input from the user, information related to the at least one image to an external device for additional processing.
A wearable apparatus is provided for identifying a person in an environment of a user of the wearable apparatus based on non-facial information. The wearable apparatus includes a wearable image sensor configured to capture a plurality of images from the environment of the user, and a processing device programmed to analyze a first image of the plurality of images to determine that a face appears in the first image. The processing device also analyzes a second image of the plurality of images to identify an item of non-facial information appearing in the second image that was captured within a time period including a time when the first image is captured. The processing device also determines identification information of a person associated with the face based on the item of non-facial information.
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
H04N 21/2668 - Creating a channel for a dedicated end-user group, e.g. by inserting targeted commercials into a video stream based on end-user profiles
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
H04N 21/422 - Input-only peripherals, e.g. global positioning system [GPS]
H04N 21/442 - Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed or the storage space available from the internal hard disk
G06F 16/9535 - Search customisation based on user profiles and personalisation
H04N 5/232 - Devices for controlling television cameras, e.g. remote control
G06F 1/16 - Constructional details or arrangements
G06K 9/46 - Extraction of features or characteristics of the image
G06Q 30/02 - Marketing; Price estimation or determination; Fundraising
H04L 29/08 - Transmission control procedure, e.g. data link level control procedure
A wearable apparatus and methods may analyze images. In one implementation, a wearable apparatus for capturing and processing images may comprise a wearable image sensor configured to capture a plurality of images from an environment of a user of the wearable apparatus and at least one processing device. The at least one processing device may be programmed to: analyze the plurality of images to identify a plurality of people; analyze the plurality of images to determine an affinity level between the user and each of the plurality of people; obtain an image representation of each of the plurality of people; and generate, based on the affinity levels, a visualization comprising the image representations.
Devices and a method are provided for providing feedback to a user. In one implementation, the method comprises obtaining a plurality of images from an image sensor. The image sensor is configured to be positioned for movement with the user's head. The method further comprises monitoring the images, and determining whether relative motion occurs between a first portion of a scene captured in the plurality of images and other portions of the scene captured in the plurality of images. If the first portion of the scene moves less than at least one other portion of the scene, the method comprises obtaining contextual information from the first portion of the scene. The method further comprises providing the feedback to the user based on at least part of the contextual information.
A61F 9/08 - Devices or methods enabling eye-patients to replace direct visual perception by another kind of perception
G06K 9/18 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints using printed characters having additional code marks or containing code marks, e.g. the character being composed of individual strokes of different shape, each representing a different code value
77.
WEARABLE DEVICE AND METHODS FOR ANALYZING IMAGES AND PROVIDING FEEDBACK
The present disclosure relates to systems and methods for selecting an action based on a detected person. In one implementation, a wearable apparatus may include a wearable image sensor configured to capture a plurality of images from the environment of the user of the wearable apparatus and at least one processing device. The at least one processing device may be programmed to analyze at least one of the plurality of images to detect the person; analyze at least one of the plurality of images to identify an attribute of the detected person; select at least one category for the detected person based on the identified attribute; select at least one action based on the at least one category; and cause the at least one selected action to be executed.
The present disclosure relates to systems and methods for constructing and indexing a database with joint profiles for persons viewed by multiple wearable apparatuses. In one implementation, a system for updating profile information based on data collected by a wearable apparatus may include a database storing a plurality of profiles and at least one processing device. The at least one processing device may be programmed to obtain identification information associated with a person detected in one or more images captured by a wearable image sensor included in the wearable apparatus; obtain, from the wearable apparatus, auxiliary information associated with the detected person; identify, in the database, a profile associated with the detected person based on the identification information; and update the identified profile based on the auxiliary information.
G06F 17/30 - Information retrieval; Database structures therefor
G01S 5/00 - Position-fixing by co-ordinating two or more direction or position-line determinations; Position-fixing by co-ordinating two or more distance determinations
H04N 1/32 - Circuits or arrangements for control or supervision between transmitter and receiver
A system for providing information to a user of a wearable apparatus is provided. The system includes at least one processing device programmed to identify a person represented in at least one image captured by a wearable image sensor included in the wearable apparatus, obtain information associated with the person represented in the at least one captured image, and obtain at least one affinity measurement representing a degree of a relationship between the user and the person. The processing device is also programmed to determine, based on the at least one affinity measurement, an information level to be disclosed to the user of the wearable apparatus and provide, to the user of the wearable apparatus, the information based on the information associated with the person and on the information level.
A wearable apparatus for registering a verbal contract is provided. The wearable apparatus includes an image sensor for capturing a plurality of images, an audio sensor for capturing audio data, and at least one processing device. The processing device is programmed to analyze the plurality of images to detect a person in the environment of the user of the wearable apparatus, obtain identification information associated with the detected person, analyze at least a portion of the audio data to identify one or more words associated with the verbal contract and spoken by the user of the wearable apparatus or the detected person, obtain at least one profile of the user, authenticate an identity of the user based on the at least one profile of the user, and, based on the authentication of the identity of the user, register the verbal contract and the identification information associated with the detected person.
G06F 17/30 - Information retrieval; Database structures therefor
G01S 5/00 - Position-fixing by co-ordinating two or more direction or position-line determinations; Position-fixing by co-ordinating two or more distance determinations
H04N 1/32 - Circuits or arrangements for control or supervision between transmitter and receiver
A wearable apparatus may include at least one image sensor configured to capture a plurality of images from an environment of a user of the wearable apparatus and at least one projector. The at least one light projector may be configured to emit a light pattern configured to visually indicate to the user of the wearable apparatus an active field of view of the at least one image sensor.
G06F 16/51 - Indexing; Data structures therefor; Storage structures
G06F 16/583 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
G06T 7/73 - Determining position or orientation of objects or cameras using feature-based methods
G06Q 50/00 - Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
G01S 5/00 - Position-fixing by co-ordinating two or more direction or position-line determinations; Position-fixing by co-ordinating two or more distance determinations
A wearable apparatus and method are provided for capturing image data. In one implementation, a wearable apparatus for capturing image data is provided. The wearable apparatus includes at least one image sensor for capturing image data of an environment of a user, wherein a field of view of the image sensor includes a chin of the user. The wearable apparatus includes two or more microphones, and an attachment mechanism configured to enable the image sensor and microphones to be worn by the user. The wearable apparatus includes a processing device programmed to capture at least one image, identify the chin of the user to obtain a location of the chin, select a microphone from the two or more microphones based on the location, process input from the selected microphone using a first processing scheme, and process input from a microphone that is not selected using a second processing scheme.
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06Q 30/02 - Marketing; Price estimation or determination; Fundraising
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
G06F 1/16 - Constructional details or arrangements
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06F 3/038 - Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
G06F 17/30 - Information retrieval; Database structures therefor
G01S 3/786 - Systems for determining direction or deviation from predetermined direction using adjustment of orientation of directivity characteristics of a detector or detector system to give a desired condition of signal derived from that detector or detector system the desired condition being maintained automatically
The present disclosure relates to systems and methods for directing the audio output of a wearable device having a plurality of speakers. In one implementation, the system may include an image sensor configured to capture one or more images from an environment of the user of the wearable apparatus, a plurality of speakers, and at least one processing device. The at least one processing device may be configured to analyze the one or more images to determine at least one indicator of head orientation of the user of the wearable apparatus, select at least one of the plurality of speakers based on the at least one indicator of head orientation, and output the audio to the user of the wearable apparatus via the selected at least one of the plurality of speakers.
The present disclosure relates to systems and methods for directing the audio output of a wearable device having a plurality of speakers. In one implementation, the system may include an image sensor configured to capture one or more images from an environment of the user of the wearable apparatus, a plurality of speakers, and at least one processing device. The at least one processing device may be configured to analyze the one or more images to determine at least one indicator of head orientation of the user of the wearable apparatus, select at least one of the plurality of speakers based on the at least one indicator of head orientation, and output the audio to the user of the wearable apparatus via the selected at least one of the plurality of speakers.
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
H04R 1/40 - Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
H04R 1/34 - Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by using a single transducer with sound reflecting, diffracting, directing or guiding means
09 - Scientific and electric apparatus and instruments
Goods & Services
Wearable computer hardware and software for capturing, analyzing, and processing data and video integrated with eyeglasses, sunglasses, spectacles, eyeglass frames, sunglass frames, and spectacle frames
86.
Systems and methods for determining and distributing an update to an inference model for wearable apparatuses
A system and a method determine a global update to an inference model and distribute the global update to a plurality of wearable apparatuses. In one implementation, the system includes a data interface and at least one processing device. The at least one processing device is programmed to receive, via the data interface and from a first one of the plurality of wearable apparatuses, first update information associated with the inference model. The at least one processing device is further programmed to receive, via the data interface and from a second one of the plurality of wearable apparatuses, second update information associated with the inference model. The at least one processing device is further programmed to analyze the first update information and the second update information to determine the global update to the inference model, and transmit, via the data interface, the global update to the plurality of wearable apparatuses.
A wearable imaging apparatus is provided for capturing and processing images from an environment of a user. In one implementation, the wearable apparatus may be configured with a memory for storing privacy mode triggers and associated automatically variable privacy mode settings, and at least one processing device. The processing device may analyze the images captured by the wearable apparatus, and recognize the presence of at least one of the privacy mode triggers within the images. After recognizing the at least one trigger, the processing device may cause one or more adjustments to the wearable apparatus based on the privacy mode settings associated with the at least one trigger.
G06F 16/22 - Indexing; Data structures therefor; Storage structures
G06F 16/51 - Indexing; Data structures therefor; Storage structures
G06F 16/583 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
G06F 16/58 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06K 9/62 - Methods or arrangements for recognition using electronic means
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
G06K 7/10 - Methods or arrangements for sensing record carriers by corpuscular radiation
H04N 1/00 - PICTORIAL COMMUNICATION, e.g. TELEVISION - Details thereof
A wearable apparatus is provided for capturing and processing images from an environment of a user. In one implementation, a system for controlling one or more controllable devices includes a transceiver and at least one processing device. The processing device is programmed to obtain one or more images captured by an image sensor included in a wearable apparatus, analyze the one or more images to identify a controllable device in an environment of a user of the wearable apparatus, analyze the one or more images to detect a visual trigger associated with the controllable device and, based on the detection of the visual trigger, transmit, via the transceiver, a command. The command may be configured to change at least one aspect of the controllable device.
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 16/22 - Indexing; Data structures therefor; Storage structures
G06F 16/51 - Indexing; Data structures therefor; Storage structures
G06F 16/583 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
G06F 16/58 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
H04W 4/029 - Location-based management or tracking services
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06K 9/62 - Methods or arrangements for recognition using electronic means
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
H04N 1/00 - PICTORIAL COMMUNICATION, e.g. TELEVISION - Details thereof
A wearable apparatus is provided for capturing and processing images from an environment of a user. In one implementation, a wearable apparatus includes a transmitter, a memory, an image sensor configured to capture a stream of images from an environment of a user of the wearable apparatus, and a processing device. The processing device is programmed to receive the stream of images from the image sensor, analyze the stream of images to detect the external device in the environment of the user, and cause the transmitter to transmit an interrogation signal that is configured to cause a change in an aspect of the external device. The processing device is further programmed to analyze the stream of images to detect the change in the aspect of the external device and, after detection of the change, store in the memory information relating to the external device.
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 16/22 - Indexing; Data structures therefor; Storage structures
G06F 16/51 - Indexing; Data structures therefor; Storage structures
G06F 16/583 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
G06F 16/58 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
H04W 4/029 - Location-based management or tracking services
A wearable apparatus is provided for capturing and processing images from an environment of a user. In one implementation, a system for facilitating collaboration between individuals includes a transceiver and at least one processing device. The at least one processing device is programmed to obtain and analyze one or more images captured by an image sensor included in a wearable apparatus. The at least one processing device is further programmed to detect, by the analysis, a visual trigger in an environment of a wearer of the wearable apparatus. The visual trigger may be associated with a collaborative action to be taken. The at least one processing device may be further programmed to use the transceiver to transmit an indicator relating to the visual trigger associated with the collaborative action to be taken.
A wearable apparatus is provided for capturing and processing images from an environment of a user. In one implementation, a server-based system includes a data interface for receiving a data stream including image-based information. The image based information may include images captured by a camera present on a particular wearable apparatus from among the plurality of wearable apparatuses worn by different users. The server-based system also includes at least one processing device programmed to receive the data stream from each of the plurality of wearable apparatuses using the data interface, and analyze the data streams to determine at least one trait common to two or more of the different users of the plurality of wearable apparatuses. The at least one processing device may also be programmed to store in a database information relating to the determined at least one trait.
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06F 17/30 - Information retrieval; Database structures therefor
G06K 9/62 - Methods or arrangements for recognition using electronic means
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
G06K 7/10 - Methods or arrangements for sensing record carriers by corpuscular radiation
H04N 1/00 - PICTORIAL COMMUNICATION, e.g. TELEVISION - Details thereof
A wearable apparatus is provided for capturing and processing images from an environment of a user. In one implementation, the wearable apparatus is used for causing a device paired to the wearable apparatus to execute a selected function. The wearable apparatus includes an image capture device, a transmitter and at least one processing device. The at least one processing device is programmed to obtain images captured by the image capture device; analyze the images to detect a contextual situation associated with images; based on the detected contextual situation, associate with the at least one image a category tag, wherein the category tag is associated with a selected function; determine image-related information associated with the detected contextual situation; and cause the transmitter to transmit the determined image-related information to the paired device to cause the paired device to execute the selected function based on the determined image-related information.
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 16/22 - Indexing; Data structures therefor; Storage structures
G06F 16/51 - Indexing; Data structures therefor; Storage structures
G06F 16/583 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
G06F 16/58 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
H04W 4/029 - Location-based management or tracking services
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06K 9/62 - Methods or arrangements for recognition using electronic means
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
H04N 1/00 - PICTORIAL COMMUNICATION, e.g. TELEVISION - Details thereof
A wearable imaging apparatus is provided for capturing and processing images from an environment of a user. In one implementation, the wearable apparatus may be configured with a memory for storing privacy mode triggers and associated automatically variable privacy mode settings, and at least one processing device. The processing device may analyze the images captured by the wearable apparatus, and recognize the presence of at least one of the privacy mode triggers within the images. After recognizing the at least one trigger, the processing device may cause one or more adjustments to the wearable apparatus based on the privacy mode settings associated with the at least one trigger.
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 16/22 - Indexing; Data structures therefor; Storage structures
G06F 16/51 - Indexing; Data structures therefor; Storage structures
G06F 16/583 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
G06F 16/58 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
H04W 4/029 - Location-based management or tracking services
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06K 9/62 - Methods or arrangements for recognition using electronic means
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
H04N 1/00 - PICTORIAL COMMUNICATION, e.g. TELEVISION - Details thereof
A wearable apparatus is provided for capturing and processing images from an environment of a user. In one implementation, a system for facilitating collaboration between individuals includes a transceiver and at least one processing device. The at least one processing device is programmed to obtain and analyze one or more images captured by an image sensor included in a wearable apparatus. The at least one processing device is further programmed to detect, by the analysis, a visual trigger in an environment of a wearer of the wearable apparatus. The visual trigger may be associated with a collaborative action to be taken. The at least one processing device may be further programmed to use the transceiver to transmit an indicator relating to the visual trigger associated with the collaborative action to be taken.
G06K 9/46 - Extraction of features or characteristics of the image
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06K 9/62 - Methods or arrangements for recognition using electronic means
H04N 1/00 - PICTORIAL COMMUNICATION, e.g. TELEVISION - Details thereof
G06F 16/22 - Indexing; Data structures therefor; Storage structures
G06F 16/51 - Indexing; Data structures therefor; Storage structures
H04N 5/232 - Devices for controlling television cameras, e.g. remote control
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
G06F 16/583 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
G06F 16/58 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
H04W 4/029 - Location-based management or tracking services
A wearable apparatus is provided for capturing and processing images from an environment of a user. In one implementation, a system for controlling one or more controllable devices includes a transceiver and at least one processing device. The processing device is programmed to obtain one or more images captured by an image sensor included in a wearable apparatus, analyze the one or more images to identify a controllable device in an environment of a user of the wearable apparatus, analyze the one or more images to detect a visual trigger associated with the controllable device and, based on the detection of the visual trigger, transmit, via the transceiver, a command. The command may be configured to change at least one aspect of the controllable device.
G05B 15/02 - Systems controlled by a computer electric
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 1/16 - Constructional details or arrangements
G06F 16/22 - Indexing; Data structures therefor; Storage structures
G06F 16/51 - Indexing; Data structures therefor; Storage structures
G06F 16/583 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
G06F 16/58 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
H04W 4/029 - Location-based management or tracking services
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06K 9/62 - Methods or arrangements for recognition using electronic means
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
H04N 1/00 - PICTORIAL COMMUNICATION, e.g. TELEVISION - Details thereof
A wearable apparatus is provided for capturing and processing images from an environment of a user. In one implementation, a server-based system includes a data interface for receiving a data stream including image-based information. The image based information may include images captured by a camera present on a particular wearable apparatus from among the plurality of wearable apparatuses worn by different users. The server-based system also includes at least one processing device programmed to receive the data stream from each of the plurality of wearable apparatuses using the data interface, and analyze the data streams to determine at least one trait common to two or more of the different users of the plurality of wearable apparatuses. The at least one processing device may also be programmed to store in a database information relating to the determined at least one trait.
A wearable apparatus is provided for capturing and processing images from an environment of a user. In one implementation, the wearable apparatus is used for identifying a contextual situation related to a wearer. The wearable apparatus includes a wearable image sensor configured to capture a plurality of images from an environment of the wearer. The wearable apparatus further includes at least one processing device. The at least one processing device is programmed to analyze the plurality of images to identify the contextual situation related to the wearer; determine information associated with the contextual situation; and cause the transmitter to transmit the determined information to a device paired with the wearable apparatus to cause the paired device to provide at least one alert to the wearer based on the determined information associated with the contextual situation.
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 16/22 - Indexing; Data structures therefor; Storage structures
G06F 16/51 - Indexing; Data structures therefor; Storage structures
G06F 16/583 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
G06F 16/58 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
H04W 4/029 - Location-based management or tracking services
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06K 9/62 - Methods or arrangements for recognition using electronic means
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
H04N 1/00 - PICTORIAL COMMUNICATION, e.g. TELEVISION - Details thereof
A wearable apparatus is provided for capturing and processing images from an environment of a user. In one implementation, a wearable apparatus includes a transmitter, a memory, an image sensor configured to capture a stream of images from an environment of a user of the wearable apparatus, and a processing device. The processing device is programmed to receive the stream of images from the image sensor, analyze the stream of images to detect the external device in the environment of the user, and cause the transmitter to transmit an interrogation signal that is configured to cause a change in an aspect of the external device. The processing device is further programmed to analyze the stream of images to detect the change in the aspect of the external device and, after detection of the change, store in the memory information relating to the external device.
G06F 21/62 - Protecting access to data via a platform, e.g. using keys or access control rules
H04N 5/232 - Devices for controlling television cameras, e.g. remote control
G06F 3/00 - Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
G06F 17/30 - Information retrieval; Database structures therefor
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
A wearable apparatus is provided for capturing and processing images from an environment of a user. In one implementation, a wearable apparatus for monitoring activities includes a wearable image sensor configured to capture a plurality of images from an environment of a user of the wearable apparatus. The wearable apparatus also includes at least one processing device programmed to analyze the plurality of images to identify in one or more of the plurality of images at least one indicator of an activity, and to transmit, to an external device, the at least one indicator of the activity.
H04N 21/2668 - Creating a channel for a dedicated end-user group, e.g. by inserting targeted commercials into a video stream based on end-user profiles
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
H04N 21/422 - Input-only peripherals, e.g. global positioning system [GPS]
H04N 21/442 - Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed or the storage space available from the internal hard disk
G06F 16/9535 - Search customisation based on user profiles and personalisation
H04N 5/232 - Devices for controlling television cameras, e.g. remote control
G06F 1/16 - Constructional details or arrangements
G06K 9/46 - Extraction of features or characteristics of the image
G06Q 30/02 - Marketing; Price estimation or determination; Fundraising
H04L 29/08 - Transmission control procedure, e.g. data link level control procedure
A system is provided for analyzing information collected by a plurality of wearable camera systems. In one implementation, the system may include a memory storing executable instructions and at least one processing device. The at least one processing device may be programmed to execute the instructions to receive information derived from image data captured by the wearable camera systems; analyze the derived information to identify a commonality related to the image data captured by at least two of the wearable camera systems; determine, based on the commonality, statistical data related to users of the at least two of the wearable camera systems; and select, based on the statistical data, at least one content item for at least one of the users of the wearable camera systems who share the commonality.
H04N 21/2668 - Creating a channel for a dedicated end-user group, e.g. by inserting targeted commercials into a video stream based on end-user profiles
G06K 9/00 - Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
H04N 21/422 - Input-only peripherals, e.g. global positioning system [GPS]
H04N 21/442 - Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed or the storage space available from the internal hard disk
G06F 16/9535 - Search customisation based on user profiles and personalisation
H04N 5/232 - Devices for controlling television cameras, e.g. remote control
G06F 1/16 - Constructional details or arrangements
G06K 9/46 - Extraction of features or characteristics of the image
G06Q 30/02 - Marketing; Price estimation or determination; Fundraising
H04L 29/08 - Transmission control procedure, e.g. data link level control procedure