A method may include obtaining a first video that includes sign language content. In some embodiments, the sign language content may include one or more video frames of a figure performing sign language. The method may also include obtaining language data that represents the sign language content in the first video and creating a second video including sign language content by altering the first video. The method may further include training a machine learning model of a translation system configured to translate between sign language and language data using the second video and the language data.
A method may include obtaining a first video data including sign language originating at a first device during a communication session, obtaining one or more features from the first video data, and determining one or more matching functions from the one or more features. The method may further include determining, using a language model, a first set of one or more symbols from the one or more matching functions, and determining a second set of one or more symbols from the first set of one or more symbols.
A communication system may include a communication device configured to receive a video stream including sign language content and any content indicators associated with the video stream during a real-time communication session within a single communication device or between a plurality of communication devices. The communication system may also include a translation engine configured to automatically translate the sign language content into word content during the real-time communication session without assistance of a human sign language interpreter. Further, the communication system may be configured to output the word content translation to a communication device during the real-time communication session.
A method to generate a contact list may include receiving an identifier of a first communication device at a captioning system. The first communication device may be configured to provide first audio data to a second communication device. The second communication device may be configured to receive first text data of the first audio data from the captioning system. The method may further include receiving and storing contact data from each of multiple communication devices at the captioning system. The method may further include selecting the contact data from the multiple communication devices that include the identifier of the first communication device as selected contact data and generating a contact list based on the selected contact data. The method may also include sending the contact list to the first communication device to provide the contact list as contacts for presentation on an electronic display of the first communication device.
H04L 67/561 - Adding application-functional data or data for application control, e.g. adding metadata
H04M 1/2757 - Devices whereby a plurality of signals may be stored simultaneously with provision for storing more than one subscriber number at a time using static electronic memories, e.g. chips providing data content by data transmission, e.g. downloading
H04M 1/72478 - User interfaces specially adapted for cordless or mobile telephones specially adapted for disabled users for hearing-impaired users
Video relay services, communication systems, non-transitory machine-readable storage media, and methods are disclosed herein. A video relay service may include at least one server configured to receive a video stream including sign language content from a video communication device during a real-time communication session. The server may also be configured to automatically translate the sign language content into a verbal language translation during the real-time communication session without assistance of a human sign language interpreter. Further, the server may be configured to transmit the verbal language translation during the real-time communication session.
A method may include obtaining a text string that is a transcription of audio data and selecting a sequence of words from the text string as a first word sequence. The method may further include encrypting the first word sequence and comparing the encrypted first word sequence to multiple encrypted word sequences. Each of the multiple encrypted word sequences may be associated with a corresponding one of multiple counters. The method may also include in response to the encrypted first word sequence corresponding to one of the multiple encrypted word sequences based on the comparison, incrementing a counter of the multiple counters associated with the one of the multiple encrypted word sequences and adapting a language model of an automatic transcription system using the multiple encrypted word sequences and the multiple counters.
G10L 15/06 - Creation of reference templatesTraining of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
G10L 15/19 - Grammatical context, e.g. disambiguation of recognition hypotheses based on word sequence rules
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
A method may include obtaining a first audio signal including first speech originating at a remote device during a communication session between the remote device and a communication device and obtaining a second audio signal including second speech originating at the communication device during the communication session between the remote device and the communication device. The method may also include obtaining a characteristic of the communication session from one or more of: the first audio signal, the second audio signal, and settings of the communication device and determining a hearing level of a user of the communication device using the characteristic of the communication session.
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
H04M 1/247 - Telephone sets including user guidance or feature selection means facilitating their use
H04M 1/72475 - User interfaces specially adapted for cordless or mobile telephones specially adapted for disabled users
A method to access a device may include obtaining, at a first device, data over a short-range wireless network from a second device. The data may originate at a remote system that sends the data to the second device through a network connection over a wide area network. The method may also include in response to a fault at the second device, obtaining, at the first device from the remote system, a maintenance command for the second device. The maintenance command may be obtained by the first device over an analog voice network. The method may also include directing, from the first device to the second device, the maintenance command over the short-range wireless network to enable the second device to perform the maintenance command.
H04L 41/0659 - Management of faults, events, alarms or notifications using network fault recovery by isolating or reconfiguring faulty entities
H04L 41/0816 - Configuration setting characterised by the conditions triggering a change of settings the condition being an adaptation, e.g. in response to network events
H04W 4/80 - Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
H04W 76/18 - Management of setup rejection or failure
H04W 24/04 - Arrangements for maintaining operational condition
A method to transcribe communications may include selecting a first transcription generation technique from among multiple transcription generation techniques for generating transcriptions of audio of one or more communication sessions that involve a user device and obtaining performances of the multiple transcription generation techniques with respect to generating the transcriptions of the audio. The method may also include monitoring comparisons between the performances of the multiple transcription generation techniques and obtaining input from the user with respect to the comparisons. The method may further include selecting a second transcription generation technique from among the multiple transcription generation techniques based on the input from the user.
A method transcript presentation may include generating, by a device, audio data using an audible audio signal that is broadcast by the device. The method may also include obtaining, at the device, transcript data. The transcript data may be generated using the audio data and may include a transcription of the audio data. The method may also include presenting, by the device, the transcript data.
A system is provided that includes a first network interface for a first network type and a second network interface for a second network type that is different from the first network type. The system also includes at least one processor configured to cause the system to perform operations. The operations may include obtaining, from the first network interface, audio from a communication session with a remote device established over the first network and obtaining an indication of a communication device available to participate in the communication session and direct audio obtained from the communication session to a remote transcription system. The operations may also include directing the audio to the second network interface for transmission to the communication device, obtaining transcript data from the remote transcription system based on the audio, and directing the transcript data to the second network interface for transmission to the communication device.
G10L 15/01 - Assessment or evaluation of speech recognition systems
G10L 15/18 - Speech classification or search using natural language modelling
G10L 15/28 - Constructional details of speech recognition systems
G10L 15/30 - Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
H04M 1/247 - Telephone sets including user guidance or feature selection means facilitating their use
H04M 1/253 - Telephone sets using digital voice transmission
H04M 3/42 - Systems providing special services or facilities to subscribers
H04M 7/00 - Arrangements for interconnection between switching centres
H04M 1/72412 - User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories using two-way short-range wireless interfaces
A method to present communications may include captioning, by a human assistant during a call between a first user using a first captioning telephone device and a second user using a second telephone device, words spoken by the second user into the second telephone device. The method may also include presenting the captioned words on a first display of the first captioning telephone device, receiving text typed into the second telephone device by the second user, and presenting the received text on the first display of the first captioning telephone device.
A method may include obtaining an indicator that a first device is in a location of a second device and in response to obtaining the indicator, sending a redirect request to a communication service provider of the first device to direct, to the second device, incoming communication requests handled by the communication service provider that are directed to the first device. The method may further include after sending the redirect request and after a communication request to a communication session is directed to the first device, obtaining, at the second device, a communication indication to participate in the communication session. The method may further include directing audio of the communication session to a transcription system and obtaining, at the second device, the transcription of the audio from the transcription system. The method may also include presenting, by the second device, the audio and the transcription.
H04W 4/80 - Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
A method of transcript presentation may include obtaining, at a device, transcript data that includes a transcription of audio data. The method may also include presenting, by a display of the device, the transcript data and obtaining, at the device, a revision to the transcript data. The method may further include obtaining an indication of a change to the presentation on the display of the device and in response to the indication of the change to the presentation, presenting, by the device, the revision.
A method of transcript presentation may include obtaining, at a device, transcript data that includes a transcription of audio data. The method may also include presenting, by a display of the device, the transcript data and obtaining, at the device, a revision to the transcript data. The method may further include obtaining an indication of a change to the presentation on the display of the device and in response to the indication of the change to the presentation, presenting, by the device, the revision.
A method of transcription communication may include obtaining, at a first device, audio data during a communication session between the first device and a remote device. The audio data may be obtained from a public switched telephone network. The method may also include directing the audio data from the first device to a second device over a short-range wireless network. The second device may host the short-range wireless network. The method may further include directing the audio data from the second device to a remote transcription system over a wireless cellular network and obtaining, at the second device, transcript data from the remote transcription system over the wireless cellular network. The method may also include limiting data transmitted and received over the wireless cellular network as authorized by the second device to data associated with transcription of audio obtained by the first device over the public switched telephone network.
H04W 4/80 - Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
A method to generate a contact list may include receiving an identifier of a first communication device at a captioning system. The first communication device may be configured to provide first audio data to a second communication device. The second communication device may be configured to receive first text data of the first audio data from the captioning system. The method may further include receiving and storing contact data from each of multiple communication devices at the captioning system. The method may further include selecting the contact data from the multiple communication devices that include the identifier of the first communication device as selected contact data and generating a contact list based on the selected contact data. The method may also include sending the contact list to the first communication device to provide the contact list as contacts for presentation on an electronic display of the first communication device.
H04M 1/2757 - Devices whereby a plurality of signals may be stored simultaneously with provision for storing more than one subscriber number at a time using static electronic memories, e.g. chips providing data content by data transmission, e.g. downloading
18.
AUTOMATIC TRANSLATION BETWEEN SIGN LANGUAGE AND SPOKEN LANGUAGE
Methods, apparatus, systems, and articles of manufacture to translation between sign language and spoken language are disclosed. An example apparatus includes processor circuitry to at least one of instantiate or execute machine readable instructions to identify a plurality of candidate signs across different frames in video; associate a respective gloss to respective ones of the candidate signs; associate a respective confidence score with the respective glosses; identify overlapping frames of the candidate signs; select one or more of the candidate signs as performed signs based on the respective confidence scores and overlapping frames; and convert the performed signs to audio data.
Communication systems, methods, and non-transitory machine-readable storage media are disclosed herein. A communication system may include a communication device configured to receive a video stream including sign language content and any content indicators associated with the video stream during a real-time communication session within a single communication device or between a plurality of communication devices. The communication system may also include a translation engine configured to automatically translate the sign language content into word content during the real-time communication session without assistance of a human sign language interpreter. Further, the communication system may be configured to output the word content translation to a communication device during the real-time communication session.
A facemask system with automated voice display is disclosed. The facemask includes a covering configured to be positioned about the face of a user. At least one microphone is positioned adjacent the covering and the facemask includes a display for displaying language spoken captured by at least one microphone. A processor is in operable communication with the microphone and the display. The facemask includes a memory that contains instructions that can be executed by the processor to perform operations. The operations include receiving signals from at least one microphone, identifying language contained within the signals, and displaying the language on the display.
G09F 9/33 - Indicating arrangements for variable information in which the information is built-up on a support by selection or combination of individual elements in which the desired character or characters are formed by combining individual elements being semiconductor devices, e.g. diodes
A system is provided that includes a first network interface for a first network type and a second network interface for a second network type that is different from the first network type. The system also includes at least one processor configured to cause the system to perform operations. The operations may include obtaining, from the first network interface, audio from a communication session with a remote device established over the first network and obtaining an indication of a communication device available to participate in the communication session and direct audio obtained from the communication session to a remote transcription system. The operations may also include directing the audio to the second network interface for transmission to the communication device, obtaining transcript data from the remote transcription system based on the audio, and directing the transcript data to the second network interface for transmission to the communication device.
G10L 15/28 - Constructional details of speech recognition systems
H04M 7/00 - Arrangements for interconnection between switching centres
G10L 15/01 - Assessment or evaluation of speech recognition systems
G10L 15/18 - Speech classification or search using natural language modelling
H04M 1/247 - Telephone sets including user guidance or feature selection means facilitating their use
H04M 1/253 - Telephone sets using digital voice transmission
G10L 15/30 - Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
H04M 3/42 - Systems providing special services or facilities to subscribers
H04M 1/72412 - User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories using two-way short-range wireless interfaces
22.
Performing artificial intelligence sign language translation services in a video relay service environment
Video relay services, communication systems, non-transitory machine-readable storage media, and methods are disclosed herein. A video relay service may include at least one server configured to receive a video stream including sign language content from a video communication device during a real-time communication session. The server may also be configured to automatically translate the sign language content into a verbal language translation during the real-time communication session without assistance of a human sign language interpreter. Further, the server may be configured to transmit the verbal language translation during the real-time communication session.
A method may include obtaining first audio data of a first communication session between a first and second device and during the first communication session, obtaining a first text string that is a transcription of the first audio data and training a model of an automatic speech recognition system using the first text string and the first audio data. The method may further include in response to completion of the training, deleting the first audio data and the first text string and after deleting the first audio data and the first text string, obtaining second audio data of a second communication session between a third and fourth device and during the second communication session obtaining a second text string that is a transcription of the second audio data and further training the model of the automatic speech recognition system using the second text string and the second audio data.
A method to present communications may include captioning, by a human assistant during a call between a first user using a first captioning telephone device and a second user using a second telephone device, words spoken by the second user into the second telephone device. The method may also include presenting the captioned words on a first display of the first captioning telephone device, receiving text typed into the second telephone device by the second user, and presenting the received text on the first display of the first captioning telephone device.
A method may include obtaining a first audio signal including first speech originating at a remote device during a communication session between the remote device and a communication device and obtaining a second audio signal including second speech originating at the communication device during the communication session between the remote device and the communication device. The method may also include obtaining a characteristic of the communication session from one or more of: the first audio signal, the second audio signal, and settings of the communication device and determining a hearing level of a user of the communication device using the characteristic of the communication session.
H04M 1/72475 - User interfaces specially adapted for cordless or mobile telephones specially adapted for disabled users
H04M 3/42 - Systems providing special services or facilities to subscribers
H04M 1/247 - Telephone sets including user guidance or feature selection means facilitating their use
G10L 15/22 - Procedures used during a speech recognition process, e.g. man-machine dialog
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
A method to transcribe communications is provided. The method may include obtaining first communication data during a communication session between a first communication device and a second communication device and transmitting the first communication data to the second communication device by way of a mobile device that is locally coupled with the first communication device. The method may also include receiving, at the first communication device, second communication data from the second communication device through the mobile device and transmitting the second communication data to a remote transcription system. The method may further include receiving, at the first communication device, transcription data from the remote transcription system, the transcription data corresponding to a transcription of the second communication data, the transcription generated by the remote transcription system and presenting, by the first communication device, the transcription of the second communication data.
A method to present a summary of a transcription may include obtaining, at a first device, audio directed to the first device from a second device during a communication session between the first device and the second device. Additionally, the method may include sending, from the first device, the audio to a transcription system. The method may include obtaining, at the first device, a transcription during the communication session from the transcription system based on the audio. Additionally, the method may include obtaining, at the first device, a summary of the transcription during the communication session. Additionally, the method may include presenting, on a display, both the summary and the transcription simultaneously during the communication session.
Systems, methods, and computer-readable mediums for autofocusing cameras in a videophone are disclosed. A videophone system may include a camera having a lens, a lens actuator configured to move the lens, and an image sensor configured to capture images received through the lens. The videophone system may also include a distance sensor configured to determine an object of focus distance between the camera lens and the object of focus. The system includes memory for storing instructions that when executed by a processor cause the camera lens to automatically focus on the object of focus. The processor includes a lookup table configured to correlate an object of focus distance determined by the distance sensor with a digital number that can be converted by a digital to analog converter to a power value sufficient to power lens actuator to move the lens to a predetermined focus position. The lookup table is configured to correlate an object of focus distance returned by the distance sensor with a digital number representing a lens position, such that the object of focus falls within a depth of field about a focal plane corresponding to the lens position.
A method may include obtaining first features of first audio data that includes speech and obtaining second features of second audio data that is a revoicing of the first audio data. The method may further include providing the first features and the second features to an automatic speech recognition system and obtaining a single transcription generated by the automatic speech recognition system using the first features and the second features.
A method may include obtaining first audio data of a communication session between a first device and a second device and obtaining, during the communication session, a first text string that is a transcription of the first audio data. The method may further include directing the first text string to the first device for presentation of the first text string during the communication session and obtaining, during the communication session, a second text string that is a transcription of the first audio data. The method may further include comparing a first accuracy score of the first word to a second accuracy score of the second word and in response to a difference between the first accuracy score and the second accuracy score satisfying a threshold, directing the second word to the first device to replace the first word in the first location as displayed by the first device.
A method may include obtaining a text string that is a transcription of audio data and selecting a sequence of words from the text string as a first word sequence. The method may further include encrypting the first word sequence and comparing the encrypted first word sequence to multiple encrypted word sequences. Each of the multiple encrypted word sequences may be associated with a corresponding one of multiple counters. The method may also include in response to the encrypted first word sequence corresponding to one of the multiple encrypted word sequences based on the comparison, incrementing a counter of the multiple counters associated with the one of the multiple encrypted word sequences and adapting a language model of an automatic transcription system using the multiple encrypted word sequences and the multiple counters.
A method may include obtaining first audio data of a communication session between a first device and a second device and obtaining, during the communication session, a first text string that is a transcription of the first audio data. The method may further include directing the first text string to the first device for presentation of the first text string during the communication session and obtaining, during the communication session, a second text string that is a transcription of the first audio data. The method may further include comparing a first accuracy score of the first word to a second accuracy score of the second word and in response to a difference between the first accuracy score and the second accuracy score satisfying a threshold, directing the second word to the first device to replace the first word in the first location as displayed by the first device.
A method may include obtaining first audio data originating at a first device during a communication session between the first device and a second device. The method may also include obtaining an availability of revoiced transcription units in a transcription system and in response to establishment of the communication session, selecting, based on the availability of revoiced transcription units, a revoiced transcription unit instead of a non-revoiced transcription unit to generate a transcript of the first audio data. The method may also include obtaining revoiced audio generated by a revoicing of the first audio data by a captioning assistant and generating a transcription of the revoiced audio using an automatic speech recognition system. The method may further include in response to selecting the revoiced transcription unit, directing the transcription of the revoiced audio to the second device as the transcript of the first audio data.
A method may include obtaining audio of a communication session between a first device of a first user and a second device of a second user. The method may further include obtaining a transcription of second speech of the second user. The method may also include identifying one or more first sound characteristics of first speech of the first user. The method may also include identifying one or more first words indicating a lack of understanding in the first speech. The method may further include determining an experienced emotion of the first user based on the one or more first sound characteristics. The method may also include determining an accuracy of the transcription of the second speech based on the experienced emotion and the one or more first words.
A method to transcribe communications may include obtaining audio data originating at a first device during a communication session between the first device and a second device and providing the audio data to an automated speech recognition system configured to transcribe the audio data. The method may further include obtaining multiple hypothesis transcriptions generated by the automated speech recognition system. Each of the multiple hypothesis transcriptions may include one or more words determined by the automated speech recognition system to be a transcription of a portion of the audio data. The method may further include determining one or more consistent words that are included in two or more of the multiple hypothesis transcriptions and in response to determining the one or more consistent words, providing the one or more consistent words to the second device for presentation of the one or more consistent words by the second device.
A method to transcribe communications may include obtaining, at a first device, an audio signal that originates at a remote device during a communication session. The audio signal may be shared between the first device and a second device. The method may also include obtaining an indication that the second device is associated with a remote transcription system and in response to the second device being associated with the remote transcription system, directing the audio signal to the remote transcription system by one of the first device and the second device instead of both the first device and the second device directing the audio signal to the remote transcription system when the second device is not associated with the remote transcription system.
A method to adjust volume may include obtaining an audio signal during a communication session between a first device and a second device. The method may also include adjusting a volume level of the audio signal so that the volume level is a particular volume level above a noise floor of the audio signal and directing the adjusted audio signal to a remote transcription system.
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
The present application includes operations related to routing communications. In some embodiments, the operations may include receiving, from a third-party communication address, an incoming communication directed toward a general communication address associated with a communication system. The operations may also include routing the incoming communication to a specific communication address associated with the communication system instead of to the general communication address based on a previous outgoing communication being from the specific communication address and being directed toward the third-party communication address.
A method to transcribe communications may include selecting a first transcription generation technique from among multiple transcription generation techniques for generating transcriptions of audio of one or more communication sessions that involve a user device and obtaining performances of the multiple transcription generation techniques with respect to generating the transcriptions of the audio. The method may also include monitoring comparisons between the performances of the multiple transcription generation techniques and obtaining input from the user with respect to the comparisons. The method may further include selecting a second transcription generation technique from among the multiple transcription generation techniques based on the input from the user.
A method may include obtaining a first audio signal including first speech originating at a remote device during a communication session between the remote device and a communication device and obtaining a second audio signal including second speech originating at the communication device during the communication session between the remote device and the communication device. The method may also include obtaining a characteristic of the communication session from one or more of: the first audio signal, the second audio signal, and settings of the communication device and determining a hearing level of a user of the communication device using the characteristic of the communication session.
H04M 1/72475 - User interfaces specially adapted for cordless or mobile telephones specially adapted for disabled users
H04M 3/42 - Systems providing special services or facilities to subscribers
H04M 1/247 - Telephone sets including user guidance or feature selection means facilitating their use
G10L 15/22 - Procedures used during a speech recognition process, e.g. man-machine dialog
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
Systems, apparatuses, and methods for enabling communication between parties are disclosed. A system may include a first communication device associated with a first user and a second communication device associated with a second user. The system may include a first communication channel between the first user and the second user for transmitting call data between the first communication device and the second communication device. The system may include a second communication channel between the first user and the second user established concurrently with the first communication channel. The second communication channel may be separate from the first communication channel. The second communication channel may be established by connection of the communication session participants to a multipoint control unit that is configured to transmit media data between the first communication device and the second communication device. The system allows transmission of one or more types of call data to be switched between the first and second communication channels.
A method may include obtaining first audio data originating at a first device during a communication session between the first device and a second device. The method may also include obtaining a first text string that is a transcription of the first audio data, where the first text string may be generated using automatic speech recognition technology using the first audio data. The method may also include obtaining a second text string that is a transcription of second audio data, where the second audio data may include a revoicing of the first audio data by a captioning assistant and the second text string may be generated by the automatic speech recognition technology using the second audio data. The method may further include generating an output text string from the first text string and the second text string and using the output text string as a transcription of the speech.
A method to generate a contact list may include receiving an identifier of a first communication device at a captioning system. The first communication device may be configured to provide first audio data to a second communication device. The second communication device may be configured to receive first text data of the first audio data from the captioning system. The method may further include receiving and storing contact data from each of multiple communication devices at the captioning system. The method may further include selecting the contact data from the multiple communication devices that include the identifier of the first communication device as selected contact data and generating a contact list based on the selected contact data. The method may also include sending the contact list to the first communication device to provide the contact list as contacts for presentation on an electronic display of the first communication device.
H04M 1/2757 - Devices whereby a plurality of signals may be stored simultaneously with provision for storing more than one subscriber number at a time using static electronic memories, e.g. chips providing data content by data transmission, e.g. downloading
A method may include obtaining audio originating at a remote device during a communication session conducted between a first device and the remote device and obtaining a transcription of the audio. The method may also include processing the audio to generate processed audio. In some embodiments, the audio may be processed by a neural network that is trained with respect to an analog voice network and the processed audio may be formatted with respect to communication over the analog voice network. The method may further include processing the transcription to generate a processed transcription that is formatted with respect to communication over the analog voice network and multiplexing the processed audio with the processed transcription to obtain combined data. The method may also include communicating, to the first device during the communication session, the combined data over a same communication channel of the analog voice network.
G10L 15/06 - Creation of reference templatesTraining of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
G10L 15/22 - Procedures used during a speech recognition process, e.g. man-machine dialog
G10L 15/30 - Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
G10L 19/00 - Speech or audio signal analysis-synthesis techniques for redundancy reduction, e.g. in vocodersCoding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
A method to access a device may include obtaining, at a first device, data over a short-range wireless network from a second device. The data may originate at a remote system that sends the data to the second device through a network connection over a wide area network. The method may also include in response to a fault at the second device, obtaining, at the first device from the remote system, a maintenance command for the second device. The maintenance command may be obtained by the first device over an analog voice network. The method may also include directing, from the first device to the second device, the maintenance command over the short-range wireless network to enable the second device to perform the maintenance command.
H04L 41/0659 - Management of faults, events, alarms or notifications using network fault recovery by isolating or reconfiguring faulty entities
H04L 41/0816 - Configuration setting characterised by the conditions triggering a change of settings the condition being an adaptation, e.g. in response to network events
H04W 4/80 - Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
H04W 76/18 - Management of setup rejection or failure
H04W 24/04 - Arrangements for maintaining operational condition
A method may include obtaining audio originating at a remote device during a communication session conducted between a first device and the remote device and obtaining a transcription of the audio. The method may also include processing the audio to generate processed audio. In some embodiments, the audio may be processed by a neural network that is trained with respect to an analog voice network and the processed audio may be formatted with respect to communication over the analog voice network. The method may further include processing the transcription to generate a processed transcription that is formatted with respect to communication over the analog voice network and multiplexing the processed audio with the processed transcription to obtain combined data. The method may also include communicating, to the first device during the communication session, the combined data over a same communication channel of the analog voice network.
G10L 25/30 - Speech or voice analysis techniques not restricted to a single one of groups characterised by the analysis technique using neural networks
G10L 15/16 - Speech classification or search using artificial neural networks
A method may include obtaining first audio data originating at a first device during a communication session between the first device and a second device. The method may also include obtaining an availability of revoiced transcription units in a transcription system and in response to establishment of the communication session, selecting, based on the availability of revoiced transcription units, a revoiced transcription unit instead of a non-revoiced transcription unit to generate a transcript of the first audio data. The method may also include obtaining revoiced audio generated by a revoicing of the first audio data by a captioning assistant and generating a transcription of the revoiced audio using an automatic speech recognition system. The method may further include in response to selecting the revoiced transcription unit, directing the transcription of the revoiced audio to the second device as the transcript of the first audio data.
According to one or more aspects of the present disclosure, operations related to selecting a transcription generation technique may be disclosed. In some embodiments, the operations may include obtaining multiple user ratings that each correspond to a different one of multiple transcriptions. Each transcription may be obtained using a first transcription generation technique and may correspond to a different one of multiple communication sessions. The operations may further include selecting, for a subsequent communication session that occurs after the multiple communication sessions, a second transcription generation technique based on the user ratings. In addition, the operations may include providing the subsequent transcription to a device during the subsequent communication session.
A method may include obtaining an indicator that a first device is in a location of a second device and in response to obtaining the indicator, sending a redirect request to a communication service provider of the first device to direct, to the second device, incoming communication requests handled by the communication service provider that are directed to the first device. The method may further include after sending the redirect request and after a communication request to a communication session is directed to the first device, obtaining, at the second device, a communication indication to participate in the communication session. The method may further include directing audio of the communication session to a transcription system and obtaining, at the second device, the transcription of the audio from the transcription system. The method may also include presenting, by the second device, the audio and the transcription.
H04W 4/80 - Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
A system is provided that includes a first network interface for a first network type and a second network interface for a second network type that is different from the first network type. The system also includes at least one processor configured to cause the system to perform operations. The operations may include obtaining, from the first network interface, audio from a communication session with a remote device established over the first network and obtaining an indication of a communication device available to participate in the communication session and direct audio obtained from the communication session to a remote transcription system. The operations may also include directing the audio to the second network interface for transmission to the communication device, obtaining transcript data from the remote transcription system based on the audio, and directing the transcript data to the second network interface for transmission to the communication device.
G10L 15/28 - Constructional details of speech recognition systems
H04M 7/00 - Arrangements for interconnection between switching centres
G10L 15/01 - Assessment or evaluation of speech recognition systems
H04M 3/42 - Systems providing special services or facilities to subscribers
H04M 1/247 - Telephone sets including user guidance or feature selection means facilitating their use
H04M 1/253 - Telephone sets using digital voice transmission
H04M 1/72412 - User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories using two-way short-range wireless interfaces
G10L 15/18 - Speech classification or search using natural language modelling
G10L 15/30 - Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
51.
Transcription of communications using multiple speech recognition systems
A method may include obtaining audio data originating at a first device during a communication session between the first device and a second device and providing the audio data to a first speech recognition system to generate a first transcript based on the audio data and directing the first transcript to the second device. The method may also include in response to obtaining a quality indication regarding a quality of the first transcript, multiplexing the audio data to provide the audio data to a second speech recognition system to generate a second transcript based on the audio data while continuing to provide the audio data to the first speech recognition system and direct the first transcript to the second device, and in response to obtaining a transfer indication that occurs after multiplexing of the audio data, directing the second transcript to the second device instead of the first transcript.
A method to present a summary of a transcription may include obtaining, at a first device, audio directed to the first device from a second device during a communication session between the first device and the second device. Additionally, the method may include sending, from the first device, the audio to a transcription system. The method may include obtaining, at the first device, a transcription during the communication session from the transcription system based on the audio. Additionally, the method may include obtaining, at the first device, a summary of the transcription during the communication session. Additionally, the method may include presenting, on a display, both the summary and the transcription simultaneously during the communication session.
A method to access a device may include obtaining, at a first device, data over a short-range wireless network from a second device. The data may originate at a remote system that sends the data to the second device through a network connection over a wide area network. The method may also include in response to a fault at the second device, obtaining, at the first device from the remote system, a maintenance command for the second device. The maintenance command may be obtained by the first device over an analog voice network. The method may also include directing, from the first device to the second device, the maintenance command over the short-range wireless network to enable the second device to perform the maintenance command.
H04W 4/80 - Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
H04W 76/18 - Management of setup rejection or failure
H04W 24/04 - Arrangements for maintaining operational condition
55.
Apparatus, method and computer-readable medium for automatically adjusting the brightness of a videophone visual indicator
A videophone apparatus and a method for automatically adjusting a videophone visual indicator are provided. The videophone includes a camera unit with an imager. Data from the imager and imager data processed by an image signal processor may be converted into a value that can be compared to a scale correlating to a brightness condition of an area in which the imager is located, allowing the brightness condition to be determined. The visual indicators of the videophone can be set to illuminate based upon the determined brightness condition. Thus, the brightness condition of a room can be determined using components already found on the videophone, without the need for additional sensors, hardware or software that increase the cost and complexity of the videophone.
H04N 5/58 - Control of contrast or brightness in dependence upon ambient light
H04M 1/57 - Arrangements for indicating or recording the number of the calling subscriber at the called subscriber's set
G08B 5/36 - Visible signalling systems, e.g. personal calling systems, remote indication of seats occupied using electric transmissionVisible signalling systems, e.g. personal calling systems, remote indication of seats occupied using electromagnetic transmission using visible light sources
A method may include obtaining audio of a communication session between a first device of a first user and a second device of a second user. The method may further include obtaining a transcription of second speech of the second user. The method may also include identifying one or more first sound characteristics of first speech of the first user. The method may also include identifying one or more first words indicating a lack of understanding in the first speech. The method may further include determining an experienced emotion of the first user based on the one or more first sound characteristics. The method may also include determining an accuracy of the transcription of the second speech based on the experienced emotion and the one or more first words.
A method may include obtaining, at a device, a first input that results in audio playback of a first stored message and establishing a network connection between the device and a remote transcription system. The method may include directing audio of the first stored message to the remote transcription system and obtaining transcriptions of the audio from the remote transcription system during playback of the first stored message. The method may include maintaining the network connection after playback of the first stored message and obtaining a second input that results in audio playback of a second stored message. The method may include directing audio of the second stored message to the remote transcription system and obtaining transcriptions of the audio from the remote transcription system during playback of the second stored message. The method may include obtaining an indication to terminate the network connection and terminating the network connection.
A method to transcribe communications is provided. The method may include obtaining first communication data during a communication session between a first communication device and a second communication device and transmitting the first communication data to the second communication device by way of a mobile device that is locally coupled with the first communication device. The method may also include receiving, at the first communication device, second communication data from the second communication device through the mobile device and transmitting the second communication data to a remote transcription system. The method may further include receiving, at the first communication device, transcription data from the remote transcription system, the transcription data corresponding to a transcription of the second communication data, the transcription generated by the remote transcription system and presenting, by the first communication device, the transcription of the second communication data.
A method to present communications is provided. The method may include obtaining, at a device, a request from a user to play back a stored message that includes audio. In response to obtaining the request, the method may include directing the audio of the message to a transcription system from the device. In these and other embodiments, the transcription system may be configured to generate text that is a transcription of the audio in real-time. The method may further include obtaining, at the device, the text from the transcription system and presenting, by the device, the text generated by the transcription system in real-time. In response to obtaining the text from the transcription system, the method may also include presenting, by the device, the audio such that the text as presented is substantially aligned with the audio.
A computer-implemented method to provide transcriptions of an audio communication session is disclosed. The method may include receiving audio data at a first device from a second device over a first wireless network connection of a first wireless network type and providing the audio data to a speaker of the first device for output of the audio data. The method may further include encoding the audio data at the first device based on a second wireless network type and transmitting the encoded audio data from the first device over a second wireless network of the second wireless network type.
H04M 3/22 - Arrangements for supervision, monitoring or testing
H04W 4/80 - Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
A method may include obtaining first audio data originating at a first device during a communication session between the first device and a second device. The method may also include obtaining an availability of revoiced transcription units in a transcription system and in response to establishment of the communication session, selecting, based on the availability of revoiced transcription units, a revoiced transcription unit instead of a non-revoiced transcription unit to generate a transcript of the first audio data. The method may also include obtaining revoiced audio generated by a revoicing of the first audio data by a captioning assistant and generating a transcription of the revoiced audio using an automatic speech recognition system. The method may further include in response to selecting the revoiced transcription unit, directing the transcription of the revoiced audio to the second device as the transcript of the first audio data.
A method may include obtaining first audio data originating at a first device during a communication session between the first device and a second device. The method may also include obtaining an availability of revoiced transcription units in a transcription system and in response to establishment of the communication session, selecting, based on the availability of revoiced transcription units, a revoiced transcription unit instead of a non-revoiced transcription unit to generate a transcript of the first audio data. The method may also include obtaining revoiced audio generated by a revoicing of the first audio data by a captioning assistant and generating a transcription of the revoiced audio using an automatic speech recognition system. The method may further include in response to selecting the re voiced transcription unit, directing the transcription of the revoiced audio to the second device as the transcript of the first audio data.
A method may include obtaining first audio data originating at a first device during a communication session between the first device and a second device. The method may also include obtaining a first text string that is a transcription of the first audio data, where the first text string may be generated using automatic speech recognition technology using the first audio data. The method may also include obtaining a second text string that is a transcription of second audio data, where the second audio data may include a revoicing of the first audio data by a captioning assistant and the second text string may be generated by the automatic speech recognition technology using the second audio data. The method may further include generating an output text string from the first text string and the second text string and using the output text string as a transcription of the speech.
A method may include obtaining first audio data of a first communication session between a first and second device and during the first communication session, obtaining a first text string that is a transcription of the first audio data and training a model of an automatic speech recognition system using the first text string and the first audio data. The method may further include in response to completion of the training, deleting the first audio data and the first text string and after deleting the first audio data and the first text string, obtaining second audio data of a second communication session between a third and fourth device and during the second communication session obtaining a second text string that is a transcription of the second audio data and further training the model of the automatic speech recognition system using the second text string and the second audio data.
A method may include obtaining first audio data of a communication session between a first device and a second device, obtaining a text string that is a transcription of the first audio data, and selecting a contiguous sequence of words from the text string as a first word sequence. The method may further include comparing the first word sequence to multiple word sequences obtained before the communication session and in response to the first word sequence corresponding to one of the multiple word sequences, incrementing a counter of multiple counters associated with the one of the multiple word sequences. The method may also include deleting the text string and the first word sequence and training and after deleting the text string and the first word sequence, training a language model of an automatic transcription system using the multiple word sequences and the multiple counters. The method is for estimating n-gram statistics based on the communication and provides data protection mechanisms for private data.
A method may include obtaining first audio data originating at a first device during a communication session between the first device and a second device. The method may also include obtaining a first text string that is a transcription of the first audio data, where the first text string may be generated using automatic speech recognition technology using the first audio data. The method may also include obtaining a second text string that is a transcription of second audio data, where the second audio data may include a revoicing of the first audio data by a captioning assistant and the second text string may be generated by the automatic speech recognition technology using the second audio data. The method may further include generating an output text string from the first text string and the second text string and using the output text string as a transcription of the speech.
A method may include obtaining first audio data of a first communication session between a first and second device and during the first communication session, obtaining a first text string that is a transcription of the first audio data and training a model of an automatic speech recognition system using the first text string and the first audio data. The method may further include in response to completion of the training, deleting the first audio data and the first text string and after deleting the first audio data and the first text string, obtaining second audio data of a second communication session between a third and fourth device and during the second communication session obtaining a second text string that is a transcription of the second audio data and further training the model of the automatic speech recognition system using the second text string and the second audio data.
A method may include obtaining first audio data of a communication session between a first device and a second device, obtaining a text string that is a transcription of the first audio data, and selecting a contiguous sequence of words from the text string as a first word sequence. The method may further include comparing the first word sequence to multiple word sequences obtained before the communication session and in response to the first word sequence corresponding to one of the multiple word sequences, incrementing a counter of multiple counters associated with the one of the multiple word sequences. The method may also include deleting the text string and the first word sequence and training and after deleting the text string and the first word sequence, training a language model of an automatic transcription system using the multiple word sequences and the multiple counters.
A method to transcribe communications may include obtaining audio data originating at a first device during a communication session between the first device and a second device and providing the audio data to an automated speech recognition system configured to transcribe the audio data. The method may further include obtaining multiple hypothesis transcriptions generated by the automated speech recognition system. Each of the multiple hypothesis transcriptions may include one or more words determined by the automated speech recognition system to be a transcription of a portion of the audio data. The method may further include determining one or more consistent words that are included in two or more of the multiple hypothesis transcriptions and in response to determining the one or more consistent words, providing the one or more consistent words to the second device for presentation of the one or more consistent words by the second device.
An electronic device may include a housing and a directional microphone. The housing may include a front surface, a rear surface positioned opposite the front surface, and a hollow cavity positioned between the front surface and the rear surface. The hollow cavity may include a front opening defined in the front surface and a rear opening defined in the rear surface. The directional microphone may include a front port and a rear port. The directional microphone may be mounted in the hollow cavity of the housing with the front port oriented toward the front opening and with the rear port oriented toward the rear opening.
H04M 1/19 - Arrangements of transmitters, receivers, or complete sets to prevent eavesdropping, to attenuate local noise or to prevent undesired transmissionMouthpieces or receivers specially adapted therefor
H04M 1/03 - Constructional features of telephone transmitters or receivers, e.g. telephone hand-sets
H04M 1/20 - Arrangements for preventing acoustic feedback
H04R 1/34 - Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by using a single transducer with sound reflecting, diffracting, directing or guiding means
H04R 1/28 - Transducer mountings or enclosures designed for specific frequency responseTransducer enclosures modified by provision of mechanical or acoustic impedances, e.g. resonator, damping means
Operations related to performing gain operations with respect to a receive-path signal of a first device may be performed. The operations may include obtaining the receive-path signal, which includes an echo speech signal and a receive speech signal originating at a second device. In addition, the operations may include identifying a portion of the receive-path signal that includes, at a particular time, a first frequency component that corresponds to the echo speech signal and a second frequency component that corresponds to the receive speech signal in which the first frequency component is different from the second frequency component. Moreover, the operations may include attenuating the first frequency component of the portion while avoiding attenuating the second frequency component of the portion based on the first frequency component corresponding to the echo speech signal and the second frequency component corresponding to the receive speech signal.
H04M 9/08 - Two-way loud-speaking telephone systems with means for conditioning the signal, e.g. for suppressing echoes for one or both directions of traffic
H04M 9/10 - Two-way loud-speaking telephone systems with means for conditioning the signal, e.g. for suppressing echoes for one or both directions of traffic with switching of direction of transmission by voice frequency
According to one or more aspects of the present disclosure, operations related to providing transcriptions may include obtaining a first transcription of first audio obtained by a first device during a communication session conducted between the first device and a second device. The operations may further include providing the first transcription for presentation of the first transcription by a display device during the communication session. In addition, the operations may include providing, in response to a transcription quality indication, a second transcription of second audio obtained by the second device during the communication session for presentation of the second transcription by the display device during the communication session.
G06F 16/683 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
G06F 16/335 - Filtering based on additional data, e.g. user or group profiles
According to one or more aspects of the present disclosure, operations related to providing transcriptions may include obtaining a first transcription of first audio obtained by a first device during a communication session conducted between the first device and a second device. The operations may further include providing the first transcription for presentation of the first transcription by a display device during the communication session. In addition, the operations may include providing, in response to a transcription quality indication, a second transcription of second audio obtained by the second device during the communication session for presentation of the second transcription by the display device during the communication session.
A method may include obtaining first audio data originating at a first device during a communication session between the first device and a second device. The method may also include obtaining a first text string that is a transcription of the first audio data, where the first text string may be generated using automatic speech recognition technology using the first audio data. The method may also include obtaining a second text string that is a transcription of second audio data, where the second audio data may include a revoicing of the first audio data by a captioning assistant and the second text string may be generated by the automatic speech recognition technology using the second audio data. The method may further include generating an output text string from the first text string and the second text string and using the output text string as a transcription of the speech.
A system is provided that includes a first network interface for a first network type and a second network interface for a second network type that is different from the first network type. The system also includes at least one processor configured to cause the system to perform operations. The operations may include obtaining, from the first network interface, audio from a communication session with a remote device established over the first network and obtaining an indication of a communication device available to participate in the communication session and direct audio obtained from the communication session to a remote transcription system. The operations may also include directing the audio to the second network interface for transmission to the communication device, obtaining transcript data from the remote transcription system based on the audio, and directing the transcript data to the second network interface for transmission to the communication device.
The present application includes operations related to routing communications. In some embodiments, the operations may include receiving, from a third-party communication address, an incoming communication directed toward a general communication address associated with a communication system. The operations may also include routing the incoming communication to a specific communication address associated with the communication system instead of to the general communication address based on a previous outgoing communication being from the specific communication address and being directed toward the third-party communication address.
The present application includes operations related to routing communications. In some embodiments, the operations may include receiving, from a third-party communication address, an incoming communication directed toward a general communication address associated with a communication system. The operations may also include routing the incoming communication to a specific communication address associated with the communication system instead of to the general communication address based on a previous outgoing communication being from the specific communication address and being directed toward the third-party communication address.
According to one or more aspects of the present disclosure, operations may include conducting, by a host device, first communication with a first device over a first network. The operations may further include conducting, by the host device based on the first communication, second communication with the first device over a second network when the first device is disconnected from the first network. The operations may also include obtaining, by the host device, a request for a communication session from a second device while the host device is conducting the second communication with the first device over the second network. In addition, the operations may include sending, by the host device, a request notification to the first device over the second network based on the host device conducting the second communication with the first device. The request notification may notify the first device of the request.
A method to transcribe communications may include obtaining audio data originating at a first device during a communication session between the first device and a second device and providing the audio data to an automated speech recognition system configured to transcribe the audio data. The method may further include obtaining multiple hypothesis transcriptions generated by the automated speech recognition system. Each of the multiple hypothesis transcriptions may include one or more words determined by the automated speech recognition system to be a transcription of a portion of the audio data. The method may further include determining one or more consistent words that are included in two or more of the multiple hypothesis transcriptions and in response to determining the one or more consistent words, providing the one or more consistent words to the second device for presentation of the one or more consistent words by the second device.
Operations related to establishing a first communication session between a first first-end device and a first second-end device. The first first-end device may obtain first device audio during the first communication session. The operations also include establishing a second communication session between a second first-end device and a second second-end device. In addition, the operations may include communicating the first device audio to the second second-end device as second session audio of the second communication session.
G06F 15/16 - Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
H04L 65/401 - Support for services or applications wherein the services involve a main real-time session and one or more additional parallel real-time or time sensitive sessions, e.g. white board sharing or spawning of a subconference
H04L 65/1059 - End-user terminal functionalities specially adapted for real-time communication
H04L 65/1069 - Session establishment or de-establishment
According to one or more aspects of the present disclosure, operations related to selecting a transcription generation technique may be disclosed. In some embodiments, the operations may include obtaining multiple user ratings that each correspond to a different one of multiple transcriptions. Each transcription may be obtained using a first transcription generation technique and may correspond to a different one of multiple communication sessions. The operations may further include selecting, for a subsequent communication session that occurs after the multiple communication sessions, a second transcription generation technique based on the user ratings. In addition, the operations may include providing the subsequent transcription to a device during the subsequent communication session.
G10L 19/00 - Speech or audio signal analysis-synthesis techniques for redundancy reduction, e.g. in vocodersCoding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
G10L 15/01 - Assessment or evaluation of speech recognition systems
G10L 15/32 - Multiple recognisers used in sequence or in parallelScore combination systems therefor, e.g. voting systems
H04M 3/42 - Systems providing special services or facilities to subscribers
According to one or more aspects of the present disclosure, operations related to selecting a transcription generation technique may be disclosed. In some embodiments, the operations may include obtaining multiple user ratings that each correspond to a different one of multiple transcriptions. Each transcription may be obtained using a first transcription generation technique and may correspond to a different one of multiple communication sessions. The operations may further include selecting, for a subsequent communication session that occurs after the multiple communication sessions, a second transcription generation technique based on the user ratings. In addition, the operations may include providing the subsequent transcription to a device during the subsequent communication session.
According to one or more embodiments of the present disclosure, operations regarding providing indications with respect to questions of a communication session may include obtaining transcript data. The transcript data may include a transcription of audio of the communication session. The operations may further include obtaining multiple questions determined to be asked during the communication session. The operations may further include determining, based on an analysis of the transcript data during the communication session, which of the questions have been asked during the communication session. In addition, the operations may include causing, during the communication session, the device involved in the communication session to indicate which of the questions have been asked.
G10L 15/22 - Procedures used during a speech recognition process, e.g. man-machine dialog
G06F 3/0484 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
G16H 80/00 - ICT specially adapted for facilitating communication between medical practitioners or patients, e.g. for collaborative diagnosis, therapy or health monitoring
H04M 3/42 - Systems providing special services or facilities to subscribers
G10L 15/18 - Speech classification or search using natural language modelling
A system may include a camera configured to obtain an image of a user, at least one processor, and at least one non-transitory computer-readable media communicatively coupled to the at least one processor. The non-transitory computer-readable media configured to store one or more instructions that when executed cause or direct the system to perform operations. The operations may include establish a communication session between the system and a device. The communication session may be configured such that the device provides audio for the system. The operations may further include compare the image to a particular user image associated with the system and select a first method of transcription generation from among two or more methods of transcription generation based on the comparison of the image to the particular user image. The operations may also include present, a transcription of the audio generated using the selected first method of transcription generation.
Communication systems and methods are disclosed for enabling a first user at a video endpoint to communicate with a far-end user at a communication device via a relay service providing translation services for the first user. The video endpoint may include a camera and may be configured to frame a view of the camera to include a signing area of a user. The video endpoint may be configured to determine the signing area of the user by taking measurements of the user's body and framing a region around the user to include the signing area based on the measurements, by monitoring a range of motion for the signing area of the user, and other methods.
A computer-implemented method to provide transcriptions of an audio communication session is disclosed. The method may include receiving audio data at a first device from a second device over a first wireless network connection of a first wireless network type and providing the audio data to a speaker of the first device for output of the audio data. The method may further include encoding the audio data at the first device based on a second wireless network type and transmitting the encoded audio data from the first device over a second wireless network of the second wireless network type.
H04M 3/22 - Arrangements for supervision, monitoring or testing
H04W 4/80 - Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
A method may include obtaining audio data originating at a first device during a communication session between the first device and a second device and providing the audio data to a first speech recognition system to generate a first transcript based on the audio data and directing the first transcript to the second device. The method may also include in response to obtaining a quality indication regarding a quality of the first transcript, multiplexing the audio data to provide the audio data to a second speech recognition system to generate a second transcript based on the audio data while continuing to provide the audio data to the first speech recognition system and direct the first transcript to the second device, and in response to obtaining a transfer indication that occurs after multiplexing of the audio data, directing the second transcript to the second device instead of the first transcript.
A method may include obtaining first audio data of a communication session between a first device and a second device, obtaining a text string that is a transcription of the first audio data, and selecting a contiguous sequence of words from the text string as a first word sequence. The method may further include comparing the first word sequence to multiple word sequences obtained before the communication session and in response to the first word sequence corresponding to one of the multiple word sequences, incrementing a counter of multiple counters associated with the one of the multiple word sequences. The method may also include deleting the text string and the first word sequence and training and after deleting the text string and the first word sequence, training a language model of an automatic transcription system using the multiple word sequences and the multiple counters.
A method to transcribe communications may include obtaining, during a communication session, audio that includes a voice of a user participating in the communication session. The communication session may be configured for verbal communication. The method may further include establishing a network connection with a transcription system and sending the audio to the transcription system. In some embodiments, the transcription system may be configured to generate a transcript of the audio. The method may also include obtaining the transcript of the audio from the transcription system during the communication session and monitoring the audio to determine when the voice is inactive. In some embodiments, in response to the voice being inactive, the method may include stopping the sending of the audio to the transcription system while maintaining the communication session.
A computer-implemented method to select a user profile is disclosed. The method may include obtaining, from a first device, a communication attempt at a user device. The method may further include obtaining first device data associated with the communication attempt and with the first device. The method may also include comparing the first device data with multiple data elements of each of multiple user profiles associated with the user device. The method may further include selecting a user profile of the multiple user profiles that includes the data element that matches the first device data in response to the first device data matching a data element of the multiple data elements. The method may also include loading an audio profile for the selected user profile on the user device.
According to one or more aspects of the present disclosure, operations may include obtaining multiple microphone signals derived from a microphone array that includes a multiple omnidirectional microphones. Each of the microphone signals may be derived from a different microphone of the microphone array. The operations may further include determining whether the microphone signals include noise, such as wind noise, based on two or more of the plurality of microphone signals. In addition, the operations may include generating an output signal based on a beamformed signal or a reduced-noise signal based on whether the microphone signals are determined to include noise.
A method to transcribe communications may include obtaining, during a communication session, audio that includes a voice of a user participating in the communication session. The communication session may be configured for verbal communication. The method may further include establishing a network connection with a transcription system and sending the audio to the transcription system. In some embodiments, the transcription system may be configured to generate a transcript of the audio. The method may also include obtaining the transcript of the audio from the transcription system during the communication session and monitoring the audio to determine when the voice is inactive. In some embodiments, in response to the voice being inactive, the method may include stopping the sending of the audio to the transcription system while maintaining the communication session.
According to one or more embodiments of the present disclosure, operations regarding providing indications with respect to questions of a communication session may include obtaining transcript data. The transcript data may include a transcription of audio of the communication session. The operations may further include obtaining multiple questions determined to be asked during the communication session. The operations may further include determining, based on an analysis of the transcript data during the communication session, which of the questions have been asked during the communication session. In addition, the operations may include causing, during the communication session, the device involved in the communication session to indicate which of the questions have been asked.
G10L 15/22 - Procedures used during a speech recognition process, e.g. man-machine dialog
G06F 3/0484 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
G16H 80/00 - ICT specially adapted for facilitating communication between medical practitioners or patients, e.g. for collaborative diagnosis, therapy or health monitoring
H04M 3/42 - Systems providing special services or facilities to subscribers
G10L 15/18 - Speech classification or search using natural language modelling
G06F 17/27 - Automatic analysis, e.g. parsing, orthograph correction
A computer-implemented method to provide transcriptions of an audio communication session is disclosed. The method may include receiving audio data at a first device from a second device over a first wireless network connection of a first wireless network type and providing the audio data to a speaker of the first device for output of the audio data. The method may further include encoding the audio data at the first device based on a second wireless network type and transmitting the encoded audio data from the first device over a second wireless network of the second wireless network type.
H04M 3/22 - Arrangements for supervision, monitoring or testing
H04W 4/80 - Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
A method to transcribe communications is provided. The method may include obtaining first communication data during a communication session between a first communication device and a second communication device and transmitting the first communication data to the second communication device by way of a mobile device that is locally coupled with the first communication device. The method may also include receiving, at the first communication device, second communication data from the second communication device through the mobile device and transmitting the second communication data to a remote transcription system. The method may further include receiving, at the first communication device, transcription data from the remote transcription system, the transcription data corresponding to a transcription of the second communication data, the transcription generated by the remote transcription system and presenting, by the first communication device, the transcription of the second communication data.
A method to transcribe communications may include obtaining audio data originating at a first device during a communication session between the first device and a second device and providing the audio data to an automated speech recognition system configured to transcribe the audio data. The method may further include obtaining multiple hypothesis transcriptions generated by the automated speech recognition system. Each of the multiple hypothesis transcriptions may include one or more words determined by the automated speech recognition system to be a transcription of a portion of the audio data. The method may further include determining one or more consistent words that are included in two or more of the multiple hypothesis transcriptions and in response to determining the one or more consistent words, providing the one or more consistent words to the second device for presentation of the one or more consistent words by the second device.
Apparatuses and methods are disclosed for automatically connecting to a relay service and establish a captioning communication session during a call between a hearing-impaired user and a far-end user. The apparatus includes communication elements configured to receive captions from a relay service, and a processor configured to automatically connect to the relay service and establish a captioning communication session during a call responsive to a determination that a connection to the relay service failed during a time in which the hearing-impaired user desired captions to be enabled. A method includes determining captions are desired by the hearing-impaired user at a first time to be enabled for display to the communication device during a call, detecting a failed connection to the relay service at a second time, and automatically connecting with the relay service and establishing a new captioning communication session at a third time during the call.
H04M 11/10 - Telephonic communication systems specially adapted for combination with other electrical systems with dictation recording and playback systems
H04M 1/247 - Telephone sets including user guidance or feature selection means facilitating their use
A communication console may include a moveable plunger including a body and an arm. The arm may connect to and extend away from the body. A first element may be coupled to the arm. The first element may include a top point and a bottom point, the axis of displacement extending between the top point and the bottom point. The first element may move with the arm along the axis of displacement. A second element may be detached from the arm and positioned adjacent to the axis of displacement. Additionally, one of the first element or the second element may be a sensing element and another may be a detectable element. The sensing element may sense a spatial position of the detectable element and output an indication of a position of the plunger based on the spatial position of the detectable element.
H04M 1/08 - HooksCradles associated with switches operated by the weight of the receiver or hand-set
G01D 5/14 - Mechanical means for transferring the output of a sensing memberMeans for converting the output of a sensing member to another variable where the form or nature of the sensing member does not constrain the means for convertingTransducers not specially adapted for a specific variable using electric or magnetic means influencing the magnitude of a current or voltage
H01H 36/00 - Switches actuated by change of magnetic field or of electric field, e.g. by change of relative position of magnet and switch, by shielding