FIELD: information technology.
SUBSTANCE: invention discloses means for adaptive shaping of neuron pattern for multichannel speech recognition. A first channel of audio data corresponding to a speech fragment and a second audio data channel corresponding to said speech fragment are received. A first set of filter parameters for a first filter based on a first audio data channel and a second audio data channel and a second set of filter parameters for a second filter based on a first audio data channel and a second audio data channel are generated using a trained recurrent neural network. Generating a single combined audio data channel by combining first channel audio data which has been filtered using first filter, and audio data of second channel, which was filtered using second filter. Audio data are introduced for a single combined channel into a neural network trained as an acoustic model.
EFFECT: high efficiency of speech recognition.
20 cl, 5 dwg
Title | Year | Author | Number |
---|---|---|---|
TECHNOLOGY FOR ANALYZING ACOUSTIC DATA FOR SIGNS OF COVID-19 DISEASE | 2021 |
|
RU2758649C1 |
METHOD FOR IMPROVING A SPEECH SIGNAL WITH A LOW DELAY, A COMPUTING DEVICE AND A COMPUTER-READABLE MEDIUM THAT IMPLEMENTS THE ABOVE METHOD | 2023 |
|
RU2802279C1 |
CUSTOMIZED OUTPUT WHICH IS OPTIMIZED FOR USER PREFERENCES IN DISTRIBUTED SYSTEM | 2020 |
|
RU2821283C2 |
METHOD FOR AUDIOVISUAL RECOGNITION OF PERSONAL PROTECTION EQUIPMENT ON HUMAN FACE | 2022 |
|
RU2791415C1 |
METHOD OF MULTIMODAL CONTACTLESS CONTROL OF MOBILE INFORMATION ROBOT | 2020 |
|
RU2737231C1 |
DEVICE, METHOD, OR COMPUTER PROGRAM FOR GENERATING AN EXTENDED-BAND AUDIO SIGNAL USING A NEURAL NETWORK PROCESSOR | 2018 |
|
RU2745298C1 |
SPEAKER VERIFICATION | 2017 |
|
RU2697736C1 |
TRAINING NEURAL NETWORKS USING LOSS FUNCTIONS REFLECTING RELATIONSHIPS BETWEEN NEIGHBOURING TOKENS | 2018 |
|
RU2721190C1 |
AUDIO DATA GENERATOR AND METHODS OF GENERATING AUDIO SIGNAL AND TRAINING AUDIO DATA GENERATOR | 2021 |
|
RU2823015C1 |
RECOGNIZING OF MIXED SPEECH | 2015 |
|
RU2686589C2 |
Authors
Dates
2019-08-22—Published
2016-12-28—Filed