FIELD: physics.
SUBSTANCE: present invention relates to computer-implemented methods and systems for assessing emotions and, more specifically, to methods for recognizing speech emotions using 3D convolutional neural network. Method of recognizing speech emotions using three-dimensional (3D) convolutional neural network, implemented using a computer and comprising two main steps, at the first stage three-dimensional tensors are provided using the reconstructed phase space of the input speech signals, and at the second step, using 3D convolutional neural network, trained based on the three-dimensional tensors and the corresponding emotion labels, performing the prediction of the emotions contained in the input speech signals, wherein the reconstructed phase space is used as follows: a one-dimensional signal is displayed in a three-dimensional space, and then the three-dimensional tensor is extracted for use as input data of 3D convolutional neural network.
EFFECT: high efficiency of recognizing speech emotions using 3D convolutional neural network.
4 cl, 6 dwg
Title | Year | Author | Number |
---|---|---|---|
METHOD OF PROCESSING A TWO-DIMENSIONAL IMAGE AND A USER COMPUTING DEVICE THEREOF | 2018 |
|
RU2703327C1 |
METHOD AND ELECTRONIC DEVICE FOR DETECTING THREE-DIMENSIONAL OBJECTS USING NEURAL NETWORKS | 2021 |
|
RU2776814C1 |
METHOD FOR AUTOMATIC CLASSIFICATION OF X-RAY IMAGES USING TRANSPARENCY MASKS | 2019 |
|
RU2716914C1 |
ARTIFICIAL NEURAL NETWORKS COMPRESSION SYSTEM BASED ON ITERATIVE APPLICATION OF TENSOR APPROXIMATIONS | 2019 |
|
RU2734579C1 |
VISUALIZATION OF RECONSTRUCTION OF 3D SCENE USING SEMANTIC REGULARIZATION OF NORMALS TSDF WHEN TRAINING NEURAL NETWORK | 2023 |
|
RU2825722C1 |
IMAGE RECOGNITION SYSTEM: BEORG SMART VISION | 2020 |
|
RU2777354C2 |
METHOD AND SERVER FOR CONVERTING TEXT TO SPEECH | 2020 |
|
RU2775821C2 |
METHOD FOR PROVIDING COMPUTER VISION | 2022 |
|
RU2791587C1 |
IMAGE SEGMENTATION METHOD USING NEURAL NETWORK | 2017 |
|
RU2720440C1 |
REPRODUCING AUGMENTATION OF IMAGE DATA | 2018 |
|
RU2716322C2 |
Authors
Dates
2024-04-03—Published
2023-03-31—Filed