<?xml version='1.0' encoding='utf-8'?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.2 20190208//EN" "http://jats.nlm.nih.gov/publishing/1.2/JATS-journalpublishing1.dtd">
<article article-type="research-article" dtd-version="1.2" xml:lang="ru" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink"><front><journal-meta><journal-id journal-id-type="issn">2518-1092</journal-id><journal-title-group><journal-title>Research result. Information technologies</journal-title></journal-title-group><issn pub-type="epub">2518-1092</issn></journal-meta><article-meta><article-id pub-id-type="doi">10.18413/2518-1092-2022-8-3-0-5</article-id><article-id pub-id-type="publisher-id">3225</article-id><article-categories><subj-group subj-group-type="heading"><subject>ARTIFICIAL INTELLIGENCE AND DECISION MAKING</subject></subj-group></article-categories><title-group><article-title>&lt;strong&gt;CLASSIFICATION OF SPEECH DATA BY EMOTIONAL BACKGROUND&lt;/strong&gt;</article-title><trans-title-group xml:lang="en"><trans-title>&lt;strong&gt;CLASSIFICATION OF SPEECH DATA BY EMOTIONAL BACKGROUND&lt;/strong&gt;</trans-title></trans-title-group></title-group><contrib-group><contrib contrib-type="author"><name-alternatives><name xml:lang="ru"><surname>Zhikharev</surname><given-names>Alexander Gennadievich</given-names></name><name xml:lang="en"><surname>Zhikharev</surname><given-names>Alexander Gennadievich</given-names></name></name-alternatives><email>zhikharev@bsu.edu.ru</email></contrib><contrib contrib-type="author"><name-alternatives><name xml:lang="ru"><surname>Chernykh</surname><given-names>Vladimir Sergeevich</given-names></name><name xml:lang="en"><surname>Chernykh</surname><given-names>Vladimir Sergeevich</given-names></name></name-alternatives></contrib></contrib-group><pub-date pub-type="epub"><year>2023</year></pub-date><volume>8</volume><issue>3</issue><fpage>0</fpage><lpage>0</lpage><self-uri content-type="pdf" xlink:href="/media/information/2023/3/ИТ_НР_8.3_5_s4Kl0bc.pdf" /><abstract xml:lang="ru"><p>In this paper, the algorithm of classification of speech data by emotional background, developed by the authors, is considered. In particular, it describes a neural network created to recognize eight different emotions in speech. To train the neural network, a training sample obtained from the RAVDESS dataset, which contains 1440 audio files, was used. These audio files contain the speech of 24 actors (12 women and 12 men) with a neutral North American accent.

The paper describes the process of training a neural network using the Keras library, including the network architecture, layer sizes, activation functions and optimization methods. The stages of preprocessing and preparation of the original audio data before training the network are also discussed.

The results of the study show that the developed neural network has high performance and the ability to recognize emotions with an accuracy of 80%.</p></abstract><trans-abstract xml:lang="en"><p>In this paper, the algorithm of classification of speech data by emotional background, developed by the authors, is considered. In particular, it describes a neural network created to recognize eight different emotions in speech. To train the neural network, a training sample obtained from the RAVDESS dataset, which contains 1440 audio files, was used. These audio files contain the speech of 24 actors (12 women and 12 men) with a neutral North American accent.

The paper describes the process of training a neural network using the Keras library, including the network architecture, layer sizes, activation functions and optimization methods. The stages of preprocessing and preparation of the original audio data before training the network are also discussed.

The results of the study show that the developed neural network has high performance and the ability to recognize emotions with an accuracy of 80%.</p></trans-abstract><kwd-group xml:lang="ru"><kwd>audio attributes</kwd><kwd>audio</kwd><kwd>audio file</kwd><kwd>audio data</kwd><kwd>emotional background</kwd><kwd>classification</kwd><kwd>model</kwd><kwd>layer</kwd></kwd-group><kwd-group xml:lang="en"><kwd>audio attributes</kwd><kwd>audio</kwd><kwd>audio file</kwd><kwd>audio data</kwd><kwd>emotional background</kwd><kwd>classification</kwd><kwd>model</kwd><kwd>layer</kwd></kwd-group></article-meta></front><back><ref-list><title>Список литературы</title><ref id="B1"><mixed-citation>Chollet F. Deep Learning with Python. 2nd interd. edition. &amp;ndash; SPb.: St. Petersburg: St. Petersburg. &amp;ndash; 576&amp;nbsp;p.&amp;nbsp;&amp;ndash; ISBN 978-5-4461-1909-7.</mixed-citation></ref><ref id="B2"><mixed-citation>Han K., Lee K., Kim H.G. Music emotion recognition using chroma feature-based probabilistic neural network. Multimedia Tools and Applications. &amp;ndash; 2017. &amp;ndash; V. №76, Issue №3. &amp;ndash; P. 3691-3710.</mixed-citation></ref><ref id="B3"><mixed-citation>Getting to Know the Mel-Spectrogram. [Electronic resource] &amp;ndash; Electronic data, 2019. &amp;ndash; URL: https://towardsdatascience.com/getting-to-know-the-mel-spectrogram-31bca3e2d9d0.</mixed-citation></ref><ref id="B4"><mixed-citation>Graves A., Schmidhuber J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Nature. &amp;ndash; 2005. &amp;ndash; V. №18, Issue №5-6. &amp;ndash; P. 602-610.</mixed-citation></ref><ref id="B5"><mixed-citation>Understanding LSTM Networks. [Electronic resource] &amp;ndash; Electronic data, 2015. &amp;ndash; URL: http://colah.github.io/posts/2015-08-Understanding-LSTMs/.</mixed-citation></ref></ref-list></back></article>