Show simple item record

AdvisorDresvyanskiy, Denisdc.contributor.advisor
AuthorSawin, Michaeldc.contributor.author
Date of accession2023-06-06T09:45:13Zdc.date.accessioned
Available in OPARU since2023-06-06T09:45:13Zdc.date.available
Year of creation2023dc.date.created
Date of first publication2023-06-06dc.date.issued
AbstractThis work examines acoustic emotion recognition by a system using only an audio file containing a person’s speech. This can be applied in several areas of human-computer-interaction (HCI). In this context, a system consisting of only a microphone and a processor is less expensive than other methods that use additional variants for emotion recognition. This thesis explores how automatic emotion recognition by a system using only speech as input works. This work provides insight into different technologies that can be applied for emotion recognition of a system. Feature extraction and computation of Mel spectrograms from audio files was investigated and their usage as input to a machine learning algorithm or a neural network. A framework was developed using a model, which is well suited to classify an emotion for a short or longer audio clip. The experiments of this thesis show the success that can be achieved using convolutional neural networks and Mel spectrograms as its input to classify an emotion based only on an audio file. This in combination with different training methods has results in a framework that can classify emotions from short but also longer audio files in a very short time. In summary, automatic classification of emotions from a system is a very interesting topic, where multiple methods can lead to different successes. With the utilization of the results of this work, a further step towards improved HCI can be taken, as systems can adapt their action to the emotion state of the user.dc.description.abstract
Languageendc.language.iso
PublisherUniversität Ulmdc.publisher
LicenseCC BY 4.0 Internationaldc.rights
Link to license texthttps://creativecommons.org/licenses/by/4.0/dc.rights.uri
KeywordAcoustic Emotion Recognitiondc.subject
KeywordSpeech Emotion Recognitiondc.subject
KeywordNeural Networkdc.subject
KeywordAn End-to-end Deep Learning Framework for Acoustic Emotion Recognitiondc.subject
KeywordFramework for Acoustic Emotion Recognitiondc.subject
KeywordDeep Learning Framework for Acoustic Emotion Recognitiondc.subject
Dewey Decimal GroupDDC 620 / Engineering & allied operationsdc.subject.ddc
LCSHNeural networks (Computer science)dc.subject.lcsh
LCSHMachine learningdc.subject.lcsh
TitleAn End-to-end deep learning framework for acoustic emotion recognitiondc.title
Resource typeAbschlussarbeit (Bachelor)dc.type
Date of acceptance2023dcterms.dateAccepted
RefereeMinker, Wolfgangdc.contributor.referee
DOIhttp://dx.doi.org/10.18725/OPARU-48933dc.identifier.doi
PPN1847534570dc.identifier.ppn
URNhttp://nbn-resolving.de/urn:nbn:de:bsz:289-oparu-49009-8dc.identifier.urn
GNDNeuronales Netzdc.subject.gnd
GNDMaschinelles Lernendc.subject.gnd
GNDDeep learningdc.subject.gnd
FacultyFakultät für Ingenieurwissenschaften, Informatik und Psychologieuulm.affiliationGeneral
InstitutionInstitut für Nachrichtentechnikuulm.affiliationSpecific
DCMI TypeTextuulm.typeDCMI
CategoryPublikationenuulm.category
Bibliographyuulmuulm.bibliographie


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record