Logo do repositório

Integrating Audio Signal Processing and Deep Learning Algorithms for Gait Pattern Classification in Brazilian Gaited Horses

dc.contributor.authorAlves, Anderson Antonio Carvalho
dc.contributor.authorAndrietta, Lucas Tassoni
dc.contributor.authorLopes, Rafael Zinni
dc.contributor.authorBussiman, Fernando Oliveira
dc.contributor.authorSilva, Fabyano Fonseca e
dc.contributor.authorCarvalheiro, Roberto [UNESP]
dc.contributor.authorBrito, Luiz Fernando
dc.contributor.authorBalieiro, Júlio César de Carvalho
dc.contributor.authorAlbuquerque, Lucia Galvão [UNESP]
dc.contributor.authorVentura, Ricardo Vieira
dc.contributor.institutionScience and Technology of Maranhão (IFMA)
dc.contributor.institutionUniversidade de São Paulo (USP)
dc.contributor.institutionFederal University of Viçosa
dc.contributor.institutionUniversidade Estadual Paulista (UNESP)
dc.contributor.institutionNational Council for Scientific and Technological Development (CNPq)
dc.contributor.institutionPurdue University
dc.date.accessioned2025-04-29T18:07:49Z
dc.date.issued2021-01-01
dc.description.abstractThis study focused on assessing the usefulness of using audio signal processing in the gaited horse industry. A total of 196 short-time audio files (4 s) were collected from video recordings of Brazilian gaited horses. These files were converted into waveform signals (196 samples by 80,000 columns) and divided into training (N = 164) and validation (N = 32) datasets. Twelve single-valued audio features were initially extracted to summarize the training data according to the gait patterns (Marcha Batida—MB and Marcha Picada—MP). After preliminary analyses, high-dimensional arrays of the Mel Frequency Cepstral Coefficients (MFCC), Onset Strength (OS), and Tempogram (TEMP) were extracted and used as input information in the classification algorithms. A principal component analysis (PCA) was performed using the 12 single-valued features set and each audio-feature dataset—AFD (MFCC, OS, and TEMP) for prior data visualization. Machine learning (random forest, RF; support vector machine, SVM) and deep learning (multilayer perceptron neural networks, MLP; convolution neural networks, CNN) algorithms were used to classify the gait types. A five-fold cross-validation scheme with 10 repetitions was employed for assessing the models' predictive performance. The classification performance across models and AFD was also validated with independent observations. The models and AFD were compared based on the classification accuracy (ACC), specificity (SPEC), sensitivity (SEN), and area under the curve (AUC). In the logistic regression analysis, five out of the 12 audio features extracted were significant (p < 0.05) between the gait types. ACC averages ranged from 0.806 to 0.932 for MFCC, from 0.758 to 0.948 for OS and, from 0.936 to 0.968 for TEMP. Overall, the TEMP dataset provided the best classification accuracies for all models. The most suitable method for audio-based horse gait pattern classification was CNN. Both cross and independent validation schemes confirmed that high values of ACC, SPEC, SEN, and AUC are expected for yet-to-be-observed labels, except for MFCC-based models, in which clear overfitting was observed. Using audio-generated data for describing gait phenotypes in Brazilian horses is a promising approach, as the two gait patterns were correctly distinguished. The highest classification performance was achieved by combining CNN and the rhythmic-descriptive AFD.en
dc.description.affiliationDepartment of Education Federal Institute of Education Science and Technology of Maranhão (IFMA)
dc.description.affiliationDepartment of Animal Nutrition and Production School of Veterinary Medicine and Animal Science University of São Paulo
dc.description.affiliationDepartment of Animal Science Federal University of Viçosa
dc.description.affiliationDepartment of Animal Science School of Agricultural and Veterinary Sciences Säo Paulo State University (UNESP)
dc.description.affiliationNational Council for Scientific and Technological Development (CNPq)
dc.description.affiliationDepartment of Animal Sciences Purdue University
dc.description.affiliationUnespDepartment of Animal Science School of Agricultural and Veterinary Sciences Säo Paulo State University (UNESP)
dc.description.sponsorshipConselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
dc.identifierhttp://dx.doi.org/10.3389/fanim.2021.681557
dc.identifier.citationFrontiers in Animal Science, v. 2.
dc.identifier.doi10.3389/fanim.2021.681557
dc.identifier.issn2673-6225
dc.identifier.scopus2-s2.0-85131139764
dc.identifier.urihttps://hdl.handle.net/11449/297825
dc.language.isoeng
dc.relation.ispartofFrontiers in Animal Science
dc.sourceScopus
dc.subjectaudio-feature
dc.subjectconvolutional neural network
dc.subjectfour-beat gaited
dc.subjecthorse gait
dc.subjectsound analysis
dc.titleIntegrating Audio Signal Processing and Deep Learning Algorithms for Gait Pattern Classification in Brazilian Gaited Horsesen
dc.typeArtigopt
dspace.entity.typePublication
relation.isOrgUnitOfPublication3d807254-e442-45e5-a80b-0f6bf3a26e48
relation.isOrgUnitOfPublication.latestForDiscovery3d807254-e442-45e5-a80b-0f6bf3a26e48
unesp.campusUniversidade Estadual Paulista (UNESP), Faculdade de Ciências Agrárias e Veterinárias, Jaboticabalpt

Arquivos