Skjul metadata

dc.date.accessioned2023-01-25T18:00:05Z
dc.date.available2023-01-25T18:00:05Z
dc.date.created2022-06-13T16:14:45Z
dc.date.issued2022
dc.identifier.citationBentsen, Lars Ødegaard Simionato, Riccardo Wallace, Benedikte Krzyzaniak, Michael Joseph . Transformer and LSTM Models for Automatic Counterpoint Generation using Raw Audio. Proceedings of the SMC Conferences. 2022
dc.identifier.urihttp://hdl.handle.net/10852/99177
dc.description.abstractA study investigating Transformer and LSTM models applied to raw audio for automatic generation of counterpoint was conducted. In particular, the models learned to generate missing voices from an input melody, using a collection of raw audio waveforms of various pieces of Bach’s work, played on different instruments. The research demonstrated the efficacy and behaviour of the two deep learning (DL) architectures when applied to raw audio data, which are typically characterised by much longer sequences than symbolic music representations, such as MIDI. Currently, the LSTM model has been the quintessential DL model for sequence-based tasks, such as generative audio models, but the research conducted in this study shows that the Transformer model can achieve competitive results on a fairly complex raw audio task. The research therefore aims to spark further research and investigation into how Trans- former models can be used for applications typically dominated by recurrent neural networks (RNN). In general, both models yielded excellent results and generated sequences with temporal patterns similar to the input targets for songs that were not present in the training data, as well as for a sample taken from a completely different dataset.
dc.languageEN
dc.publisherSociety for Sound and Music Computing
dc.rightsAttribution 3.0 Unported
dc.rights.urihttps://creativecommons.org/licenses/by/3.0/
dc.titleTransformer and LSTM Models for Automatic Counterpoint Generation using Raw Audio
dc.title.alternativeENEngelskEnglishTransformer and LSTM Models for Automatic Counterpoint Generation using Raw Audio
dc.typeJournal article
dc.creator.authorBentsen, Lars Ødegaard
dc.creator.authorSimionato, Riccardo
dc.creator.authorWallace, Benedikte
dc.creator.authorKrzyzaniak, Michael Joseph
cristin.unitcode185,15,30,0
cristin.unitnameInstitutt for teknologisystemer
cristin.ispublishedtrue
cristin.fulltextoriginal
cristin.qualitycode1
dc.identifier.cristin2031494
dc.identifier.bibliographiccitationinfo:ofi/fmt:kev:mtx:ctx&ctx_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.jtitle=Proceedings of the SMC Conferences&rft.volume=&rft.spage=&rft.date=2022
dc.identifier.jtitleProceedings of the SMC Conferences
dc.identifier.doihttps://doi.org/10.5281/zenodo.6572847
dc.type.documentTidsskriftartikkel
dc.type.peerreviewedPeer reviewed
dc.source.issn2518-3672
dc.type.versionPublishedVersion


Tilhørende fil(er)

Finnes i følgende samling

Skjul metadata

Attribution 3.0 Unported
Dette verket har følgende lisens: Attribution 3.0 Unported