Hacker Newsnew | past | comments | ask | show | jobs | submit | diibv's commentslogin

It is worth noting that AcousticBrainz was based on an open-source audio analysis library Essentia [1] that has been gradually improving since the launch of the AcousticBrainz project. It now has better ML models, with higher accuracy and generalization, than those used in AcousticBrainz in the pre-deep learning age of handcrafted signal processing features and SVMs.

See [2] for the list of models currently available. It includes updated ML models for the classifiers used in AcousticBrainz and many new models (e.g., a 400-style music classifier, models for arousal-valence emotion recognition, and feature embeddings that can be used for audio similarity or training new classifiers).

[1] https://essentia.upf.edu/ [2] https://essentia.upf.edu/models.html


Sadly all models are licensed for non-commercial use only (CC BY-NC-SA 4.0), so not Free Software.


As a part of an ongoing research collaboration between the Centre for Digital Music of the Queen Mary University of London and the Music Technology Group of Universitat Pompeu Fabra Barcelona, we are happy to announce the launch of Song Describer, a crowdsourcing initiative to collect the first open research dataset of music captions.

Song Describer is an open-source annotation platform featuring CC-licensed music that anyone can annotate with natural language descriptions. Through this platform, we aim to create the first research dataset of music-caption pairs, which will give us insights into how people describe music and support the development of audio-text ML models.

There is an increasing need for music datasets that go beyond categorical labels and we hope that our data collection efforts will promote more research on the relationship between natural language and music audio, and its many potential applications. If you're interested in this field or are curious or willing to contribute, you can head to Song Describer to read more about it and start annotating right away!

For more information about the platform's design, see the late-breaking demo presented at the ISMIR 2022 conference: https://ismir2022program.ismir.net/lbd_405.html


hledger is another open-source CLI tool for accounting that works amazingly well for me. You can add transactions editing in plain text too and there is a basic web UI on top of it.


And hledger is based on the original https://www.ledger-cli.org/ which is also a great option for CLI accounting.


You can easily modify the code to bypass MusicBrainz, and use Discogs as your source instead.


Hi, I'm another researcher at the Music Technology Group, which runs Freesound. Definitely, this one of our long-time challenges.

Facilitating search and exploration is one of our main research directions related to Freesound. We do audio ML research, in particular, audio-based auto-tagging (see the already mentioned Freesound Datasets https://annotator.freesound.org/fsd/ we've built for this purpose) but also similarity and clustering based on audio analysis and tags.

So far, Freesound API provides similarity search and audio analysis features.

Have a look at the Freesound Explorer, which is an example of a 2D sound exploration interface built on top of audio features: https://labs.freesound.org/apps/2017/08/21/freesound-explore...


The program is now online: https://webaudioconf2021.com/program/


Donald Hoffman has already been mentioned in the comments. Here's his interesting talk "Realism Is False" on Edge.org: https://www.edge.org/conversation/donald_d_hoffman-realism-i...


Can you elaborate more?


Love Discogs, have been using it for many years!

Some interesting musicological statistics can be computed using their public dumps. For example: http://hdl.handle.net/10230/32931


There's also AcousticBrainz for audio features, coupled to MusicBrainz: https://acousticbrainz.org/


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: