TOPIC: MetaBrainz Community and Development channel | MusicBrainz non-development: #musicbrainz | BookBrainz: #bookbrainz | Channel is logged; see https://musicbrainz.org/doc/IRC for details | Agenda: Reviews, Doc sprint (yvanzo), Repo review (reo)
yvanzo: before I start writing it, we don't have any docs for how to add support for a new URL domain to the code, right?
2022-01-24 02412, 2022
reosarevok
As in, a doc that explains what files to change for autoselect/cleanup, favicons, sidebar, etc
2022-01-24 02444, 2022
yvanzo
reosarevok: No, we should even probably have a separate doc explaining how to handle those tickets to check current usage in the DB, old and current links on the web…
2022-01-24 02403, 2022
yvanzo
Like HACKING-URLCLEANUP.md?
2022-01-24 02421, 2022
reosarevok
Hmm, I was going to put it on HACKING but that also seems good
2022-01-24 02438, 2022
reosarevok
We can move it to a dev docs repo if we eventually start one
i saw those in RTD tutorial first and thought to add LB docs as well.
2022-01-24 02451, 2022
alastairp
why do we have "Web root url" in the api docs? I can't see it used anywhere in the docs
2022-01-24 02410, 2022
lucifer
yeah, i think we can remove that. dosen't make sense to list non-api endpoints there in LB API docs.
2022-01-24 02419, 2022
mayhem
alastairp: got a sec for some brainstorming?
2022-01-24 02430, 2022
alastairp
yes, go for it
2022-01-24 02435, 2022
lucifer
although it could come in handy in dev/maintainer docs
2022-01-24 02457, 2022
mayhem
let me find a reference real quick. hang on.
2022-01-24 02458, 2022
alastairp
lucifer: yeah, there may be value in having it somewhere, but it feels a bit out of place right there
2022-01-24 02400, 2022
lucifer
yup makes sense. i'll remove that in another PR. the current ones depend on each other so making the change here will cause cascading merges/rebases to resolve conflict.
2022-01-24 02410, 2022
alastairp
no prob
2022-01-24 02412, 2022
akshaaatt
lucifer, A reminder about the data safety form in playstore 😃
2022-01-24 02403, 2022
mayhem
meh. can't find the paper that talks about moods derived from listens. but that paper said something along the lines of:
2022-01-24 02417, 2022
mayhem
"Extrating moods from audio is hard and doesn't work that well".
2022-01-24 02421, 2022
mayhem
*extracting
2022-01-24 02423, 2022
alastairp
the one I linked last week?
2022-01-24 02426, 2022
mayhem
yes.
2022-01-24 02450, 2022
mayhem
and in general, there is a sentiment that user behaviour is a better input to recommendation that audio files.
2022-01-24 02405, 2022
mayhem
A point that Dimi made in this defense.
2022-01-24 02420, 2022
mayhem
so, then it seems that AB is really barking up the wrong tree.
2022-01-24 02457, 2022
mayhem
not only does it not provide good results, its also very hard to do unless you have access to a huge cache of files. Easy for spotify, very hard for us.
2022-01-24 02443, 2022
alastairp
yeah, the question of "can you throw machine learning at audio signals and get some results?" has always been uncertain
2022-01-24 02420, 2022
mayhem
And I think that we fell in love with the concept; we were never fully clear on what data we wanted to collect, and we just got ourselves hornswaggled on this approach. I think it its time to step back even further than we already have and re-examine.
2022-01-24 02441, 2022
mayhem
exactly. so, how about we don't. period.
2022-01-24 02401, 2022
mayhem
instead, lets us ask ourselves: what data do we want to collect?
2022-01-24 02430, 2022
mayhem
once we answer that, we should find ways to collect those pieces of data, by the easiest means possible. meaning: what approaches fit with an open source approach.
2022-01-24 02446, 2022
alastairp
so, instead of "description of songs based on analysing audio", maybe we actually just want "description of songs"
2022-01-24 02447, 2022
lucifer
akshaaatt: thanks for the reminder, lets talk with mayhem about that after this conversation.
2022-01-24 02459, 2022
mayhem
DING! EXACTLY THAY!
2022-01-24 02400, 2022
mayhem
THAT
2022-01-24 02401, 2022
mayhem
because reading the moods from behaviour paper gives a clear insight. If transferlearning is needed -- well, we have a CF filtering setup already. If we're bolting another piece of on top of that, then this may not be that much work in the grand scheme of things.
2022-01-24 02418, 2022
mayhem
right now AB needs: DSP and ML knoweldge.
2022-01-24 02439, 2022
mayhem
the new approach only needs ML knowledge. and everyone wants to do this right now -- its the hot thing to be doing.
and we have the infrastructure for it already. so, lets use that.
2022-01-24 02402, 2022
mayhem
thanks!
2022-01-24 02433, 2022
mayhem
there are bits and pieces of AB that are clearly useful. such as creating datasets, managing them and then throwing them at ML.
2022-01-24 02451, 2022
mayhem
newAB could then throw away all the audio stuff and keep all the ML/dataset stuff.
2022-01-24 02428, 2022
alastairp
yes, right
2022-01-24 02440, 2022
mayhem
we could start with collecting mood information manually from users as part of BrainzPlayer. (listen to a track, select the mood from a dropdown, submit to newAB.
2022-01-24 02420, 2022
mayhem
once we have a body of those, we can start doing the ML as described in the paper. the paper allows us to "fill in the blanks" of the data we're missing.
2022-01-24 02430, 2022
mayhem
with me so far?
2022-01-24 02436, 2022
alastairp
oh, that's interesting
2022-01-24 02454, 2022
alastairp
working out how to collect data has always been a bit of an open question
2022-01-24 02402, 2022
mayhem
and those are our strengths OR are things that are tangibly close to being within our reach.
2022-01-24 02422, 2022
alastairp
and I think that the AB way of saying "here are some results, do they look wrong? give us feedback" was a bit weak because the data was a bit sparse
2022-01-24 02434, 2022
mayhem
so, lets focus on that. lets build an OpenML project with a focus on music.
2022-01-24 02445, 2022
mayhem
yes, indeed.
2022-01-24 02446, 2022
alastairp
but maybe integrating into BP might be a better way of getting that data
2022-01-24 02426, 2022
alastairp
the previous way that we've done it is give it to students as an assignment. "your task is to listen to these 500 songs from jamendo and categorise genre/mood/other things, then do some analysis on the results"
2022-01-24 02428, 2022
mayhem
I think BP is the outlier on this front. I think with enough effort, we can do music analysis on it. I think you're close to showing that.
2022-01-24 02447, 2022
mayhem
but, just to get BPM, do we really need to build all this infra for it?
2022-01-24 02455, 2022
mayhem
what if we wrote BPM stuff as a pluging for picard?
2022-01-24 02411, 2022
mayhem
and picard runs 3+ algs and if there is consensus, submit the data.
2022-01-24 02417, 2022
mayhem
bam, we're done.
2022-01-24 02424, 2022
alastairp
how do you mean? people use picard and tap along with the song?
2022-01-24 02429, 2022
alastairp
oh, for the algorithms. mmm
2022-01-24 02437, 2022
alastairp
one sec, let me commit some stuff
2022-01-24 02448, 2022
mayhem
yes. since it looks like there are promising algs, lets stuff them into picard and be done on that front.
2022-01-24 02406, 2022
mayhem
then focus ML in the newAB and I think we have a project might might get more people interested in helping.
2022-01-24 02412, 2022
mayhem wonders if lucifer has any ideas
2022-01-24 02458, 2022
monkey
Are "moods" a specific subset of tags?
2022-01-24 02409, 2022
lucifer
nothing to add but following along the discussion. +1 on all BP, picard and new AB ideas
2022-01-24 02438, 2022
monkey
(I've been wanting to add a tags input and display into the ListenCard on LB, perhaps another input can prompt for adding/selecting moods)
2022-01-24 02450, 2022
zas
legoktm[m]: ping
2022-01-24 02445, 2022
mayhem
monkey: moods could be tags -- yes, but I think I would build a more general infrastructure to support this in newAB.
2022-01-24 02413, 2022
mayhem
The only thing that bugs me about this is that newAB no longer deals with acoustics. thus the name is now borked.
3. If there is not, have the user select the most "salient" parts of the song. and then run analysis again. do we get a stable result now? if so, we're done.
2022-01-24 02422, 2022
mayhem
4. If not... not sure. "tap on every downbeat" and we'll calculate it?