(There were a number of diversity discussions about music festivals in Denmark last year. MB would be a perfect fit to codify this data and make reports from.)
Hobbyboy joined the channel
hugo
@fresco: here is an overview of the styles https://www.youtube.com/watch?v=I0zEAQTnUrU not the music i listen to daily, it is just folk but this radiostation just plays those sounds
drsaunders joined the channel
Sophist-UK
samj1912: Ok - I give up. I have a test_metadata.pyc file and cannot find a matching .py file anywhere.
samj1912
lol :P
I think you made it sometime and it stayed
we have a test_scripts and a test_formats
that we use for similar tests
you can add it to test_scripts maybe?
Sophist-UK
Well the .pyc file would not get deleted by git checkout, only if I do it manually. So it will hang around. But at some point I must have had a matching .py file on my PC and run tests.
Perhaps I attempted to create a set of tests for metadata and then decided it was too much effort. Who knows. It was 3 years ago and I can't remember what I had for breakfast.
Mineo
there's a module called decompyle (I think) that can turn a .pyc back into a .py file if you want to have a look at it :)
ruaok: will do my initial project pitch to a couple of profs tomorrow
I was thinking about extending MsB to a NLP based metadata extractor (I think I have a very very vague idea how it currently works on last fm data)
couldnt find much on MsB-server repo
I don't even know much about what the scope and future plans for MsB are
SothoTalker_
MusicBrainz? :D
samj1912
messybrainz
SothoTalker_
:-)
Quesito
hey gcilou! I added a few things to the UX survey doc, about to call it a night, any q's or thoughts before I go?
gcilou
Hey Quesito :) I looked over it, looks nice!
Quesito
hopefully getting somewhere....:P let me know your thoughts / comments, etc!
gcilou
Yeah, definitely! I'll look more thoroughly soon
Quesito
thank you! any advice, thoughts, etc are needed :)
Sophist-UK
sam1912: I decompiled the test_metadata.py file and it doesn't look to me like code I would have written. But relatively short - I will try to build a comprehensive test_metadata.py.
regagain has quit
samj1912
okay
samj1912 sleeps
SothoTalker_
nighty
is there any way to circumvent 503 errors, except being a paying supporter? ^^
kepstin
SothoTalker_: what are you trying to do?
SothoTalker_
just fetch stuff from the webservice
kepstin
sure, but what kind of stuff, and how much?
SothoTalker_
all releases of a label
which are 16 requests in total, to get everything.
kepstin
so, the way to not get 503 errors is to follow the rate limiting requirements...
SothoTalker_
i currently make one request every 5 seconds
kepstin
hmm, that should be ok; the actual limit is ~1/second
per ip address, for properly configured apps (set user agent, etc.)
SothoTalker_
it's just a simple script :-)
kepstin
if you're using some generic user agent rather than specifying one that identifies your tool, you hit much harsher rate limits
so that's the first thing to do - set a user agent, call it "SothoTalker's Label Scraper" or whatever, doesn't matter - then you should be able to use the webservice at 1 request/second
if you want faster than 1 req/second, then you basically have two options - look into getting an api key with raised rate limit for the musicbrainz.org ws, or run your own replicated server locally which you can query as fast as you like (or even just read info direct from the postgres db)
SothoTalker_
well, i dont want faster access, i just dont want that 503 error happen to me :D
i suspect the WS is just hit heavily, since this did not occur earlier like 2 weeks ago.
kepstin
right now I assume you're being thrown into the "anonymous" group because of the user agent on your webservice requests, and due to the global 50/s limit, that's unpredictable.
so fix the user agent first, then see if that helps :/
SothoTalker_
i tried a few ua strings, and it didnt help :)
i even made my script wait 10 seconds between each request
kepstin
hmm. I guess the webservice might just be overloaded overall, then.
SothoTalker_
that's my guess, too.
kepstin
I don't know what the current global limit is, but I thought it was raised with the switch to newhost
SothoTalker_
well, that problem does happen since about 2 weeks ago. before i had no problems.
kepstin
well, the solution is to "try again later" - If you want to encode that into your script, look at having a backoff algorithm so it's not making the overload worse.
SothoTalker_
and it's not like my script permanently hammers the server. i just run it once or twice a day to update my spreadsheet
i might have to parse the 503 error message and put a 'try again after a few seconds' code :x
kepstin
if you're doing any sort of query like that, try to run it at a random time each day rather than the same time every day
well, you shouldn't retry the entire series of requests - just have it wait for a while, then continue where it left off...
SothoTalker_
that's what i meant
kepstin
you probably want to do an exponential backoff - so if you're e.g. waiting 1s between requests, then if one fails, wait 2s before trying again, if it fails again, wait 4s, and so on.
and if the wait gets too long, maybe decide to give up with an error message ;)
SothoTalker_
heh
well, now it went through, luckily
alastairp
gcilou: I don't think alpha import should have as much focus. Ideally we want people to ignore their alpha account
gcilou
Ok! I can definitely adjust to that. Rob sent me a comment about his ideal layout for the page, so the whiteboard was following that.
alastairp
Ah, I didn't see that
Anyway, that's my opinion
gcilou
yeah that's cool. I'm trying a few things out to get a feel for it