Hi yvanzo.... sorry I got disconnect from the chat and didn't see your message. I did some digging and I think the amqp trigger retries after a failure, so I think it's just a warning, but I had some other questions about the indexer because it does seem to be taking a long time for the live indexing. I followed the thread on github, I seem to be
seeing a lot of "Index limit exceeded. Entity: recording, Total Rows: 5038988" but is it ok to bump that number up in my indexer.ini? What happens after the index limit is exceeded? It seems like activity in the indexer stops for about 10-15 minutes then processes more messages until it hits the index limit again. Just wanted to figure out what was
going on so that maybe I could debug it. Sorry for the long message! Thanks!
roger_that has quit
Lotheric has quit
adk0971 has quit
iliekcomputers
good morning!
MajorLurker joined the channel
MajorLurker has quit
Lotheric joined the channel
roger_that joined the channel
dseomn_ joined the channel
ZaphodBeeblebrox joined the channel
ZaphodBeeblebrox has quit
ZaphodBeeblebrox joined the channel
navap1 joined the channel
spuniun- joined the channel
Zhele_ joined the channel
ijc_ joined the channel
kloeri_ joined the channel
yvanzo
roger_that: It limits the number of rows queried from PostgreSQL at once. If exceeded, it aborts processing the current messages and requeues them as failed.
spuniun has quit
dseomn has quit
CatQuest has quit
Zhele has quit
ijc has quit
kloeri has quit
navap has quit
dseomn_ is now known as dseomn
It is ok to bump that number up as long as it fits allocated resources.
yvano: So if those messages fail, my search indexer can never be up to date?
I have 8 cpus, 32gb ram... is it not enough to run the live indexing?
And also... why do the failed messages cause such a big delay in processing the next messages? (Although the search server seems to be processing stuff - there's just nothing in the indexer logs for about 10-15 minutes after the failed query)
roger_that20: exactly, these queries require a lot of RAM and take a lot of time to be processed by PostgreSQL, so better reject them to allow for other queries to go through.
It does not use the full power of your CPUs/RAM because each component (indexer, solr, postgres) are limited for safety.
roger_that20
I see
yvanzo
The tricky part is to correctly allocate a balanced amount of resources to each component so it works smoothly.
roger_that20
What happens after they fail? Like I'm trying to understand what happens between the failed query and the next batch of messages that get processed
yvanzo
Rejected messages are not lost, they are in a failed queue that can be processed later on.
roger_that20
Oh ok
kloeri_ is now known as kloeri
And if I restart the containers.... the messages are saved?
Like if I wanted to change the indexer_limit
yvanzo
Yes, they are in the *_mqdata volume
adk0971 joined the channel
roger_that20
I noticed though, in that github thread... that JoshDi set his index limit to 1500000
So i wonder how those messages were getting processed
MBS-10004: inconsistency between JSON-LD and documentation
reosarevok
They kinda have a point that JSON-LD URIs probably should not be https?
yvanzo
makes sense
reosarevok
bitmap: do you remember if the https there was intentional?
yvanzo
This is the same reason we use http for VIAF URLs.
roger_that20: The only difference between 1.x and 2.x is about indexing recording’s first release date (in 2.x only).
1.x is a bit behind 2.x for a few other points but that it will be catching up with them.
reosarevok wonders about https://tickets.metabrainz.org/browse/MBS-9997 - seems that solved "itself", which usually means we fixed something in the meantime :p
ruaok: iliekcomputers: i was looking in sentry logs and saw that there are over 11m spotify errors related to same user over the past few months. any ideas?
ruaok
yeah, it been on my list to do. there are some simple things, some more complicated.
_lucifer
i see that spotipy hides the actual error message.
ruaok
I was trying to do weekly sentry cleanups for while, but got distracted by other things.
_lucifer
i can help :)
ruaok
the 400 errors are harder to clean up -- I think some users may have deleted their spotify account, but not turned off LB listen saving.
I think once we get a 400 error enough times, we need to disable that account. we used to have some logic like that, but it ended up disabled 2/3 of the accounts. ;(
please do help!3
_lucifer
makes sense
the 400 message might actually have some useful info.
i'll try to reproduce this error locally and see if we can come up with a better solution
ruaok
kewl.
_lucifer: alastairp I'm running a "not nice" query on bono. if it starts smoking and burning, you know whom to blame.