#metabrainz

/

      • xSkkarf[m]
        nvm i figured it out xD. Cursor helped me sort out some docker networking issues.
      • tux0r
        ah, docker being docker again
      • minimal has quit
      • LupinIII joined the channel
      • Jigen has quit
      • Jigen joined the channel
      • ApeKattQuest has quit
      • LupinIII has quit
      • ApeKattQuest joined the channel
      • ApeKattQuest has quit
      • ApeKattQuest joined the channel
      • vardhan_ joined the channel
      • d4rk-ph0enix has quit
      • d4rk-ph0enix joined the channel
      • d4rkie has quit
      • d4rkie joined the channel
      • vardhan_ has quit
      • jasje[m]
        <reosarevok[m]> "jasje: let me know if I really..." <- reosarevok: ^^ tqice forwarded email does not contain the users email :(
      • pite has quit
      • function1_ has quit
      • function1 joined the channel
      • KrishnaCosmic joined the channel
      • KrishnaCosmic has quit
      • SigHunter has quit
      • SigHunter joined the channel
      • MyNetAz has quit
      • MyNetAz joined the channel
      • d4rk-ph0enix has quit
      • d4rk-ph0enix joined the channel
      • nobiz has quit
      • d4rkie has quit
      • d4rkie joined the channel
      • vardhan_ joined the channel
      • vardhan_ has quit
      • nobiz joined the channel
      • d4rk-ph0enix has quit
      • d4rk-ph0enix joined the channel
      • monu8[m] has quit
      • BrainzGit
        [musicbrainz-server] 14reosarevok merged pull request #3495 (03master…mbs-12826): MBS-12826: Speed up release series reordering https://github.com/metabrainz/musicbrainz-serve...
      • d4rkie has quit
      • d4rkie joined the channel
      • [listenbrainz-android] 14hemang-mishra opened pull request #556 (03feat-playlists…PlaylistScreen): MOBILE-214 Playlist-Detail-Screen & Add-Track-To-Playlist-Sheet https://github.com/metabrainz/listenbrainz-andr...
      • reosarevok[m]
        Updating beta
      • HemangMishra[m] joined the channel
      • HemangMishra[m]
        jasje: I have made the final PR of the playlist feature. Let me know if you have any changes are required. Then we can merge it in the main branch.
      • jasje[m]
        HemangMishra[m]: Need time to test
      • HemangMishra[m] uploaded an image: (736KiB) < https://matrix.chatbrainz.org/_matrix/media/v3/download/matrix.org/jhCzeuhZjPransvQZQvwedGO/image.png >
      • HemangMishra[m]
        HemangMishra[m]: Also, should I change the gradient or rounded corners of the description section. Like Artists page and others have rounded corners but Profile>listens doesn't, so I'm not sure which one to follow.
      • reosarevok[m]
        Beta updated one
      • * Beta update done
      • rustynova[m]
        Hi! I plan to make some contributions to Melba again now that the project is a more stable state (and not bound by the rules of GoSC).
      • But I do wonder, is the project on hiatus? Or just need contributors? There's a pending PR by yellowhatpro that hasn't been touched in 4 months...
      • For now I'd just make some refactoring changes to clean things up a bit (and finally rebase my own PR).
      • yellowhatpro[m] joined the channel
      • yellowhatpro[m]
        Hi, rustynova. Yeah you can work on the project. I have been busy with some other projects outside of MetaBrainz that I am not able to work on melba.
      • vardhan_ joined the channel
      • rustynova[m]
        Well that will be a side project for me too. Something to change my mind off other projects. Will try at least a commit per week though
      • lucifer[m]
        mayhem: hi! i am doing MLHD+ chunk by chunk processing and i think we might be able to do it on the current cluster. it processes one chunk in an hour and 16 chunks in total so assuming that takes a day. however each chunk produces an output of 400G to store. and that's without replication. with replication its 3x. so the current bottleneck is disk space.
      • mayhem[m]
        ok, let me see what our options are.
      • lucifer[m]
        zas: mayhem: i see hetzner has 5TB storage boxes for $13/month. so in the worst case we keep this around for a month, its 65usd still cheaper than the vms.
      • mayhem[m]
        (but, squeee, that is exciting!)
      • the storage boxes run via SAMBA and end up being slow and unreliable.
      • lucifer[m]
        i am guessing its HDD not SSD so slower but we have 12 hours of spark free time daily so i could 3-4 chunks a days.
      • i see
      • mayhem[m]
        its for storing things, not for working things.
      • I tried to run navidrome on top of a storage box and it lasted a week before it shat itself.
      • lucifer[m]
        i can disable replication cluster wide or disable it to 2 for the duration of mlhd processing and but even then i expect at least 10 TB of disk space requirement.
      • s/disable/reduce/
      • mayhem[m]
        we could have disks put into those machines, but it would be a real pain.
      • given all this, I still think we should use new VMs to get this done.
      • lucifer[m]
        i see.
      • mayhem[m]
        and we can attach storage volumes (fast and suitable for working) to the nodes to reach the DB levels we need.
      • lucifer[m]
        is it possible to have multiple disks at a single mount point?
      • mayhem[m]
        no, unix doesn't allow that.
      • why do you need that?
      • lucifer[m]
        actually nvm, i found the way to specify multiple directories/mounts to hdfs
      • mayhem[m]
        if you wanted that you would have to use something like LVM and make a virtual drive.
      • phew.
      • lucifer[m]
        zas: would it possible for the new VMs to connect to existing spark servers?
      • i need a smaller server for running the leader, but i guess we could get a smaller vm for that and keep it all isolated from the current cluster.
      • mayhem: so i am thinking CCX53 * 3 + CCX53 * 1 + 10 TB storage volume on each.
      • zas[m]
      • Sorry, but I don't have much time today for this, still fighting back against excess traffic on MB
      • lucifer[m]
        sure we can arrange it later.
      • MyNetAz has quit
      • mayhem[m]
        I can create the servers, no problem. any objections zas
      • ?
      • zas[m]
        none :)
      • mayhem[m]
        ok, hang on lucifer, VMs coming.
      • lucifer[m]
        mayhem: we'll need to connect all the VMs in a vlan or make them accesible to each other somehow.
      • mayhem[m]
        yes, I believe running the server setup should take care of all of that.
      • lucifer[m]
        okay cool
      • mayhem[m]
        but I dont actually know how to run that. so maybe we should wait for zas. unless he can tell me how.
      • lucifer[m]
        yeah i think makes sense to wait until he's free.
      • mayhem[m]
        ok, let us know when you free up zas
      • zas[m]
        First you need to add new VMs to Ansible, since they are Hetzner VMs you can copy https://github.com/metabrainz/metabrainz-ansibl... (a VM) but you need to change values to match the network settings for this VM. This config doesn't add any virtual network. For that you need to follow https://docs.hetzner.com/cloud/networks/connect... , we didn't do it successfully yet, and it will require tricky
      • firewall settings I guess (we can use Ansible for initial deployment, and just do configs manually later on)
      • Each VM should be added to https://github.com/metabrainz/metabrainz-ansibl... (in right groups, check where is auth VMs)
      • Each machine we want to connect should be in spark-vnet (this is on Hetzner Robot)
      • For example, jermaine has:... (full message at <https://matrix.chatbrainz.org/_matrix/media/v3/...>)
      • monkey[m]
        <aerozol[m]> "I think this was about the..." <- Yeha, I think we have enough stuff to show-and-tell, that would be nice. Not sure we have any feature on the immediate horizon to justify waiting for.
      • Let me know if you want some help, and in what form. For example I can collate everything I think should go in there, and let you filter, massage and post?
      • zas[m]
        michael (which is also on this vnet) can ping jermaine over this network:... (full message at <https://matrix.chatbrainz.org/_matrix/media/v3/...>)
      • So the idea is to have your VMs in this network too.
      • mayhem[m]
        oy. I think it will be better to wait zas -- I think you'll do it much faster than me trying to learn this all.
      • zas[m]
        Actually it would be good if you add VMs to Ansible (this part is easy), and then learn how to deploy (basically run bootstrap.yml playbook then site.yml playbook). You need ssh to be properly configured (ssh root@yourvm should work for bootstrap, and then ssh youruser@yourvm should work for site)
      • You risk nothing trying (at worse we rebuild the VM and start over). Properly configuring network is a must (of course, else we lose connection)
      • for the vswitch part, that's trickier, because we don't have such setup yet, and I expect some changes will be needed regarding network/firewall settings to handle this case in Ansible.
      • Bootstrap needs root access because users aren't created yet, then site can run on user+sudo
      • mayhem[m]
        well, this sounds more interesting that the thing I am fighting now...
      • <lucifer[m]> "mayhem: so i am thinking CCX53 *..." <- this is actually not clear. one machine with 10TB storage?
      • lucifer[m]
        each.
      • mayhem[m]
        and 3 stock CCX53s?
      • so 4 CCX53 with each 10TB?
      • lucifer[m]
        oh sorry, i made a typo. 3 CCX53's and 1 CCX23. with 10 TB each yes.
      • but i did think of an alternative plan to tackle this dataset too if you want to wait.
      • mayhem[m]
        what is that alternate plan?
      • lucifer[m]
        process each file individually parallely using duckdb or somesuch and then just do the final combination step in spark.
      • mayhem[m]
        sounds like a lot of code for you to write.
      • And I can't get a dedicated CPU VM. we've reached some limit I need to have raised.
      • lucifer[m]
        hard to tell without taking a shot at it, might be a day or less or more.
      • mayhem[m]
        maybe spend an hour or 2 on it and see if that approach is promising?
      • zas[m]
        If those machines are temporary, we may not add them to Ansible at all, and just do a basic config manually
      • mayhem[m]
        yes, all temp. but we'll see if we actually need them
      • lucifer[m]
        i don't anticipate issues to show up until say 10% of the dataset is done but sure let me try.
      • zas[m]
        ok, tell me, I'll be happy to help, but I'm very busy right now at controlling this absurd traffic surge...
      • mayhem[m]
        zas[m]: focus on the surge. lucifer will try an alternate path.
      • d4rkie has quit
      • d4rkie joined the channel
      • d4rkie has quit
      • d4rkie joined the channel
      • d4rk-ph0enix has quit
      • d4rk-ph0enix joined the channel
      • vardhan_ has quit
      • leftmostcatUTC-8 has quit
      • bitmap[m]
        reosarevok: yvanzo: lucifer: hello, are we meeting now?
      • reosarevok[m]
        Hi! I thought we were! :)
      • yvanzo[m]
        I thought it was in 1h from now but I’m around already.
      • monkey[m]
        mayhem, lucifer An interesting timezone-related issue I never saw before was reported (LB-1766): `can't compare offset-naive and offset-aware datetimes`. The ticket was assigned to me but I don't know if I'm the ideal candidate. Can I assign the ticket to either of you?
      • BrainzBot
        LB-1766: Internal Server Error (500) when accessing listening history https://tickets.metabrainz.org/browse/LB-1766
      • reosarevok[m]
        IIRC we agreed on 5 PM CET which would be now
      • bitmap[m]
        you may be right, I probably got confused by the dst shift again
      • reosarevok[m]
        Oh no, it was UTC
      • My bad :)
      • Then yes, in one hour
      • Although if lucifer is around we can always start earlier
      • lucifer[m]
        i am around now if we want to start early