12:00 PM
D4RK-PH0ENiX joined the channel
12:22 PM
ruaok
zas: no wifi on train, I take it?
12:23 PM
zas
just boarded, intermittent wifi and/or 4G
12:28 PM
ruaok
perfect for doing remote work! :) :)
12:32 PM
12:32 PM
Leo_Verto
woo!
12:33 PM
ruaok
Leo_Verto: so, no luck on that problem from yesterday.
12:34 PM
Leo_Verto
hm
12:36 PM
ruaok
and even with the firewall completely turned off, it doesn't work.
12:37 PM
Leo_Verto
can you access hadoop from another container in spark-network?
12:37 PM
ruaok
yep
12:40 PM
michelv has quit
12:41 PM
Leo_Verto
okay, are you sure setting up the services on a dedicated overlay network is what you want to do?
12:41 PM
ruaok
yes, yes.
12:41 PM
are for a login?
12:41 PM
care
12:41 PM
Leo_Verto
sure
12:41 PM
ruaok
which key should I use?
12:41 PM
and login?
12:42 PM
Leo_Verto
if I wanted to make a service externally accessable I'd use a bridged/host network instead, not sure if that works with services though
12:43 PM
although I use expose instead of publish, hm
12:47 PM
ruaok
the overlay network is what enabled the nodes to easily communicate with one another.
12:47 PM
which is the whole reason why I am using docker swarm.
12:47 PM
and it works quite well from what I can see. except this whole publishing ports bit
12:48 PM
Leo_Verto
yeah, I think you might need an additional bridged network for publish to work
12:48 PM
at least that's how I got my (non-swarm) setup working
12:49 PM
ruaok
hmm.
12:51 PM
so, create bridged network, add services to both bridged network and to overlay network.
12:51 PM
and then?
12:52 PM
Leo_Verto
I mean that's how it'd work with normal containers, not entirely sure what the ingress network does here
12:54 PM
also my network graphing tool is broken, nice
12:54 PM
ruaok
except the it works for port 8080, for the spark-master service.
12:54 PM
12:55 PM
that goes into the spark service. but the hadoop service one doesn'twork.
12:55 PM
and they are both on the overlay network.
12:58 PM
Leo_Verto
if you docker inspect the individual containers you can see that not only are they on spark-network but also ingress
12:58 PM
ruaok
thats normal/correct, no?
12:58 PM
Leo_Verto
yeah, I'm pretty sure spark-network isn't really used here, the requests use ingress
13:00 PM
and running docker service inspect --format="{{json .Endpoint.Spec.Ports}}" <service> I get pretty much the same setup for both
13:00 PM
ruaok
for a single node setup it isn't used, but won't it be needed for a multi-node setup (which we use)?
13:01 PM
Leo_Verto
lateron definitely, but not right now. maybe we should try removing it to see if it interferes here?
13:02 PM
ruaok
sure.
13:02 PM
would you like to try it yourself or do you want me to make the changes?
13:02 PM
if you clone this, you can have a play yourself
13:02 PM
13:03 PM
Leo_Verto
any reason you aren't using compose? :P
13:03 PM
ruaok
compose doesn't do swarms. or at least it didn't when I started this mess.
13:04 PM
and I'm not interested in using compose much when I have to reinvent the wheel into normal run commands half the time. might as well just do run commands from the get go.
13:04 PM
ruaok snaps his unix suspenders in defiance.
13:04 PM
I just pushed the publish change from last night.
13:04 PM
but if you clone this and then run docker/start-master-service.sh
13:04 PM
you can start it yourself.
13:05 PM
Leo_Verto
okay, thanks
13:05 PM
ruaok
then start-worker-service.sh
13:05 PM
and the requisite stop scripts.
13:05 PM
a working setup should yield both 8088 and 8080 services being mapped to localhost.
13:14 PM
Leo_Verto
okay these containers take forever to pull
13:51 PM
michelv joined the channel
14:02 PM
jwf has quit
14:02 PM
CallerNo6 has quit
14:02 PM
Sophist-UK has quit
14:02 PM
yvanzo has quit
14:02 PM
CallerNo6 joined the channel
14:02 PM
yvanzo joined the channel
14:04 PM
Sophist-UK joined the channel
14:04 PM
jwf joined the channel
14:12 PM
ruaok, figured it out
14:12 PM
the webinterface on hadoop-master is binding to 10.0.0.10 for some reason
14:12 PM
and 10.0.0.0/24 is spark-network
14:20 PM
14:20 PM
ruaok
That seems ok, no?
14:20 PM
Leo_Verto
14:21 PM
well, that means you can't access the webinterface from outside spark-network, no matter what you publish
14:24 PM
ruaok
ah, the publish command doesn't bridge it?
14:25 PM
Leo_Verto
nope, the container has to be on a bridge network and the service inside must bind to that network or all of them
14:25 PM
ruaok
ok, let me try the settings you suggested.
14:25 PM
Leo_Verto
but the ingress network should work as a bridge, I think
14:26 PM
that is 10.255.0.0/24 btw
14:26 PM
ruaok
yarn.nodemanager.runtime.linux.docker.default-container-network ==> ingress
14:26 PM
?
14:27 PM
Leo_Verto
yeah
14:29 PM
ruaok
like this perhaps?
14:29 PM
14:30 PM
Leo_Verto
hmm, reading more of that article it seems like that requires using the "Linux Container Executor" to start yarn which is not something we want, right?
14:31 PM
ruaok
dunno what that exactly is. I want swarm to start the containers.
14:31 PM
Leo_Verto
yeah, th
14:31 PM
there has to be a simpler way of just changing the bind address
14:32 PM
14:33 PM
ruaok
and we're on hadoop 3.1.1. at that...
14:33 PM
Leo_Verto
oh yeah
14:36 PM
yarn.resourcemanager.hostname
14:36 PM
14:37 PM
so you should be fine setting that to 0.0.0.0 I suppose, otherwise it wouldn't be accessible from spark-network
14:38 PM
ruaok
Let me that that in a bit, afk for a moment.
14:39 PM
iliekcomputers
14:42 PM
Leo_Verto
iliekcomputers, you should see New Year's Eve here. one year we had so much smog there was only ~15m visibility
14:42 PM
iliekcomputers
So dumb.
14:43 PM
Leo_Verto
14:43 PM
iliekcomputers
Official fireworks time is supposed to be 8-10 PM
15:25 PM
Soumya joined the channel
15:25 PM
reosarevok
There's official firewords time? :D
15:25 PM
Soumya
Hi... I am a gci participant...
15:26 PM
thefar8
Hi Soumya
15:26 PM
welcome to metabrainz irc
15:26 PM
Soumya
Greetings! I need some help regarding my task..
15:28 PM
thefar8
which task is it?
15:30 PM
Soumya
I have to collect some imteresting facts aboit METABRAINZ
15:30 PM
Interesting***
15:31 PM
About***
15:45 PM
iliekcomputers
Soumya: Hi, can you link the task you're working on?
15:46 PM
reosarevok: there is! it's to stop them going all night I guess, but nobody cares ;)
15:47 PM
Soumya
15:49 PM
iliekcomputers
Soumya: ok, thanks. I can talk to you about CritiqueBrainz, ListenBrainz or AcousticBrainz if you want. I'm not a mentor of that task, so I don't know exactly what kind of facts they're expecting.
15:50 PM
Soumya
Okay.... I'll ask after a few minutes of researching...
15:50 PM
Soumya has quit
15:51 PM
CallerNo6
Soumya, also, non-mentors can't follow that link, but might have interesting input. Can you pastebin the task description?
15:51 PM
Sophist-UK has quit
15:57 PM
iliekcomputers
15:59 PM
CallerNo6
iliekcomputers, thanks!
15:59 PM
iliekcomputers
=)
16:09 PM
bukwurm has quit
16:13 PM
16:14 PM
ruaok
on your server?
16:14 PM
iliekcomputers
on my laptop, local small dump :D
16:14 PM
ruaok
k.
16:14 PM
perfect timing. because I hope to have hdfs up and running version soon
16:15 PM
Leo_Verto: that 0.0.0.0 trick worked and in hindsight makes total sense.
16:16 PM
Leo_Verto
yeah, all the time spent digging through docker docs and it's something as simple as binding to the wrong interface…
16:17 PM
ruaok
yeah...
16:20 PM
Leo_Verto
anyway, let me know when spark's running and we can get started with the jupyter notebook server :D
16:21 PM
michelv has quit
16:27 PM
ruaok
looks like it needs more config settings.
16:27 PM
yarn.resourcemanager.hostname defines the hostname which others refer to the service with.
16:27 PM
useless for connecting to the service.
16:32 PM
yarn.resourcemanager.bind-host I bet.
16:57 PM
github joined the channel
16:57 PM
github
[picard] phw opened pull request #1027: PICARD-1375: Handle empty values in metadata sanitation (master...PICARD-1375)
https://git.io/fpkkR
16:57 PM
github has left the channel
17:01 PM
ruaok
iliekcomputers: exciting, I like that PR!
17:02 PM
iliekcomputers
ruaok: yusss. =)
17:02 PM
For streaming I was thinking, a simple flask app with an API and a spark writer on one of the nodes
17:08 PM
ruaok
please elaborate on that thought -- sounds complicated.
17:09 PM
iliekcomputers
the spark nodes will probably not have a direct connection with lemmy right?
17:10 PM
run a small flask app on the nodes to which the `biquery-writer` equivalent in listenbrainz submit new listens
17:11 PM
the flask app then submits to a script running on the nodes which writes them to hdfs
17:11 PM
ruaok
> the spark nodes will probably not have a direct connection with lemmy right?
17:11 PM
they could
17:12 PM
rabbitmq makes a good isolator in case the cluster is not available.
17:12 PM
iliekcomputers
could the spark nodes access rabbitmq?
17:12 PM
ruaok
yes.