10:54 AM
yvanzo
Epic won't disappear from other projects though, it is a known thing.
10:54 AM
CatQuest
sure!
10:54 AM
I was talking solely about inst
10:54 AM
I am *curious* about how the epic ticket type functions
10:54 AM
that was it
10:54 AM
yvanzo
So, instead of "Epic" or "Mini-version", you do want "Research" (or some other name?) with labels field?
10:54 AM
CatQuest
... no
10:54 AM
i want
10:55 AM
I'd like "Mini-version" but perhaps with a labels field for it?
10:55 AM
liek the "epic" ting has
10:55 AM
shivam-kapila
iliekcomputers: I was thinking the favourite artists and releases can also be shown graphically like a circular graph. What do you think about it
10:56 AM
yvanzo
Epic is what iliekcomputers told you: an epic goal that usual depends on several other tickets.
10:56 AM
CatQuest
yea yea yea sure but it needs ot be worked with to be understood
10:56 AM
yvanzo
But what “Mini-version” means?
10:56 AM
CatQuest
I mean. "Mini-version" is something I made reo make for me. idk why it ended up no other projects too
10:56 AM
iliekcomputers
shivam-kapila: circular graph?
10:57 AM
CatQuest
it was just a "min-version" because (at the time) I didn't know what epic did. nore did i have any editing for it
10:57 AM
yvanzo
It ended up no other project because nobody understand what it is, there is no description attached to it.
10:57 AM
CatQuest
¯\_(ツ)_/¯
10:57 AM
i'm sorry abotu that
10:57 AM
yvanzo
No problem, so mini-version = epic?
10:58 AM
CatQuest
basically "Mini-version" is a ticket that collects loads of related tickets
10:58 AM
that's waht I was wodnerin if maybe was a good idea :D when i heard tis epic description
10:58 AM
!
10:59 AM
so if "Mini-version" could have the *abilities* that epic has.. lables
10:59 AM
maybe
10:59 AM
anyway i cant really do anything epic on isnt right now. i don't have that label field
11:00 AM
inst*
11:00 AM
11:00 AM
BrainzBot
INST-392: Veena
11:00 AM
CatQuest
11:00 AM
ehr
11:00 AM
if "Mini-version" is a problematic name..
11:01 AM
CatQuest needs breakfast first
11:02 AM
shivam-kapila
11:08 AM
prabal joined the channel
11:14 AM
sarthak_jain joined the channel
11:17 AM
Nyanko-sensei has quit
11:22 AM
BrainzGit
11:23 AM
MajorLurker has quit
11:28 AM
Nyanko-sensei joined the channel
11:35 AM
sarthak_jain
11:35 AM
BrainzBot
LB-473: ListenBrainz-Labs: Use PySpark SQL Module in place of SQL queries.
11:35 AM
MajorLurker joined the channel
11:39 AM
sarthak_jain
We wish to return a dictionary right, from each function?
11:39 AM
pristine__
Let us say we have a dataframe = df
11:40 AM
We want to select col = user_id from this
11:40 AM
There are two ways for that
11:40 AM
Register the dataframe and then write sql query
11:41 AM
`select user_id from df`
11:41 AM
We are following this scheme in utils.py
11:41 AM
CatQuest
Mr_Monkey: why exactly does all entities have 2 "language" selectors?
11:41 AM
pristine__
Now the other way out is don't register the dataframe but use pyspark sql module
11:42 AM
Like this : `df.select("user_id")`
11:43 AM
We want to update all the sql queries in utils.py to use pyspark module.
11:43 AM
You will have a read the pyspark sql doc a bit. I have provided the link in the task.
11:43 AM
sarthak_jain
Have we used this pyspark.sql module in any file in the listenbrainz_spark?
11:44 AM
pristine__
listenbrainz_spark/recommendations/create_dataframe.py
11:45 AM
But I don't think referring to the file is the best way. You should read the doc once, it is really good :)
11:46 AM
sarthak_jain
Yes sure. I will use the doc.
11:46 AM
One more doubt!
11:46 AM
Since we are passing a table to the function, do I need to now pass dataframe as a parameter?
11:48 AM
or maybe just use this `sqlContext.table("table")` to convert it to a dataframe ?
11:48 AM
Guest84633 is now known as c1e0
11:51 AM
Also pristine__, I dont see any queries listenbrainz_spark/recommendations/create_dataframe.py in this file. '=D
11:53 AM
pristine__
This script is calling other modules where queries are actually written
11:53 AM
You have a lot of time. Please read the code.
11:53 AM
Wait. I will give you the line no. A sec
11:54 AM
11:55 AM
> or maybe just use this `sqlContext.table("table")` to convert it to a dataframe ?
11:55 AM
I have no idea what this line of code is?
11:55 AM
> Since we are passing a table to the function, do I need to now pass dataframe as a parameter?
11:56 AM
Which function? Take your time. I will review when you make a patch :)
11:57 AM
sarthak_jain
This piece will register the table as a df so that I can use df operations
11:57 AM
Okay sure
11:59 AM
pristine__
No
12:00 PM
That's the whole point. We don't want to register tables, that is why we are directly using dataframe
12:00 PM
Did you follow the above examples?
12:01 PM
Tables are not registered as dataframe, we register a dataframe as a view/table to query it.
12:01 PM
sarthak_jain has quit
12:01 PM
Coder44 has quit
12:03 PM
sarthak_jain joined the channel
12:03 PM
sarthak_jain
12:03 PM
yes exactly
12:04 PM
What I am saying is, to perform df operations, I need to paas df as a parameter here, right?
12:04 PM
sarthak_jain has quit
12:05 PM
pristine__
Obviously!
12:06 PM
sarthak_jain joined the channel
12:07 PM
sarthak_jain
Yup cool 😅
12:07 PM
I will get started then
12:07 PM
sarthak_jain has quit
12:07 PM
pristine__
Purpose of utils.py is to hold all the queries together at one place.
12:08 PM
Yeah. Nice
12:14 PM
shivam-kapila
pristine__: iliekcomputers: For testing the stats module how much data should I download from the ftp server dumps?
12:14 PM
pristine__
Hey
12:15 PM
So we are in that process but I can suggest you a workaround
12:16 PM
Checkout to `all-changes-mapping` branch
12:17 PM
12:18 PM
replace this path to `http://ftp.musicbrainz.org/pub/musicbrainz/listenbrainz/incremental/`
12:18 PM
12:19 PM
replace this to `listenbrainz-listens-dump-99-20191107-000002-spark-incremental.tar.xz`
12:20 PM
yvanzo
CatQuest: From what I can see on Veena INST-392, you should be able to transition, link, and label it. Am I missing something?
12:20 PM
BrainzBot
12:21 PM
pristine__
12:22 PM
and run `/usr/local/spark/bin/spark-submit spark_manage.py upload_listens`
12:22 PM
CatQuest
yvanzo: sure..
12:22 PM
but I can't use a "epic link"
12:23 PM
pristine__
The configs are for full dumps which are not ideal for local env. We are in process to make small dumps for local env which will be done soon. Till then you can use incremental dumps to for stats
12:25 PM
shivam-kapila
pristine: Got it. Thanks :D
12:26 PM
pristine__
> and run `/usr/local/spark/bin/spark-submit spark_manage.py upload_listens
12:26 PM
this should be run inside listenbrainz_playground_1 container.
12:26 PM
shivam-kapila:
12:27 PM
shivam-kapila
Okay.
12:28 PM
pristine__
`docker exec -it listenbrainzspark_playground_1 bash`
12:31 PM
shivam-kapila
This is nice. No download required. Wow :p
12:34 PM
pristine__
We are trying to make it better. In time people will have to write just one command to fill in data on their local machine, and in case you want to test something on big machines, a single command to fill in big data.
12:34 PM
thank you :)
12:34 PM
antlarr has quit
12:35 PM
yvanzo
CatQuest: thanks, got it :)
12:36 PM
antlarr joined the channel
12:56 PM
amCap1712
iliekcomputers: worked on the feature. it is working well on the linux build
13:01 PM
c1e0_ joined the channel
13:03 PM
c1e0 has quit
13:07 PM
iliekcomputers
pristine__: can we document the steps you listed somewhere (for now)
13:08 PM
pristine__
Readme?
13:08 PM
shivam-kapila: worked for you?
13:11 PM
iliekcomputers
pristine__: we have a HACKING.md file that seems perfect for stuff like this
13:11 PM
13:12 PM
pristine__
Cool. Thanks.
13:27 PM
prabal has quit
13:59 PM
travis-ci joined the channel
13:59 PM
travis-ci
13:59 PM
travis-ci has left the channel
14:19 PM
Lotheric has quit
14:21 PM
iliekcomputers
Dumps are getting created again, BTW.
14:24 PM
sarthak_jain joined the channel
14:30 PM
Lotheric joined the channel
14:48 PM
shivam-kapila has quit
15:06 PM
yvanzo
CatQuest: should work now, but the migration is going to be cumbersome.
15:30 PM
rahul24 joined the channel
15:36 PM
rahul24 has quit
15:38 PM
rahul24 joined the channel
15:45 PM
rahul24 has quit
15:47 PM
rahul24 joined the channel
15:55 PM
rahul24 has quit
16:00 PM
rahul24 joined the channel
16:05 PM
prabal joined the channel
16:13 PM
sarthak_jain has quit
16:15 PM
rahul24 has quit
16:16 PM
reosarevok