[00:05:57] madhuvishy: back [00:06:14] nuria: I put a print in the sql handler [00:06:21] aham [00:06:59] nuria: like [00:07:02] https://www.irccloud.com/pastebin/F0YAKuGl/ [00:07:09] now this shows up [00:07:12] for everything [00:07:28] except this specific revision of the analytics schema [00:07:37] it makes no sense [00:07:40] madhuvishy: jaja [00:07:43] madhuvishy: boy ... [00:08:01] madhuvishy: let's do a different test [00:08:09] madhuvishy: let's drop the table for old schema [00:08:17] madhuvishy: and send events for that table [00:08:21] *that schema [00:08:23] nuria: am i missing something when I update a schema? I changed the wiki page - got a new revision number - and tried to send events to the new revision [00:08:33] nuria: alright [00:08:39] madhuvishy: no, i saw your events on all-events [00:08:48] nuria: yes! and kafka [00:08:50] anyway [00:08:56] madhuvishy: 1 by 1 [00:09:04] madhuvishy: let's test autoincrement table creation [00:09:13] nuria: ya okay [00:10:55] madhuvishy: try that and let me know [00:19:04] nuria: no table [00:19:12] madhuvishy: and no logging? [00:21:19] madhuvishy: makes no sense cause i dropped a table just recently to test toku db in beta (just last week) [00:21:27] madhuvishy: and it got created np [00:21:44] nuria: I see logs from the handler [00:21:48] but not from jrm.py [00:22:08] nuria: yeah! I sent data to a non existent table yesterday [00:22:11] madhuvishy: did you log with print? [00:22:14] and the table got created [00:22:20] no - i used logging [00:22:28] in both places [00:22:29] so [00:23:14] madhuvishy: ok let's rollback your change & build & stop and restart [00:23:29] madhuvishy: and try to create table again [00:23:37] nuria: it may be the mysql tokudb thing [00:23:38] hold on [00:23:48] k [00:26:52] nuria: okay [00:26:55] it was that [00:27:00] aham [00:27:04] default_storage_engine inside the file [00:27:14] default-storage-engine is the command line option [00:27:15] my.cnf [00:27:22] ah ok [00:27:24] i did not know that [00:27:33] so there was an innodb vs tokudb [00:27:36] nuria: although - the dropped table got created [00:27:48] i think the default in my.cnf before was Aria [00:27:49] in 300 ms [00:28:06] but - I dont see the new revision table being created [00:28:07] yet [00:28:31] or showing up in the logs [00:28:35] so that part is weird [00:28:41] ok so table with old schema got created ok [00:28:48] yes [00:28:54] with autoincrement id [00:29:13] yup! [00:29:20] let me look at schema versions, what is schema AnalyticsTest? [00:29:39] https://meta.wikimedia.org/wiki/Schema:Analytics [00:29:45] https://meta.wikimedia.org/wiki/Schema:Analytics [00:29:46] yes [00:29:59] nuria: I added the height field today [00:30:07] 15332607 is the new revision [00:30:24] 13317883 is the old one [00:31:58] boy i ahve no explanation [00:32:00] *have [00:33:27] nuria: me too - and it seems to not even be tied with this autoincrement change [00:33:53] madhuvishy: yaaaa, not only that events are duplicated in beta labs, but that could be varnish [00:34:18] nuria: ah yes [00:34:25] hmmmm [00:35:08] madhuvishy: but i already verfied msgs are duplicated in kafka topic [00:35:16] eventlogging_valid_mixed [00:37:40] nuria: ah yes i also see some repetition in the logs - but this new revision events not showing up is weirdd [00:37:52] madhuvishy: but they show up in logs [00:38:10] nuria: not in the mysql handler logs though [00:38:19] they get inserted everywhere but mysql [00:38:23] madhuvishy: right, right, they show up on all events log [00:38:30] yup [00:38:53] they get all the way eventlogging_Analytics - and for some reason the mysql consumer ignores them [00:40:00] madhuvishy: i think i know [00:40:07] nuria: oh? [00:40:14] {"clientIp": "08148e4cbe16649292dfe542124ee80292070454", "event": {"height": 31, "name": "yC"}, "recvFrom": "29", "revision": 15332607, "schema": "Analytics", "seqId": 121643, "timestamp": 1455050642, "userAgent": "\"-\"", "uuid": "718a4f5eda685c5c909bf54a2e10166d", "wiki": "Xe"} [00:40:19] height doesn't have quotes [00:40:32] nuria: why should it? [00:41:01] nuria: also if it's a validation issue why will it be valid anyway [00:41:04] madhuvishy: ah no, it is a number [00:41:13] madhuvishy: ya, validation according to schema is ok [00:41:13] yes [00:41:45] madhuvishy: but even then the json of the schema needs to be parsed as text to be inserted [00:42:16] madhuvishy: but no, as long as it is a number (which it is) [00:42:19] that is fine [00:44:44] madhuvishy: and you also have logging to table lookup [00:46:31] madhuvishy: there are exceptions on the logging though: [00:46:33] https://www.irccloud.com/pastebin/cHP2XwuX/ [00:46:46] that will kill the thread [00:48:03] madhuvishy: see: 2016-02-09 21:56:49,144 (MainThread) Driving [00:48:04] kafka:///deployment-kafka02.deployment-prep.eqiad.wmflabs:9092?topic=eventlogging-valid-mixed&blacklist=^Analytics|CentralNoticeBannerHistory$&zookeeper_connect=deployment-zookeeper01.eqiad.wmflabs/kafka/analytics-deployment-prep&auto_commit_enable=True&auto_commit_interval_ms=1000&auto_offset_reset=-1&identity=mysql-m4-master -> [00:48:04] mysql://eventlogging:68QrOq220717816UycU1@127.0.0.1/log?charset=utf8&statsd_host=labmon1001.eqiad.wmnet&replace=True.. [00:48:21] analytics is on the blacklisted list? [00:50:32] madhuvishy: ok, that must be it [00:51:19] madhuvishy: did your removed teh blacklist? [00:52:41] nuria: really? no i dint but the old revision events went through [00:53:17] madhuvishy: analytics table event #1 is | 1 | 1e0405e0bb865a529ac26958f1b9fbcb | a6fc10fe8efa9e47f5ecf23e8491229bb866ed72 | 20160209025919 | "-" | NULL | yv | 67 | YY | [00:53:30] with date 20160209025919 [00:54:07] let's drop that table again, how were you generating events? [00:54:11] nuria: right - the old one [00:54:50] nuria: i'm using the eventlogging-load-tester script [00:55:00] madhuvishy: k [00:55:02] like [00:55:05] python bin/eventlogging-load-tester -s 'Analytics:15332607:1' --optional-values always 10 [00:55:13] this is for new revision [00:55:18] nuria: this table doesn't exist [00:55:28] so i agree its part of the blacklist [00:55:35] right [00:55:45] but i dont understand how it got created for the old revision yesterday [00:56:17] madhuvishy: ya... [00:56:51] madhuvishy: i dropped table, it should not be created even if you run the load tester [00:57:19] madhuvishy: try to run it and let's wait 3 mins [00:57:33] madhuvishy: after we will remove the blacklist and re-start [00:59:58] madhuvishy: let me know when you have run your scripts [01:00:01] nuria: [01:00:04] yes [01:00:05] it got created [01:00:08] whatttt [01:00:11] Analytics_13317883 is backkk [01:00:13] ha ha ha ha [01:00:21] now its a new problem [01:00:40] madhuvishy: GOOD JOB Finding that bug [01:00:46] jajaja [01:01:03] apart from Donald Trump leading republican new hampshire primaries [01:01:21] nuria: ha ha [01:01:35] i'm going to try to revise some other schema that is not blacklisted [01:02:01] i ay ay [01:03:16] madhuvishy: ya, ok, i think autoincrement changes are fine [01:03:26] nuria: yup i think so too [01:03:42] madhuvishy: but do test with one schema that is not on blacklist [01:03:47] madhuvishy: feel free to drop tables [01:03:55] madhuvishy: that is why we have beta [01:04:18] nuria: yup I think I tested that with SendBeaconReliability [01:04:20] it's fine [01:04:52] madhuvishy: k, just file a ticket for the blacklist.. do a new batch of events go on.. i wonder if it is only table creation [01:05:04] madhuvishy: we have 1281 on table [01:05:09] madhuvishy: can you run script again? [01:05:24] nuria: sent some more [01:05:30] it does go [01:05:45] so much for that blacklisting. [01:05:48] he he [01:05:55] the regex is off [01:05:56] it works for new revisions but not old? [01:05:59] makes no sense [01:06:05] ya i'll file a bug [01:06:10] madhuvishy: well... i think taht is another problem [01:06:25] let me change config [01:06:30] i think is teh "^" [01:07:43] madhuvishy: try again [01:07:47] i just dropped table [01:08:15] madhuvishy: and changed regex [01:08:23] nuria: okay i sent some events [01:08:39] and there's the table [01:09:07] nuria: i wonder if it's cached somewhere - something is offf [01:09:37] madhuvishy: all right.. there goes my theory [01:09:45] madhuvishy: let's file a bug for this one [01:09:50] nuria: yup [01:10:07] madhuvishy: and when you are done testing autoincrement we can deploy to prod (tomorrow) [01:10:19] nuria: i think it's good [01:11:38] madhuvishy: k, let's deploy tomorrow, write to SAL etc [01:11:46] madhuvishy: and coordinate with jaime [01:12:26] nuria: cool. also i see the git remote for /srv/deployment/eventlogging in beta is http://deployment-bastion.eqiad.wmflabs/eventlogging/eventlogging/.git [01:12:36] is there a method to deploy to beta? [01:13:22] madhuvishy: one that works? ... jaja, no. i have always had to pull [01:13:35] madhuvishy: since teh dawn of the times [01:13:38] nuria: ahhokay [01:29:25] Analytics, Analytics-EventLogging: Some blacklist matching schemas are being consumed by Eventlogging {oryx} - https://phabricator.wikimedia.org/T126410#2013814 (madhuvishy) NEW [01:30:27] Analytics: Prepare for lightning talk on Last access uniques - https://phabricator.wikimedia.org/T126411#2013821 (madhuvishy) NEW a:madhuvishy [05:03:42] Analytics-Wikistats, operations, Regression: [Regression] stats.wikipedia.org redirect no longer works ("Domain not served here") - https://phabricator.wikimedia.org/T126281#2014030 (Peachey88) Is the redirected used all that much? Perhaps someone from #operations could have a look at the stats for it. [08:28:49] o/ [08:49:28] Analytics, ContentTranslation-Analytics, MediaWiki-extensions-ContentTranslation, operations, and 4 others: schedule a daily run of ContentTranslation analytics scripts - https://phabricator.wikimedia.org/T122479#2014195 (Arrbee) Open>Resolved [08:51:16] (PS1) DCausse: Ensure that the schema_repo git submodules are available before packaging [analytics/refinery/source] - https://gerrit.wikimedia.org/r/269627 [09:19:20] Hi elukey :) [09:38:00] dcausse: good morning ! [09:38:04] dcausse: you here? [09:49:35] joal: hi! [09:49:52] Hey dcausse : just sent an email [09:50:09] dcausse: Wanted to let you know that everything is fine on camus this morning [09:50:31] joal: thanks! [09:50:40] np dcausse :) [09:50:47] sorry for the mess yesterday [09:56:20] joal: hey np, it happens. it's our fault because we should monitor this data more carefully [09:57:04] dcausse: shared error (I had a bad feeling while deploying with ottomata about changing jars that don't actually change) [09:57:27] dcausse: I'll add that to my today's discussion sheet :) [09:57:34] :) [10:05:19] (PS1) Joal: Correct bug in last_access_uniques jobs [analytics/refinery] - https://gerrit.wikimedia.org/r/269645 [10:55:43] Analytics-Kanban: Add the possibility to introduce new email templates to Burrow - https://phabricator.wikimedia.org/T126008#2014419 (elukey) Change has been merged! [10:55:53] Analytics-Kanban: Add the possibility to introduce new email templates to Burrow - https://phabricator.wikimedia.org/T126008#2014420 (elukey) Open>Resolved [10:56:31] Analytics-Tech-community-metrics: Korma: Incorrect ticket count on top contributors page - https://phabricator.wikimedia.org/T126328#2014421 (He7d3r) and now it shows only 7 tickets: | Rank | Name | IRC | MW Rev | Tickets | Source Rev | ML | ----- | ----- | ----- | ----- | ----- | ----- | ----- | 38 |... [11:55:17] (CR) DCausse: "Well.. in fact the exec plugin that runs git submodule update is mandatory, the release plugin will clone a fresh repo from the tag when b" [analytics/refinery/source] - https://gerrit.wikimedia.org/r/269627 (owner: DCausse) [13:31:48] Analytics-Tech-community-metrics: Korma: Incorrect ticket count on top contributors page - https://phabricator.wikimedia.org/T126328#2014709 (Aklapper) Open>Invalid a:Aklapper The algorithm takes contributor **positions** (emphasis by me) in each data source activity. It is a rank not a number of ti... [14:13:41] Analytics-Tech-community-metrics: Korma: Incorrect ticket count on top contributors page - https://phabricator.wikimedia.org/T126328#2014788 (He7d3r) Oh, cool! I must have read that description too fast and incorrectly... Maybe it would be helpful to add a tooltip to the numbers, as in "This contributor is r... [14:15:22] Is it expected that only the top 10 users are ranked for participation on mailling lists? [14:15:22] http://korma.wmflabs.org/browser/top-contributors.html [14:15:41] (for the others the page shows a "-" instead of a rank) [14:16:52] What is the difference between the columns "Mediawiki Reviews" and "Source code review"? [14:55:06] joal, mforns: any idea --^ [14:55:18] I am a bit ignorant on the subject :) [14:58:47] no idea elukey [15:00:11] I'm completely ignorant on the subject :) [15:02:18] Helder: woud you mind to wait for US people to come online? They will surely have a better understanding [15:08:57] I'll probably be here [15:28:08] joal / elukey (not pinging otto 'cause he's sleeping): I was wondering what you thought about backfilling pageviews back to May 2015: http://dumps.wikimedia.org/other/pageviews/ [15:28:23] right now we only have projectviews back past October [15:29:05] it would be quite a bit of work, because the files are hourly. I'm thinking we could do it in Erik Zacthe's compressed format and those would be daily [15:29:35] milimetric: no idea about the work to do sorry :( [15:29:44] milimetric: If we do it with oozie and parallelize nicely (not too much), should be ok to have it with current format [15:29:53] elukey: no sok, I can explain [15:30:17] elukey: so all these jobs are done, they just got started with a start date of October so we don't overload the cluster [15:30:24] It would take some time to get done, but would certainly get done :) [15:30:33] so I'm just asking for capacity and Jo seems to think we have it [15:31:04] milimetric: On the look run we have capacity, it's just the matter of not being bursty :) [15:31:09] elukey: here's the job and its start: https://github.com/wikimedia/analytics-refinery/blob/master/oozie/pageview/hourly/coordinator.properties#L36 [15:31:12] uh... that says May :) [15:31:14] wait a minute [15:31:24] k, makes sense [15:31:29] yeah, so we have to make sure we launch it nicely [15:31:44] milimetric: We still would have to make a dedicated job to only dump pageviews, not recompute them [15:32:25] cause for the moment the job extracts pageviews from webrequest and then dump them - we only would keep the dump part of it on existing extracted data [15:32:50] mmm, right, so a trimmed workflow and a new coordinator [15:32:59] yessir [15:33:02] I can do that, that's fine [15:33:21] thx, good to know about capacity if we take care to not burst [15:33:22] milimetric: Don't bother too much -- It would be a run-and-garbage code [15:33:29] yeah, 'course [15:33:49] milimetric: it also depends on the complexity of the runs and the size of the data, obviously :) [15:34:35] milimetric, joal: if it is not too urgent I'd like to do the work [15:35:04] it's not urgent, no, and I can show you how I do it, maybe I can learn it better in the process [15:35:07] elukey, milimetric : Can I let you prepare that, and possibly review before launch ? [15:35:07] I'll file a task [15:35:19] it would be interesting to know 1) how to build a oozie job, 2) how to check the cluster's capacity [15:35:25] super :) [15:36:13] atm I still need to reimage 16 servers with jessie for ops, so I'll probably get to it on Friday/next week [15:36:29] joal: are you happy that I'll bother you a lot during the next days? :P [15:36:32] kidding :) [15:36:56] elukey: I'm off for end-of-wee, so you can try :-P [15:37:07] milimetric: would you mind to check Helder's questions posted before? [15:37:15] ahhhhh [15:38:38] a-team, need to go and grab Lino from the creche [15:38:47] Will see you when he's asleep [15:41:16] Analytics: Back-fill pageviews data for dumps.wikimedia.org to May 2015 - https://phabricator.wikimedia.org/T126464#2015119 (Milimetric) NEW a:elukey [15:41:41] \o/ [15:41:47] :) [15:42:30] so you can bother me with building an oozie job but I'm not super sure how to check on the cluster health, I'd just stumble through the docs on wikitech [15:42:42] I'm *awesome* at building an oozie job [15:43:04] [15:43:21] :) [15:53:52] milimetric, yt? [15:53:58] yes, hi :) [15:54:18] hey! do you have 10 mins to troubleshoot d3 with me? [15:54:38] I'm facing a poltergeist bug [15:55:38] to the batcave! [15:55:41] :] [16:00:21] milimetric, I lost you [16:06:42] mforns: nuria's not around yet, so we can keep chatting if you want [16:06:52] milimetric, ok [16:06:56] batcave? [16:07:01] yep [16:18:52] dcausse: hihiii [16:18:53] morning [16:18:59] well, late morning for meeee :) [16:29:20] ottomata: o/ [16:30:54] hiyaa [16:32:07] ottomata: hi! [16:32:52] dcausse: ok, i think we ahve standup starting now buuUuuut [16:32:56] things mostly looking good [16:33:07] its a little messy, so and i'm worried about data in a couple of hours on feb 3 [16:33:17] but, most of the data from last week is all coming back now [16:33:29] overnight i imported to my home dir [16:33:39] am now moving it back in place, and then letting the normal camus do its thing [16:38:25] Analytics-Kanban, Patch-For-Review: camus-wediawiki job should run in production (or essential?) queue {hawk} [1 pts] - https://phabricator.wikimedia.org/T125967#2015242 (Milimetric) a:JAllemandou [16:48:16] mforns: oops, we should've stayed :) [16:48:20] dcausse: ok [16:48:24] but maybe just push the code so I can start checking it out too [16:48:32] milimetric, hehehe, omw [16:48:35] I'm gonna grab lunch soon, but I'll look at it [16:49:00] hmmm, ok yeah, i am missing data from hours 14 and 15 on feb 3 [16:49:14] earliest ts in kafka is Wed Feb 3 13:26:24 UTC 2016 !!! [16:49:15] hahah [16:49:16] ah! [16:49:20] i have 30 mins to save this data..... [16:49:21] ON IT [16:51:56] (PS1) Mforns: Add hierarchy visualization [analytics/dashiki] - https://gerrit.wikimedia.org/r/269713 (https://phabricator.wikimedia.org/T124296) [16:52:41] (PS2) Mforns: Add hierarchy visualization [analytics/dashiki] - https://gerrit.wikimedia.org/r/269713 (https://phabricator.wikimedia.org/T124296) [16:53:23] haha (not true about 30 mins...actually i think kafka could delete this data at any time :/) [16:53:25] importing now [17:04:16] i remember something about eventlogging data being available in hive, but not finding it. Where could i find it? Looking to do some heavier analytics/session reconstruction/etc on the data we collect about search [17:05:11] https://wikitech.wikimedia.org/wiki/Analytics/EventLogging#Access_data_in_Hadoop [17:05:12] ebernhardson: ^ [17:05:25] ahha, i should have looked harder in wikitech :) [17:05:32] :) [17:05:36] i find it easier to access in spark [17:05:41] since you don't need to create a table first [17:05:41] ottomata: sorry was in meeting [17:06:04] ottomata: yea i think spark will actually be easier for session reconstruction and whatnot as well [17:07:04] dcausse: oh, whoa [17:07:09] just read your comment on that patch [17:07:17] so mvn release + submodules don't work eh? [17:07:36] ottomata: yes.. it was not really your fault, even with a proper git submodules update the build would have been broken anyways [17:07:56] huh, how did we release 0.0.23 with them there??? [17:07:58] yes mvn release:perform will checkout a fresh repo [17:08:10] it was not a git submodule at that time [17:08:16] ahhh [17:08:18] hm [17:08:31] ok, i think we need a task to figure out a better way to share these schemas! [17:08:38] we kinda want that for eventbus too, but maybe that is a separate task [17:08:38] hm [17:08:39] best of both world maven release and git submodule :) [17:09:17] dcausse: do the schemas have to be present at build time? [17:09:30] could we make it take a local dir path, and it'd just look them up at run time? [17:09:52] ottomata: I think it would be way better to not have them in the jar yes [17:10:07] yeah, and we can easily deploy event-schemas on its own [17:10:08] and the build does not care about specific schemas [17:10:12] aye [17:10:27] it would only if you were actually using specific class bindings in your code for it [17:10:35] and it would allow to do schema upgrade without builder refinery-camus [17:10:39] ja [17:11:00] ottomata: no we don't use class bindings, only generic stuff [17:11:03] aye [17:12:30] so, in thory we could deploy the event schemas to hdfs, or maybe just to analyitcs machines via trebuchet? [17:15:34] ja for sure [17:25:59] (CR) Ottomata: [C: 2 V: 2] Correct bug in last_access_uniques jobs [analytics/refinery] - https://gerrit.wikimedia.org/r/269645 (owner: Joal) [17:27:34] sp dcausse should I even merge that refinery pom change? [17:27:45] or, should we just create a ticket to fix this? [17:30:36] (CR) Ottomata: "Do we still need this? Should we abandon?" [analytics/camus] - https://gerrit.wikimedia.org/r/242907 (owner: Nuria) [17:32:55] ottomata: I don't know... it depends if you plan to make future release ? [17:33:20] if a release is needed like a hotfix we should merge it before [17:34:14] hmm, but i thought it won't work anyway? [17:36:11] my patch? [17:36:51] maybe I was not clear, without the pom.xml patch all releases will be broken [17:37:52] oh, i thought you were saying that mvn release wouldn't work even with your patch [17:38:24] no I think it works and is actually needed [17:39:09] oh ok [17:39:10] cool [17:39:11] merging then [17:39:20] thanks much for that [17:39:28] (CR) Ottomata: [C: 2 V: 2] Ensure that the schema_repo git submodules are available before packaging [analytics/refinery/source] - https://gerrit.wikimedia.org/r/269627 (owner: DCausse) [17:40:00] ottomata: thanks, I'll cleanup everything when we move these schemas out of the jar [17:42:29] ok cool [17:57:15] (PS3) Milimetric: [WIP] Add hierarchy visualization [analytics/dashiki] - https://gerrit.wikimedia.org/r/269713 (https://phabricator.wikimedia.org/T124296) (owner: Mforns) [18:07:51] logging off a-team! talk with you tomorrow! [18:08:25] elukey: night! [18:08:30] laters! [18:08:31] later Luca, have a nice night :) [18:09:41] Hi ottomata, quick heads up on camus for media-wiki [18:10:10] I looked into camus this morning, it looked fine, so I assumed everything was back in place - Sorry for the false joy dcausse :( [18:10:19] its all cool now! :) [18:10:33] Thanks for having triple checked ! [18:10:37] yes all data is here just need to run oozie [18:11:03] right dcausse ... It's a pain, but should work [18:14:43] ottomata: You've had to re-import everything using a fresh camus for mediawiki ? [18:15:36] joal: batcave real quick? [18:15:40] sure ottomata [18:31:14] oop, sorry joal! [18:31:22] were you just saying sometihng? [18:38:16] nah, nothing ottomata :) [18:38:40] k laters, heading to cafe...back in a bit [18:42:59] madhuvishy: mobrovac was interested in your dockerization of wikimetrics, because they have dockerized some mediawiki stuff as well [18:43:15] talk amongst yourselves :) [18:43:24] milimetric: ah yes we chatted about it a little bit at docker meetup yesterday :) [18:43:24] yup, madhuvishy mentioned it yesterday at the docker meet-up :) [18:43:30] haha [18:43:33] :D [18:43:40] ah cool [18:44:31] madhuvishy: https://github.com/wikimedia/mediawiki-containers [18:44:34] mobrovac: I wrote a little bit here - https://github.com/wikimedia/analytics-wikimetrics#development-environment - the compose file and Dockerfile are in the root too [18:44:44] gr8 [18:44:45] thnx [18:44:57] mobrovac: cool thanks :) [18:45:12] maybe we could integrate the two [18:45:56] mobrovac: wikimetrics doesn't need mediawiki though [18:46:12] true [18:46:27] it's just a flask app with celery/redis and mysql [19:03:53] halfak: Hello o/ [19:04:30] o/ joal [19:04:41] you looked for me yesterday [19:05:20] Oh! Yeah. It was regarding the location of the jar file we needed on the IA cluster [19:05:23] But we found it :) [19:05:33] Horray for old etherpads :) [19:05:33] Arf :) [19:05:40] For sure ! [19:05:51] script is working ? [19:06:06] Not quite yet. Last I left it in schana's hands. [19:06:16] Mwarf :( [19:06:29] We were having trouble getting subprocess to pass arguments to your java stuff in ways that it understands. [19:06:32] Weirdly enough. [19:06:41] Weird indeed ! [19:07:07] Do you know if your script should handle something like: --timeout '500'? [19:07:07] I have a branch that I switched over to argparse (for type safety stuff), but haven't tested it yet [19:07:25] Wait. python argparse? [19:07:30] yes [19:07:32] I sent a pull request with a first version of dump_downloader [19:07:38] Boo [19:08:00] Argparse is a close second to docopt IMO. [19:08:08] But close enough that it's OK. [19:08:19] * halfak <3's the self-documenting nature of docopt. [19:08:34] idk, once you start doing a bunch of casting all over the place, I think argparse pulls ahead :) [19:08:43] schana: one trick to try if not yet done: separate the arg flag and its value in two separeted values in the arg list [19:08:57] it was like that initially [19:09:18] Mwarf [19:09:51] schana, that's cool. We'll end up butting heads a bit, but I do like argparse. It's the best modular library for argument parsing available. [19:10:51] I don't like losing the self-documenting bit, and was more-so doing it to test if it had any bearing on subprocess (which I kinda doubt it will) [19:11:20] That's really weird this argument problem :( [19:11:44] schana, if only we could have some sort of hybrid doc-parse or something that could capture the best of both. [19:13:45] I think overall I like the pattern of generating documentation from code and not the other way around. The other way I could think to link them is (I think it's called) doctest showing the --help option [19:15:55] Analytics, Discovery: Look into encrypting logs sent between mediawiki app servers and kafka - https://phabricator.wikimedia.org/T126494#2015899 (EBernhardson) NEW [19:15:57] who's talking about docopt?! [19:16:07] schana: can you explain a bit more on how argparse and hadoop job not running in java ? [19:16:27] Ah ottomata, you're back :) [19:16:39] I have something I'd like to discuss around kafka [19:16:45] oh ja? [19:16:48] if not now, next week :) [19:17:21] Yeah, the issue you solved with elukey yesterday (disk full) --> SHouldn't we augment the number of partitions for big webrequest topics ? [19:17:58] joal: yesterday we were getting errors regarding the arguments being passed through subprocess being of the wrong type when making the call to hadoop [19:19:59] schana: hadoop complaining ? [19:20:20] schana: if so, could be a bug in my hadoop job implementation ! [19:21:05] joal - I don't have the stack trace, so I'll have to reproduce it before I know (trying now) [19:21:58] ottomata: 4 lines above ;) [19:22:26] joal: not sure if that would help? [19:22:35] it could possibly make the disks more balanced [19:22:35] milimetric, do we have a phab ticket for data? [19:22:44] but the issue we solved yesterday would still happen [19:22:54] kafka still would wait too long to purge old data [19:22:55] ottomata: high probability that data got spread on more disks, no ? [19:22:59] after a boker recovery [19:23:13] Analytics-Kanban, Editing-Analysis, Patch-For-Review, WMF-deploy-2016-02-02_(1.27.0-wmf.12), WMF-deploy-2016-02-09_(1.27.0-wmf.13): Edit schema needs purging, table is too big for queries to run (500G before conversion) {oryx} [8 pts] - https://phabricator.wikimedia.org/T124676#2015961 (Jdlrobson... [19:23:18] joal: we think the better solution for the disk balancing is to use raid [19:23:19] ottomata: worst case scenario: 14 days of partition to store [19:23:29] joal: indeed [19:23:41] raid would also help us slightly reduce downtime [19:23:53] in that a single disk failure wouldn't bring a broker down [19:24:07] and would help with that issue of consumers getting bad data for non fatal disk problems [19:24:50] makes sense for raid [19:25:04] but ja, that is a longer term change [19:25:13] might not change it until we replace kafka hardware [19:25:19] which we didn't budget for this year [19:25:31] could always adjust budget allocations, we will see [19:25:35] ottomata: The other (less important reason then) is to have more parallel readers [19:25:52] ja indeed, if we notice that is a bottleneck anywhere, then for sure [19:25:55] we can increase [19:26:14] But yeah, if we have no raid soon, then I think adding partitions is an easy mitigation [19:26:56] true [19:27:47] ok ottomata, you know my view on that :) [19:27:59] You're our ops master, I let you prioritise that (or not) [19:28:01] :) [19:29:17] :) [19:29:21] i think its not a bad idea [19:29:39] joal, its also related to more topics [19:29:46] i'd really like webrequest to be split up into more topics [19:30:02] maybe. [19:30:02] heh [19:30:10] as is, it would make oozie wrangling harder [19:30:12] ottomata: yeah, particularly more homogeneous topics in term of data size [19:30:13] if we need more coordinators [19:30:28] but, maybe there's a way we can make oozie smarter [19:30:30] true as well [19:30:31] about the topics [19:30:33] dunno [19:30:35] something to think about [19:30:39] yup [19:30:50] hm, ottomata : [19:30:57] joal: maybe in terms of datasize, maybe not, maybe just more with more logical groupings. project perhaps? who knows [19:31:08] or maybe both! [19:31:34] ottomata: if topics on the kafka side are not interesting for users on the data side, then we can have a job taking multiple path as inputs and put everything in a single output folder [19:31:47] yeah, don't know [19:31:50] :) [19:32:26] Ok a-team, my time for today :) [19:32:48] schana, halfak: I leave and will be off the end of week, I'll catch next week :) [19:33:09] Have a good (US) holiday joal! [19:33:26] Remember that the "office" is closed on Monday :) [19:33:33] Thanks halfak :) [19:33:37] Ah, forgot about that [19:33:47] I'll be in my home office nonetheless :) [19:33:57] Holidays are Thursday and Friday for me :) [19:34:40] Gotcha. Enjoy your time away. :) I'll probably see you around on Monday. [19:34:46] But maaaayybeee not :D [19:35:23] :D [19:35:26] Bye bye ! [19:35:53] Analytics, Discovery: Look into encrypting logs sent between mediawiki app servers and kafka - https://phabricator.wikimedia.org/T126494#2016008 (Ottomata) TLS is in Kafka 0.9, and we aren’t planning on upgrading soon. Depending on your volume, it may be possible to use the main (non analytics) Kafka clu... [19:36:23] laters! [19:40:52] Analytics: Load Avro schemas from configurable external path - https://phabricator.wikimedia.org/T126501#2016018 (Ottomata) NEW [19:47:46] (PS1) Catrope: Follow-up 52ac1f78d1: actually add beta feature graph to dashboard [analytics/limn-flow-data] - https://gerrit.wikimedia.org/r/269761 [20:07:11] yurik: you should be more specific, what about data? [20:07:42] milimetric, some phab ticket that tracks all the data-related stuff we have been talking about - like storing data blob in a wiki page [20:10:44] Hey folks. I'm trying to work out how the JSON in schemas gets converted to the cool visual representation you can see here: https://meta.wikimedia.org/wiki/Schema:Analytics [20:10:58] It looks like this table has the class "mw-json". [20:11:10] So it might be custom code for MediaWiki/Schemas [20:11:35] I'd like to use something similar with JSON blobs re return via ORES. [20:11:53] See our simplistic UI here: http://ores.wmflabs.org/ui/ [20:12:12] halfak: yeah it's an extension [20:12:25] I don't know much - but let me see if i can find the code [20:12:58] I'm hoping that I can find a snippet of JS that renders the table and provides styling. [20:13:04] * halfak doesn't hold his breath. [20:13:06] Could be in PHP [20:13:14] Thanks for your help madhuvishy :) [20:16:19] halfak: https://github.com/wikimedia/mediawiki-extensions-EventLogging i think [20:16:54] not sure [20:17:03] "schemaschema" :)))) https://github.com/wikimedia/mediawiki-extensions-EventLogging/blob/master/schemas/schemaschema.json [20:17:34] :) [20:17:46] ori would be the person to ask about it though [20:18:12] Gotcha [20:19:36] (PS2) Catrope: Follow-up 52ac1f78d1: actually add beta feature graph to dashboard [analytics/limn-flow-data] - https://gerrit.wikimedia.org/r/269761 (https://phabricator.wikimedia.org/T114111) [20:20:26] Hello analytics people [20:20:32] Could I get a quick review of https://gerrit.wikimedia.org/r/#/c/269761/ ? (one-line patch) [20:21:18] milimetric: ^ [20:23:09] i've noticed dbstore1002.eqiad.wmnet has been disconnecting me more often lately on expensive queries, am i right to guess that is the server protecting itself and i should either reformulate the queries, or run them in hive/spark (its against eventlogging data) [20:24:54] (CR) Milimetric: [C: 2 V: 2] Follow-up 52ac1f78d1: actually add beta feature graph to dashboard [analytics/limn-flow-data] - https://gerrit.wikimedia.org/r/269761 (https://phabricator.wikimedia.org/T114111) (owner: Catrope) [20:25:14] Thanks milimetric [20:26:07] RoanKattouw: looks like the data's not there yet, but the graph is added: http://flow-reportcard.wmflabs.org/ [20:26:46] RoanKattouw: ah, it's because it's broken down by wiki: http://datasets.wikimedia.org/limn-public-data/flow/datafiles/flow_betafeature/ [20:28:59] (by broken down by wiki I mean: https://github.com/wikimedia/analytics-limn-flow-data/blob/master/flow/config.yaml#L68) lemme know if you are confused :) [20:30:27] halfak: https://github.com/wikimedia/mediawiki-extensions-EventLogging/blob/master/includes/JsonSchemaContent.php#L137-L178 [20:30:45] that's where the meat of it lives [20:30:57] Thanks ori [20:31:46] halfak: np; legoktm knows this part of the stack quite well too, so he's a good person to ask questions about how to re-use / adapt / extend that code [20:33:10] milimetric: Oh, I see. Is there a way to graph that in a nicer way? [20:33:35] ContentTranslation has some sort of dashboard for their beta feature uptake, but it looks different [20:33:37] RoanKattouw: you want all the wikis on the same graph? [20:33:50] That could work [20:34:04] :) it's harder - I need to know what you need first [20:34:15] so is one number important or the number measured across wikis [20:34:15] Looking at what CX has [20:34:21] https://language-reportcard.wmflabs.org/#empty [20:34:25] sorry [20:34:26] https://language-reportcard.wmflabs.org/ [20:34:26] Each wiki is interesting separately [20:34:29] Yeah [20:34:35] ok, so then you have two options [20:34:41] I had https://language-reportcard.wmflabs.org/#projects=ptwiki,idwiki,eswiki,viwiki,ukwiki,dewiki,trwiki,ruwiki,frwiki/metrics=Content%20Translation in my history but the hash part doesn't work [20:34:51] 1.) make a custom limn graph instead of letting it infer from the datafile you pass it [20:34:59] 2.) make a separate dashboard [20:35:08] (i.e. the sum across all wikis is not a particularly interesting number; not totally valueless but not really what we care about) [20:35:21] OK [20:35:26] and I guess 3.) change the flow_betafeature.sql script to just select one column per wiki [20:35:38] 3.) is the easiest [20:35:52] but also the hardest to add a new wiki when you want to do that (not super hard, just harder) [20:36:03] Re #3, what do you mean? [20:36:09] It's only selecting date and ount [20:36:11] *count [20:36:37] Or do you mean rewriting it as SELECT 'blah' as wiki, date(...) as date, count(*) as count from .... ? [20:36:49] no, more like: [20:37:12] uh, oh wait, this is from different dbs [20:37:13] nasty [20:37:19] I've gotta run, I'll be back in 30 [20:46:59] (Abandoned) Nuria: [WIP] Changes to camus to debug/test avro in 1002 [analytics/camus] - https://gerrit.wikimedia.org/r/242907 (owner: Nuria) [21:31:58] (PS1) Milimetric: Suggest ugly sql [analytics/limn-flow-data] - https://gerrit.wikimedia.org/r/269834 [21:32:06] RoanKattouw_away: ^ that's the suggestion for the ugly sql [21:32:16] you can see how it'll get worse as you add more wikis [21:32:31] that's what Lang. team was doing before they just said ew and made the new dashboard [21:32:40] the new dashboard is actually not that hard, I'm happy to set it up for you [21:32:59] and that should work with your data the way you have it now [21:35:39] That would be cool too [21:35:47] I tried to set this up based on the language stuff anyway [21:36:08] Oh jeez, I just saw the SQL [21:36:13] Yeah let's do the separate dashboard thing [21:36:22] That lets you toggle individual wikis too, right? [21:36:32] In case we ever have a very large number of wikis there [21:47:53] Quarry: Make available more options for number of shown rows of resultset (Quarry) - https://phabricator.wikimedia.org/T126540#2016800 (XXN) NEW [21:50:15] yes, you can toggle [21:50:46] k, doing [21:51:06] Quarry: Add page navigation on top also (Quarry) - https://phabricator.wikimedia.org/T126542#2016825 (XXN) NEW [22:21:37] (PS1) Milimetric: Move flow_betafeature to limn-language-data [analytics/limn-flow-data] - https://gerrit.wikimedia.org/r/269851 [22:21:58] (PS1) Milimetric: Add flow beta enablements [analytics/limn-language-data] - https://gerrit.wikimedia.org/r/269854 [22:22:14] (CR) Milimetric: [C: 2 V: 2] Move flow_betafeature to limn-language-data [analytics/limn-flow-data] - https://gerrit.wikimedia.org/r/269851 (owner: Milimetric) [22:22:21] (CR) Milimetric: [C: 2 V: 2] Add flow beta enablements [analytics/limn-language-data] - https://gerrit.wikimedia.org/r/269854 (owner: Milimetric) [22:31:05] RoanKattouw_away: those 4 patches above, some wiki configs, and you got https://flow-by-wiki.wmflabs.org [22:31:16] nuria, milimetric: someone should respond at https://en.wikipedia.org/wiki/Wikipedia:Village_pump_(technical)#Pageview_Stats_down_again [22:31:56] milimetric: Dude that's awesome [22:31:56] RoanKattouw_away: sadly I had to add it to the limn-language-data repo, and that caused me to rethink my "simple" convention [22:31:58] Thank you so much [22:32:12] wtf why [22:32:12] In order to reuse shared code? [22:32:36] well, the other way would've needed a new repo [22:32:43] and a puppet merge that I don't have rights for [22:33:03] because of a limitation of reportupdater to only write to one output folder [22:33:10] so I'm filing a task to fix it, no worries [22:33:22] I will, ori [22:34:53] thanks [22:35:18] Analytics: reportupdater to support multiple output folders per config - https://phabricator.wikimedia.org/T126549#2016977 (Milimetric) NEW [22:35:23] RoanKattouw_away: https://phabricator.wikimedia.org/T126549 [22:35:58] milimetric: could you include in your response an indication of what is the best way to report issues to the team? (phab, presumably) [22:38:13] ori: this is about stats.grok.se, so my response will basically be that it's not going to be maintained any more [22:38:20] and that there are plans to replace it, foremost this one: https://phabricator.wikimedia.org/T120497 [22:40:05] milimetric: yeah, +1 [22:41:48] milimetric: Hmm I think I also submitted similar code for the cross-wiki beta feature, lemme find that [22:41:52] Cause that'd need similar treatment [22:43:57] milimetric: With https://gerrit.wikimedia.org/r/#/c/264700/ I think I was hoping to make something appear at http://ee-dashboard.wmflabs.org/dashboards/enwiki-features# , but it looks like there's lots of definitions of things there that are not in the repo? [22:44:27] I don't care much about making that graph (also a preference tracking graph, just a different preference) appear at that particular URL, but I'd like it to be somewhere [22:45:22] k, I'll see what I can do after I answer more satisfactorily to these poor people starving for page stats [22:45:57] Sure, no rush [22:46:09] This preference doesn't even exist yet outside of testwiki, and won't until next week [22:54:38] ori: omg, how do people interact in that kind of text. I'm literally dizzy and sick from trying to read in that format where everyone's yelling and starting random threads that say the same thing [22:55:02] milimetric: ya i was just reading it [22:55:06] and am very confused [22:55:08] milimetric: hah! It's really validating to hear you say that, because that is exactly how I feel [22:55:13] as to where and how to respond [22:55:26] I responded in two places, hopefully people find it in that mess [22:55:31] ha ha [22:55:40] are you a Wikipedia administrator? [22:55:42] I think I need a shower now :) [22:55:55] I am the *furthest* possible from an admin [22:56:12] if I saw an admin, I think a black hole would instantaneously form and swallow both of us [22:56:26] milimetric: https://tools.wmflabs.org/pageviews/#start=2016-01-21&end=2016-02-09&project=en.wikipedia.org&platform=all-access&agent=user&pages=Cat|Dog [22:56:29] is interesting [22:56:37] yeah, that's what I linked to in my response [22:56:38] Someone has forked our demo [22:56:43] and made something nicer [22:56:44] yep, marcel knows about it [22:56:50] oi? [22:56:53] ah okay cool :) [22:56:57] musik guy's think [22:57:00] *thing [22:57:04] milimetric: I think your response is great [22:57:11] k, good [22:57:35] I wanted this crazy early period where people jumped to make their own tool [22:57:53] madhuvishy, yes, I saw that, it's awesome :] [22:57:56] because I secretly think it holds us honest to making good data APIs and data releases if we answer to client authors [22:58:15] but I maybe underestimated a bit how much this decision might affect wikipedia [22:58:19] so I regret it a little bit [22:58:34] and I'm happy to just buckle down and write a production-ready site over the next week [22:58:55] whatever everyone thinks (it would undercut the musik dude and Ainali's students) [22:59:13] hmmm [23:03:51] Analytics-Kanban, Editing-Analysis, Patch-For-Review, WMF-deploy-2016-02-02_(1.27.0-wmf.12), WMF-deploy-2016-02-09_(1.27.0-wmf.13): Edit schema needs purging, table is too big for queries to run (500G before conversion) {oryx} [8 pts] - https://phabricator.wikimedia.org/T124676#2017121 (Neil_P._Qu... [23:04:25] (PS1) Milimetric: Add flow crosswiki beta enablements [analytics/limn-language-data] - https://gerrit.wikimedia.org/r/269866 [23:04:56] (PS1) Milimetric: Remove crosswiki beta enablements [analytics/limn-ee-data] - https://gerrit.wikimedia.org/r/269867 [23:05:02] milimetric: do you have couple minutes for a python question? https://github.com/wikimedia/eventlogging/blob/master/eventlogging/handlers.py#L191 - Is there any reason you see for this continue not working - because I think it isn't [23:05:14] (CR) Milimetric: [C: 2 V: 2] Add flow crosswiki beta enablements [analytics/limn-language-data] - https://gerrit.wikimedia.org/r/269866 (owner: Milimetric) [23:05:21] (CR) Milimetric: [C: 2 V: 2] Remove crosswiki beta enablements [analytics/limn-ee-data] - https://gerrit.wikimedia.org/r/269867 (owner: Milimetric) [23:05:27] one sec :) [23:09:22] RoanKattouw_away: https://flow-by-wiki.wmflabs.org/#projects=testwiki/metrics=Crosswiki%20Beta%20Enables [23:10:07] Nice! Thanks milimetric [23:10:20] Analytics: reportupdater to support multiple output folders per config - https://phabricator.wikimedia.org/T126549#2017137 (Milimetric) [23:10:30] ok madhuvishy what's up? :) [23:10:34] milimetric: Why is there only one wiki there? The query runs against three wikis (testwiki, test2wiki, mediawikiwiki) [23:10:47] I don't think the feature is available on mediawikiwiki yet so that'd be 0, but it should be >0 on test2wii [23:10:51] RoanKattouw_away: you can try selecting those, but from what I could tell there's no data for them [23:11:13] yeah, RoanKattouw_away https://flow-by-wiki.wmflabs.org/#projects=testwiki,test2wiki/metrics=Crosswiki%20Beta%20Enables [23:11:20] (just search in the top left for whatever wiki you want) [23:11:33] it'll add it to the location.hash when you select it [23:11:47] milimetric: I am finding behavior that the code keeps executing beyond that continue and writes to kafka anyway (https://github.com/wikimedia/eventlogging/blob/master/eventlogging/handlers.py#L191 ) [23:12:08] lol, wait. what? :P [23:12:12] Yes! [23:12:17] I have no idea why [23:13:06] milimetric: Weird cause the query returns 8 on test2wiki [23:14:15] RoanKattouw_away: why's that weird, that's what the graph shows for test2wiki? [23:14:16] 8 [23:14:32] good night a-team! [23:14:39] here, if you look at just test2wiki: https://flow-by-wiki.wmflabs.org/#projects=test2wiki/metrics=Crosswiki%20Beta%20Enables [23:15:56] madhuvishy: so you're seeing "blah is blacklisted, not writing event, and then it writes that same exact event? [23:16:05] milimetric: yes [23:16:09] I put some prints [23:17:11] https://www.irccloud.com/pastebin/itqMkxgM/ [23:17:21] and i see [23:17:24] https://www.irccloud.com/pastebin/b4807b9v/ [23:17:34] madhuvishy: a print() right after the continue? [23:17:44] milimetric: Oh I didn't realize I had to add it through the interface, apparently URL manipulation doesn't work [23:18:01] ottomata: yes [23:18:06] in the same block though [23:18:09] that's what i did [23:18:11] oh [23:18:13] ya [23:18:15] let me [23:18:25] it shouldn't print anyway but [23:18:51] madhuvishy: also, print the event out [23:18:56] so you can be sure its the same event [23:18:59] from the uuid [23:19:21] yeah, you should print the exact event with its guid like on the "is blacklisted" line [23:19:29] ottomata: I'm pretty sure - it says schema Analytics - which is blacklisted [23:19:41] but i will print the event anyway [23:19:46] yeah, but maybe one event is matching, but some other isn't [23:19:47] for some reason [23:20:04] ottomata: I sent 4 events - I see the same set of prints for all [23:20:08] ah [23:20:10] k [23:20:11] hm [23:20:21] i'd still print it out, but ja that is weird [23:25:47] oook i'm out, laters all! [23:31:29] (PS1) Milimetric: Add beta-enables dashboard to edit-analysis [analytics/dashiki] - https://gerrit.wikimedia.org/r/269874 [23:31:40] (CR) Milimetric: [C: 2 V: 2] Add beta-enables dashboard to edit-analysis [analytics/dashiki] - https://gerrit.wikimedia.org/r/269874 (owner: Milimetric) [23:33:08] nite! o/ [23:41:18] milimetric: super weirddd [23:41:21] https://www.irccloud.com/pastebin/zdsOJ0BH/ [23:41:44] It writes the event before the entering the blacklist block [23:44:22] a little clearer prints [23:44:25] https://www.irccloud.com/pastebin/KZ535FZc/ [23:48:52] Analytics-Wikistats, operations, Regression: [Regression] stats.wikipedia.org redirect no longer works ("Domain not served here") - https://phabricator.wikimedia.org/T126281#2017268 (Dzahn) With all due respect, i think the Analytics team is the right one to do analytics of traffic on the statistics site.