[09:16:56] 10Analytics, 10Analytics-Cluster, 06Operations, 13Patch-For-Review: Zookeeper heap usage patterns - https://phabricator.wikimedia.org/T157968#3024296 (10elukey) Moritz completed the restarts and the Heap usage pattern changed on all the nodes, so this is probably something to expect with the current settin... [09:21:32] probably noone will notice at this time, but quick headsup that I'll install PHP security updates on bohrium/piwik in a minute [09:23:43] super [09:24:29] done. elukey, are you able to make a quick check that all still works as expected? [09:25:37] moritzm: it seems working fine:) [09:33:08] ok :-) [09:38:01] 10Analytics-Tech-community-metrics: Updated data in mediawiki-identities DB not deployed onto wikimedia.biterg.io? - https://phabricator.wikimedia.org/T157898#3024337 (10Lcanasdiaz) @Aklapper I confirm this is broken right now. I'm appliying it manually today and talking with the guys to code the fix. [10:01:58] (03CR) 10Addshore: [C: 032] Remove minutely SPARQL script from CRON [analytics/wmde/scripts] - 10https://gerrit.wikimedia.org/r/336208 (https://phabricator.wikimedia.org/T146468) (owner: 10Addshore) [10:02:03] (03CR) 10Addshore: [C: 032] Remove minutely SPARQL script [analytics/wmde/scripts] - 10https://gerrit.wikimedia.org/r/336209 (https://phabricator.wikimedia.org/T146468) (owner: 10Addshore) [10:02:06] (03Merged) 10jenkins-bot: Remove minutely SPARQL script from CRON [analytics/wmde/scripts] - 10https://gerrit.wikimedia.org/r/336208 (https://phabricator.wikimedia.org/T146468) (owner: 10Addshore) [10:02:12] (03Merged) 10jenkins-bot: Remove minutely SPARQL script [analytics/wmde/scripts] - 10https://gerrit.wikimedia.org/r/336209 (https://phabricator.wikimedia.org/T146468) (owner: 10Addshore) [11:00:54] 06Analytics-Kanban, 06Operations: Periodic 500s from piwik.wikimedia.org - https://phabricator.wikimedia.org/T154558#3024623 (10elukey) Just checked with Manuel if `thread_pool_size` was available for mysql 5.5 but it seems that it needs a proprietary extension to work. I executed `SET GLOBAL max_connections=... [11:43:04] hey team! [11:44:11] Hi mforns :) [11:44:55] mforns: I feel bad - I didn't really provided help on the pagecounts thing :( [11:45:21] joal, ??? you gave most of the ideas! [11:45:47] mforns: hm, I didn't really support though [11:45:49] anyway :) [11:46:06] mforns: how is it going now? [11:46:22] np, it took me a bit more time, but this way I won't forget, heheh [11:46:40] going well, I managed to read the output with an externat table [11:47:01] Yay :) [11:47:03] also, found some parsing errors, there are some rows in the original pagecounts-raw data that are corrupt [11:47:08] mforns: This outuput is CSV, right? [11:47:14] tsv, yes [11:47:16] Arf mforns :( [11:47:20] compressed with snappy [11:47:20] ok [11:47:38] but I don't thiks we can do anything, they are missing columns [11:47:41] mforns: I think GZ compression would be better for this special case [11:47:45] anyway it's a small fraction [11:47:48] k [11:47:59] joal, OK, I can change to gz [11:48:17] but the files are quite small [11:48:18] mforns: Let's do it once the big rewriting is done :) [11:48:42] mforns: It's not about data size, it;s more about format copnsistency for archived data [11:49:03] mforns: I think in archive we try to use GZ for the data to be easily exportable on other platforms [11:49:32] I see, I read snappy was a faster codec (less compression), that's why I used it, given that files are small [11:49:38] but consistency makes sense [11:50:24] mforns: and more than consistency, if we want to export data using rsync for instance, it'll be way easier to read on other systems if it's gz [11:50:29] joal, however, this is not going to be in archive, right? it's in the form of a table... [11:50:39] sure, makes sense [11:51:00] mforns: I think we'll keep them in archive folder, no? [11:51:19] joal, OK makes sense as well [11:51:24] mforns: We actually could have them in 2 places: archive in TSV.gz, core in parquet.snappy [11:51:36] That's what we do with pageviews [11:52:17] joal, yes, but the files are so small, that the performance improvement in parquet+snappy won't be so high I guess [11:52:29] mforns: true :) [11:52:39] Maybe not needed :) [11:52:40] maybe it's not worth duplicating the data [11:52:43] ok [11:52:51] At least for the projectcounts ones [11:52:58] let's see if the archive version + external table is enough [11:53:02] If we get to pagecounts, it might be different [11:53:15] sure, the pagecounts will be wild [12:09:41] * elukey lunch! [12:18:58] Hi zareen [12:19:22] hey joal [12:19:32] zareen: The job I had launched failed :( [12:19:43] zareen: I have launched a new one with more workers and smaller data [12:20:04] ahhh, do we know why? [12:20:06] zareen: https://yarn.wikimedia.org/proxy/application_1486634611252_14350/ [12:20:34] zareen: too many partitions for too few workers [12:21:17] i see, thanks for launching another job [12:22:18] zareen: no problem, it's my fault the previous one didn't work [12:25:29] zareen: The job I pasted above is for the end of dec 2016 [12:25:42] zareen: I'll go with Jan 2017 after, then feb [12:28:22] joal: okay, so it will probably be at least another day or 2 before this table is ready with the latest data? [12:28:30] (03CR) 10Erik Zachte: [C: 032] Update banner with design consultation [analytics/wikistats] - 10https://gerrit.wikimedia.org/r/337452 (owner: 10Milimetric) [12:28:43] zareen: I think you're right [12:29:19] 06Analytics-Kanban: Hive code to count global unique devices per top domain (like *.wikipedia.org) - https://phabricator.wikimedia.org/T143928#3024816 (10JAllemandou) a:03JAllemandou [12:29:28] (03CR) 10Erik Zachte: [V: 032 C: 032] Update banner with design consultation [analytics/wikistats] - 10https://gerrit.wikimedia.org/r/337452 (owner: 10Milimetric) [12:30:14] hm, i'm tempted to re-run the monthly partitioned table from last week (with the new changes made) just in case this job doesn't run in the timeframe expected/fails [12:30:42] (as a separate table) [12:32:19] zareen: I think this is not needed [12:32:49] joal: the job launched friday has 4 reducers, and the job launched today has 16. this might be a silly question, but what is the process if the number of reducers isn't set? (i've never specified number of reducers) does it just use all available resources? [12:33:53] zareen: hive decides itself based on estimated data size [12:35:09] zareen: while it's usually good for queries, for specific patterns like dynamic partitioning it might lead to unwanted results (many small files) [12:35:39] So for our specific case, it's a tradeoff between volume of data treated at once and number of files at the end [12:35:53] (03CR) 10jenkins-bot: Update banner with design consultation [analytics/wikistats] - 10https://gerrit.wikimedia.org/r/337452 (owner: 10Milimetric) [12:36:14] joal: i see. okay, i'll monitor this job throughout the day. [12:36:37] zareen: it'll finish soon, I'll launch the Jan one after [12:37:04] joal: great, thanks for the help. [12:52:03] joal: o/ - I added new alarms for Hadoop heap usage and there is a warning for an1055 - https://grafana.wikimedia.org/dashboard/db/analytics-hadoop?panelId=17&fullscreen [12:52:07] looks good :) [12:52:42] elukey: looks like it works indeed :) [12:52:50] elukey: that's great [12:53:06] elukey: And with the NM patch, restart when alarm, no problemo :) [13:39:50] 10Analytics-Cluster, 06Analytics-Kanban, 13Patch-For-Review, 15User-Elukey: Monitor Hadoop cluster running out of HEAP space with Icinga - https://phabricator.wikimedia.org/T88640#1016773 (10elukey) [13:41:25] 10Analytics-Cluster, 06Analytics-Kanban, 06Operations, 13Patch-For-Review, 15User-Elukey: Zookeeper heap usage patterns - https://phabricator.wikimedia.org/T157968#3025004 (10elukey) [14:03:37] joal: sweet, dec 2016 job is complete :) [14:05:50] zareen: launching Jan 2017 [14:06:16] zareen: As exepected, tradeoff - sorry for having been on the worng side for the first one [14:06:26] joal, qq: how can I mvn package refinery/source with scala 1.6? [14:06:37] joal: o/ [14:07:21] hm, mforns: you mean spark 1.6 right? [14:07:30] joal, sorry yes [14:07:33] ok :) [14:07:40] I am not sure I understand your question [14:07:46] change in the pom? [14:07:49] Yes :) [14:07:53] aha [14:08:29] mforns: I suggest for you to look at the Mediawiki History patch - I has this change in the pom IIRC [14:08:38] joal, OK thanks! [14:08:43] np :) [14:18:19] joal, I don't manage to use it... should I use: import com.holdenkarau.sql.functions.input_file_name, instead of import org.apache.spark.sql.functions.input_file_name? [14:18:38] anyway that does not work either [14:23:41] joal, never mind, I was confused [14:25:40] mforns: works now? [14:25:44] no... [14:25:45] mforns: sorry I missed your ping [14:25:47] Arf [14:25:49] np [14:25:57] mforns: batcave? [14:26:04] sure joal [15:01:12] elukey: thought about piwik, for me I generally agree with what you say, this part is where there's grey area though: "a service needs to be supported properly or it should be dismissed" [15:01:30] so normally that's true [15:01:44] but ops specifically did not want to support piwik properly, and we brought it in anyway [15:02:26] so I'm all for doing things the best way you see fit, and I agree that's using the mariadb module and letting the DBAs know about it [15:03:12] but I also want to defend ops's right to reject supporting piwik, as they did initially [15:03:28] otherwise there's this loophole where we keep sneaking in work for ops [15:03:58] (03PS1) 10Mforns: Add spark job to aggregate historical projectviews [analytics/refinery/source] - 10https://gerrit.wikimedia.org/r/337593 (https://phabricator.wikimedia.org/T156388) [15:05:45] joal, meeting? [15:06:25] milimetric: morning :) [15:06:39] I discovered that we have the 'mysql' module, that might help [15:06:58] basically it handles for you all the things like custome /etc/mysql/conf.d erb template, etc.. [15:07:24] milimetric: when was the last time that we discussed "ops specifically did not want to support piwik properly" ? [15:07:37] I doubt that the statement still holds [15:07:50] (especially since me and Andrew will take care of it :P) [15:11:44] mforns, milimetric: joining now [15:12:47] elukey: for some reason I didn't get your pings, the ops opinion was expressed in late 2015 [15:13:09] elukey: I totally agree that you joining changes the equation a lot, so that's why this is mostly up to you [15:19:17] milimetric: it really depends who are the users and what is their use case [15:19:26] it seems to me that the service has been running for a while [15:19:37] so we now should have a good idea about the use cases [15:25:55] 10Analytics, 10DBA, 13Patch-For-Review: Set up auto-purging after 90 days {tick} - https://phabricator.wikimedia.org/T108850#3025412 (10jcrespo) a:05jcrespo>03None [15:29:40] 10Analytics-Tech-community-metrics, 06Developer-Relations (Jan-Mar-2017): Merge detached SCM and SCR identities in korma DB if they have the same email address but not the same uuid - https://phabricator.wikimedia.org/T156283#3025416 (10Aklapper) > and then checking for items listed more than once. That manual... [15:43:42] (03PS6) 10Joal: Update sqoop script with labsdb specificity [analytics/refinery] - 10https://gerrit.wikimedia.org/r/334042 (https://phabricator.wikimedia.org/T155658) [15:44:05] (03PS19) 10Joal: Add mediawiki history spark jobs to refinery-job [analytics/refinery/source] - 10https://gerrit.wikimedia.org/r/325312 (https://phabricator.wikimedia.org/T141548) [15:44:48] a-team, I'll be late in standup, need to spend some time with workers now - hopefully I'll be back before the end [15:45:28] elukey: yes, use cases for piwik: [15:45:32] * us on our dashboards [15:45:43] * annual report, sites like 15.wikipedia.org, etc. [15:45:47] * iOS app [15:46:47] (also transparency report goes in that second bullet) [15:47:20] heyo all, not sure if you're the the right people to be asking, but I've been looking at getting a % of English Wikipedia administrators who haven't enabled 2FA (discussed from a security point of view at https://phabricator.wikimedia.org/T157500) [15:47:23] I know more details about each one, and I guess except for the iOS app the rest of the use cases are totally fine with a lot of error [15:47:56] hi samtar, where is 2FA enabling stored? Do you know? The db somewhere in user_options? [15:48:38] milimetric: I'm afraid I'm going to be useless here and say I have no idea - https://phabricator.wikimedia.org/T157500#3009105 sort of mentioned how we'd get at the data [15:49:36] gotcha, ok, samtar, I'll take a look and see if I can run a quick query [15:49:44] don't hang your hopes on me though :) [15:50:03] milimetric: it'd be great if you could, thank you :) if not, a shove in the right direction would be much appreciated [15:50:16] shame it's not an EventLogging schema [15:50:47] 10Analytics, 10Analytics-EventLogging, 10DBA: Potentially decrease db1046's InnoDB buffer pool - https://phabricator.wikimedia.org/T125829#3025480 (10jcrespo) a:05jcrespo>03None [15:52:19] milimetric: now that I've increased max_connections, this seems to be the main issue https://github.com/piwik/piwik/issues/6398 [15:53:01] elukey: awesome, yeah that was happening before [15:53:05] so you got rid of the 500s?! [15:53:13] well, the _other_ 500s? [15:53:22] 10Analytics: Render swagger spec docs as service endpoint - https://phabricator.wikimedia.org/T158066#3025501 (10Ottomata) [15:53:35] milimetric: I don't see the weird apache logs, at least it didn't show up during the last hours [15:53:39] this doesn't mean that it is gone :D [15:53:41] samtar: do you happen to know what user groups you want me to check? Here are the ones listed in english wiki's table: [15:53:42] 10Analytics: Create EventStreams swagger spec docs endpoint - https://phabricator.wikimedia.org/T158066#3025513 (10Ottomata) [15:53:43] https://www.irccloud.com/pastebin/2YczElst/ [15:54:01] elukey: yaaaay it's gone, you're awesome [15:54:11] 10Analytics: Create EventStreams swagger spec docs endpoint - https://phabricator.wikimedia.org/T158066#3025501 (10Ottomata) [15:54:17] milimetric: `sysop` only please :) [15:54:21] samtar: k [15:54:38] milimetric: but for example, now that I'd need to make the change permanent in the mysql config, with the current puppet configuration I'd need to put a file in /etc/mysql/conf.d/etc.. meanwhile a module would add that via erb template :P [15:54:49] ottomata: HIIIIiiiiIIIIiiii [15:54:52] how are you today? [15:55:31] HIIIII [15:55:36] ok, pretty phlegmy but ok! [15:55:38] how are you? [15:55:45] elukey: but there's a template for mysql config there... [15:56:02] piwik was taking mysql params from an ini file [15:57:01] milimetric: where is the mysql template? I might have missed it [15:57:05] do you mean in puppet? [15:57:12] yeah [15:58:48] elukey: I mean, wait, there should be... the research password is in there and everything [16:01:35] elukey: that ini file is /etc/piwik/config.ini.php [16:01:44] but you're right I'm not seeing a template for that?! [16:01:51] this I definitely agree with you should be in puppet [16:02:46] oooooh https://github.com/wikimedia/operations-puppet/blob/9b4bc170fba7d1623032d47a42ccf878ba382750/modules/role/manifests/piwik/server.pp#L6-L22 [16:05:19] 10Analytics-Cluster, 06Analytics-Kanban, 13Patch-For-Review, 15User-Elukey: Investigate if Node Managers can be restarted without impacting running containers - https://phabricator.wikimedia.org/T156932#3025571 (10Nuria) 05Open>03Resolved [16:07:18] 06Analytics-Kanban, 13Patch-For-Review: CDH 5.10 upgrade - https://phabricator.wikimedia.org/T152714#3025576 (10Ottomata) Current plan: do this February Tues 28th. I will send out announcement and schedule downtime. [16:13:57] 06Analytics-Kanban: Check abnormal pageviews for XHamster - https://phabricator.wikimedia.org/T158071#3025609 (10Nuria) [16:14:48] milimetric: yep but there is no configuration file for mysql :D [16:15:00] say you need to set max_connections [16:15:20] elukey: yeah, that's what the comment says, supposedly why it's impossible to have one with piwik [16:15:32] so it pulls the config from pwstore somehow locally maybe? [16:16:31] milimetric: nono that part is mysql related, shouldn't be related to the piwik web installer [16:16:45] I am talking about files like /etc/mysql/conf.d/etc.. [16:17:06] it is fine to have to bulk load the database etc.. [16:17:07] right, that'd be for the mysql for the box, you're right, we could stick config there [16:17:37] it's dedicated to piwik anyway [16:18:24] so I agree, instead of requiring just the package you would import the module and configure the threadpool, right? [16:18:43] yep exactly :) [16:38:51] elukey: should we review network acls real quick [16:39:17] milimetric: did you want me to assign that task to you btw? [16:39:34] samtar: no, I'd only be doing a small part of it [16:39:47] Ah no worries ^^ [16:39:57] samtar: in a bunch of meetings for the next few hours, but trying to cobble the query together, should get it by end of day today [16:40:05] [fdans] is such a millennial, making everyone go into slack [16:40:11] ahaha [16:40:15] No rush milimetric! :D [16:48:23] nuria: when we moved retro it conflicts with Asaf's meeting [16:48:26] which I would like to attend [16:48:41] milimetric: sorry, i do not have that meeting [16:48:53] nuria: no, it's ok, just letting you know [16:49:10] it's in 12 min., it sounds like a useful thing to have at least one of us at [16:49:35] what is meeting about? [16:50:32] nuria: [16:50:35] https://www.irccloud.com/pastebin/Evn1agKb/ [16:52:14] milimetric: sorry it conflicts. Due to poststandup we always miss the first 10 mins of retro thus i though i would move it so we could have a full meeting [16:53:05] true. We could also just hard-end standup after 20 minutes on retro days [16:53:20] a-team: moved retro to tomorrow, looks like milimetric had an scheduling conflict [16:53:29] ok [16:53:33] it has become a bit drawn-out lately, I actually was going to mention that at retro :) [16:54:05] milimetric: well, the teams we need to discuss do not magically disappear but maybe we can address that in adifferent way [16:54:05] oook [16:54:42] k [16:55:09] joal since we're not doing retro, do you have a few minutes a la grotte aux chauve-souris? [16:55:11] sure, but we have to weigh that against really long standups like 1 hour. That starts defeating the purpose of standup: quick short energizing meeting [16:55:13] hey chasemp, are you around by any chance? [16:55:28] sure fdans ! A la grotte aux chauves souris ! [16:56:12] joal: kind of :) just getting out of a meeting and pre-lunch [16:57:40] chasemp: I'm after the process to ask you (or our beloved DBAs) to update the views with 2 new fields [17:01:36] joal: I last ran it lasts week or so but I can do so again [17:01:58] chasemp: Oh ? [17:02:34] joal: https://phabricator.wikimedia.org/T153743#3009402 [17:03:01] looks like a few new things since then possibly [17:03:30] 06Analytics-Kanban: Check abnormal pageviews for XHamster - https://phabricator.wikimedia.org/T158071#3025805 (10Nuria) a:03Nuria [17:03:33] chasemp: the 2 fileds I'm talking about are ar_content_model and ar_content_format in archive talbe [17:03:41] 06Analytics-Kanban, 13Patch-For-Review: Add "Damn Small XSS Scanner" (DSXS) to list of known bots - https://phabricator.wikimedia.org/T157528#3025806 (10Nuria) 05Open>03Resolved [17:04:10] joal: unf that same permission bug seems to have surfaced 'pymysql.err.OperationalError: (1142, "CREATE VIEW command denied to user 'maintainviews'@'localhost' for table 'abuse_filter_action'")' [17:04:16] I'll have to ping the dba crew [17:05:37] 06Analytics-Kanban, 10Analytics-Wikistats: Add "Interwicket" to the list of bots - https://phabricator.wikimedia.org/T154090#2900990 (10Nuria) a:03Nuria [17:07:15] 06Analytics-Kanban: Check abnormal pageviews for XHamster - https://phabricator.wikimedia.org/T158071#3025815 (10MusikAnimal) A few things I found: * It seems the number of unique IPs that visited `/wiki/XHamster` is significantly less than articles that received around the same number of pageviews. * The top re... [17:07:35] 06Analytics-Kanban, 10Analytics-Wikistats: Add "Interwicket" to the list of bots - https://phabricator.wikimedia.org/T154090#3025828 (10Nuria) anything "bot" should already be caught [17:08:03] 06Analytics-Kanban, 06Operations, 10netops: Review ACLs for the Analytics VLAN - https://phabricator.wikimedia.org/T157435#3025831 (10Ottomata) > term udplog { + 1 > Remove IPs the term analytics-publicIP-v4: +1 > Review the IPs in term ssh Don't know anything about this, but also not sure why we have speci... [17:09:55] ottomata: I think that the kubernetes1003 reference in prelabsdb-mysql is an IP that got recycled :( [17:10:00] hm ok [17:10:12] elukey: these are all on the analytics vlan port? [17:10:35] yep [17:15:22] ottomata: I am getting cert errors downloading from archiva: [17:15:27] https://www.irccloud.com/pastebin/y4u37ge3/ [17:16:01] bwaaah? [17:17:42] ottomata: yessir [17:18:01] https kaput? [17:18:19] iinteresting, i get that too [17:18:36] https should be resolved by the wildcard cert [17:18:39] so that's weird [17:19:03] nuria i gotta run to make a bus to get home real quick, will investigate more when I'm back [17:19:11] ottomata: np, no rush at all [17:19:47] k back in a bit... [17:57:19] * elukey afk! Talk with you tomorrow team! [17:57:22] byyeeee o/ [18:00:35] Bye elukey [18:05:41] (03PS2) 10Mforns: Add spark job to aggregate historical projectviews [analytics/refinery/source] - 10https://gerrit.wikimedia.org/r/337593 (https://phabricator.wikimedia.org/T156388) [18:30:07] nuria: did you upgrade java on your laptop recently? [18:30:09] (I did) [18:30:14] wonder if others who haven't get this error [18:30:17] no, no [18:35:14] wikimedia/mediawiki-extensions-EventLogging#634 (wmf/1.29.0-wmf.12 - 04e3fe4 : Translation updater bot): The build has errored. [18:35:14] Change view : https://github.com/wikimedia/mediawiki-extensions-EventLogging/compare/wmf/1.29.0-wmf.12 [18:35:14] Build details : https://travis-ci.org/wikimedia/mediawiki-extensions-EventLogging/builds/201609654 [18:42:33] (03CR) 10Joal: "Some comments inline :)" (0310 comments) [analytics/refinery/source] - 10https://gerrit.wikimedia.org/r/337593 (https://phabricator.wikimedia.org/T156388) (owner: 10Mforns) [18:42:43] thanks :] [18:42:49] np mforns :) [18:43:02] I love the power of the mapPartition trick :) [18:43:33] nuria: pretty sure it is a problem with something on our laptops [18:43:42] refinery/source builds fine on stat1002 [18:49:48] joal, you comment that the files are stored in monthly folders, but in the end I went for no partitions in the hive table, so all files will be in the root folder, you see a problem in that? [18:50:08] mforns: absolutely not :) [18:50:25] ok [18:50:37] mforns: something we didn't consider that could be interesting would be to sort the output [18:51:00] joal, yes, makes sense, I struggled a bit to vet the data, it would have helped. Will do [18:51:04] Since it is TSV (not parquet), sorted will be easier to read for users [18:51:12] totally [18:51:17] mforns: last thought (sorry, forgot to mention in the CR) [18:51:24] no no go ahear [18:51:28] *ahead [18:51:57] I think since those files are ready-for-export (in some ways), outpuing a single file is better [18:52:19] I'd still use the number of partitions to reduce by more efficiently, but then would coalesce to 1 [18:53:26] But, in that case, since gz is not splittable, it means hadoop cannot parallelise work... hmmmm [18:53:52] I think I'd actually prefer a partition by year with single files in each [18:53:57] For the reasosn above [18:54:02] makes sense mforns | [18:54:03] ? [18:55:31] nuria: i'm taking it back, i think there might be something not right with the cert [18:58:53] 10Analytics, 10MediaWiki-extensions-WikimediaEvents, 10The-Wikipedia-Library, 10Wikimedia-General-or-Unknown, 13Patch-For-Review: Implement Schema:ExternalLinksChange - https://phabricator.wikimedia.org/T115119#3026457 (10Samwalton9) [18:59:58] joal, a year of projectcounts-raw is around 50MB [19:00:32] mforns: is it a good file-size, or to big even? [19:00:51] joal, I think it's good [19:01:33] mforns: Do you see my point about how to access / sort etc? [19:01:48] but then the writing of the yearly files will be done "by hand", because I don't think there's a way of defining hive partitions in a dataframe [19:02:21] mforns: I think best could possibly be to have 1 gz file per year, with everything in a single folder for hive to parse (or with parttions, I don't miond) [19:02:32] so something like: for year in years: write("/year=2007/" + outputPath) [19:02:38] mforns: There definitely are [19:02:59] mforns: the problem is more about using CSV formatter with partitions [19:03:25] mforns: http://stackoverflow.com/questions/37509932/write-spark-dataframe-as-csv-with-partitions [19:03:42] joal, but if gzip is not splittable, how are we having multiple gzipped files in a folder? will hive be able to read them? [19:04:07] mforns: hive has no problem reading multiple files [19:04:28] mforns: issue with non-splittable files is that a single mapper needs to read them (instead of potentially more than one) [19:04:44] Ah! I see [19:04:57] But with 50mb file size, we're very ok :) [19:05:19] ok [19:12:12] 06Analytics-Kanban: Document the difference in aggregate data on wikistats and wikistats 2.0 - https://phabricator.wikimedia.org/T150963#3026551 (10Elitre) I also checked after a week and provided some advice, our intervention during this Phase 1 is not anticipated. [19:12:30] nuria: i think i have a fix [19:12:31] can you try? [19:18:10] nuria: https://wikitech.wikimedia.org/wiki/Archiva#Using_new_Let.27s_Encrypt_SSL_certs [19:34:32] joal: see if your views are there now? [19:34:42] there is some auth bug that got in the way but in theory we overcame [19:35:07] 10Analytics, 10Wikimedia-Stream: Update EventBus RCFeed config to use newly refactored settings - https://phabricator.wikimedia.org/T158106#3026676 (10Ottomata) [19:35:15] chasemp: specific wiki? [19:35:37] it should be fine everywhere the wikidb exists and the field [19:35:43] it was a mass update of existing things [19:35:43] chasemp: It's not actually views, it's specific fields of a an existing view [19:35:47] ah [19:35:54] maybe I misunderstood [19:35:57] And, not changed [19:36:05] is this a field that does not exist in a view that is created? [19:36:07] I'm sorry if I've been unclear [19:36:16] correct chasemp [19:36:21] no worries, it would been a quick circle back if not for the bug etc [19:36:39] mwarf [19:37:08] ah, so the process is largely I think the same for adding fields to a view (create a task, we put up a proposed change to the views yaml definition file, and bug one of the security folks to ok it) [19:37:18] ottomata: trying [19:37:24] I can't add fields to the view, or at least don't feel empowered to without that [19:37:27] or more to the point [19:37:29] chasemp: details: in revision table, there are rev_content_model and rev_content_format, but in archive, no ar_content_format nor ar_content_model [19:37:38] hm [19:38:22] chasemp: I create a new task and flag it with labs-infra and DBA ? [19:38:41] ottomata: you were totally right, i upgraded to java 8, but was a long time ago, i think it is been a while since i needed new deps [19:38:58] we can do that dance together, yeah, make a task, and upload a proposed changeset to /etc/maintain-views.yaml [19:39:07] and we'll try to wrangle whoever can say it's ok to expose [19:39:09] nuria: i don't think it was java 8 [19:39:16] i think it was the lets encrypt ca chnage [19:39:17] it's a pita but it's mostly for good reasons I think [19:39:30] and largely as it's been mostly unmanaged I think so can only get better? :) [19:39:41] huhu chasemp :) [19:40:04] the /etc/maintain-views.yaml file is in puppet? [19:40:32] 'tis [19:40:32] modules/role/templates/labs/db/views/maintain-views.yaml [19:40:58] here is whta I see for the view now [19:41:00] select rev_id, rev_page, if(rev_deleted&1,null,rev_text_id) as rev_text_id, if(rev_deleted&2,n$ [19:41:00] as rev_comment, if(rev_deleted&4,null,rev_user) as rev_user, if(rev_deleted&4,null,rev_user_te$ [19:41:01] rev_timestamp, rev_minor_edit, rev_deleted, if(rev_deleted&1,null,rev_len) as rev_len, rev_par$ [19:41:02] if(rev_deleted&1,null,rev_sha1) as rev_sha1, rev_content_model, rev_content_format [19:41:11] chasemp: awesome - I'm gonna leave for tonight but will submit a patch later on [19:41:12] which is not easy to reason about for me honestly [19:41:14] ok [19:41:22] thanks again chasemp ! [19:41:27] a-team, gone for tonight :) [19:42:11] laters! [19:47:29] ottomata: yours truly had some ahem... mix of java versions [19:47:36] ottomata: all good [19:47:38] now [19:48:50] (03PS1) 10Nuria: Add "Interwicket" to the list of bots [analytics/refinery/source] - 10https://gerrit.wikimedia.org/r/337632 (https://phabricator.wikimedia.org/T154090) [19:49:44] 10Analytics-Dashiki, 06Analytics-Kanban, 13Patch-For-Review: Add Map component to dashiki layout - https://phabricator.wikimedia.org/T153921#3026716 (10Nuria) 05Open>03Resolved [19:50:09] 06Analytics-Kanban, 10Analytics-Wikistats: Visual Language for http://stats.wikimedia.org replacement - https://phabricator.wikimedia.org/T152033#3026719 (10Nuria) [19:50:11] 06Analytics-Kanban: Involve the community in Wikistats 2.0 design. Stage 1. - https://phabricator.wikimedia.org/T156287#3026718 (10Nuria) 05Open>03Resolved [19:50:47] great [20:08:35] 06Analytics-Kanban: Document the difference in aggregate data on wikistats and wikistats 2.0 - https://phabricator.wikimedia.org/T150963#3026824 (10Milimetric) Thanks Elitre. For Phase 2, we added the planned March 3rd - March 10th consultation on this page: https://meta.wikimedia.org/wiki/Community_Engagement/... [20:10:26] ottomata: for cluster upgrade ... wouldn't it be better middle of the month so when it comes back to live it doesn't have to face 1st thing the major amount of work that are monthly jobs? [20:10:48] 10Analytics, 10MediaWiki-extensions-WikimediaEvents, 10The-Wikipedia-Library, 10Wikimedia-General-or-Unknown, 13Patch-For-Review: Implement Schema:ExternalLinksChange - https://phabricator.wikimedia.org/T115119#3026829 (10Milimetric) @Samtar: that delay is most likely due to event logging replication cat... [20:11:45] milimetric: about? [20:12:11] i geuss that is the next day.. [20:12:15] but nuria no good time other than that [20:12:20] unless we wait longer [20:12:29] (just to discuss T115119 if you have a mo) [20:12:29] T115119: Implement Schema:ExternalLinksChange - https://phabricator.wikimedia.org/T115119 [20:12:33] ottomata: k, your call [20:13:10] samtar: sure [20:14:33] samtar: do you have a question? [20:14:48] or nuria ^^ - so milimetric has said that a delay of an hour or so is to be expected on stat1003, now this might just be my lack of understanding (and I would *not* be surprised if that's all it is!) but aren't EventLogging events logged straight to stat1003? [20:15:00] samtar: no, they are not [20:15:05] samtar: that is a DB replica [20:15:18] hm, why am I missing 1/2 my pings all of a sudden [20:15:22] sorry nuria, I'm around [20:15:53] yeah, samtar, it's a replica with batched custom replication [20:15:53] nuria: well then there you go :) so for example I'm looking at ExternalLinksChange - they come from production to stat1003? [20:16:03] samtar: for all events [20:16:13] and it's that replication which is the delay? [20:16:14] (for correctness, stat1003 is not a db replica :) ) [20:16:15] samtar: data is logged to prod (master) and replicated [20:16:30] ottomata: sorry, not a replica, but not-prod either [20:16:36] there is no db on stat1003 [20:16:38] right, analytics-store is the replica [20:16:46] which can be accessed from stat1003 [20:16:49] :) [20:16:54] samtar: delay depends on how busy dbs are [20:17:13] samtar: if some maintenace is going on, delays might be big [20:17:17] heh, though technically analytics-store is just a domain name for I don't know what actual server [20:17:35] Okay, because the delay is one thing, but every now and then it just plainly misses some events - this could well be down to the logging code itself [20:17:51] but to summarise, a delay of an hour is normal? [20:18:10] or is that a little high? [20:18:27] (03CR) 10Milimetric: "I commented on a small preference for the sqooping function, take a look and self-merge after that whichever way you decide (not a strong " (031 comment) [analytics/refinery] - 10https://gerrit.wikimedia.org/r/334042 (https://phabricator.wikimedia.org/T155658) (owner: 10Joal) [20:18:42] (I realise that's probably a `how long is a piece of string` style question :P) [20:19:32] samtar: delay would not make it miss events, that is , as you said, related to your client side code [20:20:04] samtar: makes sense? [20:20:22] nuria: yes definitely, thank you for the clarification :) [20:22:00] samtar: but i would test for events on beta, rather than prod [20:23:13] samtar: the missing of events on that schema is due to the implementation. It seems to be a very tricky issue that Lego couldn't figure out, related to an interaction with another extension [20:23:37] joal: looks like January 2017 job is done! :) https://yarn.wikimedia.org/cluster/app/application_1486634611252_14631 could you please launch Feb 2017 job when you get a chance? [20:24:01] but in general, if you do event X, Y, and Z, in that order, and you only see event X and Z in the database after a while, then it's 99% likely Y never got sent [20:24:19] milimetric: interesting.. that seems to be what we're getting [20:24:32] samtar: check out Lego's possible explanation on that task [20:24:36] appreciate it's not analytics' issue though! [20:24:58] I don't know enough about mediawiki to help, but happy to brain-bounce if you know how things work and just need someone to talk it over [20:25:05] I've been rooting for that issue to be resolved for like a year [20:32:12] (03PS1) 10Ottomata: Add --partition-type option to refinery-drop-hourly-partitions script [analytics/refinery] - 10https://gerrit.wikimedia.org/r/337638 (https://phabricator.wikimedia.org/T146915) [20:32:56] (03CR) 10Ottomata: [V: 032 C: 032] Add --partition-type option to refinery-drop-hourly-partitions script [analytics/refinery] - 10https://gerrit.wikimedia.org/r/337638 (https://phabricator.wikimedia.org/T146915) (owner: 10Ottomata) [20:33:34] ottomata: we should add "The WMF Analytics team code reviews jobs that publish data here" to https://analytics.wikimedia.org/datasets/README [20:34:47] milimetric: send me full paste of whatever you want in there, and I will change it! [20:35:55] https://www.irccloud.com/pastebin/Zy4aQSsj/ [20:35:59] ottomata: ^ [20:36:01] thanks [20:48:32] (03PS1) 10Milimetric: [WIP] DO NOT MERGE [analytics/dashiki] - 10https://gerrit.wikimedia.org/r/337642 (https://phabricator.wikimedia.org/T125854) [20:53:04] 10Analytics-Cluster, 06Analytics-Kanban, 06Operations: Move cloudera packages to a separate archive section - https://phabricator.wikimedia.org/T155726#3027062 (10Ottomata) a:03Ottomata [20:54:20] ottomata: :( we forgot that reportupdater does its own syncing now [20:54:30] milimetric: ? [20:54:40] it has an rsync_to: thorium::/srv/limn-public-data [20:54:52] so it's going to come on top of the rsync scripts you made [20:55:32] nuria milimetric fdans mforns emailed you first draft for dashboard visd [20:55:36] please send feedback [20:55:41] ashgrigas: k [20:55:44] ashgrigas, will do! thanks :] [20:55:51] hmm, ok, well, we can disable that later, right? if we make it write out, or sync to published-datasets [20:55:52] that was fast, thanks ash [20:55:53] that should be fine [20:55:54] right? [20:56:02] or milimetric if it is easiest, it can just do a local rsync_to: [20:56:12] right /srv/published-datasets [20:56:13] hm [20:56:16] rsync_to /a/published-datasets/ ... should be ok [20:56:17] yeah [20:56:21] (it is /a on stat1002 [20:56:25] unfortunetly [20:56:30] right, i'm on 1003 [20:56:32] oh ok [20:56:33] then yeah [20:56:39] (that's almost all the RU jobs) [20:57:18] yeah that seems fine, though it adds yet more delay on the reports getting out [20:57:23] 'cause they're rsync-ed twice now [21:02:22] ottomata: can I rsync it to thorium::/srv/analytics.wikimedia.org/datasets/periodic/ [21:03:09] hmm, not really milimetric beacuse of the rsync --delete stuff [21:03:21] hmm lemme think [21:03:23] ok [21:03:24] is that true? [21:03:38] yeah, it won't exist on 1002 or 1003 so it'll delete it, you're right [21:03:48] I think we've created a monster [21:03:56] yeah no, that won't work [21:04:01] haha [21:04:08] i mean, it can go to a different place [21:04:13] analytics.wm.org/reports [21:04:13] ? [21:04:40] milimetric: why does it have to rsync at all now though? [21:04:45] can't it just write out to /srv/published-datasets locally [21:04:46] ? [21:05:15] it's just the way RU works, it writes to $base/reportupdater/output/ [21:07:12] ottomata: can we make another directory on thorium under /srv/published-datasets-rsynced? [21:08:42] hm [21:09:17] milimetric: wonder if ln -s $base/reportupdater/output /srv/published-datasets/reports would work [21:10:22] yeah, that's kind of how it works now, limn-public-data is a link to RU/output [21:10:31] but we'd have to remove rsyncing from RU's puppet module [21:10:47] and I forget why we added it in the first place [21:11:21] also, small thing, I was thinking of calling it "periodic" instead of "reports" because it's more descriptive and I want people to use reportupdater for periodic jobs [21:11:29] (so we can review their code) [21:12:48] ottomata: Hiii you have some commits unmerged on puppet-master, can i merge them? :) [21:15:54] oop! [21:16:13] yes madhuvishy thank you [21:16:16] i thought i merged those! [21:16:17] cool [21:16:35] merging ;) [21:19:11] 06Analytics-Kanban, 06Discovery-Analysis, 07Browser-Support-Apple-Safari: Visits/searches from Safari 10 location bar search suggestions - https://phabricator.wikimedia.org/T157796#3027160 (10debt) p:05Triage>03Normal [21:23:22] (03PS3) 10Mforns: Add spark job to aggregate historical projectviews [analytics/refinery/source] - 10https://gerrit.wikimedia.org/r/337593 (https://phabricator.wikimedia.org/T156388) [21:23:39] (03CR) 10Mforns: Add spark job to aggregate historical projectviews (0310 comments) [analytics/refinery/source] - 10https://gerrit.wikimedia.org/r/337593 (https://phabricator.wikimedia.org/T156388) (owner: 10Mforns) [21:23:57] ok [21:24:11] milimetric: let's try the symlink thing then [21:24:23] 06Analytics-Kanban, 10Analytics-Wikistats, 13Patch-For-Review: Add "Interwicket" to the list of bots - https://phabricator.wikimedia.org/T154090#3027193 (10Nemo_bis) >>! In T154090#3026709, @gerritbot wrote: > Change 337632 had a related patch set uploaded (by Nuria): > Add "Interwicket" to the list of bots... [21:24:34] ottomata: https://gerrit.wikimedia.org/r/#/c/337672/ take a look and let me know what you think [21:24:44] and also what the best way to do the ln -s is [21:24:48] (in puppet) [21:25:29] man I ate one of the new beyond meat burgers, and my body is seriously confused [21:25:57] haha [21:25:58] what is that? [21:26:55] milimetric: you know how to declarea file in puppet? [21:27:00] same way, except you want to add [21:27:02] ensure => 'link [21:27:03] ' [21:27:04] and [21:27:05] 06Analytics-Kanban, 13Patch-For-Review: Clean up datasets.wikimedia.org - https://phabricator.wikimedia.org/T125854#3027202 (10Milimetric) Those last two patches should hopefully be all the code changes we need. Now we need to: 1. fix and merge the puppet one 2. test that dashboards work with it 3. merge the... [21:27:14] target => '/path/to/source/file' [21:27:16] so [21:27:40] k, no I know that way, just making sure it's correct in that context and you don't think we should factor it out into reportupdater or anything [21:28:10] (I'm talking about the two TODO lines in that puppet patch) [21:28:34] oh ok [21:28:35] hm [21:28:39] no i think there is good [21:28:41] hm [21:28:43] bye a-team, see you tomorrow! [21:28:48] might be nice to not hardcode the published-datasets path though [21:29:33] where are you defining it in the first place, for the rsync script? [21:29:48] the burger is this thing: https://www.youtube.com/watch?v=4-PT0VDKno0 [21:30:04] (03CR) 10Nuria: "@EBernhardson, if you clear merge conflicts we can probably merge this one. Thank you" [analytics/refinery] - 10https://gerrit.wikimedia.org/r/335158 (owner: 10EBernhardson) [21:30:07] milimetric: mind if iI amend your patch? [21:30:13] not at all, go for it [21:30:51] milimetric: so beyond meet == good? [21:31:46] it's not meat, but it tricks Tiger Lily and if you take a quick bite you might not notice. Definitely light years ahead of tofu and stuff like that [21:32:40] haha, this dude in the video says it has no odor, it definitely does, so don't trust whatever he says on taste [21:39:57] (03PS5) 10EBernhardson: Adjust drop hourly script to handle mediawiki log partitions [analytics/refinery] - 10https://gerrit.wikimedia.org/r/335158 [21:40:59] (03PS6) 10EBernhardson: Adjust drop hourly script to handle mediawiki log partitions [analytics/refinery] - 10https://gerrit.wikimedia.org/r/335158 [21:41:47] ebernhardson: doesn't camus create directories with hourly/ ... [21:41:47] ? [21:42:08] ottomata: yes, but its already in the table spec [21:42:18] ? [21:42:35] oh hourly is in your table exteranal LOCATION? [21:42:58] ottomata: yea [21:43:00] i think we should probably rename the 'raw' partition type to camus [21:43:09] and drop the 'hourly' part of partition_globs in the script [21:43:18] afaik there isn't anybody currently using partition type raw [21:43:21] at least its not puppetized.. :) [21:43:31] and just rely on --table-location flag [21:43:31] always [21:43:36] --location [21:43:37] hmm, thats reasonable if nothing is auto-detecting [21:44:41] milimetric: https://gerrit.wikimedia.org/r/#/c/337672/ [21:45:01] yep, I checked it out, thanks ottomata, commented [21:45:07] basically +1 but let's wait till tomorrow [21:45:18] we can do it tomorrow morning [21:45:43] k [21:48:30] (03PS7) 10EBernhardson: Adjust drop hourly script to handle mediawiki log partitions [analytics/refinery] - 10https://gerrit.wikimedia.org/r/335158 [21:51:04] in that case though, it seems we could almost skip changing the script and specify location without the 'hourly' part [21:51:40] (03CR) 10Ottomata: [V: 032 C: 032] Adjust drop hourly script to handle mediawiki log partitions [analytics/refinery] - 10https://gerrit.wikimedia.org/r/335158 (owner: 10EBernhardson) [21:51:46] thanks ebernhardson [21:56:04] ottomata: actually i just grabbed the latest copy of puppet, and it has a new refinery-drop-wdqs-extract-partitions that doesn't specify a partition type, so it's getting raw currently [21:57:18] oh nvm i can't read, it's using hive partition type [21:59:03] :) [21:59:06] yeah just short options [21:59:09] i just committed that today [22:01:11] 10Analytics-Cluster, 06Analytics-Kanban, 06Operations, 13Patch-For-Review: Move cloudera packages to a separate archive section - https://phabricator.wikimedia.org/T155726#3027384 (10Ottomata) TODO after CDH upgrade, remove old cloudera/thirdparty updates. [22:01:31] Gone for now a-team [22:01:36] see you tomorrow [22:01:53] good night^morning joal [22:41:18] 10Analytics, 06Research-and-Data: geowiki data for Global Innovation Index - https://phabricator.wikimedia.org/T131889#3027569 (10leila) @Milimetric I ran the following ``` select country, sum(edits) from staging.erosen_geocode_country_edits where ts like '2016-%' group by country order by sum(edits) desc; `... [23:32:28] 06Analytics-Kanban, 13Patch-For-Review: Add "Damn Small XSS Scanner" (DSXS) to list of known bots - https://phabricator.wikimedia.org/T157528#3027726 (10Tbayer) Thanks for the quick merge! To document the effect on total pageviews for later reference, here is a plot of the daily percentage (for the timespan wh...