[01:17:42] ori: out of idle historical curiosity, what does "millisec" mean in this context when the actual timestamp values only have a resolution of entire seconds? https://meta.wikimedia.org/w/index.php?title=Schema:EventCapsule&diff=5222982&oldid=5152617 [02:45:43] Hive seems to be kaput throwing tons of errors like: [02:45:47] https://www.irccloud.com/pastebin/BM0fGRTJ/ [02:50:40] ah no, maybe my tablescan was too large? [04:26:18] nuria: it looks like you were querying a partition that was just being written into [04:26:35] i see this happen in that case sometimes [07:18:44] Analytics: Compile a request data set for caching research and tuning - https://phabricator.wikimedia.org/T128132#2143580 (Danielsberger) [07:27:20] Analytics: Compile a request data set for caching research and tuning - https://phabricator.wikimedia.org/T128132#2143590 (Danielsberger) [07:44:36] (CR) Mforns: [C: 2 V: 2] Adding piwik reporting to browser-reports [analytics/dashiki] - https://gerrit.wikimedia.org/r/278960 (https://phabricator.wikimedia.org/T130653) (owner: Nuria) [09:01:29] Analytics: Upgrade aqs* to nodejs 4.3 - https://phabricator.wikimedia.org/T123629#2143751 (MoritzMuehlenhoff) Upgrade procedure: - Depool one of the aqs servers via conftool - Stop restbase - nodetool drain && systemctl stop cassandra - upgrade nodejs - start Cassandra and check whether it fully rejoined -... [09:04:34] Analytics, Operations: Upgrade aqs* to nodejs 4.3 - https://phabricator.wikimedia.org/T123629#2143752 (MoritzMuehlenhoff) [09:35:36] dcausse: around ? [09:35:46] joal: yes [09:35:50] Hi :) [09:35:52] hi! :) [09:36:07] dcausse: do you want us to merge / deploy this change https://gerrit.wikimedia.org/r/#/c/255105/ ? [09:36:17] In other words: has in been tested and all ? [09:36:49] joal: I think so, this patch has been here for a long a time :/ [09:36:55] It has indeed [09:37:18] I'm deployong today, so reviewing patches that have been left around in refinery-source [09:37:38] joal: do you deploy the new ApiAction schema? [09:37:38] Ok, so I merge and deploy that today - Is there anything to restart with that ? [09:37:48] dcausse: yes [09:38:08] joal: ok, everything has been updated to comply with this patch I think [09:38:34] dcausse: nothing happens with ApiAction for the moment IIRC [09:39:02] ok [09:39:05] I'll deploy refinery after source, so at that moment, if we start the new oozie import job, then yes, [09:39:33] also dcausse, there have been a big refactor from bd808 on oozie - I don't know if you've looked at it [09:39:52] no I only looked at the avro schema [09:39:59] It would be interesting for you and your team to read it :) [09:40:09] ok :) [09:40:15] Analytics-Cluster, Analytics-Kanban, Patch-For-Review: Add automatic failover to the Hadoop's name node - https://phabricator.wikimedia.org/T129838#2143788 (elukey) The last code review should add the necessary code, but I'd prefer to get the puppet-compiler working with submodules first (https://phab... [09:40:23] basically: generalisation of the mediawiki import jobs to include not CirrusSeasrch but other channels [09:40:30] dcausse: --^ [09:40:44] joal: nice! [09:40:49] dcausse: Merging youre patch :) [09:41:10] (CR) Joal: [C: 2] "Merging for deploy." [analytics/refinery/source] - https://gerrit.wikimedia.org/r/255105 (owner: DCausse) [09:41:10] joal: when will you mvn release? [09:41:17] dcausse: now ;0 [09:41:29] ok, +2 a schema change then [09:41:35] sure dcausse [09:43:45] (Abandoned) Joal: Revert "Update uap-core with latest code from upstream" [analytics/ua-parser] - https://gerrit.wikimedia.org/r/277306 (owner: Joal) [09:44:40] joal: merged, let me know when the artefact is deployed to archiva, I'll check that the new schema has been included [09:45:13] dcausse: submodule doesn't say it needs an update: weird [09:45:36] Have you updated refinery-source with new submodule sha? [09:45:40] https://gerrit.wikimedia.org/r/#/c/274312 [09:46:49] joal I don't know this part, I thought it would just update to the latest automatically [09:49:31] dcausse: trying to release, we'll check [09:50:28] (PS1) Joal: Update changelog.md for v0.0.28 release [analytics/refinery/source] - https://gerrit.wikimedia.org/r/279093 [09:50:34] elukey: around ? [09:53:42] joal: o/ [09:54:00] Do you mind having a look at --^ and merging ? [09:55:39] (CR) Elukey: [C: 2] Update changelog.md for v0.0.28 release [analytics/refinery/source] - https://gerrit.wikimedia.org/r/279093 (owner: Joal) [09:56:43] joal: is it --^ [09:56:54] elukey: Thanks mate :) [09:56:59] (PS1) DCausse: Update schema_repo submodule to latest [analytics/refinery/source] - https://gerrit.wikimedia.org/r/279094 [09:57:27] right dcausse, I assumed th [09:57:29] joal: you were right... I should have updated the submodule ref [09:57:33] :) [09:57:37] ok, merging that [09:57:51] (CR) Joal: [C: 2] "Merging before deploy." [analytics/refinery/source] - https://gerrit.wikimedia.org/r/279094 (owner: DCausse) [09:58:20] (CR) Joal: [V: 2] "Actually merging." [analytics/refinery/source] - https://gerrit.wikimedia.org/r/279094 (owner: DCausse) [10:00:39] (PS1) Joal: Update changelog.md before v0.0.28 release. [analytics/refinery/source] - https://gerrit.wikimedia.org/r/279096 [10:01:23] (CR) Joal: [C: 2] "Self merging for deploy." [analytics/refinery/source] - https://gerrit.wikimedia.org/r/279096 (owner: Joal) [10:08:39] (CR) Joal: [V: 2] Update changelog.md before v0.0.28 release. [analytics/refinery/source] - https://gerrit.wikimedia.org/r/279096 (owner: Joal) [10:11:01] !log Release refinery-source v0.0.28 to archiva [10:26:14] joal: refinery-camus-0.0.28 looks good, thanks! [10:54:49] great dcausse, thanks ! [11:05:58] (CR) Joal: [C: 2 V: 2] "Merging for release." [analytics/refinery] - https://gerrit.wikimedia.org/r/278325 (https://phabricator.wikimedia.org/T130187) (owner: Joal) [11:11:47] (PS1) Joal: Add v0.0.28 jar artifacts [analytics/refinery] - https://gerrit.wikimedia.org/r/279105 [11:12:17] elukey: if still around, couple of reviews to come :) [11:12:32] suuureeee [11:12:37] Thanks man :) [11:13:06] joal: I am working with Giuseppe to fix the puppet-compiler to work with submodules change, after that I'll test the HA change on analytics1001 [11:13:18] this is why I didn't ask you to stop HDFS this morning [11:13:18] Sounds great ! [11:13:31] np elukey, I am in no rush :) [11:18:20] (PS1) Joal: Bump jar version and record version in refine job [analytics/refinery] - https://gerrit.wikimedia.org/r/279107 [11:21:42] joal: I have no idea (shame on me) about https://gerrit.wikimedia.org/r/#/c/279105/1, but if you checked all the hash and you are sure I'll merge [11:22:19] don't also know about what the "record version" mean in https://gerrit.wikimedia.org/r/279107, but if it is ok the rest looks fine :) [11:22:23] elukey: I've add no issues so far with that process, but I didn't double check hashes [11:23:28] Analytics: Fix phab script to gather stats also from point field , not only from title - https://phabricator.wikimedia.org/T130543#2143962 (mforns) a:mforns [11:23:41] Analytics-Kanban: Fix phab script to gather stats also from point field , not only from title - https://phabricator.wikimedia.org/T130543#2138954 (mforns) [11:28:19] joal: would it be something to do or we can proceed anyway? [11:29:35] elukey: I think we're fine :) [11:29:52] okok! [11:29:57] Merging [11:30:28] (CR) Elukey: [C: 2] "Discussed with Joseph on IRC!" [analytics/refinery] - https://gerrit.wikimedia.org/r/279107 (owner: Joal) [11:30:41] (CR) Elukey: [V: 2] "Discussed with Joseph on IRC!" [analytics/refinery] - https://gerrit.wikimedia.org/r/279107 (owner: Joal) [11:31:25] (CR) Elukey: [C: 2 V: 2] "Discussed with Joseph on IRC" [analytics/refinery] - https://gerrit.wikimedia.org/r/279105 (owner: Joal) [11:32:01] ---^ joal [11:32:35] great, a last one if you don't mind https://gerrit.wikimedia.org/r/279107 [11:36:14] elukey: --^ [11:36:37] already merged joal! [11:36:57] Oh sorry elukey ! [11:37:00] Thanks you :) [11:37:39] joal: going afk for ~1hr, will brb! [11:37:45] Sure elukey, np [11:46:51] !log Deploy refinery [11:50:34] !log stop/restart load and refine jobs [12:27:55] !log fix bug in load job manually [12:30:42] team, will be back before standup :] [12:30:50] Bye mforns_afk [12:34:18] Analytics-Engineering, Analytics-Wikimetrics: Removing usernames from a cohort doesn't work - https://phabricator.wikimedia.org/T130714#2144078 (Halibutt) [12:36:59] Analytics-Engineering, Analytics-Wikimetrics: Can't correct names in a cohort - https://phabricator.wikimedia.org/T130715#2144091 (Halibutt) [12:40:55] (PS1) Joal: Correct bug in error email text in load job [analytics/refinery] - https://gerrit.wikimedia.org/r/279111 [12:41:23] (CR) Joal: [C: 2 V: 2] "Self-merging, already corrected manually in prod." [analytics/refinery] - https://gerrit.wikimedia.org/r/279111 (owner: Joal) [12:51:43] back! [13:37:59] a-team, I'm AFK, will be back for standup [13:38:10] hi ok! [13:38:42] ottomata: hellooooo [13:38:50] hiiya! [13:39:13] do you have time for a puppet-compiler discussion? [13:39:38] I am trying to add submodules support to it, but my brain is kinda twisted [13:44:32] ohhh [13:44:33] sure yeah [13:44:34] sounds fun [13:44:42] i don't know puppet compiler well [13:44:46] but i do know puppet-merge well [13:44:53] and i bet the issue might be silmiar [13:45:00] elukey: ^ [13:45:01] bat-cave? [13:45:05] ok [13:45:08] one sec lemme get headphones [14:09:16] ottomata: mmmm git submodules update does get the new commits and ends up in a detached state.. I tried rebase and it does the same [14:09:32] the submodules detached? [14:09:33] ja, makes sense [14:09:40] it just checks out annamed sha [14:09:43] unnamed [14:09:46] which == detached, no? [14:10:05] iyeah [14:10:08] on my local ops puppet [14:10:11] cd modules/nginx [14:10:12] git status [14:10:16] HEAD detached at e74a5c7 [14:10:18] that's normal [14:11:15] all right but didn't you end up having the code on the host (testing it too)? [14:11:31] maybe I am missing something trivial [14:12:05] detached state is the default for git submodules [14:12:32] elukey: detached just means that the checked out commit doesn't have a name [14:12:33] i don't even think it's possible to make a submodule track a specific branch [14:12:34] like a branch or a tag [14:13:24] ottomata: sure sure but all the new commits gets fetched and merged theoretically, and you end up testing all of them rather than the one that you want no? [14:15:00] elukey: you can do whatever you want to the submodule, e.g. cd submodule && git cherry-pick your_sha1_here or whatever [14:15:10] then just git checkout -- submodule_dir [14:15:14] to get it back [14:15:29] elukey: ? [14:15:41] the new commits will only get merged for submodules that actually had sha changes [14:18:41] ottomata: ok I checked git submodule status after the git submodule update and the sha is sound [14:19:22] I am probably confused because if I go in 'cd modules/cdh' I can see the new detached state [14:19:43] yeah detached is 100% normal [14:19:50] mobrovac: sorry you are right, I seem a bit crazy but I am trying to add submodules support to puppet-compiler [14:20:16] elukey: you are? then i'm afraid you are indeed crazy :D [14:20:36] good luck! [14:20:38] mobrovac: thank you! I like supporting people [14:20:40] :P [14:20:43] hahaha [14:21:00] fwiw it shouldn't be too hard [14:21:19] git submodule update --init is a no-op if no changes were made to them [14:23:19] Analytics-Kanban, Wikipedia-Android-App-Backlog: Count requests to RESTBase from the Android app - https://phabricator.wikimedia.org/T128612#2144362 (Nuria) > should be sending requests with "x-analytics: preview=1". Can android devs confirm (via wireshark or similar) that app is sending the header? [14:24:06] mobrovac: yeah I was talking with andrew about the use case of having multiple modules changed (so alredy merged in their repo) but only testing a code review to merge in the main puppet repo the new sha (like https://gerrit.wikimedia.org/r/#/c/279091/) [14:24:28] OHH, elukey ah, that's where you are confused, i see [14:24:34] a merge in a submodule's repo does nothing to ops/puppet [14:24:48] no no yes I know tha! [14:24:51] *that! [14:24:51] oh ok [14:25:12] the only thing that will change ops puppet is if the submodule's sha changes [14:25:15] in ops/puppet [14:25:21] whihc takes a commit to ops/puppet [14:25:23] which is what you want to test [14:34:14] ottomata: you wanna deploy this thing? [14:34:19] (the dumps change) [14:36:38] oh ok! [14:36:39] yeah [14:37:13] milimetric: i lost link, you should add me as reviewer [14:37:54] done: https://gerrit.wikimedia.org/r/#/c/269696/ [14:38:25] a-team: all ok with my draft on the etherpad? [14:38:28] https://etherpad.wikimedia.org/p/analytics-email-drafts [14:39:43] ok milimetric should I merge? [14:40:05] ottomata: sure, and then we'll deploy and revert if needed [14:40:16] ottomata: https://gerrit.wikimedia.org/r/#/c/279117/3/puppet_compiler/prepare.py <-- this is basically what we discussed [14:40:20] best way to test evah :/ [14:41:09] looks good elukey! [14:44:07] milimetric: merged and run [14:44:13] checking [14:44:21] http://dumps.wikimedia.org/analytics [14:44:35] ooops [14:44:37] :( [14:44:43] why 404? I didn't set it up right? [14:44:51] where is it supposed to go? [14:44:54] there [14:45:13] https://gerrit.wikimedia.org/r/#/c/269696/3/modules/dataset/manifests/html.pp [14:45:26] OH sorry [14:45:29] i didn't read puppet output [14:45:30] ja [14:45:33] analytics dir doesn't exist [14:45:40] you gotta create it if you want to put something in it [14:45:43] aha [14:45:47] oops, sorry [14:47:37] ottomata: are you making the dir or should I submit a patch? [14:48:55] milimetric: submit a patch [14:48:59] i can make the dir real quick to fix [14:49:10] but puppet should make the dir [14:49:18] milimetric: sorry, am also chatting in other rooms [14:49:54] oh milimetric [14:49:58] maybe it goes in dataset::dirs? [14:50:04] (this module is nuts!) [14:50:44] milimetric: ^ oh man this module makes no sense [14:51:01] uh :) [14:51:03] why the heck would you always create these dirs, and then conditionally create index files?! [14:51:15] anyway, i guess just be consistent with the module, add the dir in dataset::dirs class [14:51:39] k [14:57:39] ok, thanks ottomata, all looks good now, sending the email [14:58:43] cool! [14:59:43] joal, milimetric: is aqs stateless, can I simply restart the service via "systemctl restart aqs.service"? (need to restart it for an openssl update) [14:59:59] moritzm: yes, but would you please restart one node at a time? [15:00:12] that way we can still serve requests while it's coming up [15:00:33] sure, will then start in 10 mins or so [15:00:37] thx! [15:01:05] and thanks for keeping us up to date, moritzm! much appreciated [15:13:43] Analytics-Cluster, Discovery, Maps, RESTBase-Cassandra, Patch-For-Review: Create separate Kibana dashboards for production Cassandra clusters - https://phabricator.wikimedia.org/T130393#2144505 (Eevans) Open>Resolved Separate dashboards for Maps and AQS have been created using the new `c... [15:13:59] Analytics-Cluster, Discovery, Maps, RESTBase-Cassandra, Patch-For-Review: Create separate Kibana dashboards for production Cassandra clusters - https://phabricator.wikimedia.org/T130393#2144507 (Eevans) p:Triage>Normal [15:21:02] aqs has been restarted on all aqs* hosts [15:22:29] Thanks moritzm :) [15:31:10] Analytics-Kanban, EventBus, Patch-For-Review: Add --topic-prefix config to eventlogging-service - https://phabricator.wikimedia.org/T130562#2139475 (Ottomata) [15:53:33] Analytics, Pageviews-API, Services, RESTBase-API: Document that wikimedia pageviews API is blocked by ad blockers - https://phabricator.wikimedia.org/T126947#2144588 (Milimetric) Does anyone have any contact with the block list maintainers? I can poke randomly around the internet otherwise :) [15:58:10] Analytics-Kanban: Back-fill pageviews data for dumps.wikimedia.org to May 2015 - https://phabricator.wikimedia.org/T126464#2144609 (Milimetric) @elukey: I forgot to mention, the process is that @Nuria is the only one who closes tasks as resolved. That way she can "accept" that they're done. [16:01:05] (PS1) Ottomata: Remove single use of x_forwarded_for in insert_hourly_pagecounts.hq [analytics/refinery] - https://gerrit.wikimedia.org/r/279132 (https://phabricator.wikimedia.org/T118557) [16:01:16] joal: ^ [16:02:19] (CR) Joal: [C: 2 V: 2] "LGTM !" [analytics/refinery] - https://gerrit.wikimedia.org/r/279132 (https://phabricator.wikimedia.org/T118557) (owner: Ottomata) [16:04:30] joal: fyi [16:04:33] looks fine [16:04:36] hive (otto)> select uri_path, x_forwarded_for from otto.webrequest_null_xff; [16:04:36] OK [16:04:36] uri_path x_forwarded_for [16:04:36] /w/load.php NULL [16:05:03] awesome ottomata [16:05:11] I'd have expected that, good to confirm [16:06:09] ottomata: shall we go and merge bd808 change? [16:06:20] its merged, no? [16:06:27] just not deployed? [16:06:54] the x_forwarded_for one ? [16:18:20] Analytics: Evaluate deprecation of udp-like files produced on oozie (legacy tsvs) - https://phabricator.wikimedia.org/T130725#2144649 (Nuria) [16:35:59] Analytics-Kanban, Patch-For-Review: Add piwik reporting to browser reports [1] - https://phabricator.wikimedia.org/T130653#2144710 (Nuria) [16:50:47] Analytics: Check if we can deprecate legacy TSVs production (same time as pagecounts?) - https://phabricator.wikimedia.org/T130729#2144772 (JAllemandou) [16:52:40] Analytics: Make webrequest load and refine jobs a single bundle - https://phabricator.wikimedia.org/T130731#2144802 (JAllemandou) [16:53:59] Analytics: Standardise naming in oozie jobs (particularly for top level ones), and/or force top-level jobs to be bundles - https://phabricator.wikimedia.org/T130732#2144819 (JAllemandou) [16:55:09] Analytics: Check if we acn merge maps partition into misc partition at varnishkafka level - https://phabricator.wikimedia.org/T130733#2144836 (JAllemandou) [16:56:46] Analytics: Document a proposal for bundling other than load-refine jobs together (see refine/diagram) - https://phabricator.wikimedia.org/T130734#2144849 (JAllemandou) [16:57:07] a-team: five new tasks created, please review :) [16:57:29] :) [16:58:29] ottomata: about bd808 change, I think this one https://gerrit.wikimedia.org/r/#/c/253474/ has not been merged [16:58:33] Maybe I'm wrong [17:05:57] nuria, the phab script now seems to work, but as we are marking the tasks in Done as resolved in a more batched manner, the script returns distorted results [17:07:14] nuria, we should either: count the points of the tasks that have been moved to Done, instead of marked as resolved; or we mark the tasks in Done as resolved as soon as possible [17:07:30] mforns: on meeting can talk later [17:07:35] sure, sorry [17:11:17] I have data! `kafkacat -b kafka1012 -t mediawiki_ApiAction -c 1` is pulling records down on stat1002 [17:11:57] joal: I don't think that patch from bblack is directly related to my metrics. [17:16:38] Analytics, MediaWiki-API, Reading-Infrastructure-Team, MW-1.27-release-notes, and 4 others: Publish detailed Action API request information to Hadoop - https://phabricator.wikimedia.org/T108618#2144937 (bd808) >>! In T108618#2132875, @bd808 wrote: > New checklist: > [x] [[https://gerrit.wikimedia.o... [17:17:11] Analytics, MediaWiki-API, Reading-Infrastructure-Team, MW-1.27-release-notes, and 4 others: Publish detailed Action API request information to Hadoop - https://phabricator.wikimedia.org/T108618#2144939 (bd808) [17:19:31] Analytics, MediaWiki-API, Reading-Infrastructure-Team, MW-1.27-release-notes, and 4 others: Publish detailed Action API request information to Hadoop - https://phabricator.wikimedia.org/T108618#2144970 (bd808) Verified that data is reaching Kafka from MediaWiki via `kafkacat -b kafka1012 -t mediawi... [17:24:23] Analytics, Hovercards: Capture hovers as previews in analytics - https://phabricator.wikimedia.org/T129425#2145023 (dr0ptp4kt) @bearND do you have the code samples handy for header enrichment for previews? [17:32:40] lunchin! back in a bit [17:32:55] ah joal [17:33:01] that is b bblack, not bd808 that's why i was confused [17:33:05] wasn't sure what you were talking about [17:33:12] yes, that has not been merged yet [17:37:01] (ah still here for a bit...) [17:38:03] joal: i think xff vk patch can be merged even if we haven't restarted the legacy tsv job [17:38:09] or xff = '-' will just do nothing [17:38:12] since it will always be null [17:38:29] bd808: awesome! [17:39:05] milimetric: i think you don't care where the actual analytics/index.html lives, right? [17:39:11] as long as the main other page links to it? [17:39:26] so, add redirect in nginx if you like, but i think the link being other/analytics/index.html is fine, no? [17:39:48] yeah, I just did that with my new patch [17:39:55] so I both moved it and added the rewrite [17:40:30] ah ok [17:40:53] oh, hm [17:41:06] wait milimetric what's the point of doing a rewrite [17:41:07] ? [17:41:08] permanent [17:41:16] i thought you were maybe going to do it transparently [17:41:20] so the file lives at other/analytics [17:41:24] but the link is still /analytics [17:41:37] this link is new enough that i htink you can leave off a full permanent redirect [17:41:38] if you want [17:41:40] I just copied what was there for fundraising, it seemed to be the same thing, I don't really know nginx [17:41:45] the way you have it [17:41:52] it will cahnge the url in the browser to other/analytics [17:42:04] yea, that [17:42:07] you want that? [17:42:08] why [17:42:08] that's ok, right? [17:42:09] ? [17:42:11] sure it sok [17:42:14] but what's the point? [17:42:36] it just seemed closer to what Ariel would like? [17:42:52] i think ariel just wants you to keep files out of root [17:42:54] that's it [17:43:05] a perm redirect would be for if you needed to move it,but wanted to keep something backwards compatible [17:43:09] but this is new and you dont' need that [17:43:18] aha, ok, so what's the code for the transparent kind of rewrite? [17:43:26] a transparent rewrite might be nice if you wanted it, so that the url would be /analytics, but the file would be other/analytics [17:43:28] not sure! [17:43:29] heheh [17:43:47] i gotta run to lunch, sorry! but i'd say don't worry about it, if all the other non xml dumps are in /other/ [17:43:52] might as well just use /other/analytics [17:43:54] no? [17:44:08] my opinion: don't do any rewrite, just use other/analytics [17:44:12] unless you really want the url to be /analytics [17:44:24] ok, gotta go, will merge after lunch, sok? or you can get ariel or someoe else to merge [17:44:40] Analytics, Hovercards: Capture hovers as previews in analytics - https://phabricator.wikimedia.org/T129425#2145178 (bearND) @Jdlrobson, @dr0ptp4kt We just add a header like this: ``` x-analytics: preview=1 ``` This is for both [[ https://phabricator.wikimedia.org/diffusion/APAW/browse/master/app/src/ma... [17:44:44] oh, ottomata I would agree except the email went out and it has /analytics in there [17:44:54] so I figure if that's how people click through they shouldn't see a 404 [17:45:25] ottomata: have a good lunch, no worries or rush, I think this is a non-breaking style change, we can merge it when Ariel's ok with it [17:46:12] ook ok! [18:01:39] bd808: many apologizes for the name confusion :) [18:02:30] bd808: I deployed today a new version of refinery and refinery source [18:02:46] bd808: Are you hove/oozie jobs ready to be running\ ? [18:05:56] Analytics, Operations, hardware-requests, Patch-For-Review: eqiad: (3) AQS replacement nodes - https://phabricator.wikimedia.org/T124947#2145301 (RobH) The systems that can be used for this were ordered today on T130738. I'm now assigning this task to @Mark.Otaris @Mark: Please review the above... [18:13:17] milimetric: When exactly do stats for a new day become available via the pageviews API? Right after midnight UTC? midnight some other time zone? Some time later? [18:13:20] Analytics-Cluster, Operations, hardware-requests: eqiad: New Hive / Oozie server node in eqiad Analytics VLAN - https://phabricator.wikimedia.org/T124945#2145344 (RobH) Sorry about that @ottomata, this was assigned to @mark so I missed your question. System WMF4541 is an older spare pool system with... [18:13:59] kaldari: it's not a fixed time, and we haven't figured out the average or maximum [18:14:31] kaldari: we're continuously ingesting data, and sometimes there's loss, duplication, etc. so cleaning it takes longer. Sometimes jobs have to be restarted, and that takes longer too [18:14:43] milimetric: any rough guess? [18:14:50] and it's a somewhat complicated pipeline so things could go wrong in 3 or 4 places. [18:15:20] kaldari: from what I've seen, data for the previous day is usually available within 5 hours of midnight. [18:15:33] but that's not very scientific [18:15:51] Analytics-Cluster, Operations, hardware-requests: eqiad: New Hive / Oozie server node in eqiad Analytics VLAN - https://phabricator.wikimedia.org/T124945#2145365 (mark) Approved. [18:16:38] milimetric: and that's midnight UTC right? [18:16:46] yes [18:16:58] Analytics-Cluster, Operations, hardware-requests: eqiad: New Hive / Oozie server node in eqiad Analytics VLAN - https://phabricator.wikimedia.org/T124945#2145374 (RobH) a:mark>RobH [18:17:14] a-team: talk with you tomorroooowwww o/ [18:17:20] bye elukey ! [18:17:42] bye elukey [18:40:01] Analytics-Cluster, Operations, hardware-requests: eqiad: New Hive / Oozie server node in eqiad Analytics VLAN - https://phabricator.wikimedia.org/T124945#2145490 (Ottomata) Thanks! [19:19:21] Analytics, Hovercards: Capture hovers as previews in analytics - https://phabricator.wikimedia.org/T129425#2145687 (Nuria) Loopin @Tbayer here so he knows what to expect, please do not do any code changes quite yet as we are meeting to talk about this in the next couple days. [19:20:19] Analytics, Hovercards: Capture hovers as previews in analytics - https://phabricator.wikimedia.org/T129425#2145690 (Nuria) >The moment more than one key-value pair starts to be sent for this header, a more deluxe solution will be required to ensure different client side authors don't trump each other's c... [19:25:51] (PS1) Mforns: Fix script after Phabricator layout changes [analytics/limn-analytics-data] - https://gerrit.wikimedia.org/r/279177 (https://phabricator.wikimedia.org/T130543) [20:17:39] I am about to deploy a new versions of browser-reports with some piwik code cc milimetric [20:29:56] joal: I think that there should be data for the ApiAction oozie job now. I'm not sure how to check to see if camus is pulling the data from kafka into hdfs though. [20:30:14] bd808: checkin /wmf/data/raw/mediawiki... [20:31:54] ottomata: /mnt/hdfs/wmf/data/raw/mediawiki/mediawiki_ApiAction/hourly exists but is empty [20:33:20] bd808: in meeting, with you afer [20:33:27] np. thx [20:44:13] (PS1) Madhuvishy: TEST COMMIT DO NOT MERGE [analytics/refinery/source] - https://gerrit.wikimedia.org/r/279187 [20:44:37] (Abandoned) Madhuvishy: TEST COMMIT DO NOT MERGE [analytics/refinery/source] - https://gerrit.wikimedia.org/r/279187 (owner: Madhuvishy) [20:46:05] (PS1) Madhuvishy: Test commit do not merge [analytics/refinery/source] (release) - https://gerrit.wikimedia.org/r/279188 [21:00:24] bd808: ok hey [21:00:25] looking [21:04:10] hmm, bd808 [21:04:10] java.lang.RuntimeException: null record [21:04:24] at com.linkedin.camus.etl.kafka.mapred.EtlRecordReader.nextKeyValue(EtlRecordReader.java:295) [21:07:03] ottomata: hmm... does that mean my Avro schema is wrong or ?? [21:07:31] * bd808 is multi-tasking with RfC meeting [21:08:35] no tsure [21:09:11] i'm looking at code to see if i could see a path for that [21:13:30] ottomata: looks like it might be schema related -- https://github.com/linkedin/camus/blob/master/camus-etl-kafka/src/main/java/com/linkedin/camus/etl/kafka/mapred/EtlRecordReader.java#L139-L156 [21:16:11] yeah, maybe, but there's only one version of that schema out there, no? [21:16:15] the one in event-schemas? [21:16:28] unless..maybe refinery-camus hasn't been rebuilt and deployed with that schema [21:16:30] checking [21:18:18] gotta run in 3 mins tough... [21:18:38] ottomata: if we have data by the end of the week I'll be happy :) [21:18:43] hmm, i tlooks there [21:18:44] ok [21:18:47] bd808: yeah sorry [21:18:51] no worries [21:18:58] i'm unfortunetly not familiar with the avro/camus integration... [21:19:05] madhuvishy: miiiight be able to help you [21:19:21] fwiw i was able to consume an offending offset that caused that error in camus from kafka, and then read the data out of it using that schema [21:19:30] bd808: it looks to me like its there in refinery-camus.jar [21:19:34] 1464 Wed Mar 23 11:20:44 UTC 2016 schema_repo/avro/mediawiki/ApiAction/101453221640.avsc [21:19:45] was deployed this morning i think by joal maybe [21:19:49] but could have been there longer too [21:20:05] I'll try to decode a packet in a bit. If I get really stuck I'll poke some discovery folks who have dealt with this pipeline. [21:20:37] btw, in case you need the schemas, they are now also checked out on stat1002 at /srv/event-schemas [21:20:50] ok, laters!