[00:09:27] Analytics-Engineering, Analytics-Cluster: automate deletion of old logs in stat1002:/a/squid/archive/edits - https://phabricator.wikimedia.org/T87246#1002843 (kevinator) [00:15:33] milimetric: yt? [00:16:14] nuria: do you know how to reboot pentaho? [00:17:14] kevinator: resurrect you mean, cause reboot is not going to do it but yeah [00:17:33] ahh [00:17:52] yeah, it still has a pulse (502 bad gateway message) [00:18:22] kevinator: let me take a look [00:19:10] thanks [00:37:14] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Vital Signs user clicks "Add Metric" then Daily Pageviews (New Definition) - https://phabricator.wikimedia.org/T86141#1002897 (kevinator) [00:38:26] kevinator: sorry but something is up with labs now, machine is real slow and unresponsive (This is unrelated to pentaho) [00:39:18] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Vital Signs user clicks "Add Metric" then Daily Pageviews (New Definition) - https://phabricator.wikimedia.org/T86141#962128 (kevinator) [00:39:54] kevinator: something is up with labs machine is very slow and unresponsive [00:40:12] nuria: ahh [00:40:13] kevinator: (nothing to do with pentaho per se) [00:40:20] ok [00:40:45] kevinator: i cannot even use the command line so i guess this is going to have to wait [00:41:12] Carolynne was wanting to use it. I’ll let her know it will not get resolved tonight and we’ll see if labs is better tomorrow [00:55:34] ottomata: yt? [00:58:51] yes but deep in some code! [01:18:58] ottomata: ok, lemme know when you are out of the deepness [01:22:20] hm, nuria go ahead [01:22:34] ottomata: looked at data for mobile jobs [01:22:55] https://www.irccloud.com/pastebin/umVJWz8s [01:23:31] ah ya? [01:23:43] hm, [01:23:44] the 0 pad didn't come out, eh? [01:23:49] and the only file that looks right is 2015-01-28.gz and i am wondering what is up with "/wmf/data/archive/mobile_apps/uniques/daily/2015-1-29.gz" cause that looks is from an "older" job [01:23:56] no, it did for the 28th [01:24:18] the 28th has "good" data [01:24:29] ottomata: rest of files look from prior runs [01:25:04] hm, btw, nuria, do you have a bunch of test coordinators left in running state for this? [01:25:22] ottomata: no, i should have kill them all [01:25:35] $ oozie jobs -jobtype coordinator -filter status=RUNNING | grep mobile_apps_daily_uniques-coord [01:25:35] 0016981-150116220452628-oozie-oozi-C mobile_apps_daily_uniques-coordRUNNING 1 DAY 2015-01-28 00:00 GMT 2015-01-31 00:00 GMT [01:25:35] 0004041-150116220452628-oozie-oozi-C mobile_apps_daily_uniques-coordRUNNING 60 MINUTE 2015-01-08 00:00 GMT 2015-01-30 02:00 GMT [01:25:35] 0004028-150116220452628-oozie-oozi-C mobile_apps_daily_uniques-coordRUNNING 60 MINUTE 2015-01-08 00:00 GMT 2015-01-30 02:00 GMT [01:25:35] 0003988-150116220452628-oozie-oozi-C mobile_apps_daily_uniques-coordRUNNING 60 MINUTE 2015-01-08 00:00 GMT 2015-01-30 02:00 GMT [01:25:35] 0000265-150116220452628-oozie-oozi-C mobile_apps_daily_uniques-coordRUNNING 60 MINUTE 2015-01-08 00:00 GMT 2015-01-30 02:00 GMT [01:25:35] 0000015-150116163703781-oozie-oozi-C mobile_apps_daily_uniques-coordRUNNING 60 MINUTE 2015-01-08 00:00 GMT 2015-01-30 02:00 GMT [01:25:36] 0026318-141210154539499-oozie-oozi-C mobile_apps_daily_uniques-coordRUNNING 1 DAY 2015-01-16 00:00 GMT 2015-01-31 00:00 GMT [01:25:39] I kill them with " oozie jobs -jobtype coord | grep mobile |awk '{print $1}' | xargs oozie job -kill" [01:26:41] ok, not sure how those survived but they shoudl be killed and we should have just 1 running [01:26:54] if I test i will prefix them with "TESTING" [01:27:14] yeah, hm, i think maybe i didn'g kill one too? [01:27:23] i see a daily one running since 01-15 [01:27:27] just killed that [01:27:36] that looks like one I submitted? [01:27:37] hm [01:28:08] I'm going to kill them all, and restart this job, and remove the 01-28 data and start it from there [01:28:10] see how it does [01:28:13] s'ok? [01:28:15] ottomata: I do not see those doing oozie jobs -jobtype coord | grep mobile [01:28:46] ottomata: ok [01:28:46] i just killed them [01:28:46] i think [01:29:15] ottomata: ok, then data as of tomorrow should be good [01:29:24] i'll start it for yesterday [01:29:27] so it launches immediately [01:29:57] ottomata: sounds great [01:30:17] nuria [01:30:17] 0023327-150116220452628-oozie-oozi-C [01:30:24] fingers crossed [01:30:43] ottomata: sorry about the old jobs but I thought all displayed whne doing: oozie jobs -jobtype coord | grep mobile [01:30:48] oh cool, it launched yesterday and today [01:30:55] ottomata: of course! [01:30:58] i would think that would work, nuria [01:30:58] dunno [01:31:34] ottomata: thank you. will look at data in a bit and see how it looks [01:39:32] k cool [02:29:59] it always disturbs me when I appear to have fixed a bug but don't understand why. [06:10:45] (PS10) Ananthrk: Added Geocode core class that uses Maxmind-V2 Added GeocodedCountryUDF to get country code from IP address Added GeocodedDataUDF to get Geo data from IP address Added new dependency maxmind-geoip2 in refinery-core Added Maxmind's test databases to resourc [analytics/refinery/source] - https://gerrit.wikimedia.org/r/183551 [06:39:52] (PS1) Ananthrk: Added IpUtil class for IP address related utility methods Added ClientIpUDF to get client IP address from source IP and X-Forwarded-For header values [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187651 [08:15:16] Analytics-Engineering: Add time range selection to Limn dashboards (or new Dashiki dashboards) - https://phabricator.wikimedia.org/T87603#1003892 (Tgr) This would be useful to double-check assumptions, amongst other things. Currently creating a graph is expensive so one typically makes one graph for a given i... [15:25:43] Analytics-Engineering, Analytics-Wikimetrics: Wikimetrics should run latest puppet w/o local changes - https://phabricator.wikimedia.org/T88116#1004281 (kevinator) p:Triage>High [15:26:52] Analytics-Engineering, Analytics-Wikimetrics: Wikimetrics should run latest puppet w/o local changes - https://phabricator.wikimedia.org/T88116#1004284 (Nuria) Remember: git pull --rebase when pulling [15:46:29] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Vital Signs user clicks "Add Metric" then Daily Pageviews (New Definition) - https://phabricator.wikimedia.org/T86141#1004318 (Nuria) There are 4 different steps: #1 Pageview data goes into refined tables (done) #2 oozie job to compute pageview agr... [15:49:38] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Vital Signs user clicks "Add Metric" then Daily Pageviews (New Definition) - https://phabricator.wikimedia.org/T86141#1004322 (Nuria) Some oozie info: https://wikitech.wikimedia.org/wiki/Analytics/Cluster/Oozie Code for other oozie jobs: https://git... [15:52:05] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Vital Signs user clicks "Add Metric" then Daily Pageviews (New Definition) - https://phabricator.wikimedia.org/T86141#1004324 (Nuria) More docs: https://wikitech.wikimedia.org/wiki/Analytics/Pagecounts-all-sites Code for step #3 (python) https://gi... [16:39:24] Wikidata, Analytics: active user statistics that have less lag than wikistats - https://phabricator.wikimedia.org/T88121#1004364 (JanZerebecki) NEW [17:36:50] nuria, pulling with --rebase is only for production? or is it also for staging? [17:36:59] mforns: both [17:37:11] mforns: staging connects to prod enwiki just teh same [17:37:21] it's just that the wikimetrics db is different [17:37:26] aha [17:38:11] so I pulled with --rebase in staging, but I can not see any file with changes... [17:38:24] what files should be different from origin production, nuria ? [17:38:52] mforns: there should be a commit on top with passwords i believe, [17:38:58] so you cann connect to db [17:39:06] as passwords are not on puppet [17:40:20] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Analytics Engineer has an oozie job to aggreate page views by time - https://phabricator.wikimedia.org/T88125#1004447 (kevinator) NEW [17:40:27] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Analytics Engineer has an oozie job to aggreate page views by time - https://phabricator.wikimedia.org/T88125#1004447 (kevinator) Some oozie info: https://wikitech.wikimedia.org/wiki/Analytics/Cluster/Oozie [17:41:27] nuria, thanks, anyway I don't see the files, I'll have a look at production and see if I can find them [17:41:47] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Analytics Engineer has a python job to aggregate by project - https://phabricator.wikimedia.org/T88127#1004465 (kevinator) NEW [17:44:24] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Vital Signs user has access to the new page view data - https://phabricator.wikimedia.org/T88128#1004474 (kevinator) NEW [17:44:49] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Vital Signs user has access to the new page view data - https://phabricator.wikimedia.org/T88128#1004474 (kevinator) [17:47:26] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Analytics Engineer has a python job to aggregate by project - https://phabricator.wikimedia.org/T88127#1004484 (kevinator) [17:54:26] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Vital Signs user clicks "Add Metric" then Daily Pageviews (New Definition) - https://phabricator.wikimedia.org/T86141#1004505 (kevinator) [17:54:32] nuria, found them [17:55:23] mforns: ok [17:59:34] hey mforns [17:59:42] hi DarTar [18:00:01] safely back in paradise island? [18:00:07] xD, yes [18:00:13] good [18:00:17] a bit jet-lagged [again] [18:00:33] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Analytics Engineer has an oozie job to aggreate page views by time - https://phabricator.wikimedia.org/T88125#1004517 (kevinator) [18:01:36] quick one: [18:01:51] I noticed that the latest cube (v.0.4) in Pentaho has garbled country names for non ASCII characters [18:04:03] mforns: ^ [18:04:22] sorry, DarTar [18:04:37] I see [18:05:16] havin a look DarTar [18:05:33] cool thx [18:17:18] ottomata: The cron that cleans Hive data will kill the data for 2015-01-01, but I could not yet backfill the glam_nara tsvs, as the geocoding udf is not yet merged. [18:17:35] Do you think it'd be ok if I switched the 31 deletion interval to 35 days? [18:18:03] (That way the udf could get more time to review/merge, and I could still backfill afterwards) [18:18:20] Meh ... s/35/40/ [18:21:43] qchris, i think that's fine? or, i am ready to merge that thing today i think [18:21:46] sorry, just talking to toby right now [18:21:55] we don't have the IP udf yet though [18:22:07] The glam_nara does not use the IP stuff. So that's ok [18:22:27] (I know that this is a broken approach ... but that's what glam files currently do) [18:22:52] Happy chatting with tneg-rin. I'll upload something to puppet then for review. [18:22:53] Thanks [18:23:19] MediaWiki-ContentHandler, Analytics-EventLogging: [Regression 1.25wmf14] HTML of JSON blocks (e.g. Schema pages) appears garbled - https://phabricator.wikimedia.org/T86748#1004574 (Jdforrester-WMF) [18:24:48] (PS1) Gilles: Add missing post-merge script [analytics/multimedia] - https://gerrit.wikimedia.org/r/187712 [18:25:06] (CR) Gilles: [C: 2 V: 2] Add missing post-merge script [analytics/multimedia] - https://gerrit.wikimedia.org/r/187712 (owner: Gilles) [18:31:02] (PS1) Gilles: Fix lack of metric wildcard [analytics/multimedia] - https://gerrit.wikimedia.org/r/187715 [18:31:18] (CR) Gilles: [C: 2 V: 2] Fix lack of metric wildcard [analytics/multimedia] - https://gerrit.wikimedia.org/r/187715 (owner: Gilles) [18:31:23] (Merged) jenkins-bot: Fix lack of metric wildcard [analytics/multimedia] - https://gerrit.wikimedia.org/r/187715 (owner: Gilles) [18:33:17] nuria: in terms of viz, I was kind of leaning towards sunburst: http://bl.ocks.org/kerryrodden/7090426 [18:33:27] lemmee seeee [18:33:40] NICE!!!! [18:33:40] with filters above, where people could filter down dates, wikis, etc. [18:33:52] i am such a sucker for this kind of thing!!! [18:34:10] :) I like to find the perfect graph, so I'm still a little torn [18:34:33] Iteration man, when they see this they will LOVE you [18:39:53] Analytics-Dashiki, Analytics-Engineering, Analytics-Cluster: Analytics Engineer has a python job to aggregate by project - https://phabricator.wikimedia.org/T88127#1004629 (Ottomata) Does this have to be a "Python job"? Maybe we should remove that as a requirement. [18:45:13] (CR) Ottomata: [C: 2 V: 2] Added Geocode core class that uses Maxmind-V2 Added GeocodedCountryUDF to get country code from IP address Added GeocodedDataUDF to get Geo [analytics/refinery/source] - https://gerrit.wikimedia.org/r/183551 (owner: Ananthrk) [18:45:17] merged! [18:45:20] qchris: % :D [18:45:21] ^ [18:45:26] \o/ [18:45:28] Thanks. [18:49:43] nuria: thinking about the 1000 minimum bucket size rule of thumb [18:49:52] so in the case of a funnel [18:50:05] do you think the first step of the funnel should have at least 1000? Or the last step? [18:50:08] I was thinking first step [18:51:50] milimetric: I think the last step… especially if the drop off from 1st step to 2nd step is huge (10% make it to the second step) [18:52:12] kevinator: then we would be discounting the vast majority of data [18:52:24] because it's mostly granular and the drop-off is pretty big [18:52:28] or is that not what you meant [18:54:15] I’m not sure what you mean by “discounting" [18:58:23] kevinator: sorry, was making sure I can join the VE QR [18:58:40] kevinator: ok, so if we require 1000 hits or more to the last step of the funnel [18:59:08] then that means we have to not count, or discount the majority of wikis [18:59:30] all besides frwiki, itwiki, and two others I forget [18:59:49] and on some days, even those don't have enough data [19:00:20] ahhh, ok [19:01:54] DarTar, the country names are fixed [19:02:39] milimetric: are the events sampled? [19:03:33] mforns: awesome, thanks [19:03:55] DarTar, yw! [19:06:56] milimetric: sorry i missed the ping [19:07:11] milimetric: what are we going to calculate with teh funnel data [19:07:24] / [19:07:31] a rate of that sort? [19:07:36] kevinator: not sure at what rate [19:07:44] nuria: yes, for now [19:08:44] milimetric: the "bucket" rules i have used when calculating means and percentiles, in the case of rates we just have to decide what is a statistically significant dataset [19:09:30] i guess for now I can just display the absolute counts along with the rates, and we can implement a "filter out lower than X" feature [19:10:04] milimetric: but we should not show a wiki with 100 points, that data is meaningless [19:10:27] ok, so i'll make that a minimum for now [19:11:11] (PS1) QChris: Bump version to 0.0.3 to include Geocoding UDFs [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187721 [19:11:12] (PS1) QChris: Bump working version to 0.0.5-SNAPSHOT, now that 0.0.4 is released [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187722 [19:12:14] leila: yt? [19:12:21] (CR) QChris: [C: -2] Bump version to 0.0.3 to include Geocoding UDFs [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187721 (owner: QChris) [19:12:36] yup, wanna talk nuria? [19:12:42] qchris: -2? also, commit message wrong version? [19:12:45] leila; one side question [19:12:52] (PS2) QChris: Bump version to 0.0.4 to include Geocoding UDFs [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187721 [19:12:54] (PS2) QChris: Bump working version to 0.0.5-SNAPSHOT, now that 0.0.4 is released [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187722 [19:13:00] ottomata: Yes :-) [19:13:08] nuria: sure. [19:13:14] for milimetrics work with VE with similar to yours with wikigrok ^leila [19:13:17] (CR) QChris: Bump version to 0.0.4 to include Geocoding UDFs [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187721 (owner: QChris) [19:13:25] yes, I realized that nuria. [19:13:25] ottomata: now it should be good. [19:13:52] leila, milimetric : we are calculating rates like "number of users on edit step"/"number of users on save step" [19:13:56] qchris: there is also now a Webrequest class and corresponding UDFs [19:13:58] I wasn't sure if VE's problem was specific to them but milimteric's email suggested similar anomalies in their data. [19:14:04] that do things like extract xanalytics values, etc. [19:14:07] add that to changelog? [19:14:23] I did add that. Didn't I? [19:14:26] leila: in order to extrapolate a raye seems to me we need a minimum statistically significant dataset [19:14:37] oOOp [19:14:37] leila, milimetric : a rate [19:14:38] you did [19:14:40] "Also UDFs to get webrequest's access method, extract values from the X-Analytics header and determine whether or not the request came from a crawler..." [19:14:44] sorry, it was a long sentence :p [19:14:48] Yes :-( [19:14:55] Feel free to rephrase. [19:14:56] nuria: the data-set should be large enough for that kind of analysis, right? [19:15:01] tha'ts fine, i would have added multiple items for that [19:15:01] right [19:15:04] but whatever, i don't really care [19:15:05] cool! [19:15:14] leila, milimetric : i normally have used buckets sizes yo calculate means / percentiles and such [19:15:29] (CR) Ottomata: [C: 2 V: 2] Bump version to 0.0.4 to include Geocoding UDFs [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187721 (owner: QChris) [19:15:36] Oh ... I thought ... meh. it seems I mis-understood your changelog format then. [19:15:36] leila, milimetric : something like: for a percentile 90 you need at least 1000 samples [19:15:53] leila: should we apply similar rules here? [19:15:53] (CR) Ottomata: [C: 2 V: 2] Bump working version to 0.0.5-SNAPSHOT, now that 0.0.4 is released [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187722 (owner: QChris) [19:15:57] nuria: are you worried that you don't have enough data? [19:16:03] qchris: there is no particular format [19:16:06] cuz otherwise, you can use all the data we have, nuria [19:16:25] :-D [19:16:30] leila, milimetric : i am worried there is very little data for small wikis for a daily value [19:16:33] qchris: you gonna tag, etc? you just reminded me to make a phab task for something... [19:16:54] Yes, I'll tag. But could you take care of uploading to archive? (I have no account there) [19:17:00] nuria: wikigrok was only on in enwiki so we're fine for that one with using the whole data. [19:17:20] Editing team's data is different since they have it on everywhere. [19:17:23] leila, milimetric: but for VE for some wikis there is less than a 1000 points per day right milimetric ? [19:17:38] (tagging done) [19:17:53] Analytics-Engineering, Analytics-Cluster: Streamline analytics/refinery/source release and deployment process - https://phabricator.wikimedia.org/T88137#1004755 (Ottomata) NEW [19:17:57] nuria, leila: yes, some wikis only have like 2 points per day [19:17:57] qchris: yes [19:18:01] Thanks. [19:18:12] if I can login....its been hanging on that recently [19:18:42] Why cannot Java Webservers behave and be responsive? :-/ [19:19:14] nuria: in this case, with 2 points you can't do much. I suggest that you divide the effort in two parts: 1) for wikis with many data points, use your gestimate approach, 1000 data points provide n% confidence. For smaller wikis, either ignore them for now, or see how much confidence you can get from the data we have and report results with confidence attached to them. [19:19:21] qchris: http://en.wikipedia.org/wiki/The_Scorpion_and_the_Frog :) [19:20:02] leila, milimetric: ok, then let's go with bucket sizes than allow us a higher confidence, let's start at a 1000 [19:20:37] leila, nuria: this seems a bit arbitrary to me now... why not 100? [19:21:12] milimetric: because a 1000 interval gives you a confidence to calculate a percentile90 over your time-based dataset of rates [19:21:18] because 100 will give you a very low confidence, milimetric. [19:21:40] milimetric, leila: There are other strategies , for example" If more than 20 percent (or one out of five) of the test-execution results appear not to be similar to the others, something is generally wrong with the test environment, the application, or the test itself." [19:22:12] milimetric, leila: this is also a rule of thumb i used before [19:22:49] yeah, nuria. nuria, milimetric, let's start with wikis that have at least 1000 data points. If figuring out the issue based on those doesn't explain the issue for smaller wikis, then we will revisit to dig more? [19:23:00] leila: ok, so back to my initial question :) [19:23:03] milimetric, leila: much agreed. [19:23:13] if we're computing a simple funnel, A -> B [19:23:25] and there are 1200 people that get to A [19:23:29] and 900 people that get to B [19:23:33] do we use that bucket? [19:24:01] milimetric, leila: i would say we need at least 1000 points on "final" step to be able to "infer" anything [19:24:18] the answer to this depends on which part of the funnel you're suspicious about, milimetric. [19:24:41] leila: we're computing rate of A->B, so we need % got to B [19:24:52] If you have no idea, I'd go with nuria's suggestion of requiring 1000 points on the final step [19:25:01] leila: it's a two step funnel for now [19:25:07] leila: so simple [19:25:10] it's not [19:25:15] i'm simplifying [19:25:17] sorry, simpler [19:25:26] can we go to the batcave, nuria, milimetric? [19:25:29] sure [19:25:33] i'm sadly in a different meeting [19:25:41] ok, later then [19:25:45] ooki. should we talk after your meeting milimetric? [19:25:51] you ping us [19:26:53] ok, it'll be a while, I'll ping [19:27:06] leila: regarding wikigrok [19:27:16] yes, yes, nuria. [19:27:39] the # of events with "widget-impression" w/o "page-impression" for say one day (just picked the 19th of Jan) [19:27:44] it's pretty large [19:27:55] yes! [19:27:58] yes, nuria. [19:28:19] of " 886720" widget-impression only 320222 have a "page-impression" [19:28:26] I'm also digging to see if there is something wrong with taskToken. [19:28:41] i tested that there is no event drops and there is not, what is on teh db is on the logs [19:28:48] hokay, that's huge, nuria. :-| [19:29:18] i am going to test couple more things and will let you know but I think their tokens are not doing what they would expect [19:29:18] I see, nuria. [19:30:08] cause the "volume" makes sense. see totals (for 1 day) : "page-impression": 1612105 "widget-impression": 886720 [19:30:22] so "twice as many people on step #1 than step#2" [19:30:28] ^leila [19:30:38] but after what doesn't match is the sequence [19:31:06] i am going to do another match (base on something that is not the user token) and will let you know [19:31:29] yes, the volume is not alarming. [19:32:15] but tokens are wrong, we should tell kaldari when he gets in [19:32:29] nuria, if you group_concat(event_action) and group by event_userToken, you see a lot of events with more than one type of action, mostly page-impression, in them. [19:32:47] that's a problem, nuria, and it's most probably what you say. the userToken doesn't work properly. [19:33:17] I'll ping him to loop in [19:33:38] leila: will spend another hour on this and let you know [19:33:45] okay. thanks! [19:37:00] !log released refinery 0.0.4 [19:39:39] (PS1) Ottomata: Update refinery artifacts to 0.0.4 [analytics/refinery] - https://gerrit.wikimedia.org/r/187729 [19:40:10] (CR) Ottomata: [C: 2 V: 2] Update refinery artifacts to 0.0.4 [analytics/refinery] - https://gerrit.wikimedia.org/r/187729 (owner: Ottomata) [20:21:38] !log deployed refinery 0.0.4 [20:21:42] hm, i guess logbot is not listening [20:21:45] analytics-logbot: hello? [20:21:45] I am a logbot running on tools-exec-01. [20:21:45] Messages are logged to www.mediawiki.org/wiki/Analytics/Server_Admin_Log. [20:21:45] To log a message, type !log . [20:21:55] oh it is [20:21:58] just no ack. [20:21:58] cool [21:00:14] leila: yt? [21:00:22] yes, nuria [21:01:09] ok, so i do not have much to add to what i said before if you look at events by userToken they do not match [21:01:48] i looked at the code for the token and it is clearly labeled with 'do not trust this to be unique' [21:01:53] taskToken or userToken, nuria? [21:02:00] userToken [21:02:14] doh! then the sampling will also be problematic, nuria, right? [21:02:36] batcave nuria? [21:02:41] (with a link) [21:02:56] if it is purely random like if i< 3 and i = rand(0,3) nop [21:03:10] give me 5 mins [21:03:13] sure [21:03:17] me, too [21:06:09] leila, batcave whenever you are ready: https://plus.google.com/hangouts/_/wikimedia.org/a-batcave [21:15:36] mforns: are you working on this task: T88116 [21:18:47] milimetric: we're in batcave to discuss sampling [21:24:56] kevinator, yes [21:25:30] mforns: do you mind assigning the bug to yourself? It helps when me when I’m triaging bugs :-) [21:27:01] kevinator, sure [21:27:20] Analytics-Engineering, Analytics-Wikimetrics: Wikimetrics should run latest puppet w/o local changes - https://phabricator.wikimedia.org/T88116#1005106 (mforns) a:mforns [21:30:33] leila, found the bug about randomness: https://phabricator.wikimedia.org/T78449 [21:30:45] thanks nuria [21:31:48] leila: also, let me re-state, what is on the logs is on the db so there are no validation issues or dropped event issues [21:32:13] yes, nuria. thanks! :-) [21:32:54] leila: also tasktoken is also generated from this method: taskToken: mw.user.generateRandomSessionId(), [21:33:04] so both will suffer teh same issues [21:33:34] I see. nuria, do you know who can work on this issue? [21:33:44] leila, nuria, coming [21:34:28] oh, you're not there anymore :) [21:35:38] nuria: can you come back to batcave, kevinator, join as well, if you're interested [21:36:38] leila: no, not really, kaldari would know, but note that this might account for lack of uniqueness of ad most say 10/100 (i found 1/1000 but being super conservative here cause also the mobile browser base it's more skewed towards safari, it's probably less than 10/100) [21:36:52] leila: let me estimate that [21:38:34] nuria: wanna jump in the hangout, we're here [21:38:40] k [22:07:14] leila: so the lack of randomness is actually more important in mobile as a LOT of traffic in enwiki comes from safari, so actually it is a bigger problem in mobile (wikigrok) than desktop [22:07:30] leila: need to eat, will try to ctach kaldari later [22:08:11] ottomata: holaaa, can we delete the old data for the mobile jobs? [22:08:34] oh yes [22:08:42] want me to re run some of them? [22:08:45] i can launch a backfill job [22:09:09] ottomata: that will be awesome, 28th and 29th look good [22:09:17] everything else can be scrapped [22:09:36] ottomata: how do i do so those files are available through the mount in 1002? [22:09:57] ls /mnt/hdfs/wmf/data/archive/mobile_apps/uniques/daily [22:10:25] ooohhhhh MAGIC! [22:10:38] ottomata: ok, will check new jobs when i came back from lunch [22:10:44] ottomata: thank you! [22:12:20] nuria: oozie job -info 0025430-150116220452628-oozie-oozi-C [22:15:49] nuria: I'll go eat, too, and then dig a bit in theory. will touch-base later in the day [22:15:52] ciao. [22:18:33] (PS1) Ottomata: Add use of versions-maven-plugin and maven-scm-plugin [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187820 (https://phabricator.wikimedia.org/T88137) [22:44:34] (PS2) Ottomata: Add use of versions-maven-plugin and maven-scm-plugin [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187820 (https://phabricator.wikimedia.org/T88137) [22:45:42] woops [22:46:10] did not mean to do that! [22:50:16] (PS1) Ottomata: Oops, reverting to 0.0.5-SNAPSHOT [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187826 [22:50:37] (CR) Ottomata: [C: 2 V: 2] Oops, reverting to 0.0.5-SNAPSHOT [analytics/refinery/source] - https://gerrit.wikimedia.org/r/187826 (owner: Ottomata) [22:54:23] (PS2) Jdlrobson: Update scripts in light of recent changes [analytics/limn-mobile-data] - https://gerrit.wikimedia.org/r/181428