[00:14:01] yoyoooo, dschoon, any luck? [00:14:33] looked a little bit. [00:14:58] finishing up questions for diederik first, tho [00:15:07] will be on it in maybe 30? [00:15:59] that ok, ottomata? [00:16:19] ja [00:19:19] erosen: question [00:19:37] erosen: about matplotlib [00:20:11] erosen: I have 4G memory on this laptop [00:20:31] and while plotting with matplotlib I get 30% up to 70% gulped by matplotlib [00:20:55] erosen: I would be interested if you encountered this before [00:21:05] they're just regular barcharts [00:21:15] on stat1 I don't think it will be a problem since it has plenty of memory [00:21:29] erosen: https://github.com/wikimedia/metrics/commit/8c404e2bf9e132483d78fd973bb4c96ce6a70826 [00:21:45] drdee: did you get big memory consumption when using matplotlib ? [00:21:45] ottomata, new hue is cool and job browser works! [00:21:56] average_drifter: not that i recall [00:21:59] ok [00:22:02] new hue? [00:22:14] yes new hue [00:22:18] but shit i spoke too quickly [00:23:16] erosen: ok, moving to stat1 for a full run nov=>dec every day of it [00:23:17] did we upgrade? [00:23:23] i thought we were already on 2.2.0 [00:23:30] hm [00:23:34] yeah, definitely has new stuff. [00:23:38] impala! <3 [00:23:39] ottomata upgraded to cdh 4.2 [00:23:44] yes impala as well [00:23:47] ah, yes [00:23:58] didn't mean to upgrade! [00:24:05] somehow half of the nodes were already upgraded [00:24:09] so I just finished the job [00:24:13] boo, \u002D in names [00:24:15] but now i'm stuck on this hostname problem, which I didn't have before [00:24:19] http://localhost:8888/jobsub/ [00:24:25] job browser somewhat works but throws a 500 error if you dive into stuff [00:24:41] later guys [00:24:55] probalby for the same reason these other things aren't working, job browser would have to talk to resourcemanager [00:24:58] actually, drdee [00:25:01] how do you get a 500? [00:25:25] average_drifter: great [00:26:06] AH HA [00:26:10] dschoon, look at this! [00:26:10] Mapreduce.jobtracker.address [00:26:10] analytics1010.wikimedia.org:8032 [00:26:13] tsk tsk tsk [00:26:21] try http://localhost:8888/jobbrowser/jobs/job_1363208055308_0021 [00:26:25] what! [00:26:28] nope. [00:26:29] i think I need to set that explicitly [00:26:29] how! [00:26:34] i dunno, i'm not setting it [00:26:42] so probably the default is somehow picking that up [00:26:42] no idea how [00:26:49] foudn that ind a job conf [00:26:49] http://localhost:8888/jobbrowser/jobs/job_1363208055308_0001 [00:26:53] ahhh [00:26:53] yay for job browser! [00:26:56] i see why [00:26:58] oh? [00:27:09] erosen: http://bit.ly/ZAUssn [00:27:12] [00:27:12] mapreduce.jobtracker.address [00:27:12] localhost:8021 [00:27:14] mapred-site.xml [00:27:16] [00:27:20] *localhost* [00:27:29] erosen: have a look please . I tried to make a small barchart class [00:27:30] https://issues.apache.org/jira/browse/MAPREDUCE-3033 [00:27:40] erosen: then iterating through all features and making charts for them [00:27:43] average_drifter: in a meeting, but looking [00:27:55] erosen: we end up with around ~170 charts afterwards [00:27:56] change that to analytics1010.eqiad.wmnet? [00:28:05] erosen: oh, alright [00:28:11] ottomata: ^^ [00:28:16] also [00:28:20] i'm sure you've seen this [00:28:20] http://wiki.apache.org/hadoop/UnknownHost [00:28:45] but [00:28:47] yeah i've seen that [00:28:49] but [00:29:28] hm [00:29:49] hmm, that's weird though [00:29:51] that's 8021 [00:29:59] that's not 8032 [00:30:20] either way [00:30:26] change it to be the internal name? [00:30:28] see what happens? [00:30:33] but that isn't the value that is being picked up by the job [00:30:42] so, i actualy am explicitly setting that in mapred-site.xml [00:30:47] since the old version of hadoop required it [00:30:56] i'm going to first see what happesn if I remove it altogether [00:32:44] dschoon: i updated https://mingle.corp.wikimedia.org/projects/analytics/cards/92 [00:32:57] section Custom UserAgents [00:33:11] will look, drdee [00:33:16] ottomata: what were you setting it to? [00:33:40] before you start hunting for user agent strings, there is a wiki for that, and i added the link in the card [00:35:01] dschoon, i'm going to take it out of mapred-site.xml and leave it to whatever the mapred-default does [00:35:05] drdee: I have a suggestion to make related to custom user agents [00:35:32] drdee: it might be overkill, but I think it would be interesting(maybe good) to find a way to insert them in the dtree data structures [00:35:32] you might be missing the context in which i was using it [00:35:38] aiight [00:35:54] drdee: yeah, i just had that bookmarked from a while ago [00:35:57] figured i'd add it [00:35:59] ok [00:36:17] * drdee is off to the airport [00:40:18] !log restarting hadoop without mapred.job.tracker set to localhost:8021 [00:40:20] Logged the message, Master [00:41:35] can I use the logbot too ? [00:41:39] yup! [00:41:54] !log started charting for each feature on stat1 (500M bump related) [00:41:56] Logged the message, Master [00:42:14] ottomata: can I ammend a log message ? [00:43:27] i think you can edit the wiki page maybe [00:43:33] https://www.mediawiki.org/wiki/Analytics/Server_Admin_Log [00:43:35] ottomata: [00:43:50] [00:43:50] mapreduce.jobtracker.address [00:43:50] local [00:43:52] mapred-default.xml [00:43:53] [00:44:00] that seems... inauspicious [00:44:10] yup, same error without it set [00:44:15] ok going to set it now, but i'm not sure what to set it to [00:44:22] i'm starting to doubt this is related [00:45:35] well [00:45:40] !log upgraded to cdh4 4.2.0 :) :/ [00:45:41] it might not be related, but we have to fix it [00:45:42] Logged the message, Master [00:45:46] why? it might not matter [00:45:50] we aren't using a jobtracker [00:45:59] because jobs vomit errors, yes? [00:46:12] yeah but maybe not related to jobtracker address [00:46:12] ps: i vote analytics1010.eqiad.wmnet [00:46:15] there is no jobtracker [00:46:20] that doesn't even really make any sense [00:46:23] btw, are you setting this everywhere? [00:46:30] hmm, i guess it could [00:46:30] yes [00:46:35] but what port? [00:46:37] the resource manager port? [00:46:41] there's no jobtracker runnign [00:46:56] like, it requires updating conf/mapred-site.xml on every goddamn machine, doesn't it? [00:48:25] yes [00:48:29] sed :) [00:48:32] + dsh [00:48:42] (actually i'm doing for loop + ssh, dsh never works 100% for me) [00:49:26] heh [00:51:46] ottomata: hmmm [00:52:13] i think oozie might be the one interacting with the non-existent jobtracker [00:52:25] http://oozie.apache.org/docs/3.3.1/AG_HadoopConfiguration.html [00:52:27] ottomata: ^^ [00:54:29] ottomata: do we have an oozie-site.xml? [00:56:09] it probably wouldn't hurt to add oozie.service.HadoopAccessorService.hadoop.configurations=*=hadoop-conf [00:56:17] or rather: [00:56:50] yes we do [00:57:19] *=/etc/hadoop/conf [00:57:24] thats what it is right now [00:57:38] huh [00:57:48] it doesn't show up on http://localhost:8088/conf [00:58:14] that's in oozie-site.xml [00:58:41] :8088 is the ResourceManager, oozie doesn't run on that node [00:58:48] oozie runs on an27 [00:58:56] iiiinteresting [00:58:56] which has the same confs [00:59:00] same hadoop confs [00:59:04] but also has /etc/oozie/conf [01:00:06] hmmm. [01:00:21] i'm pretty sure this isn't an oozie problem though [01:00:31] i thougth it was at first [01:00:33] and to test [01:00:44] I manually added an /etc/hosts entry for analytcis1010.wikimedia.org on an27 [01:00:55] oozie submitted the job just fine then [01:01:09] but when hadoop tried to run the job, I got the same error from one of the job containers on one of the nodemanager/datanodes [01:01:18] so this is def a more global thing [01:02:57] lol [01:03:01] i have a terrible question [01:03:10] when confs are changed, what do you restart? [01:03:26] all hadoop services [01:03:37] namenode, resourcemanager, datanode, nodemanager [01:03:53] well there are more hadoop services, history server, don't thikn I restart those [01:04:12] ...do you restart oozie? [01:04:18] * dschoon ducks [01:04:22] https://github.com/wikimedia/kraken/blob/master/bin/khadoop [01:04:25] yes [01:04:36] yeah I restart oozie [01:04:41] not with the khadoop script [01:04:52] i'm 95% sure this is not an oozie problem [01:05:01] i guess I could run a pig script, entirely out of oozie [01:05:04] and see if I get the same problem [01:05:05] interesting. [01:05:06] okay. [01:05:16] so let's list theories: [01:05:25] - actual, honest-to-god DNS problem [01:05:34] average_drifter: hey, I'm back [01:05:45] (not totally crazy because it'd be outside our control, and it could appear without us doing anything) [01:05:46] bar chart looks good [01:05:47] oh shit! [01:06:04] ottomata: any chance you've heard anything recently from mark about the network ACLs? [01:06:27] like, say, that we're no longer in the eqiad DNS zone? [01:07:02] have not heard that, but getCanonicalHostName still resloves to .eqiad [01:07:13] hmm. [01:07:18] shouldn't it be .eqiad.wmnet? [01:07:35] yes yes [01:07:41] i just didn't type that out :) [01:07:56] kk [01:07:57] ok before I re-add mapreduce.jobtracker.address to the configs [01:08:02] i'm going to try running a pig script [01:08:03] hmm. [01:08:04] let's see.. [01:08:45] erosen: ah cool :) [01:09:07] average_drifter: do you have the charts sitting around somethwere? or did I miss that form earlier? [01:10:41] erosen: the script is running now, so they'll be available in the next hours [01:10:48] cool [01:11:42] question [01:12:38] ottomata: why is analytics1010.wikimedia.org hardcoded into your job conf? [01:12:39] https://github.com/wikimedia/kraken/blob/master/oozie/webrequest_mobile_country_device_vendor/coordinator.properties#L3 [01:12:42] ... [01:12:51] ......... [01:12:54] ….......>>>>>.... [01:12:57] ... [01:12:58] .. [01:12:59] ...... [01:12:59] . [01:13:03] [01:13:08] so! [01:13:17] this is the part where we all pretend this never happened [01:13:18] The program 'mobile_report.py' received an X Window System error. [01:13:19] This probably reflects a bug in the program. [01:13:19] The error was 'BadAlloc (insufficient resources for operation)'. [01:13:19] (Details: serial 279515 error_code 11 request_code 53 minor_code 0) [01:13:21] (Note to programmers: normally, X errors are reported asynchronously; [01:13:23] must be a copy paste error [01:13:24] AGGHGHGHGHGHG [01:13:24] that is, you will receive the error a while after causing it. [01:13:27] To debug your program, run it with the --sync command line [01:13:29] option to change this behavior. You can then get a meaningful [01:13:30] and the problem was actually really subtle and impressive [01:13:32] backtrace from your debugger if you break on the gdk_x_error() function.) [01:13:33] average_drifter: i've seen that [01:13:38] erosen: hmm [01:13:45] that is because you need to set the backend to something which doesn't try to display something [01:13:47] one sec [01:13:51] * erosen digs for code [01:13:55] erosen: ok [01:13:59] (i bet i know the origin -- when we switched an01 -> an10; an01 had a real .wikimedia.org address) [01:16:23] aghhhh eyah i copied this from died's stuffa ggghh [01:16:25] thank youuuuu [01:16:27] i think nobody wants to admit what % of "debugging" time is shit like this [01:16:36] i know i do this ALL THE TIME [01:16:39] average_drifter: http://matplotlib.org/users/customizing.html [01:16:39] god save pair programming [01:16:51] backend : PDF [01:18:15] on the plus side [01:18:19] you were totally right, ottomata [01:18:23] it wasn't oozie :) [01:18:49] haha [01:18:58] well it kinda was, it wasn't oozie configs [01:19:05] but, weiiiird that I didn't see this error utnil I upgraded cdh4 [01:23:05] interesting. [01:23:09] i have a theory [01:23:17] the typo was in coordinator.properties [01:23:38] but ${jobTracker} isn't used in coordinator.xml [01:23:46] only in workflow.xml [01:24:36] maybe jobTracker and nameNode have defaults, and coordinator.properties wasn't loaded for workflow.xml interpolation until the upgrade? [01:25:25] average_drifter: or import matplotlib [01:25:26] matplotlib.use('PDF') [01:25:35] [travis-ci] master/04b2381 (#86 by Andrew Otto): The build has errored. http://travis-ci.org/wikimedia/kraken/builds/5488226 [01:28:38] how's it look now, ottomata? [01:33:40] the workflow ran! [01:33:52] i'm trying to see if I can edit the device coordinator job properties [01:34:40] edit? [01:34:44] oh, via the UI? [01:37:32] via mysql :p [01:37:35] actually [01:37:59] i really wanted to know if workflow confs get recreated from coordinator properties when you rerun a job [01:38:00] they do! [01:38:01] so that's good [01:38:08] heh [01:38:13] and I edited the coordinator conf in mysql :p [01:38:14] so [01:38:16] ah, good [01:38:17] should be good [01:38:25] i don't have to go and edit all of the pending workflows [01:46:43] couldn't you just cancel them all? [01:46:47] and resubmit? [01:47:15] i think that would make it rerun all of the jobs, even for the ones that have alrady succeeded [01:49:30] grrr, I think I do have to update all the workflows, crap, they are coordinator actions [01:49:33] aghhhh [01:49:40] hm. what to do... [01:49:55] i could try to submit a new coordinator, many of the tasks will die due to existing data [01:49:57] i guess that is ok? [01:50:05] hm [01:50:42] i woudln't mind updating the individual workflow confs, except that the confi s stored as text xml [01:50:50] not individual fields [02:03:00] lmk what you decide to do [02:03:06] i realize i kinda skipped lunch [02:03:12] and i am getting light-headed, heh [02:04:56] i dunno, i'm trying to update them now, unngh [02:05:01] i need to peace out again real soon [02:05:04] going to give this 20 more mins [02:05:11] aiight [02:05:16] i'll probably get into it after dinner [02:05:37] could you send a progress update email to the group? [02:05:49] sure [02:05:53] this blocker was a Source Of Worry earlier, so it'd be nice to say we figured it out [02:06:07] (or at least, took another step toward figuring it out) [02:07:10] wtf, now all the coord actions are updated, and I didn't change them [02:07:18] hm! maybe oozie got smart after a while [02:07:18] HMM [02:09:54] ahh, no it didn't, doh [02:10:41] oh god now i'm kicking myself because ottomata said like a while back - why the hell is it hitting analytics1010.wikimedia.org, that setting's not showing up anywhere [02:10:56] i knowwww! [02:10:59] and I saw that and failed to say anything [02:11:00] ugh [02:11:02] so much slef kicking right now [02:11:04] awww [02:11:08] team self kick man [02:11:14] :) [02:11:18] yay for david [02:11:39] i'd do some sort of self-aggrandizing backpatting [02:11:40] also - dschoon- that's why you always get sick brother [02:11:43] gotta eat man! [02:11:45] EAT FOOD [02:11:52] but i've done that WAY too often myself [02:36:46] heading out [02:38:26] drdee: i'm almost done with some questions/clarifications for the mobile features [02:38:35] i'll send after eating :) [02:38:44] back in a bit [05:17:11] [travis-ci] master/d738fab (#87 by Diederik van Liere): The build has errored. http://travis-ci.org/wikimedia/kraken/builds/5491263 [13:34:20] morning [13:34:42] morning drdee [13:34:56] any news on the iOS user agent string for mobile apps?, maybe you can add it to http://www.mediawiki.org/wiki/Mobile/User_agents [13:35:15] drdee: I responded to you yesterday :) [13:35:20] drdee: but you disconnected righta fter, I think [13:35:21] let me find it [13:35:27] oh yes, I'll just add it there [13:35:27] oh ughh [13:35:31] thx! [13:36:45] drdee: Android is accurate, and I'm not sure if the claims about iOS are right or wrong [13:36:47] let me verify [13:41:43] drdee: I *think* iOS should also be setting it properly [13:41:44] (the UA) [13:41:53] drdee: I think that that bug is only for *master*, not the last released version [13:42:09] drdee: can you grep the logs for 'WikipediaMobile/3.2.1' and see if iOS UAs turn up? [13:42:10] they should [13:43:29] drdee: oh, nevermind. [13:43:35] drdee: so, iOS app is not setting any UA :| [13:43:43] I should shoot brion when I see him, I think? [13:44:14] drdee: I don't think there's an easy way to pick up iOS app stuff now :| [13:47:19] ok good to now [13:47:32] and old iOS stuff, how do we pick that up? [13:51:48] drdee: until 3.2.1, you can pick that up via 'WikipediaMobile/3.2.1' prefix [13:51:55] to the UA [13:52:16] drdee: that page says you can still detect it because it lacks the 'Safari' tag, but that's going to catch a few other apps too [13:52:20] ok [13:52:29] drdee: I'll poke brion into setting out a release for this [13:52:32] can you update the wiki accordingly [13:52:36] ? [13:52:38] ok [13:52:55] drdee: oh, it is up to date, I tihnk [13:53:17] just added exact version info [13:53:32] coffee shop's closing down, am out. bb [13:53:33] brb [13:53:41] cool [14:08:12] drdee: hi [14:13:00] hi [14:13:07] charts ready? [14:13:19] drdee: yes they are [14:13:37] show me :) [14:20:50] drdee: check e-mail please [14:21:52] ty [14:28:12] morning [14:28:20] monring [14:30:20] I have an idea [14:30:25] but it's crazy [14:30:31] can I write it even though it's really crazy ? [14:31:19] yes [14:31:31] ok, here goes [14:32:21] the english language is spoken in a few major parts of the world, which means most pageviews come from those places [14:32:43] * milimetric braces himself [14:32:44] U.S. , UK, australia, new zealand, [14:33:01] now, each of those countries have different timezones, let's take UK [14:33:06] it has only one(I think) [14:33:28] that means pageviews align in a periodic pattern on the time-series [14:33:35] day-night-day-night-day-night... [14:33:51] that means we have a small wave [14:34:09] when counting pageviews for the english language, they might come from many parts of the world [14:34:23] let's study the charts first :) [14:34:25] but all of these parts have different waves like this. and those waves are shifted depending on timezone [14:34:43] right, with you so far [14:34:48] but this is exactly about the charts :) [14:34:52] the mobile page views chart looks like an ocean [14:35:03] so here comes the weird idea => [14:35:04] (the one broken down by countries) [14:35:25] ... [14:35:30] we can do Fourier analysis on those things, and find out the different waves that compose the big time-series of pageviews [14:35:38] i object [14:35:38] * average_drifter shuts up [14:35:40] sorry [14:35:41] hehe [14:35:46] drdee: please object [14:35:55] (sheepishly) I actually did that a while back for fun [14:43:03] hahaha YES [14:43:13] dschoon: yes? [14:43:40] i am of the firm belief that if you have a hard problem, the application of an FFT can only improve things [14:43:50] drdee: Evan suggested that we now go for combinations of features that show spikes on 14-december-2012 [14:44:01] yes that makes sense [14:44:07] yes [14:44:16] and if that somehow doesn't help, blindly applying SVD comes next [14:44:29] finally, as a last resort, one might try to understand the problem [14:44:31] al those varnish boxes that you mention are the mobile cache servers [14:44:41] but experience has taught me that road is long and fraught with peril [14:45:02] dschoon: lol, ok [14:45:11] average_drifter: let's look more closely at the mime types that are '-' [14:45:57] ok [15:01:10] drdee: what you wrote about varnish boxes above was for me? [15:01:10] dschoon: i answered most questions on https://mingle.corp.wikimedia.org/projects/analytics/cards/61 [15:01:21] sweet! [15:01:28] average_drifter: yes responding to your email [15:01:34] i'ma eat a food [15:01:34] drdee: cool, thanks [15:01:40] dschoon: q's about visualization i am relaying to jon [15:01:56] should get them answered today [15:28:52] ottomata: I've been assuming you've been on top of the packet loss alerts that have been pretty persistent the past few days, but I suppose I'd better check that assumption [15:32:44] dschoon, answered https://mingle.corp.wikimedia.org/projects/analytics/cards/240 [15:33:11] with the exception of the use case, waiting for reply from maryana [15:39:55] robla, yeah, but only halfways, it seemed there was a buncha network weirdness yesterday due to pope? [15:40:05] all of the analytics udp2log instances had the same problem [15:40:08] all at the same time [15:41:21] otto, drdee, average_drifter, do you know what private data is stored on stat1001? [15:41:35] ottomata: how about this morning? [15:41:48] milimetric: yes [15:42:12] cool, that'd be useful since Marc's reviewing our request [15:42:54] milimetric: is this related to exposing stuff on stat1 ? [15:43:04] it's all good, drdee is filling me in [15:43:41] oh ok, I hope I can rely on my /home/spetrea/www on stat1 from some point onwards [15:44:27] yes, public data is all good, just trying to figure out what private data is on there so that ops can do a proper review [15:44:38] haven't been through all my email yet [15:44:54] ah, locke had a flap, for 30 mins [15:44:54] hm [15:45:03] ok [15:49:33] morning [15:49:48] morning kraigparkinson [15:50:11] mornin [15:50:39] ¡hola! drdee [15:50:45] fancy! [15:50:50] :) [15:51:28] MK cafe time, be back on in a bit [15:51:40] hey analytics people [15:51:45] hey jwild [15:51:53] :) do we have page views for Meta? [15:53:38] jwild: http://stats.wikimedia.org/wikispecial/EN/TablesWikipediaMETA.htm [15:54:46] If I wanted to know specific number of page views (e.g., for all the FDC applications and their talk pages), could that be excavated? [15:57:37] jwild we don't have that currently, we could do it by importing meta request logs into kraken [15:59:05] millimetric - ok, thanks. and I know I should know this, but is that true for all projects besides Wikipedia? [16:00:19] it's an interesting question [16:00:33] we'll add to the backlog [16:00:58] haha thanks dschoon :) [16:06:30] average_drifter, I'll be there soon. [16:08:15] dschoon: updated https://mingle.corp.wikimedia.org/projects/analytics/cards/92 [16:08:57] drdee: looks like we have mobile meta counts, but not plain meta counts [16:09:07] drdee: does that seem plausible? [16:09:08] meta.m Special:WhatLinksHere/Wikikultur/zh-hans 1 9754 [16:09:12] kraigparkinson: oh, [16:09:15] I have to open my laptop [16:09:22] erosen: where / what data source? [16:09:33] http://dumps.wikimedia.org/other/pagecounts-raw/2013/2013-03/ [16:09:52] webstatscollector [16:09:58] erosen: that does not seem plausible [16:10:23] either meta is in there for both mobile and desktop or it's not in there at all [16:10:33] and IIRC meta should be in webstatscollector [16:10:34] drdee: hmm [16:10:57] drdee: nvm [16:10:59] drdee: it's there [16:11:04] not sure why I didn't see it before [16:11:22] ok, jwild could use that to answer her questions [16:11:25] grep '^meta ' pagecounts-20130301-000000 | wc -l [16:11:25] 98 [16:11:56] milimetric: jwild we don't have that currently, we could do it by importing meta request logs into kraken [16:12:02] milimetric; that's incorrect [16:12:18] webstatscollector does have page views per article also for meta [16:12:36] doh! [16:12:41] shows what i know about webstatscollector [16:19:44] thanks drdee, milimetric, and erosen! [16:19:50] np [16:19:57] jwild: what was the answer! [16:20:07] we have it [16:20:13] drdee: webstatscollector does have page views per article also for meta [16:20:44] heading into the office now [16:20:45] ttys [16:24:40] drdee: while we're on the subject of meta page counts, there still seems to be something fishy going on: [16:25:01] why does stats.grok.se have meta.m in the url: http://stats.grok.se/meta.m/latest60/FDC [16:25:23] dschoon: updated https://mingle.corp.wikimedia.org/projects/analytics/cards/244 [16:26:07] erosen: that should be mobile requests to meta [16:26:18] exactly [16:26:28] but if you go to http://stats.grok.se/meta/latest60/FDC (drop the .m) [16:26:30] there is nothing [16:26:32] drdee: ^^ [16:26:54] but is the data in the raw datafiles? [16:27:03] sort of [16:27:06] :) [16:27:10] is that a yes or a no [16:27:15] ? [16:27:16] there are some [16:27:35] erosen: grep '^meta ' pagecounts-20130301-000000 | wc -l98 [16:27:40] iirc, stats.grok.se only shows page views for the wikipedia projects and not for all projects in the data files [16:27:53] that's another reason why we should host the stats.grok.se code ourselves [16:27:58] yeah, it is just weird that they have meta.m and not meta [16:28:09] yup [16:28:47] drdee: here is something interesting: [16:28:48] evanrosen@trantor-2:tmp$ grep '^meta ' pagecounts-20130301-000000 | wc -l [16:28:48] 98 [16:28:49] evanrosen@trantor-2:tmp$ grep '^meta.m ' pagecounts-20130301-000000 | wc -l [16:28:50] 6371 [16:29:15] mmmmm [16:29:49] that seems counter intuitive [16:29:53] indeed [16:30:10] i have a hard time imagining people searching for meta regularly on their mobile devices [16:30:50] maybe it's a random spike, that's just one hour and remember we host all banners from meta, not sure if any of those banners show on mobile but could be the case [16:31:25] interesting [16:33:54] drdee: are the pagecounts files sitting somewhere handy? [16:34:17] on dumps.wikimedia.org :D [16:34:23] or you need access to dataset2 [16:38:30] that would be useful for this sort of thing... [16:38:45] drdee: trying dataset2 now [16:39:18] i guess you don't have accesss [16:44:32] drdee: correct--got distracted [16:59:24] erm scrum ? [17:00:57] getting headphones! [17:01:24] I'll be there in 4-5 [17:28:14] drdee: come baaaaack [17:28:16] drdee: wanna re-join pl ? [17:28:17] *pls [17:28:19] https://plus.google.com/hangouts/_/2da993a9acec7936399e9d78d13bf7ec0c0afdbc [17:32:47] erosen: https://plus.google.com/hangouts/_/d8ffe77174815951419842a71265d0909ad09874 [17:42:03] drdee in a meeting right now. :) [17:59:13] milimetric: when you have some time, can you show me how to data to one of the available limn instances ? [17:59:44] milimetric: I've got all these charts here http://stat1.wikimedia.org/spetrea/embr_py_mobile_pageviews/r3-resized-charts/ [18:00:00] milimetric: and it would be easier that instead of .png they would be imported into limn [18:00:22] milimetric: for me it wouldn't be hard to get data into limn format but I need some directions to know the limn instance to import them in [18:01:04] milimetric: currently they're generated with matplotlib but Evan showed me limnpy so I could convert them to Limn format [18:01:20] average_drifter: if dan doesn't get back to you, i'll help out after this meeting [18:01:21] brb! [18:02:07] dschoon: awesome ! :) thanks [18:02:47] hey average_drifter, let's talk [18:03:15] so should we make a new instance for this [18:03:22] call it wikistats.wmflabs.org for now? [18:04:03] ok, battery dying, running back to find outlet before planning [18:07:14] milimetric: yes please [18:07:23] is that a good name or you'd rather something else? [18:07:33] milimetric: can you call it spetrea ? [18:07:48] milimetric: or mobile-pageviews [18:08:05] milimetric: call it wikistats.wmflabs.org for now? [18:08:08] i would push for a more general name which can be usef for one off analysis [18:08:14] well, if it's mobile pageviews, maybe it'd be best to just add to the limn-mobile-data repository [18:08:18] I'd rather not call it spetrea [18:08:21] not sure, there are already too many wikistats domains [18:08:25] same reason as erosens [18:08:25] milimetric: ok, wikistats.wmflabs.org as drdee says [18:08:43] no i didn't say that :D [18:08:44] i think he was saying not to do that [18:08:45] oh sorry [18:08:46] I misread [18:09:00] it's ok, he quoted me funny [18:09:01] uhm [18:09:12] so is it *just* mobile pageviews? [18:09:12] mobile-pageviews ? [18:09:19] yes [18:09:27] i'm confused [18:09:31] ok, then put them here: https://github.com/wikimedia/limn-mobile-data/ [18:09:32] that's what I'll be putting in it [18:09:43] is this for the stuff we were talking about average_drifter ? [18:09:51] and coordinate with YuviPanda and Juliusz [18:10:06] erosen: yes [18:10:32] average_drifter isn't it more general debugging of page view counts? [18:10:39] erosen: it is [18:10:49] hm!!! [18:10:53] i have an idea [18:10:54] so i thnk "mobile page views" would be misleading [18:10:58] how about debugging.wmflabs.org [18:11:04] i like that [18:11:04] milimetric: sounds good to me [18:11:05] otto's packet loss [18:11:08] etc. [18:11:10] exactly [18:11:12] cool [18:11:12] done [18:11:15] *doing [18:14:13] oh - erosen, average_drifter, drdee, some of the data we use to debug might potentially be private, right? [18:14:26] yes [18:14:44] so this would be a public instance, we just have to be careful [18:14:51] you cannot copy log files to labs [18:16:08] as in, it's not possible? or one should not do it? [18:17:11] milimetric: i think he means we should not do it [18:17:19] should not do it [18:17:24] milimetric: also, I'm not sure that there will be private data [18:17:34] milimetric: this will mostly be aggregated data [18:17:48] milimetric: and it needs to be public [18:18:45] k [18:19:07] … so we can share it [18:27:28] [travis-ci] master/e4afd2e (#88 by Andrew Otto): The build has errored. http://travis-ci.org/wikimedia/kraken/builds/5506803 [18:30:26] yo dudes, unaggregated zero request logs coming in now, starting for 03-14 [18:30:26] http://localhost:8888/filebrowser/#/wmf/public/webrequest/zero_carrier_country [18:30:39] I decided to put files in subdirs, rather than just hyphenated dirs, [18:30:46] i'd like to do that for the incoming raw data too [18:31:01] it would make it much easier to run stats on smaller datasets with pig, easier to specify [18:32:38] hey dschoon [18:32:49] dschoon: meeting [18:32:51] https://plus.google.com/hangouts/_/73b44aa86baa38958253c22a9da22fc35e20a9a6 [18:33:07] oh [18:34:55] [travis-ci] master/d9988de (#89 by Diederik van Liere): The build has errored. http://travis-ci.org/wikimedia/kraken/builds/5506936 [18:42:18] [travis-ci] master/85fd9e7 (#90 by Andrew Otto): The build has errored. http://travis-ci.org/wikimedia/kraken/builds/5507114 [18:49:29] gah [18:49:39] robla snookered me with his charm! [18:51:27] as predicted, dschoon is throwing me under the bus for his lack of time management. :-P [18:56:29] dschoon / milimetric: if either of you has a spare cycle for a JS code review, I could use one on https://gerrit.wikimedia.org/r/#/c/53505/. no biggie if not. [18:56:41] our pleasure ori-l [18:56:45] in a meeting atm, but soon [18:57:02] cool, thanks [18:57:51] [travis-ci] develop/5aecbdb (#114 by milimetric): The build passed. http://travis-ci.org/wikimedia/limn/builds/5507571 [19:00:00] [travis-ci] master/4d28bad (#91 by Andrew Otto): The build has errored. http://travis-ci.org/wikimedia/kraken/builds/5507537 [19:31:05] talking to jgonera in 30 minutes about card 61 [19:33:02] average_drifter [19:33:03] http://debugging.wmflabs.org/ [19:34:35] https://github.com/wikimedia/limn-debugging-data [19:34:57] that labs instance is now serving that github repo [19:35:21] you can use limnpy to put stuff into the limn-debugging-data repo and then i'll show you how to deploy it to kripke [19:35:44] it'll just be fab debugging deploy.only_data from the directory where you clone limn-deploy.git [19:35:53] average_drifter, ping me with questions anytime [19:40:42] http://localhost:8888/filebrowser/#/wmf/public/webrequest/mobile_country_device_vendor [19:42:20] dschoon, milimetric, skyping [19:42:30] yes [19:42:32] that is our goal. [19:42:33] correct. [19:42:49] i am attempting to add you, ottomata [19:42:50] i am calling you david [19:42:55] hmm [19:42:56] ok [19:43:01] stoooop [19:43:06] don't cross the streams! [19:43:13] wait [19:44:20] hokay. [19:44:29] maybe skype is also a no-go. [19:44:30] suggestions, ottomata? [19:44:50] one sec [19:44:51] trying different network [19:46:18] hmm [19:46:21] i tried skyping you [19:46:24] wanna call me again? [19:46:26] dschoon? [19:46:35] or hangout? [19:46:35] https://plus.google.com/hangouts/_/2da993a9acec7936399e9d78d13bf7ec0c0afdbc [19:47:44] hhlaallooo? dschoon, milimetric? [19:47:49] yay helo [19:47:52] ok comin [19:47:56] hokay [19:47:57] sec [20:07:55] drdee: [20:07:56] http://stats.wikimedia.org/kraken-public/webrequest/zero_carrier_country/2013-03/14/ [20:07:57] and [20:08:06] http://stats.wikimedia.org/kraken-public/webrequest/mobile_country_device_vendor/2013-03/14/ [20:08:14] in meeting will look soon [20:10:30] brb [20:22:52] back but eating [20:22:54] how're we doing? [20:23:01] oh. no otto. hokay. [20:24:04] dschoon: got a minute for a JS sanity check on a patch-set? [20:24:27] sure! [20:24:32] any background i need, ori-l? [20:24:48] that's https://gerrit.wikimedia.org/r/#/c/53505/, right? [20:25:00] yup [20:25:18] this explains the metrics http://meta.wikimedia.org/wiki/Schema:NavigationTiming [20:25:36] the talk page has some additional points [20:25:45] it works well and we have a deployment window now so i was hoping to try it out [20:27:17] dschoon, milimetric: i updated https://mingle.corp.wikimedia.org/projects/analytics/cards/61 [20:27:26] milimetric: i have a limn related question for you [20:27:49] shoot [20:28:01] i've been reading it ori-l [20:28:20] i'm stuck trying to figure out if random(10) === random(10) will be true 1/10 times :) [20:28:34] shoot was directed to drdee [20:28:36] jgonera suggested to add a drop down menu to the geo world map so you can pick the dimension that you want to visualize [20:28:38] as in shoot the question [20:28:46] right [20:28:56] like the legend but for maps [20:29:06] hm. [20:29:20] the idea of a legend for geo-world is good [20:29:24] yes [20:29:31] but i don't think that fits in scope for the 20th [20:29:48] can we find a scaled-back solution for this sprint? [20:31:05] the scaled back version would just to visualize page views per country without any additional breakdown [20:34:17] cool, hadn't seen that trick ori-l, does it have any advantage to just getRand(factor) === 0? [20:35:07] scaled back version of what drdee? [20:35:22] the legend/ drop down idea [20:35:56] milimetric: no, but doing that makes me worry about a weird unrelated error screwing things up [20:36:02] but that wouldn't meet the requirements right? [20:36:05] it's easy to have an error that produces 0 [20:36:11] gotcha [20:42:14] ori-l: given the title of the extension, to what extent do you want pithy comments about performance? [20:42:59] go nuts [20:43:13] dschoon, milimetric: card https://mingle.corp.wikimedia.org/projects/analytics/cards/240 was updated as welll [20:43:35] keep in mind that onload -> window.setTimeout( ..., 0 ) means it'll execute when the UI thread stops being busy, so it shouldn't affect page load [20:43:36] i mean, is it helpful? [20:43:52] and it only executes once, right? [20:43:58] so really, 1ms doesn't matter [20:44:01] right [20:44:34] if you're gonna be pithy about performance, 1ms will probably have to matter - not a lot of holes in this code [20:47:14] it's stupid. [20:47:15] http://jsperf.com/math-floor-vs-bit-twiddling [20:47:15] could we timebox the sanity check, ori-l? :) [20:47:19] yeah. [20:47:23] the code is solid. [20:47:47] thanks [20:47:57] kraigparkinson: sure, sorry [20:49:30] dschoon: if you could clean up card 61 right now then I can send the email to the mobile mailing list and we can unblock this asap [20:49:46] One moment. [20:49:52] no worries. we're overdue on quite a few cards, so we're just trying to stay focused. [20:50:00] wait -- cleaning up the card is a blocker? [20:50:01] I don't mind asking, just want to keep it limited if possible. [20:50:30] dschoon: cleaning up is a blocker for me to send out the email to mobile [20:50:48] sorry, didn't mean to distract. if i can offset by removing some blocker for you let me know [20:53:49] thanks ori-l [21:02:34] fixing card now, drdee. done in a sec [21:02:41] awesome [21:04:53] kraigparkinson: meeting? [21:07:09] drdee: on "Device Type" [21:07:15] iPhone, Android, etc (_not_ manufacturer) [21:07:23] is it actually Device OS? [21:07:24] yep, be right there. [21:07:31] what other values are there? [21:07:43] i think it's called parentId in openddr language [21:07:48] i will add a link to doc [21:07:53] ty [21:09:12] hm, drdee [21:09:24] Output format doesn't make sense to me. [21:09:33] if you still having questions just add them [21:09:33] ok [21:09:33] i am in meeting with kraig [21:09:38] add questions [21:09:41] i will respond asap [21:10:11] can you make up three sample lines from the output? [21:10:17] yes [21:19:36] err, drdee -- "Visualize data using the geo worldmap. Use drop-down to select dimension to visualize." [21:20:16] milimetric, do you think that's is doable by the 20th? [21:20:35] brb in a couple hours - will be working later tonight [21:20:43] waaait! [21:20:49] yeah, definitely [21:20:55] just want a quick gut-check on that [21:20:58] okay. [21:21:02] we can hack anything in like 30 minutes [21:21:06] (another tweaks node for geo?) [21:21:08] i did not promise this [21:21:08] *nod* [21:21:09] aiight [21:21:12] it just might not be pretty :) [21:21:15] hmm. [21:21:16] well [21:21:20] i told julius i would ask you guys [21:21:22] it's under "requirements", drdee [21:21:27] so in my mind the hard part will be putting the data together for card 61 [21:21:53] right now i think we're dramatically overestimating the amount of data processing that we *could* or *should* do with Limn [21:22:19] if the data's there in a format that Limn already understands, the visualization's easy [21:22:28] agreed. [21:22:29] i'll eat my underwear or work till midnight if i'm wrong on that [21:22:40] whichever of those two you gentlemen prefer [21:22:45] it will all be done by hadoop, unless we hit a blocker [21:29:33] erosen: is https://mingle.corp.wikimedia.org/projects/analytics/cards/256 different than https://mingle.corp.wikimedia.org/projects/analytics/cards/94 [21:30:05] drdee: yes [21:30:25] could you elaborate on 256 in the card itself? [21:30:30] yup [21:30:31] will do [21:30:53] is 256 more or less urgent than 94? [21:30:57] less [21:31:17] I was just talking with milimetric about where limnpy should fit in [21:31:45] and we decided that limn should create something like a default graph for each datasource [21:32:19] so that the person providing the data deosn't need to supply a graph just in order to visualize a datasource (and have it at a persisten url) [21:32:22] drdee: ^^ [21:32:39] erosen: is 256 a request for jessie? [21:32:54] it really just came out of a conversation dan and I had the other day [21:33:02] it isn't high priority, but it is really simple [21:33:19] erosen: sounds good, make a mingle card :) [21:33:22] and would reduce the overhead and getting data into limn [21:33:31] overhead of* [21:45:01] drdee: https://mingle.corp.wikimedia.org/projects/analytics/cards/256 [23:03:59] jwild, are you meeting with Erik M now? :) [23:17:51] jwild, you froze... [23:31:30] dschoon: can you save your edits to https://mingle.corp.wikimedia.org/projects/analytics/cards/61 so I can take over? [23:32:11] done [23:32:23] i left you a few FIXMEs [23:32:35] mostly around what "pageviews" is counting [23:32:47] (i've read the source and it doesn't seem meaningful to me) [23:33:10] COUNT = FOREACH (GROUP LOG_FIELDS BY (day_hour, country, device_os, device_os_version, apple_info, is_tablet, wikimedia_app) PARALLEL 10) GENERATE FLATTEN(group), COUNT(LOG_FIELDS.country) as num; [23:33:21] specifically COUNT(LOG_FIELDS.country) as num [23:33:56] i don't really have any good ideas, either, as practically everything that's not the compound key of all other fields is going to be misleading [23:34:01] brb [23:39:55] back [23:39:58] drdee: ^^ [23:44:56] about: specifically COUNT(LOG_FIELDS.country) as num [23:45:04] it doesn't really matter which field you count right? [23:45:22] (once you group by obviously) [23:47:20] uh [23:47:23] shouldn't it? [23:50:13] you count the number of log lines that constitute a page view for a particular coutry/device combination [23:50:35] i think it depends on what you group on [23:50:38] but i'll check in the shell [23:50:42] HIIIIII [23:50:43] so you can count any field in the log line that satisfies the conditions [23:50:43] HIHIHIHIHI [23:50:48] just sayin hi [23:50:50] HI [23:50:55] HI [23:50:56] HI [23:50:58] dschoon: i updated card 61 please have another look [23:50:58] HIHIHI [23:51:00] kk [23:51:06] you guys and your cards [23:51:36] 1 note for clarification: we have only 1 metric: Pageviews [23:52:01] please clarify pageviews [23:52:01] the other ones are 'dimensions' to break down the page view metrics [23:52:10] ...but they don't do that [23:52:12] The number of webrequests that constitute a pageview. See https://raw.github.com/wikimedia/metrics/master/pageviews/kraken/pageview_definition.png for a definition of a pageview [23:52:31] pageviews by ? [23:52:37] i know what a pageview is :P [23:52:45] the question is what grouping of them are we counting? [23:52:51] no it has a specific meaning for us [23:52:59] see flow diagram [23:53:06] i know [23:53:09] but i'm saying: [23:53:12] page views by country by device etc [23:53:46] is it COUNT( (ts, a2, device_os) )? [23:54:07] once you have the device and the os version the other properties are fixed, browser name, device resolution, javascript enabled don't change [23:54:21] is it COUNT( (ts, a2, device_os) )? yes, that seems right [23:54:36] gonna talk to kraig now [23:54:40] (actually, browser can change -- i have chrome, firefox, and safari on my iPhone.) [23:54:49] ok :) [23:54:49] browser version definitely changes [23:55:22] about If we want to provide a “Total” for each metric, aggregated across all countries, it’s going to have to be in a different file. It just doesn’t make sense for the data shape. [23:55:31] ja [23:55:40] totally open to cheats about that :) [23:55:40] so there is only 1 metric, and we should have probably a separate file for totals [23:55:44] me too :D [23:55:59] well, i was thinking [23:56:25] totals makes sense for "total pageviews by OS, by OS Version, by Device Class, (etc)" [23:58:19] yes and country [23:58:34] btw, okay if i send card 61 for comments to mobile mailinglist [23:58:34] ? [23:59:29] gimme 10 minutes to check in grunt [23:59:36] then we can roll if it looks ok [23:59:36] ja?