[00:00:35] Hey. I just clicked a Wikidata link and was taken to http://imm.io/18VQ9 [00:00:44] legoktm: ^ [00:01:02] ugh [00:01:05] er [00:01:09] is that a screenshot? [00:01:27] legoktm: yes [00:01:30] of what I got [00:01:39] when I clicked the ru link at https://www.wikidata.org/wiki/Q4451759 [00:01:55] Sven_Manguard: yes, aude has a patchset waiting deployment [00:02:04] legoktm: well, it gets worse [00:02:21] I went to ru.wikipedia and put Тарасовка (Киево-Святошинский район) into the searchbar and got that too [00:02:37] Sven_Manguard: Yeah. [00:02:54] what is causing that? [00:03:31] idk, but https://gerrit.wikimedia.org/r/67900 is the hotfix [00:03:34] it's not happening when I click the uk.wiki link on the same page [00:03:57] maybe a template is loading Lua which parses the item? [00:04:00] idk [00:04:44] and it's taking out a page on Russian Wikipedia? [00:04:48] dammmmn [00:05:09] ¬_¬ [00:05:26] i bet if you remove the coordinate property itll fix it [00:05:41] oh, damn [00:05:43] that's lua [00:06:34] * aude fixing [00:06:52] :D [00:06:59] I don't even have Scribuntu installed on my test setup [00:07:12] * hoo heads off to do that [00:13:44] New patchset: Aude; "Handle failure to deserialize entity in lua" [mediawiki/extensions/Wikibase] (mw1.22-wmf5) - https://gerrit.wikimedia.org/r/67932 [00:13:59] hoo: does that look sane to you? [00:14:16] * aude has lua [00:14:24] I've just installed it [00:14:31] ok [00:17:24] Jenkins is crying about that change :/ [00:17:33] ignore [00:17:41] :D [00:17:42] it's testing with datavalues master and old branch of wikidata [00:17:48] :/ [00:17:51] wont' work [00:18:26] It looks sane and I guess you tested it?! [00:18:48] I can force the merge, if ou want [00:19:32] iam testing [00:22:26] it's good [00:22:46] sure the error handling could be more elegant but it works fine [00:23:28] or it's my lua script that's not perfect [00:23:30] aude: why is http://imm.io/18VYe happening? [00:23:48] Sven_Manguard: that i don't know [00:24:07] well it's an unnecessary precision loss [00:24:13] agree :/ [00:24:17] aude: Will merge... just looked a bit through the Scribunto code and this looks good [00:24:20] i can't do anythign tonight [00:24:23] hoo: thanks [00:24:30] i'll backport and poke someone [00:25:12] New review: Hoo man; "Looks sane and follows the Scribunto documentation." [mediawiki/extensions/Wikibase] (mw1.22-wmf5); V: 2 C: 2; - https://gerrit.wikimedia.org/r/67932 [00:25:13] Change merged: Hoo man; [mediawiki/extensions/Wikibase] (mw1.22-wmf5) - https://gerrit.wikimedia.org/r/67932 [00:25:37] aude: also it would be nice if Wikidata could take the curly apostrophes and quote marks from geohack (see upper left of https://tools.wmflabs.org/geohack/geohack.php?pagename=Kesteren_railway_station¶ms=51_55_34_N_5_38_13_E_type:railwaystation_region:NL ) and convert them into ones that Wikidata can use [00:26:12] if you straight cut and paste in from geohack, Wikidata won't accept it until you manually replace the curly apostrophes and quote marks with straight ones [00:27:47] Sven_Manguard: Are you suggesting that the Javascript should escape and replace the characters? [00:28:53] Sky2042: doesn't it already do that with other things? autoconvert? [00:29:18] Sven_Manguard: I think it does, I was just wondering if that's what you had in mind. [00:29:57] I mean, on the plus side, I never would have discovered that the decimal to minute/second conversion was broken if it weren't for the inability to just straight cut/paste in the geohack curly apostrophes [00:41:56] aude: in answer to my earlier question, the reason is that apparently the automatic setting was to round to the arcminute, and I wanted to express a value that had arcseconds [00:42:05] so I overwrote the default [00:42:14] hmmmm [00:42:17] but that's still an issue [00:42:39] default needs to be at least arcsecond, if not 1/10th arcsecond [00:42:51] and that's the only time I had that issue [00:43:36] seems reasonable [00:51:11] aude: well it's doing it again [00:51:37] whenever I cut paste in from geohack and then swap out the curly for the straight apostrophes [00:54:33] rschen7754: ping [00:54:38] hi [00:55:05] when you added the geocoords to https://www.wikidata.org/wiki/Q2504328 how did you do it? [00:55:12] it's a known issue [00:55:15] it was super annoying [00:55:16] the apostrophe [00:55:27] first i copy-pasted and it didn't work [00:55:28] rschen7754: did you cut and paste? [00:55:49] then i dropped all the degrees and added it manually [00:55:59] all the minutes and seconds that is [00:56:02] yeah, what I'm getting at is that the location you entered is 32°55'27"N, 96°45'49"E. The interchange is at 32°55'27"N, 96°45'49"W. [00:56:06] dms should certainly work [00:56:18] I wanted to know if it was the software or you that made that mistake... [00:56:24] but decimal degrees appear to work better at the moment [00:56:34] i think they might still get rounded though [00:56:51] unless you pick the right precision [00:57:05] https://bugzilla.wikimedia.org/show_bug.cgi?id=49386 about the apostrophe [00:59:30] Sven_Manguard: no idea [00:59:46] aude: yeah i ran into that issue [00:59:55] is this live on test.wikidata/ [01:01:40] rschen7754: should be [01:01:48] should be same version of everything now [01:02:02] hm, so if i could just get PC on test.wikidata... [01:02:06] go ahead [01:02:15] rschen7754: I'll give it [01:02:17] oh i can't give you that though [01:02:22] * aude does not have that right :o [01:02:53] {{done|color=pink}} [01:02:53] How efficient, color=pink! [01:02:55] oO [01:03:08] thx [01:03:09] aude: The devs. should get 'crat access there IMO [01:03:24] just creating properties is fine [01:03:40] should i go file a bug to un-restrict PC on test? [01:03:40] hoo: agreed, but I can't do anything about that [01:03:44] https://test.wikidata.org/wiki/Property:P10 [01:03:48] or can someone here just do it? [01:04:16] I don't think we'd need some kind of consensus, do we? :P [01:04:16] Reedy could [01:04:24] probably not tonight [01:04:25] But he's probably sleeping now [01:04:32] hoo: idk. the reason for restricting it doesn't really exist on test. [01:04:33] the stewards can grant access [01:04:33] not thaaaat important [01:04:43] indeed [01:04:57] aude: yeah, but it is kinda annoying that people keep on having to request it [01:05:05] * aude wants bot rights on test and will need to request [01:05:18] just so i can test everything works fine [01:05:27] I'm going to ping Reedy tomorrow... I guess he's fine with giving a couple of +crats out [01:07:18] now i can't add the property to the item :/ [01:07:45] oh, it doesn't work over https [01:08:04] ewk [01:09:22] *sigh* [01:09:28] what? [01:10:19] ok, i think i figured it out… one of my scripts was conflicting [01:10:26] rschen7754: Works for on test with https? Or did you mean wikidata.org [01:10:27] ah ok [01:10:54] i have my global js disabled on regular wikidata [01:11:28] rschen7754: Me too ... I was having problems with it IIRC [01:11:33] ah, ok [01:12:04] hoo: How nice :) I'm a ... "user" ;) [01:12:19] rschen7754: Yay for using my javascripts :D [01:12:24] Hazard-SJ: mh? [01:12:25] :D [01:13:10] hoo: I think the [save | cancel] things were replaces with ||| or something like that [01:13:33] Sven_Manguard: https://test.wikidata.org/wiki/Q14 [01:13:43] I'm not going to investigate that now... but rather my nice and warm bed :D [01:14:07] Jasper_Deng: rschen7754 what [01:14:16] err, how'd Jasper wind up there [01:14:24] requested [01:14:37] yeah, that's how it looked when i entered it [01:14:47] drops the seconds automatically though [01:15:26] okay, so go to options [01:15:31] err [01:15:32] aude: Good night... tomorrow I'll have time to hack on the client move thing :) [01:15:41] advanced adjustments [01:15:59] and change percision to arcsecond [01:16:24] hoo: thanks for your help :) [01:16:32] yeah, it works when you do that [01:16:45] but the seconds are lost forever if you don't do that right off the bad [01:16:47] *bat [01:16:53] rschen7754: yes, which is an issue [01:20:10] aude: I found another weird issue [01:20:24] go to https://www.wikidata.org/wiki/Q2504328 and try to edit the coordinate there [01:21:35] when I tried to change E for W it changed the value from 32°55'N, 97°14'W to 32°55'27"N, 96°45'49"E [01:21:46] err other way around, from 32°55'27"N, 96°45'49"E to 32°55'N, 97°14'W [01:22:33] or 32°55'27"N, 97°14'11"W after I force it into arcsecond [01:23:41] it absolutely can't handle itself dealing with the direction west [01:23:43] it's freaky [01:26:14] grrrr [01:27:04] alright, it's almost getting to be morning here ! [01:27:17] * aude hears the birds chirping [01:27:22] * aude sleep :) [01:28:41] aude: you're not in NY anymore? [01:28:55] no, in berlin [01:29:00] o.O [01:29:19] * aude invades NY sometimes [01:46:51] aude: the deeper I go, the more problems I find with the way that geocoords are being handled [01:52:53] aude: rschen7754: legoktm: https://bugzilla.wikimedia.org/show_bug.cgi?id=49415#c2 [01:58:36] so basically we cant do anything west? [02:01:57] shouldn't west be a negative number? [02:02:35] (from a basic OOP programmer's perspective, it sounds like not all the fields of the location object are getting migrated accordingly) [02:07:28] 02:07, 11 June 2013 Sven Manguard (talk | contribs | block) deleted page Property:P625 (This propery is being temprarily deleted (by admin consensus via emergency IRC chat) due to errors in the datatype causing false outputs) [02:07:47] legoktm: west or south, so yes, it's a negative number issue [02:07:54] possibly / likely [02:08:21] ok, you should probably update the bug summary then [02:11:01] https://bugzilla.wikimedia.org/show_bug.cgi?id=49415#c2 [02:16:12] !log Properties P625 and P626 have been temporarily deleted following a conversation in wikidata-admin and the escalation of the bug's complexity and scope [02:16:51] :O [02:31:18] Pink|TV: awesome of you to use my image for the testwikidata [02:31:39] well thanks for creating it :) [02:31:54] :D [02:32:08] I never thought I'd be used, considering that it's an inverse [02:32:49] need some way to signify it's a crazy alternate universe ;) [02:33:18] i had someone whip up a black logo like for test.wp, but Reedy didn't like it [02:33:58] but yeah, it's a really cool image :) [02:34:00] have you ever seen the image on my enWiki user page? [02:34:29] I have a consistent aesthetic when it comes to those kinds of things. I like parallel rectangles doing interesting things [03:24:01] I'm having fun with the Wikidata logo http://imm.io/18WOm http://imm.io/18WPN http://imm.io/18WQs http://imm.io/18WR4 http://imm.io/18WS6 http://imm.io/18WSY [03:24:10] if you want any, take em [03:24:18] All my work is CC-BY-SA [03:37:25] TCN7JM: http://imm.io/18WWg [03:37:51] That's a bit too blurry for me. [03:58:39] TCN7JM: http://imm.io/18X0N and http://imm.io/18X0S then [03:59:38] Second one is great. [04:01:28] TCN7JM: okay, here's a fixed version then http://imm.io/18X1o [04:01:38] Define fixed. [04:02:07] the reason it was blurry in the beginning is that the distortion left lines [04:02:23] I thought I got them all in 18X0S but I missed a few [04:02:55] Ah. Cool. [07:35:15] what's the official way to run tests in wikibase? [07:35:28] I only tried to create a suites.xml myself [07:54:35] hello :-D [07:54:39] oh men, hmm [07:54:41] Guten Tag [07:59:05] liangent: tobi should be here soon and able to help you with that [08:23:45] liangent: wikibase tests don't run standalone but as part of the mediawiki test suite [08:24:16] liangent: go to tests/phpunit in your installation root. run php phpunit.php --group Wikibase [08:58:32] DanielK_WMDE: hangout? [09:24:46] New patchset: Daniel Kinzler; "Fix string length check in StringtValidator." [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67956 [09:24:52] Change abandoned: Daniel Werner; "seems to work again with current MW core" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67574 [09:28:49] New review: Daniel Kinzler; "(1 comment)" [mediawiki/extensions/Wikibase] (mw1.22-wmf5) - https://gerrit.wikimedia.org/r/67932 [09:31:36] https://bugzilla.wikimedia.org/show_bug.cgi?id=49425 looks bad but i am not sure who'd need to take a closer look [09:33:38] New review: Daniel Kinzler; "(1 comment)" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67862 [09:33:45] New review: Daniel Werner; "(1 comment)" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67404 [09:34:41] New review: Aude; "(1 comment)" [mediawiki/extensions/Wikibase] (mw1.22-wmf5) - https://gerrit.wikimedia.org/r/67932 [09:34:46] DanielK_WMDE: you haz comment [09:34:56] lua does not need to serialize the entity? [09:35:13] aude: i don't know - does it? what for? [09:35:20] i did not chang ethat [09:35:28] anyway, it's serializing but the error sais deserializing... c onfusing :) [09:35:38] you are right, but it was 2am [09:35:56] aude: sure, not a big deal. I'm just wondering why it is serializing at all. [09:36:04] * aude can correct that in patch for wmf6 and master [09:36:04] perhaps ask jens when you see him around [09:36:07] sure [09:36:18] would be good to have that fixed in master, i don't care about wmf6 [09:36:20] * aude needs to poke at lua more  [09:36:22] it's a minor inconsistency [09:36:28] we need it in wmf6 [09:36:30] just in case [09:36:39] although wmf6 knows about globe [09:36:49] Henning_WMDE: Danwe_WMDE: https://bugzilla.wikimedia.org/show_bug.cgi?id=49425 <- please have a look - looks bad [09:37:01] * aude won't be around on thursday either, btw [09:37:19] we need the error handling, sure. i meant that it'S not too horrible to be mistakingly referring to "deserialzation" in an error message [09:37:30] DanielK_WMDE: you are right :) [09:37:36] Lydia_WMDE: known issue, patch on gerrit [09:37:40] will link the bugs [09:37:43] thx [09:37:46] i'll just call it serialize-error [09:37:48] which it is [09:37:55] right [09:38:59] you'll be around on thursday evening right [09:39:00] ? [09:39:25] there should be no submodule changes at that point, but just in case there are any problems (unlikely now) [09:40:40] aude: yes, i'll be in the office [09:40:58] Lydia_WMDE: https://gerrit.wikimedia.org/r/#/c/67816/ [09:41:14] aude: perhaps that should be backported ----^ [09:41:24] should be possible without the dependencies [09:41:45] DanielK_WMDE: can look [09:41:47] thanks [10:00:57] Tobi_WMDE: Henning_WMDE which patches? [10:01:22] https://gerrit.wikimedia.org/r/#/c/67257/ [10:01:26] k [10:02:52] New patchset: Aude; "Fix for globeCoordinate's toDegree()" [mediawiki/extensions/DataValues] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67959 [10:06:10] aude, Henning_WMDE: why are we using degree/minute/seconds internally at all? [10:06:19] wouln't two floats be much simpler? [10:06:50] or use the system that was originally used by the user... [10:07:01] which would probably mean storing the original string [10:07:32] * aude would prefer a field for latitude and one for longitude [10:07:44] too much change though from status quo [10:07:46] New patchset: Nikerabbit; "Fix string length check in StringValidator." [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67956 [10:08:17] https://gerrit.wikimedia.org/r/#/c/67816/ does not cherry pick cleanly [10:08:20] DanielK_WMDE: [10:08:23] * aude merging the other one [10:08:25] aude: not really, the DataValue just needs to accept the old and the new array representation [10:08:32] that's not pretty, but easy enough [10:08:51] DanielK_WMDE: sure [10:08:54] aude: just ignore the test [10:09:01] Change merged: Aude; [mediawiki/extensions/DataValues] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67959 [10:09:18] the test depends on fixing the test infrastructure for data values, which was pretty broken [10:09:23] but we don't need to backport the test [10:09:24] would you make a patch for the time validation in wmf6 [10:09:36] sure, just patch the time value thing [10:09:54] since we already backporting in datavalues, this is not a problem [10:10:00] you mean git checkout wmf6 && git checkout wmf5 -- path/to/TimeValue.php? [10:10:01] sure :) [10:10:09] we don't need wmf6 [10:10:13] err, wmf5 [10:10:15] just wmf6 [10:10:32] right sorry, should be "master" there, not wmf5 [10:10:47] right [10:20:09] New patchset: Daniel Kinzler; "(bug 49264, bug 48965) validate time format." [mediawiki/extensions/DataValues] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67962 [10:20:15] thanks [10:20:47] aude: i have only unit tested this, not tried to supply garbage via the api [10:21:15] noted [10:25:07] Lydia_WMDE, aude: i have a GSoC call in my calendar for 2pm... is that going to happen? Who is goping to attend? [10:25:25] DanielK_WMDE: yes it is going to happen [10:25:28] * aude be there  [10:25:30] attendees: gsoc students, you, kaite and me [10:25:35] *katie [10:30:48] Lydia_WMDE: ah, ok. [10:33:42] DanielK_WMDE: your patch is no good :( [10:33:52] it for example does not let me remove values [10:34:03] aude: err? [10:34:27] i have no idea how that would be related [10:34:37] maybe i had badly formatted dates [10:35:06] and the validation prevents them from being removed? [10:35:10] I suppose that could happen [10:35:13] yes it did [10:35:23] * aude not convinced my dates are badly formatted either [10:35:36] * aude removes one added via the interface [10:35:45] is the ui putting in badly formatted dates? [10:35:48] basically, you cannot instantiate a TimeValue with an invalid string, so you can't instantiate a PropertyValueSnak for a broken date [10:35:50] and my bot? [10:35:58] then the UI is doing it wrong [10:36:01] i have no idea [10:36:03] * aude thinks [10:36:16] * aude does https://www.mediawiki.org/wiki/Extension:Wikibase/API#wbcreateclaim [10:36:19] perhaps change the error to include the actual value [10:36:22] that's what my bot does and it does not like it [10:36:54] i can't save dates [10:36:59] via the UI [10:37:08] wtf [10:37:12] ok, we can investigate but not backporting [10:37:14] * DanielK_WMDE is not going to get any work done today [10:37:19] :/ [10:37:25] better the way it is, as much as it sucks [10:37:59] New review: Aude; "i can't add dates via the UI or with my bot following the example https://www.mediawiki.org/wiki/Ext..." [mediawiki/extensions/DataValues] (mw1.22-wmf6) C: -1; - https://gerrit.wikimedia.org/r/67962 [10:38:15] if this is in master, i don't understand [10:38:36] does it really work in master? [10:39:11] dev repo seems fine [10:40:00] i have not tested extensively using the UI or the API [10:40:10] just seems incompatible with the branch [10:40:14] o_O [10:40:18] i'll check that again [10:40:19] * aude confused [10:40:23] the regex itself seems fine [10:40:33] it's in master so must be on dev repo [10:40:39] try print ( preg_match( '!^[-+]\d{1,16}-(0\d|1[012])-([012]\d|3[01])T([01]\d|2[0123]):[0-5]\d:([0-5]\d|6[012])Z$!', '+00000002010-01-02T00:00:00Z' ) ) ? "yes\n" : "no\n"; [10:40:40] El búfer 01 está vacío. [10:41:10] yes [10:41:22] the patch doesn't really do much else [10:41:39] * aude nods [10:41:41] i don't get it [10:42:03] http://www.wikidata.org/wiki/Wikidata:Project_chat#deleted_by_127.0.0.1_without_deletion_log.3F <- any idea what is going on there? [10:42:06] * aude imho would also validate the calendars [10:42:07] btw [10:42:33] easy for bots to put in the wrong thing like the /wiki/Qxxx page [10:42:40] instead of /entity/Qxxx [10:42:43] aude: there's DataValue valiudation, and there'S DataType validation [10:42:51] mmmm, ok [10:43:06] it is part fo the value [10:43:09] part of [10:43:10] DataValue is very basic, doesn't know about config or properties or high level types [10:43:33] yes, but as far as the DataValue is concerned, globe is a string. could be anything. [10:43:36] i suppose it gets combined at some point? [10:43:42] we could restrict the syntax a bit, but that wouldn't help [10:43:44] * aude talking atbout calendar [10:43:45] about [10:43:54] same thing [10:44:00] this might be at the api level or somewhere else then [10:44:01] globe and calendar [10:44:04] not for today [10:44:09] on the DataValue level, these are just arbitrary identifiers [10:44:15] but icky that the api accepts whatever [10:44:20] ok [10:44:24] validation of the URIs can only happen on a higher level [10:44:28] ok [10:44:39] i'm working on that, but it needs quite a bit of infrastructure and refactoring [10:44:44] fine [10:44:50] (that'S what i originally intended to do today...) [10:44:53] anyway :) [10:45:02] * aude working on https://bugzilla.wikimedia.org/show_bug.cgi?id=49417 [10:45:07] that needs another backport [10:45:50] * aude thinks code freeze and branching on wednesday is not enough time for deployment on monday [10:46:15] would be if we had actual copmmunity testing [10:47:45] anyway, i'll update test with henning's patch [10:47:52] and fix 49417 [10:49:33] aude: i can created, edit and remove dates via the UI with my patch applied to wmf6# [10:49:43] hmmmm [10:59:18] DanielK_WMDE: http://dpaste.com/1228627/ [10:59:24] now when just viewing an item [10:59:35] * aude wonders how i can have such different results on my wmf6 branch [11:14:50] DanielK_WMDE: that still fails.. [11:16:13] previously I had to put ../../extensions/Wikibase/lib/tests/phpunit in suites.xml to run it, so with --group=Wikibase, how does it know where tests are? [11:17:29] New patchset: Tobias Gritschacher; "Fix precision detection in globe coordinate parser" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67967 [11:18:42] New patchset: Tobias Gritschacher; "Fix precision detection in globe coordinate parser" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67967 [11:19:55] DanielK_WMDE: Hi! [11:19:57] I've just made a change to add support of unicode quotes that are used in Wikipedias. https://gerrit.wikimedia.org/r/#/c/67966/ [11:19:59] Is it possible to deploy it at the same time as the other bug fix for the globeCoordinate in order to avoid the bad experience to be unable to copy/paste from Wikipedia? [11:20:07] aude: this solves the precision issue (cutting off seconds): https://gerrit.wikimedia.org/r/#/c/67967/ [11:20:43] aude: can you please check with your examples and confirm it is solve. we can then backport this change as well. [11:21:15] at least we have two main issues solved then. [11:21:39] ok [11:22:05] Tpt_: Tobi_WMDE will have a look after lunch [11:22:07] thank you! [11:22:27] Tpt_: thx, will have a look at it! [11:22:40] Lydia_WMDE, Tobi_WMDE: Thanks! It was a very annoying bug. [11:23:24] [03:42:03 AM] http://www.wikidata.org/wiki/Wikidata:Project_chat#deleted_by_127.0.0.1_without_deletion_log.3F <- any idea what is going on there? <-- translate extension, known bug [11:23:55] sounds like translate [11:24:26] Lydia_WMDE: An off topic question: is Wikivoyage support is on scope of the next too weeks or will happen later? [11:25:59] New patchset: Henning Snater; "Improved GeoCoordinate parsers' coordinate detection" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67968 [11:27:51] New patchset: Aude; "(bug 49417) Skip any change in client change handling that failed for some reason" [mediawiki/extensions/Wikibase] (mw1.22-wmf5) - https://gerrit.wikimedia.org/r/67969 [11:29:02] DanielK_WMDE: https://gerrit.wikimedia.org/r/#/c/67969/ [11:29:36] i'm sure this could be more elegant but think this is the place to be sure to catch anything that goes wrong [11:36:07] DanielK_WMDE: i am getting the Exception Caught: $time needs to be a valid ISO 8601 date when saving a geocoordinate :/ [11:36:18] * aude has evil stuff in my item [11:36:35] can't remove it [11:36:58] * aude has stuff with "The value does not comply with the property's definition." [11:37:03] can't remove those [11:37:11] * aude is on master [11:37:51] *gives coffee to aude* [11:38:01] :) [11:38:12] aude: just kill the item [11:38:23] yes, but we have items like that in wikidata [11:38:30] how do we kill those? [11:38:34] aude: "like that"... like what? [11:38:35] maybe via the api [11:38:38] sure we do [11:38:55] "The value does not comply with the property's definition." means that the property definition changed [11:38:56] sure there is garbage in wikidata [11:39:01] i don't think we have that on wikidata [11:39:14] no, it's a valid property with valid snaks [11:39:34] * aude looks for corrupt item [11:39:36] complaining about time when saving a geo-coord also points towards confusion about property types [11:40:02] well, it's saving the entire entity right? [11:40:06] trying to [11:40:07] aude: can you export the item in question, so i can look at it? [11:40:11] legoktm: thx [11:40:12] one sec [11:40:31] ok, you saw https://bugzilla.wikimedia.org/show_bug.cgi?id=49425 [11:40:35] yes, it's saving the entire entitiy... but i don't think it's instantiating all the snaks [11:40:37] it shouldn't [11:40:58] with http://www.wikidata.org/wiki/Q1825747 after we put your patch in / move to the next branch [11:41:06] when editing a claim, the other claims should stay in the json form and not be touched [11:41:18] that is unremovable, unless maybe by bot somehow [11:41:38] DanielK_WMDE: but the other claims became invalid [11:41:43] "ununserializable" :P [11:41:48] they were accepted before [11:41:48] what kind of evil hack is that [11:42:08] we saw that before [11:42:08] "they"? [11:42:31] invalid time obviously gets accepted [11:42:53] with your patch, seems there is issue with removing it [11:42:54] aude: i just added a description to that item. worked fine. [11:43:00] so a broken snak doesn't block all editing [11:43:00] that's on wikidata [11:43:04] without your patch [11:43:14] with your patch, it's a problem appartently [11:43:17] apparently [11:43:44] aude: via the UI? yes, it would be [11:44:01] i don't really see a way to make this work without a lot of special case stuff [11:44:08] a bot could reset the item without the corrupt part [11:44:11] it should be possible though to remove it via the api [11:44:18] yes [11:44:22] yes removing needs to be possible [11:44:48] well then, tell the UI guys that we need to be able to remove something without entering edit mode :o) [11:44:52] they are going to love it... [11:45:20] as far as i can tell, it'S just a question of how the UI does this [11:45:25] it has to do with edit entity i think [11:45:42] i don't think so [11:45:49] EditEntity works fine for other stuff [11:46:00] edit entity the api module [11:46:06] to be clear :) [11:46:30] yea, i know [11:46:35] anyway - not sure what i'm debugging here [11:47:07] the patch makes sure no invalid data calues can be created [11:47:33] this is independant of the question of how to remove snaks that contain broken data [11:47:36] it's a good thing [11:47:51] anyway - gsoc call in 15 minutes [11:48:02] * aude shall make a bug [11:48:07] Lydia_WMDE: is there anything we should discuss before the call? [11:48:22] DanielK_WMDE: i think we're good [11:48:36] Lydia_WMDE: ok... do you think we should start with a round of introductions? [11:48:51] sounds good [11:49:04] just a few sentences to give some background [11:49:09] ok [11:49:17] * DanielK_WMDE isn't very well prepared [11:51:04] New review: Daniel Kinzler; "I can create, edit and remove dates via the UI just fine..." [mediawiki/extensions/DataValues] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67962 [11:53:36] Lydia_WMDE: i see no hangout link in the calendar... where is it? [11:54:02] DanielK_WMDE: sykep [11:56:20] Lydia_WMDE: o_O [11:56:23] seriously? [11:56:34] i don't even know whether that's gonna work on this box :/ [11:56:39] let me try... [11:56:59] also... why?! [11:58:37] Lydia_WMDE: hm, scheint zu funzen. hoffen wir's mal. das bedeutet kein video, oder? [12:02:05] because that is what the students wanted ;-) [12:02:07] ok [12:02:09] ready here too [12:02:11] let's see [12:02:57] Lydia_WMDE: Introducing the GSOC students? [12:03:22] hoo: jep [12:13:54] Change merged: jenkins-bot; [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67967 [12:19:34] Reedy: Hi. Can you apply https://dl.dropboxusercontent.com/u/7125981/wikidata.ico for test wikidata favicon? It is some times needed [12:20:21] (it is two layer icon, 16x16 & 32x32, just like original one) [12:26:35] New review: Tobias Gritschacher; "PS 2: removed console output." [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67966 [12:27:18] Change merged: jenkins-bot; [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67966 [12:48:50] New patchset: Tobias Gritschacher; "Fix precision detection in globe coordinate parser" [mediawiki/extensions/DataValues] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67972 [12:51:56] Change merged: jenkins-bot; [mediawiki/extensions/DataValues] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67972 [12:54:19] Change merged: jenkins-bot; [mediawiki/extensions/DataValues] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67973 [12:56:06] New review: Aude; "on master, with the invalid value specified in bug 49417, I get an exception and stack trace and can..." [mediawiki/extensions/DataValues] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67962 [12:57:25] Hello all [12:58:00] DanielK_WMDE: https://bugzilla.wikimedia.org/49433 [12:58:33] New review: Aude; "http://dpaste.com/1228902/" [mediawiki/extensions/DataValues] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67962 [12:59:23] New review: Aude; "imho, enough issues for it not to go in the branch" [mediawiki/extensions/DataValues] (mw1.22-wmf6) C: -2; - https://gerrit.wikimedia.org/r/67962 [12:59:33] I have seen your data is CC-0 thats very cool :-) [12:59:45] "All structured data from the main and property namespace is available under the Creative Commons CC0 License;" [13:00:06] is there more about this anywhere ? [13:00:19] Tobi_WMDE: are you doing the backports for geocoordinates> [13:00:21] ? [13:00:23] what is "the main and property namespace" ?* [13:00:33] -* [13:00:51] popeye: all the items and stuff in Property:P### [13:00:56] items = Q### [13:01:27] aude: Thank you :-) [13:02:11] aude: done already [13:02:17] yay! [13:02:25] so, if you want you can update the test system [13:02:34] can we run all the tests again? [13:02:55] not that tests catch everything, but added check [13:03:58] test repo updated [13:04:13] aude: all three changes are only affecting the JS globe-coordinate parser. we don't have selenium tests for that and it cannot break something else. I've run the qUnit tests and they're fine. [13:04:21] hmmmm, ok [13:04:33] oh, i need...... [13:04:43] aude: ? [13:04:43] https://gerrit.wikimedia.org/r/#/c/67969/ [13:04:50] if DanielK_WMDE is not looking at it [13:05:25] aude: that's for wmf5? [13:05:28] i assume aaron disabled the change notification jobs because of that [13:05:28] yes [13:05:43] aude: that won't affect wmf6 then [13:06:02] well, we should have this or another patch that covers this in wmf6, just in case [13:06:05] oh, the clients [13:06:08] Hi! I'm trying to install Wikidata, but get this error message: http://codepad.org/gGdNltZb on Special:Version [13:06:13] and any other client patches [13:06:17] are they on 5 or 6? [13:06:22] 5 [13:06:24] ok [13:06:32] before we update tonight, let's get all the wmf5 patches into wmf6 [13:06:43] * aude cherry pick them [13:06:58] running maintenance/update.php works neither: http://codepad.org/cTSXU3vr [13:07:11] although wmf6 knows about globecoordinates..... should not be so much a problem [13:07:12] how can I fix this? [13:07:52] Granjow: hmmm, new to me [13:08:44] do you have entityPrefixes in your local settings? [13:08:50] or not (and use the defaults) [13:09:33] I have possibly run update.php in the wrong order, i.e. with require Wikibase.php but without WikibaseLib.php [13:09:43] Does not tell me anything, what are they? [13:10:11] Granjow: add this to your LocalSettings.php: $wgWBSettings['entityNamespaces'][CONTENT_MODEL_WIKIBASE_ITEM] = NS_MAIN; [13:10:12] i think wikibase repo should attempt and automatically load wikibase lib [13:10:43] Granjow: and $wgWBSettings['entityNamespaces'][CONTENT_MODEL_WIKIBASE_PROPERTY] = WB_NS_PROPERTY; [13:10:53] how to know what is the Q### for a special page ? [13:11:13] s/to/do i/ [13:11:33] popeye: special pages don't have Q#### [13:11:38] Tobi_WMDE, Before or after the require? [13:11:49] if i understand.... [13:11:50] Granjow: after [13:12:17] aude: ok [13:12:22] ok. Both times this backtrace: http://codepad.org/slprA4ya [13:12:27] i'm disapointed ^^ [13:13:58] popeye: ? [13:14:18] * aude probably confused [13:15:48] aude: :-) no i think i dont have enough knowlegde [13:16:00] If i want to know who is the mayor off all towns in my region [13:16:17] what should i need to do ? [13:16:33] oh, we don't have complex queries yet [13:16:50] i'm not sure when we'll have that type of query [13:17:14] Granjow: what version of wikibase did you get? [13:17:32] Tobi_WMDE, HEAD [13:17:41] aude: ok, thank you [13:17:41] hm.. ok [13:18:06] is there a howto explaining what is possible at the moment ? [13:18:43] Tobi_WMDE, I first tried the snapshot, which gave me error messages even earlier. Did this maybe mess up my configuration? [13:18:45] popeye: some of the external tools might be able to do more http://www.wikidata.org/wiki/Wikidata:Tools [13:19:02] for wikidata itself, the wikipedias can access properties of the "connected item" [13:19:08] like for infoboxes [13:19:19] New review: Daniel Kinzler; "(1 comment)" [mediawiki/extensions/Wikibase] (mw1.22-wmf5) C: -1; - https://gerrit.wikimedia.org/r/67969 [13:19:26] popeye: you could use http://toolserver.org/~magnus/ts2/wdstats/ [13:20:04] DanielK_WMDE: !!!!! [13:20:05] Granjow: if you use HEAD and also latest version of mediawiki then it should work [13:20:15] change handler jobs are broken! [13:20:18] Granjow: you'll also need latest version of DataValues and Diff. [13:20:22] do you want a stack trace? [13:20:29] aude: how and where? the stack traces do not show anything related. [13:20:41] yes, stack trace please :) [13:20:42] aaron says our jobs are flooding the logs [13:20:52] * aude easily reproduced [13:21:02] aude: yes, failures of the parser functions in render jobs. not change notifications. [13:21:07] (as far as i can see) [13:21:26] no [13:21:28] no no no [13:21:39] http://dpaste.com/1228946/ [13:22:02] i would love a more precise solution to handle unknown data types [13:22:16] aude: Tobi_WMDE cool stuffs, thank you [13:22:28] i think we need something at this level though to handle invalid changes and log them somewhere [13:22:49] aude: ok, that's a different issue. the patch is good for that. please change the commit message accordingly, the bug reference is misleading. and use wfLogWarning instead of wfDebug, as noted in the comment. [13:22:51] unknown data types can be in main snaks [13:22:57] qualifiers, referneces etc [13:23:06] ok, updating [13:23:14] Tobi_WMDE, Updated Diff (DataValues was git already). Same error. [13:23:26] aude: i'm not opposed to the patch in general, it just as a minor issue, and it doesn't fit the bug in the summary. [13:23:33] ok [13:23:44] the bug covers "unknown data type" [13:23:55] generally, which appeared in the parser function, lua and here [13:24:06] all essentially the same issue [13:24:10] then please attache the respective stack traces to the bug report to make that clear [13:24:16] ok [13:24:22] that's where i searched when looking for a rationale for the patch [13:24:28] New patchset: Henning Snater; "Reordered GeoCoordinateValue constructor parameters" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67975 [13:24:28] New patchset: Henning Snater; "Added precision detection to the GeoCoordinate parsers" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67976 [13:26:14] aude: as to a "nice" solution: the differ should have graceful handling of unknown properties and data types [13:26:30] * DanielK_WMDE adds a bug for that [13:26:46] Granjow: try those LocalSettings: http://pastebin.com/crpQggzB [13:26:51] that's what I'm using [13:27:16] New patchset: Aude; "(bug 49417) Skip changes in client change handling that with unknown data type" [mediawiki/extensions/Wikibase] (mw1.22-wmf5) - https://gerrit.wikimedia.org/r/67969 [13:27:56] Tobi_WMDE, Interesting ... that works! [13:28:13] aude: oh, it's not a differ issue, it happens when deserializing diffs. i guess that does need to be handled in the ChangeHandler. [13:28:13] well [13:28:17] Granjow: well, yeah [13:28:19] but the differ thing is true also [13:28:24] with a tiny detail that it hangs in an infinite loop in update.php [13:28:30] Processed 1 pages up to 0. [13:28:31] it's a bunch of stuff but complicated [13:28:39] we can't just unset some partof the change [13:28:59] we need to analyse it to see if unknown data type exists in snaks, references, qualifiers, etc [13:29:06] New patchset: Henning Snater; "Added precision detection to the GeoCoordinate parsers" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67976 [13:29:07] New patchset: Henning Snater; "Improved GeoCoordinate parsers' coordinate detection" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67968 [13:29:07] New patchset: Henning Snater; "Reordered GeoCoordinateValue constructor parameters" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67975 [13:29:10] Aborted it, restarted it, did not happen anymore. [13:29:17] if we want to filter at that level [13:29:19] Granjow: strange [13:29:29] Yep [13:30:17] aude: well, it's a broken change. that should raise an error. the decision how to handle that has to be made at the level of the job (that's the top level for our purposes). [13:30:32] alright [13:30:35] that's what i think [13:30:40] skipping broken changes and continuing seems reasonable [13:30:45] sure [13:30:57] we can log them, although i would love to see such logs :) [13:31:29] Ok, the extension works. Thanks for the help! [13:32:18] New patchset: Aude; "(bug 49417) Skip changes in client change handling that with unknown data type" [mediawiki/extensions/Wikibase] (mw1.22-wmf5) - https://gerrit.wikimedia.org/r/67969 [13:32:36] DanielK_WMDE: the patch now uses wfLogWarning [13:32:49] And since you are online, I would like to have this on Wikibase and, on the long run, as a Wikimedia project if possible: http://ec2-54-234-76-44.compute-1.amazonaws.com/filter.php [13:33:58] It is my master thesis I'm working on, an identification tool for plants, but can easily be used for other forms of life. [13:34:09] http://ec2-54-234-76-44.compute-1.amazonaws.com/editor.php is the editor [13:38:24] New patchset: Aude; "Handle failure to deserialize entity in lua" [mediawiki/extensions/Wikibase] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67977 [13:38:47] Wikibase would be ideal for that, since the characters are language independent, and the descriptions can be localised. And it is an approach to identify things on all over the world. Current keys are usually not editable and only focus on a small geographic region. [13:40:05] New review: Daniel Kinzler; "Ok, with the new info in the bug report, I now see what this patch addresses. Seems fine and should ..." [mediawiki/extensions/Wikibase] (mw1.22-wmf5) C: 2; - https://gerrit.wikimedia.org/r/67969 [13:41:07] aude: CR+2; can you work with tobi to get this tested for the V+2? [13:41:28] i can ask him to run tests yet [13:41:46] although tested it with wmf6 repo and wmf5 client and globe coordinates [13:41:55] Tobi_WMDE: can we run tests please? [13:42:35] * aude can run phpunit tests [13:42:43] Tobi_WMDE: if it works, merge it :) [13:43:32] * aude look forward to be done with these bug fixes and probably go home/do coding [13:43:41] and poke someone to deploy our stuff [13:45:40] Is there a document showing where and how Items and Properties are stored? [13:46:53] Granjow: from data model perspective, http://meta.wikimedia.org/wiki/Wikidata/Data_model [13:47:23] everything is stored in json blobs (in same place that wikitext is stored) [13:50:29] aude, Ok. And this can be parsed already? E.g. to create parent/child relationships? [13:50:52] i think some people are doing that, so yes [13:51:19] http://www.wikidata.org/wiki/Wikidata:Tools#External_tools [13:51:39] the geneawiki one and denny's tree of lif [13:51:40] e [13:52:17] DanielK_WMDE: so with my patch, tests are all okay [13:52:19] phpunit [13:52:20] Okay, thanks! Will look into it later, need to leave now. Later! [13:52:29] aude: DanielK_WMDE: what do you want me to test? https://gerrit.wikimedia.org/r/#/c/67969/ ? [13:52:30] ok [13:56:40] * aude sees the good changes in recent changes [13:57:05] Tobi_WMDE: do you want to verify the patch? [13:57:24] * aude thinks i tested it extensively, manually, phpunit etc [13:58:01] it filters out the stuff with unknown data types w/o affecting valid stuff [13:58:15] aude: can run the selenium client tests [13:58:18] ok [13:58:23] * aude would also like https://gerrit.wikimedia.org/r/#/c/67977/ approved [13:58:34] but I think we don't have test cases where we include time or coord [13:58:36] prefer over self merge [13:58:39] right [13:58:50] it's kind of hard to test that [13:58:56] automatically [13:59:03] why? [13:59:23] we could just try to include a time property [13:59:25] have some unknown data type from the repo to the client [13:59:29] and see if it fails [13:59:32] the issue is unknow n data [13:59:38] ah [13:59:38] time is known now [13:59:39] ok [13:59:41] geo is not [14:00:00] i mean making an automated test for that because next branch, client knows about it [14:04:19] !admin I'd like something to be removed... [14:04:29] at your service [14:04:51] PM :) [14:04:55] kk [14:07:31] i think i missed something [14:07:45] how can i find a simple way to know the Q### of an item ? [14:08:51] thnx again :) [14:09:11] popeye: Special:ItemByTitle ? [14:09:25] to lookup an item by the wikipedia site link/page [14:11:07] sounds great :-) [14:11:17] but i missed the context ^^ [14:11:45] sorry for my noob questions [14:12:20] for example https://www.wikidata.org/wiki/Special:ItemByTitle/enwiki/Japan will take you to Q17 [14:12:41] because it has a "site link" entry to english wikipedia article Japan [14:13:15] http://www.wikidata.org/wiki/Q17#sitelinks [14:15:08] aude: You saved the day :-) [14:15:13] yay! [14:15:25] * aude tries to avoid jargon [14:16:09] notenote that i think it still is case sensitive [14:16:23] e.g. Special:ItemByTitle/enwiki/Barack obama would not match [14:16:46] in that case, you could also try the search although it is imperfect also [14:16:48] aude: Is the topic title normalization? [14:16:52] hoo: [14:17:08] i think we havei think there is a setting that we could enable it [14:17:25] but would mean an api call to the client wiki if exact match is not found [14:17:44] grrr, "i think there is a setting...." [14:17:54] Yes... the client widget still doesn't work with redirects :/ [14:18:15] But I'm not sure that the API even would normalize with the setting set [14:18:42] aude, Tobi_WMDE: would could put some dummy type into the repo's config, so we can easily test for unknown types in the client [14:19:08] DanielK_WMDE: something like that..... [14:19:56] * aude puzzled how time worked fine last time, although i patched a couple things [14:20:01] lua was perfect [14:20:06] DanielK_WMDE: Do you think we should normalize titles in wbgetentieties or hack around that and normalize manually? [14:21:05] For queries like https://www.wikidata.org/w/api.php?action=wbgetentities&sites=enwiki&titles=Berlin&languages=en I mean [14:21:29] New patchset: Henning Snater; "(bug 38201) Added wb.Site.prototype.getGroup" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67289 [14:22:12] New patchset: Henning Snater; "(bug 38201) Refactor Sites related wikibase tests in frontend" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67290 [14:22:19] New patchset: Henning Snater; "(bug 38201) Added JavaScript wikibase.getSiteGroups()" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67291 [14:22:37] aude: tests were fine [14:22:45] ok [14:23:01] so, merge? [14:23:07] yes [14:23:25] and https://gerrit.wikimedia.org/r/#/c/67977/ [14:23:48] that one is already in wmf5, so just for sanity [14:24:02] and probably should be in master [14:24:46] Shall I merge it? Still looks as sane as yesterday :D [14:24:59] DanielK_WMDE: do we want the change handler patch in wmf6 also [14:24:59] hoo: sure [14:25:01] Change merged: Tobias Gritschacher; [mediawiki/extensions/Wikibase] (mw1.22-wmf5) - https://gerrit.wikimedia.org/r/67969 [14:25:08] * aude prefers not to self merge when others are around [14:25:25] New review: Hoo man; "Still looks as sane as it did yesterday :D" [mediawiki/extensions/Wikibase] (mw1.22-wmf6); V: 2 C: 2; - https://gerrit.wikimedia.org/r/67977 [14:25:30] thanks hoo [14:25:47] Yeah... self merging is bad [14:25:51] this way, thursday there are no submodule changes [14:26:51] aude: yes, can go into wmf6 and master [14:26:56] DanielK_WMDE: ok [14:27:02] Change merged: jenkins-bot; [mediawiki/extensions/Wikibase] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67977 [14:27:12] i'll do wmf6 and then master later/tomorrow [14:28:02] New patchset: Aude; "(bug 49417) Skip changes in client change handling that with unknown data type" [mediawiki/extensions/Wikibase] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67985 [14:28:12] https://gerrit.wikimedia.org/r/#/c/67985/ [14:28:34] that's the last one, except for ensuring everything is in master [14:28:51] hello people :) [14:29:06] aude: wha are we not doind the other way around? first in master and then the backports? [14:29:13] let's do it next time [14:29:14] Tobi_WMDE: last night, no [14:29:17] no time for that [14:29:58] aude: I mean that one: https://gerrit.wikimedia.org/r/#/c/67969/ [14:30:12] maybe... [14:30:37] * aude can submit it to master now [14:30:58] Change merged: Tobias Gritschacher; [mediawiki/extensions/Wikibase] (mw1.22-wmf6) - https://gerrit.wikimedia.org/r/67985 [14:31:03] New patchset: Henning Snater; "(bug 38201) Added JavaScript wikibase.getSiteGroups()" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67291 [14:31:09] New patchset: Aude; "(bug 49417) Skip changes in client change handling that with unknown data type" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67986 [14:31:15] New patchset: Henning Snater; "(bug 38201) Fixed assumption of SiteLinksEditTool of single edit tool per page" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67303 [14:32:01] New patchset: Aude; "Handle failure to deserialize entity in lua" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67987 [14:32:09] i've got a question: i've heard about wikimedia on the radio some time ago and was asking myself if it was possible to use it for a private wiki that i have with a couple of friends. is there a way to 'get' this technology and deploy it for private wikis? [14:32:49] New patchset: Aude; "(bug 49417) Handle failure to deserialize entity in lua" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67987 [14:34:03] New patchset: Henning Snater; "(bug 38201) Display site-link group specific heading per SiteLinksEditTool in JS" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67308 [14:34:04] hi malvina_... do you want to do that and use the data from Wikidata.org or an own data storing wiki? [14:34:12] aude: what's your recipe for injecting broken time values? [14:34:25] malvina_: See also http://semantic-mediawiki.org/ [14:34:27] i guess i can try with the ApiSandbox [14:34:28] with my bot [14:34:41] that you build yourself? [14:34:41] just an api request [14:34:43] hoo: own data [14:34:44] could be done with js [14:34:49] in the console [14:34:55] DanielK_WMDE: yes :) [14:35:00] i'm looking for something that is not going to take me more than 5 minutes :) [14:35:02] * aude uses it to abuse my test wiki [14:35:06] use the js console [14:35:12] hm... export, edit, import? yea, that could work. [14:35:13] mw.user.getTokens or whatnot [14:35:14] malvina_: Then you can of course install the Wikibase extensions and use them... if you only got on Wiki you probably want to look at http://semantic-mediawiki.org/ [14:35:20] or be anon [14:35:48] It's another project, but also includes (like the name says) the ability to store data in a semantic way [14:36:03] aude: the problem exists on master, you say? items with bad time data can't be viewed/edited? [14:36:11] hoo: do you have a link for me or shall i google out the way myself? :) [14:36:23] malvina_: Link for what? [14:36:27] DanielK_WMDE: yes [14:36:34] i see a stack trace only [14:36:39] hoo: for the extensions [14:36:43] ok, will check this out [14:36:49] if i put the one from the bug report in [14:37:04] removing it, i could investigate more and try more ways [14:37:08] but it should just work [14:37:28] can't see it in the api also with wbgetentities [14:37:33] stack trace [14:37:35] malvina_: Oh sure: http://semantic-mediawiki.org/ that's semantic MediaWiki, take a look at ... and Wikibase can be found on https://www.mediawiki.org/wiki/Extension:Wikibase (that's probably harder to install) [14:37:41] * take a look at it [14:39:03] hoo: thanks a lot!!! i will have a look at both. if i don't get it at all, i shall return :P [14:39:11] :D ok [14:39:30] thanks again and have a great day! [14:39:48] Thanks, you're welcome [14:40:06] New patchset: Tobias Gritschacher; "Improved GeoCoordinate parsers' coordinate detection" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67968 [14:41:30] aude: book. ok, can reproduce. [14:41:39] it'S actually behaving as designed. [14:41:50] it's just a question of making the relevant code a biot more robust against failures :) [14:41:56] DanielK_WMDE: sure [14:42:23] New patchset: Henning Snater; "(bug 38201) Added JS wikibase.getSitesOfGroup" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67393 [14:42:32] just makes it a bit unusable [14:43:25] Change merged: jenkins-bot; [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67968 [14:43:55] aude: indeed :) [14:45:37] New patchset: Tobias Gritschacher; "Reordered GeoCoordinateValue constructor parameters" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67975 [14:55:54] aude, Danwe_WMDE: the only clean way i see is to introduce a new snak type: PropertyBadValueSnak. [14:56:07] Reedy: Done with some dark magics :D http://test.wikidata.org/wiki/MediaWiki:Common.js [14:56:11] Reedy: https://gerrit.wikimedia.org/r/#/c/67992/ and https://gerrit.wikimedia.org/r/#/c/67993/ are ready anytime [14:56:22] the alternative is to *silently* ignore claims with bad values. [14:56:26] then we should also be good for thursday with no submodule changes [14:56:49] DanielK_WMDE: could be a good approach [14:56:53] Change merged: jenkins-bot; [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67975 [15:13:45] hmmmm, so WikibaseQuery does not automatically get included with WikibaseLib or repo? [15:13:56] DanielK_WMDE: do you know? [15:14:09] e.g. i have to explicitly require it [15:24:28] thanks Reedy [15:26:20] Lydia_WMDE: updates are deployed [15:26:30] \o/ [15:26:33] * aude can cut and paste 40° 44′ 54.36″ N, 73° 59′ 8.36″ W into test.wikidata [15:26:33] thx aude and Reedy [15:26:36] and it works [15:26:50] without rounding, with the apostrophes, etc [15:26:50] i'll take care of the chat and mailinglist in a bit (in a call for the next hour or so) [15:26:54] ok [15:27:15] http://test.wikidata.org/wiki/Q17 [15:27:50] 40.748433, -73.985656 works [15:28:36] sweet [15:55:33] http://toolserver.org/~magnus/ts2/wdstats/?q={%22r%22:[{%22p%22:374,%22i%22:484170},{%22p%22:6,%22i%22:3342}]} [15:55:36] i dont understand why it doesnt work ? [15:56:33] this page got these informations https://www.wikidata.org/wiki/Q16121 [15:56:42] what did i do wrong ? [16:13:17] aude: hi :) [16:18:59] aude: i can't see a way to make DataValue validation work without breaking the UI for any item with an invalid data value [16:19:13] handling this needs quite a bit of work - not something i'd want to backport [16:19:34] ...and it ties in with refactoring EntityView :) [16:26:34] New review: Daniel Kinzler; "The jenkins failure seems unrelated... hwy is it happening here & now?" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/66122 [16:34:00] New patchset: Henning Snater; "Added precision detection to the GeoCoordinate parsers" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67976 [16:34:11] i got another noob question : is the data retrieved via https://en.wikipedia.org/w/api.php also in CC-0 [16:34:13] ? [16:38:10] New patchset: Henning Snater; "Added precision detection to the GeoCoordinate parsers" [mediawiki/extensions/DataValues] (master) - https://gerrit.wikimedia.org/r/67976 [16:38:47] popeye1: why cc-0 for enwp? [16:40:02] liangent: the data are not the same as wikidata ? [16:41:32] I am asking because i find it easier to use [16:42:40] liangent: remember i'm a noob so i may be wrong :-) [16:42:55] popeye1: does enwp api serve data from wikidata? [16:43:14] except if, for example, you feed {{#property:p1}} to action=parse [16:43:20] or you mean sitelinks [16:44:53] liangent: I'm lost ^^ [16:45:17] is https://www.wikidata.org/w/api.php the same ? [16:46:06] not the same [16:46:13] my need is : getting the name of the mayors of towns from a region [16:47:34] and other things like population, density, etc [16:47:44] and that's not systematically doable on enwp api [16:48:06] ok [16:48:29] maybe possible in wikidata api, depending on whether there're related data already [16:48:49] https://www.wikidata.org/w/api.php ? [16:49:32] yea [16:49:46] ok [16:50:16] if the info is in the infobox on wikipedia it would be ok ? [16:52:24] i'm asking because this is not the case with this page : http://en.wikipedia.org/wiki/Beaune and the Q### page https://www.wikidata.org/wiki/Q16121 [16:54:19] popeye1: it won't give you data stored in wikipedia infobox directly and there's no systematical way to extract data from wikipedia infoboxes no matter which api you use [16:54:38] migrating data in infoboxes to wikidata.org might be on some roadmap [16:57:42] liangent: ok so if i get those data in any of these ways, is this data under CC-0 ? [16:57:56] popeye1: only if it is from wikidata [16:58:26] benestar: ok [16:58:44] however, we are importing data mainly from wikipedia... [16:58:58] popeye1: which data do you need? [16:59:50] benestar: getting the name of the mayors of towns and other things like population, density, etc from a region [17:00:54] benestar: actually are those fact data eligible for copyright [17:02:57] typically things in the infoboxes [17:23:31] New patchset: Henning Snater; "(bug 38201) Introduction of "allowedSites" option in SiteLinksEditTool" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67394 [17:28:24] New patchset: Henning Snater; "(bug 38201) Introduction of "allowedSites" option in SiteLinksEditTool" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67394 [17:29:30] New patchset: Henning Snater; "Refactoring of SiteIdInterface test" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67559 [17:39:46] New patchset: Henning Snater; "Refactoring of SiteIdInterface test" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67559 [17:53:04] New patchset: Henning Snater; "less error prone jQuery.wikibase.siteselector matching character highlighting" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67560 [18:22:02] aude: ping [18:24:29] Lydia_WMDE: can u help me? [18:24:55] benestar: hey [18:25:00] what's up? [18:25:26] I installed my local wiki again just now with the latest version from gerrit [18:25:32] ok cool [18:25:41] however if I try to create an item , it fails :( [18:25:46] outsch [18:25:49] what's the error? [18:26:27] Fatal error: Call to undefined method Wikibase\ItemDiff::toArray() in <...>DiffChange.php on line 131 [18:27:04] ok [18:27:13] do you have the newest version of all the other repos too? [18:27:22] diff, datavalues and so on [18:27:31] ah, perhaps that's the issue... [18:27:39] let me get the latest versions [18:27:41] likely :) [18:31:54] Lydia_WMDE: great, it worked :) [18:33:12] \o/ [18:34:24] Lydia_WMDE: another question [18:34:39] the deleteAllData.php file does not exist in the latest revision... [18:35:05] ah, sorry. found it [18:35:12] k :) [18:35:16] has been moved to lib ;) [20:16:00] New patchset: Bene; "fixed error message in Special:SetSiteLink" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/68043 [20:16:40] Lydia_WMDE: I have fixed the non parsed error message [20:16:52] cool! [20:17:19] did you already poke someone for the without description special page? [20:18:59] @ Lydia_WMDE [20:19:29] uff what was this about again? [20:19:44] sorry this dropped of my radar it seems [20:19:46] Lydia_WMDE: https://gerrit.wikimedia.org/r/#/c/67634/ Special:EntitiesWithoutDescription [20:20:04] ohh [20:20:05] that one [20:20:25] no not today because everyone was busy with fixing yesterday's deployment [20:20:45] doesn't matter [20:20:58] I am only excited to see it in action ;) [20:21:02] :P [20:21:08] you should be [20:29:46] New patchset: Bene; "Special page for entities without description" [mediawiki/extensions/Wikibase] (master) - https://gerrit.wikimedia.org/r/67634 [20:50:28] addshore: Do you still do imports? Dutch Wikipedia has about 300.000 items now that still need to be imported to Wikidata [20:50:41] (thanks to some insect loving bot owners) [20:52:16] legoktm: Or did you do that too? :-) [20:52:42] My import bot is effectively dead, I don't have the time to maintain it. [20:52:52] Plus now people have a lot more advanced bots than mine [20:53:46] Ricordisamoa has one too [20:55:05] my import bot is also effectivly dead [20:56:21] Im still checking wikipages and removing links when they are on wikidata [20:56:24] but not importing :/ [20:56:34] Oh, that reminds me [20:56:45] theres just under a million articles left with links on (i think) [20:56:50] We have a bot task ready for importing more area info, but this time it comes from enwiki [20:57:08] Is there specific stuff we should be doing there, that the community prefers or requires? [20:57:44] (we made a test run and we think it looks fine, but maybe you want stuff done in some specific way or, well, dunno :) ) [21:00:33] reosarevok_: is it on request for approval / permissions? [21:00:35] :> [21:01:14] addshore: Can you just do a simple run at nlwp? I have a list of articles. Just import the ones that don't have any links at all [21:01:44] hmm *checks his code* [21:02:20] multichill: list? [21:02:27] (link) [21:03:06] multichill: I will make my bot run over the 21,000 i have in my db [21:03:21] addshore: http://toolserver.org/~multichill/temp/queries/wikidata/articles_not_wikidata.txt [21:03:59] If you scroll to it you'll see a lot of binomial names [21:04:17] That's all new articles imported by a bot in the last two months [21:04:35] oh, they are new in the last 2 months to nlwiki? [21:05:13] yup [21:05:30] We have a bug problem :P [21:05:57] Some of them are properly linked to other wiki's [21:06:02] Hi Amir1 [21:06:33] multichill: Hi [21:06:51] what's up [21:07:59] Nothing much, just trying to figure out who can import the recent batch of newly created articles in Dutch. I hope addshore or benestar can help me out here ;-) [21:08:32] multichill: only importing or also removing local langlinks? [21:09:03] Whatever you're best at. Removing the local links would be nice [21:10:18] multichil: there is code reza has written it, Jand is using this code too, let me search about it It import all of articles and remove local lang links too [21:13:32] not Jand, JAn Dudík. [21:13:35] his name is hard [21:13:45] http://www.wikidata.org/wiki/User:Reza1615/BOT/new_interwiki.py [21:13:50] this is the code [21:23:47] multichill: Im running my bot accross them now [21:24:01] many of the first dont seem to have iwlinks [21:25:32] addshore: No interwiki link means that you start a new item? [21:25:58] ohhh [21:26:11] so [21:26:14] you dont want to import links from them [21:26:19] but create items for them? :P [21:26:44] Yeah, would be a bit hard to import links without an item..... [21:27:26] heh, yes, but my code currently only creates an item if it finds a link to import also ;p [21:28:11] So you need at least 2 langs? [21:28:28] well, i can modify the code to create the items now ;p [21:29:24] I hope you can figure something out. Thanks for the help! Going to bed now, had a long day.... [21:30:21] multichill: addshore, If you don't want to or don't have time [21:30:24] i can do it [21:30:38] until tomorrow [21:31:22] bye [21:31:39] o/ [21:31:45] legoktm: hi [21:31:47] o/ [21:31:52] er [21:31:53] hi* [21:31:53] :P [21:31:55] how you're doing [21:32:01] pretty good, you? [21:32:13] not bad I got full marked in the exam [21:32:14] LP [21:32:18] :P [21:32:23] nice! [21:33:07] but i got B+ in electronics It's a disaster [21:33:08] :D [21:33:19] B+ = Asian F [21:33:39] heheheh [21:33:45] legoktm: about migrating to bz [21:33:51] yurik doesn't know [21:34:13] a script to export bugs of sf and import it bz [21:34:16] hm [21:34:24] Amir1, yurik is strugling with ip filterings in varnish :( [21:35:24] Amir1: looks like they have an API http://sourceforge.net/apps/trac/sourceforge/wiki/API [21:35:46] we can write it [21:36:18] yee Amir1 I will leave it to you :) [21:36:46] addshore: ok :) [21:36:50] http://sourceforge.net/api/project/name/pywikipediabot/json [21:37:28] before writing the whole script [21:37:41] let check brion [21:37:47] or ^demon knows or not [21:38:05] <^demon> Knows what? [21:38:30] ^demon: we're looking for a script that will let us move from sourceforce --> bugzilla [21:38:31] * addshore really should get on with learning pywikibot [21:38:31] ^demon: hi [21:38:48] someone had to write one for mediawiki a looooong time ago [21:38:59] <^demon> Not me :) [21:39:21] ^demon: maybe someone else you know? [21:40:17] <^demon> brion did that back in the day I suppose. Lots of things have changed since then, so he may say I don't remember :) [21:41:24] legoktm: does bz support API or importing via file? [21:41:43] Amir1: not sure. [21:45:31] legoktm: I'm asking in #mediawiki [21:45:44] ok [21:48:00] legoktm: do you think we should import closed bugs? [21:48:22] probably not….. [21:48:31] maybe we should discuss this in #pywikipediabot? [21:48:46] yes [21:49:10] I agree [22:12:06] Hey! [22:12:17] bye [22:12:51] hi [22:20:43] Lydia_WMDE: Hey. Glad to see you all moved quickly on fixing that datatype error. What was the cause of it? [22:22:11] it's fixed? [22:25:53] rschen7754: says so in the update on project chat [22:25:57] yay [22:27:41] aude: poke [22:28:24] Reedy: poke alos [22:28:26] also* [22:44:50] ? [22:45:21] addshore: yep, https://www.wikidata.org/wiki/Wikidata:Requests_for_permissions/Bot#MBAreaBot_3 (sorry, was gaming) [22:48:24] Reedy: can you give me op rights on test.wikidata? [22:51:26] We don't have operator rights onwiki [22:54:07] Reedy: I'm pretty sure Pink|nap has administrator there [22:54:23] You asked for op which is usually short for operator [22:54:30] it's not wiki vocab [22:55:03] op is irc vocab ;) [22:55:29] oh fine [22:55:51] but op is also short for sysop, which is used as a stand in for admin sometimes [22:56:06] sure [22:56:13] Will you please give me admin rights on test.Wikidata? [22:56:16] sure [22:57:01] Done [23:02:06] thanks [23:02:22] Reedy: can you confirm, BTW, that geographic coordinate is fixed? [23:02:28] I've no idea [23:03:12] I think it might be, as people were testing it after i deployed some bugfixes aude had queued up [23:06:03] :> [23:17:54] * Reedy blames addshore for breaking it [23:18:01] :> [23:18:10] mehehe [23:26:21] addshore: do you know what went wrong? [23:31:12] geohacks just went for a loop on me [23:31:38] The clue is in the title [23:31:45] Wait so, we're ready to use coordinates now? [23:31:53] legoktm: apparently [23:31:55] :D [23:31:59] Great, let me set up my bot [23:32:09] https://tools.wmflabs.org/geohack/geohack.php?pagename=Massachusetts_Institute_of_Technology¶ms=42.35982_N_-71.09211_E_region:US_type:edu <--- anyone else getting a 500 server error [23:32:10] legoktm: is it approved?| [23:32:24] "To add claims" [23:32:45] legoktm: is your phase 2 bot working again? [23:33:10] rschen7754: tonight it will be :) [23:33:14] legoktm: I wanna copy! [23:33:25] * Sven_Manguard wishes he could program [23:33:28] legoktm: get ready for a ton of US road requests in the next few days :D [23:33:34] Wait, what? [23:33:39] rschen7754: do me a favor and click on https://tools.wmflabs.org/geohack/geohack.php?pagename=Massachusetts_Institute_of_Technology¶ms=42.35982_N_-71.09211_E_region:US_type:edu please [23:33:42] does it crash? [23:33:53] nope [23:33:54] I'm getting "The server encountered an internal error or misconfiguration and was unable to complete your request." [23:34:06] Sven_Manguard: Same here. [23:34:17] it's working for me now [23:35:01] I can confirm that geocoordinates can now handle the concept of "west" without going bonkers [23:35:35] Woo! [23:35:50] I can also confirm that I make excellent fruit smoothies, although this is less relevant on this channel [23:35:53] We and ready are subjective [23:36:07] is subjective? [23:36:12] and subjective? [23:36:20] ain't subjective? [23:41:31] legoktm: what language wiki do you use as your base? [23:41:42] enwiki [23:41:51] er, wait, for which bot? [23:42:01] my general phase 2 bot can run from any wiki [23:43:29] legoktm: yeah, but what do you set it on regularly, enwiki? [23:44:22] Well I normally do enwiki because I speak english, but requests come in from all 'pedias [23:44:48] https://www.wikidata.org/wiki/User_talk:Legobot/properties.js/requests [23:44:55] legoktm: and does this bot have a UI of any sore or is it console based? [23:45:05] console based. [23:45:11] fu [23:45:17] I've been working on a web version, but never got around to it [23:45:19] buttttt [23:45:23] because, you know, I'd really love to run a copy of it through Svenbot [23:45:44] https://www.wikidata.org/wiki/Wikidata:Tools/Array_properties_gadget [23:45:47] I would totally be willing to pay you for your time (and buy it into open-sourceness [23:45:56] it already is open source [23:46:13] CC-0 :) [23:46:15] publicly viewable source code* [23:46:24] https://github.com/legoktm/wikidata/blob/master/wikidata_properties.py [23:46:39] but that is out of date [23:46:48] ill post the updated version when i clean it up tonight [23:47:22] it was written before pwb had a proper api [23:47:23] legoktm: I will give you a "pittance of cash" if you walk me through setting up my own deployment of that [23:47:44] wait [23:47:47] if you want to run it [23:47:50] just use claimit.py [23:48:03] https://www.mediawiki.org/wiki/Manual:Pywikipediabot/claimit.py [23:48:13] see, this is why I'd be willing to pay you, I literally have no idea what you just said [23:48:47] ok, this will take like 10 minutes to set up. do you have an account on labs or ts? [23:48:55] I have a labs account, yes [23:49:11] haven't touched it since... when was DC Wikimania? [23:49:16] errrm... then [23:50:05] ok, log into wikitech.wikimedia.org with your labsconsole account [23:50:16] then go to https://wikitech.wikimedia.org/wiki/Special:FormEdit/Tools_Access_Request and just say "I want to run a bot" [23:51:49] legoktm: how much sql of toolserver is behind? [23:52:00] last time was 6 days [23:52:18] is there any way to run a sql request in wmflabs? [23:52:19] Amir1: im not sure, havent checked in a while [23:52:39] Amir1: yes, if you ssh to tools-login, just run $ sql wikidatawiki_p [23:52:47] oh ok [23:52:48] but its missing all of s7 [23:53:03] legoktm: thatll be up soon :> [23:53:09] addshore: how soon? [23:53:09] "soon" :) [23:53:19] i want my centauth! XD [23:53:22] few days perhaps :P [23:53:27] night all :> [23:53:33] gnite [23:54:03] silly me being up at 1am [23:55:50] Sven_Manguard: ping me once you're done [23:59:54] * addsleep points at his new botrequest before sleep [23:59:56] sleep now :> [23:59:57] night