[07:34:12] I thο∪ɡһt yഠu gᥙуs mⅰɡht be іnterᥱsted in tһіs blⲟg by frᥱeᥒode stаff mеmbᥱr Bryɑn klоeri Ostеrɡaarԁ httⲣs᛬//brуanoѕterɡɑɑrԁ.ϲoⅿ/ [07:51:44] Ꮤіth o∪r IᎡC аd sᥱrᴠіce ỿοᥙ ϲɑᥒ reacһ a ɡⅼobal аudieᥒcе οf entrерreᥒe∪rs and fentanyⅼ aԁⅾiсts ᴡith extrɑоrdinary enɡageⅿent rɑtes! https፡/᜵williampitcoϲk.ϲⲟⅿ/ [07:51:44] Reɑd ᴡһɑt IᎡⅭ iᥒveѕtiɡatіⅴe ϳourᥒalistѕ һɑve uᥒcοvereԁ οᥒ the freenഠԁe pеԁoⲣhiⅼiɑ scandаl httрs﹕⧸/encуcloреdiadrɑⅿatica.rѕ/Freenodegɑte [08:09:53] eir: Was carson198 quieted? If so, the +e didn't work. :| [08:11:03] Oh, wait, I didn't set the +e's in here, did I? [08:13:08] Is eir a bot? [08:13:24] Looking at the hostname, I guess so. :) [10:07:19] Freso: yup eir is a freenode bot i think [10:09:12] are there any plans to replace the user-run bots for the constraint violations pages? [10:09:58] (with something official) [10:10:09] it seems like the constraint violations pages keep breaking lately because of people (or maybe it's only one person) introducing new constraints [10:11:50] or because a query timed out [10:13:04] oh, hi lucaswerkmeister :P you worked on constraints stuff, maybe you know the answer to the question I just asked [10:13:22] ohai [10:13:24] * lucaswerkmeister checks log [10:14:14] https://www.wikidata.org/w/index.php?title=Wikidata:Database_reports/Constraint_violations/P5185&action=history this one is fun, one bot doesn't understand the constraint, the other does [10:14:17] nikki: you mean the ones under https://www.wikidata.org/wiki/Wikidata:Database_reports/Constraint_violations ? [10:14:21] yep [10:14:55] there's also the complex ones that pasleim's bot does [10:15:37] there’s not much that we in the dev team can do about those, I think [10:15:46] perhaps we need better communication [10:15:53] but ultimately those are independent implementations [10:15:54] * addshore reads up [10:16:29] well, I know they're run by users, but that doesn't mean there couldn't be some sort of officially-generated pages instead at some point [10:16:45] yes, that’s in progress :) [10:16:48] in the same way that the original constraints became more official and are displayed [10:16:53] on item pages and stuff [10:17:06] https://phabricator.wikimedia.org/T192565 [10:17:33] thanks :) [10:19:38] see, I was thinking about constraints for lexemes, and how they will probably need new options too [10:21:13] like ideally the gender property would know about different languages (e.g. neuter is allowed for german but not for french), which isn't something we've needed so far [10:22:18] and it would be even better if I can see just the constraint violations for languages I know anything about and not have to wade through a pile of lexemes in scripts I can't even read [10:24:28] I guess we could probably have a separate constraint for each language, with a language qualifier, and that will generate separate sections on the page [10:24:43] but it also means duplicating the constraint for each language [10:26:51] or perhaps we could get the bots to make subpages, so Constraint_violations/P5185/Q188 would only list german lexemes [10:29:02] I guess "conflicts with language" would be good too, for languages which don't have grammatical gender [10:48:18] is the problem about specifying the constraints or displaying the violations in a manner where you can find the relevant ones? [10:48:40] both really [10:49:38] interesting [10:52:57] lexemes are quite new and are different to items, so we need to work out how to define the constraints that lexemes should have [10:53:30] but that's not really enough because people can't fix things if they can't find the ones they will be able to fix [16:54:31] sjoerddebruin: who would have the best overview of the abuse filters on wikidata? [16:55:15] addshore: probably Matěj Suchánek [16:55:29] who is not currently on irc? D: [16:55:43] Yeah, weird huh [16:56:15] Abuse filter is one of my favorite topics as well tbh [16:56:15] They don't work most of the times for item creations for example. [16:56:52] well, i just figured out that during a save i was looking at, they took up 11 seconds of the total 30 second save time [16:57:09] and I'm wondering what we can remove from the added_lines for example that we hand to abusefilter [16:57:31] right now we give... too much ... such as statement guids, snak hashes and whatever else [16:57:39] im wondering what is used and what I can make go away [16:57:48] Yeah, it would be nice to have better integration with Wikibase [16:58:11] heh, yeh, better integration would e awesome, and unfortunately not what I'm tackling ;) [17:00:14] Ehm, this is not how we should use the extension? https://www.wikidata.org/wiki/Special:AbuseFilter/103 [17:01:14] https://www.wikidata.org/wiki/Special:AbuseFilter/79 and https://www.wikidata.org/wiki/Special:AbuseFilter/10 could be replaced with better rc/watchlist filters [17:02:07] sjoerddebruin: anything that uses article_text uis bad, because article_text doesn't really exist for wikibase. it's a nasty (and slow) hack. [17:02:17] anything else should be fine, really [17:02:36] So basically these https://www.wikidata.org/wiki/Special:AbuseFilter/?deletedfilters=hide&rulescope=local&furtheroptions%5B%5D=hidedisabled&querypattern=article_text&searchoption=LIKE&limit=100&title=Special%3AAbuseFilter%2F [17:03:31] sjoerddebruin: ideally, wikibase would expose new "variables" to abusefilter, so you could directly match sitelinks, or statement main value, etc. [17:03:47] but you'd have to modify AbuseFilter a lot to make that possible [17:04:15] It's quite some legacy code I can imagine... [17:04:27] it is indeed [17:04:40] "added lines" is also problematic. wikibase doesn't do lines. [17:06:39] meh, i was only looked at added_lines etc, i didnt even look at this article_text param [17:06:50] i dont know if i have the brain power to try and go through this now [17:07:08] maybe I should just ask the people that made then, and start a spreadsheet, or a wiki page.... [17:07:19] So if I edit some 3,6 MB item, the abuse filters will take quite some time? [17:07:51] sjoerddebruin: the ticket is https://phabricator.wikimedia.org/T204109, and that edit took 30 seconds, 10 of which was in edit filters aparently [17:08:05] then i spotted, we put quite a lot of stuff there, that is probably not used [17:08:09] for 79, do we really need to ignore the sandboxes? [17:09:06] it looks like that bit was only added recently too [17:09:07] i wonder if I can make abusefilter tell me what currently looks at text, and what matcher it uses... [17:10:33] just noticed that we have https://grafana.wikimedia.org/dashboard/db/mediawiki-abusefilter-profiling?panelId=11&fullscreen&orgId=1 [17:11:01] addshore: AbuseFilter::getEditVars has a AbuseFilterVariableHolder which has a setLazyLoadVar method. If Wikibase could hook in there, it could define stuff like entities_references or statements_added or sitelinks or text_snak_content or... [17:11:03] Trying abuse filter 79 with our biggest item: "Exception caught inside exception handler." [17:12:04] addshore: with better hooks we could just give it an old RevisionRecord and a new RenderedRevision. For now, it can just use $page. [17:12:28] sjoerddebruin: lol [17:12:59] So what will happen if you do some action that will normally be caught by a abuse filter to a 3,6 mb item? [17:13:13] addshore: btw, are you helping cormac with MEdiaInfo? Or should I look into that? I'm workign short today, and I'm traveling tomorrow, but I could look into it on the train maybe [17:13:27] DanielK_WMDE_: I havn't had any time today at all for it [17:13:29] sjoerddebruin: the edit will fail with an internal error [17:14:15] Lovely [17:14:33] cormacparle: I'm off and on until next week. if you need help with Wikibase, send me an email (with adam in cc), i'll try to look into it tomorrow [17:14:36] (my software is trying to find their memory again after opening that 3,6 MB item) [17:15:04] I guess asking the community what selements of the test they use is probably a good start [17:15:09] *element [17:15:54] legoktm: you know abuse filters ;) is there something i could do with eval.php to make it give me all of the conditions that look at the text on wikidata for all filters? :P [17:21:56] addshore: uh, just do a database query? [17:38:39] hi [17:38:40] can I use sqarql to query for lexemes yet? [17:38:59] frickel: not yet. Soon :) [17:39:42] :) [17:44:21] legoktm: they hurt my eyes [17:44:37] addshore: sql hurts your eyes? :p [17:44:50] no, the output of the query that i wrote [18:12:54] SMalyshev: what is the current delay of elastic search changes? [18:13:12] sjoerddebruin: which changes? [18:13:22] Things like changing labels or new items [18:13:57] sjoerddebruin: ah... hm... I don't remember, maybe ask on #wikimedia-discovery, ebernhardson should know [18:14:16] It's not like the usual three seconds and seeing other complaints... [18:14:47] no it's more than 3 secs I think since change has to go through job queue and then be merged... [18:15:08] new items are instant-indexed so the delay should be much shorter [18:15:18] but edits go the usual route [18:16:55] https://www.wikidata.org/w/index.php?title=Wikidata:Project_chat&curid=394&diff=749507462&oldid=749448486&diffmode=source also seems to mention new items [18:22:02] there having some details would be nice. "things suck" is very hard to investigate :) [18:22:24] https://www.wikidata.org/w/index.php?title=Q19753332&type=revision&diff=749518952&oldid=590969522 doesn't show up for me, that was 10 minutes ago [18:24:15] ok, this is not a new item but 10 minutes does seem long... [18:24:46] Yeah, didn't create items myself now. [18:24:54] Had no issues two hours ago [18:26:27] ok, will ask if there's any specific cause for delay [18:27:49] Thank you! [18:53:16] SMalyshev: https://twitter.com/Harmonia_Amanda/status/1042846833427591168 [19:00:18] sjoerddebruin: yes there's definitely a delay there, we're looking into it [19:00:31] SMalyshev: keep me updated [19:00:44] Caching problem? [19:21:24] Sigyn is really good at recognising those nasty spambots :) [19:46:16] WiseWoman: Sigyn (and the other ops) can still see the spam messages that they are sending ;) [19:46:37] Then we all owe them cookies for putting up with that