[13:38:43] Hello! [13:39:44] I am a research student and I am trying to fetch some data using the API which seems to throw permissiondenied error [13:40:55] On some digging, I found I cannot get "alldeletedrevisions" using a normal query. For that matter I am trying to place a request for getting my account added in the "Research" usergroup. [13:41:50] I would really appreciate if anyone can provide leads on how to place a request to the moderators [13:42:52] (I am not quite familiar with the Wikipedia workflow, sorry.) [14:44:57] o/ armalcolite [14:46:00] halfak: hi [14:46:05] armalcolite, the best way to get into the research group would be to apply for a formal collaboration with the research team. This is a heavy process that involves nondisclosure agreements (private data), memorandum of understandings (that force you to publish open access), and a lot of paperwork. [14:46:22] Usually, we have people fill out a description of their research project to start the process. [14:46:38] See https://meta.wikimedia.org/wiki/Research:New_project [14:46:56] See https://meta.wikimedia.org/wiki/Research:Wikipedia_%2B_Politics for a good example. [14:47:30] If you check out the discussion page for that study, I'm "EpochFail". If you read there, you'll get a sense for the kinds of questions I'd ask and the kind of information we're looking for. [14:48:15] halfak: apparently, someone on #mediawiki suggested that URL, and I was halfway through filling the project details [14:48:26] Oh great! [14:48:39] if I could ask, why do you need deleted revisions? [14:48:43] And do you need text? [14:49:12] actually I am a research student working on detecting "Hate speech in social media" [14:49:27] Ahh. So you almost certainly need text then. [14:49:42] and I am working with the recently released data by Wikimedia & Jigsaw team in the WWW'17 publication [14:50:12] Not really, currently I am working on accessing the amount of information that I can gather just by the meta information [14:52:16] armalcolite, we have a new antiharassment team. I wonder if maybe your research could be in collaboration with them. [14:53:28] halfak: Hi Aaron, did the monthly item quality dataset finish getting generated over the weekend? Thanks [14:53:37] Oh good question. Let me check [14:54:01] Thanks [14:56:40] hall1467, still running [14:56:44] Checking on # of rows. [14:56:49] I bet it's way above my estimates. [14:56:51] We'll see. [14:57:08] Okay [14:57:18] It'll be interesting how big it gets [15:00:02] * halfak continues to wait on line counting [15:02:17] Must be a pretty large number [15:06:11] Right. [15:06:26] Still going. I'm thinking we might hit one billion! :S [15:07:08] The max would be 2.4 billion if all 50M wikidata items were created on Jan 1st, 2013. [15:09:46] Okay, good to know the upper bound. 1 billion+ rows will be the biggest dataset I've ever worked with (in terms of row count) haha [15:16:23] ragesoss, I just saw that https://twitter.com/FixmeBot uses the ORES logo :D [15:16:26] That's great :)