Fork me on GitHub

Wikimedia IRC logs browser - #wikimedia-tech

Filter:
Start date
End date

Displaying 134 items:

2018-03-28 01:54:16 <Sir_Designer__> Demp
2018-03-28 13:25:59 <TK-999> so since technically it's event time... who would be a code owner for the components covered in https://gerrit.wikimedia.org/r/c/419441/ ?
2018-03-28 15:00:26 <Nudin_WMDE> (i can't change the topic) eeting starting now in channel #wikimedia-tech, hosts: @amir1 & @Lucas_WMDE - all questions welcome, more infos: https://www.mediawiki.org/wiki/Technical_Advice_IRC_Meeting
2018-03-28 15:00:38 <Amir1> o/
2018-03-28 15:01:53 <Lucas_WMDE> o/
2018-03-28 15:02:07 <sjoerddebruin> o/
2018-03-28 15:02:39 <Nudin_WMDE> sry, part of my message was eaten by the unstable terminal I'm using atm
2018-03-28 15:03:38 <Csisc> Hello, I am Houcemeddine Turki, member of Wikimedia TN User Group. I have many questions. However, let me begin the most relevant one. My first question is how to benefit from Wikidata entities to enrich the newly created lexicographic data.
2018-03-28 15:05:20 <Csisc> In fact, Wikidata entities (Q) can be used as senses (S) and labels can be used as lexemes (L).
2018-03-28 15:06:33 <Csisc> Hello, I am Houcemeddine Turki, member of Wikimedia TN User Group. I have many questions. However, let me begin the most relevant one. My first question is how to benefit from Wikidata entities to enrich the newly created lexicographic data.
2018-03-28 15:06:34 <Lucas_WMDE> Csisc: Hi! I’ll try to answer your question, though I’m not sure if I’m the best person to talk to…
2018-03-28 15:06:51 <Lucas_WMDE> but while there is a relationship between items and senses, they’re not the same thing
2018-03-28 15:07:15 <Lucas_WMDE> it certainly makes sense (no pun intended) to link a Sense to an Item, via a statement (Senses can have statements)
2018-03-28 15:07:32 <Lucas_WMDE> but they’re not identical
2018-03-28 15:07:59 <Lucas_WMDE> it would be wasteful to create a separate item for all the subtle differences between senses of lexemes
2018-03-28 15:08:44 <Csisc> I see Mr. Lucas. Is there any tool that can be created to automate the process of enriching Lexicographic data from Wikidata entities
2018-03-28 15:09:30 <Lucas_WMDE> I’m not sure what you mean
2018-03-28 15:09:37 <Lucas_WMDE> but in general, I would be surprised if anything existed already
2018-03-28 15:09:48 <Lucas_WMDE> I expect the community will start working on tools once we’ve launched lexeme support :)
2018-03-28 15:11:47 <Csisc> I see. I am asking about it as I added several common nouns in Arabic dialects as labels to Wikidata entities and I would like to have them automatically implemented to Lexicographic data.
2018-03-28 15:12:58 <Csisc> However, we can talk about that after April.
2018-03-28 15:13:29 <Lucas_WMDE> I don’t think that will be possible
2018-03-28 15:13:37 <Lucas_WMDE> loads of Wikidata items have labels that aren’t lexemes
2018-03-28 15:14:05 <sjoerddebruin> How much nouns are we even talking about...
2018-03-28 15:14:27 <Csisc> 500 words
2018-03-28 15:14:47 <Csisc> in four Arabic dialects...
2018-03-28 15:15:30 <Csisc> But, for languages like Chinese or Japanese, they are around 10000 words
2018-03-28 15:15:43 <sjoerddebruin> And that is currently in Wikidata? Weird.
2018-03-28 15:16:20 <sjoerddebruin> I don't think the lexiographical data should be one-to-one copies of what is available on Wikidata.
2018-03-28 15:18:49 <Csisc> Yes, I agree. I think of creating a tool that can be used by the Wikimedia community of a given language to verify if labels are lexemes or not. If they are lexemes, they are automatically added to Lexicographic data. A kind of Wikidata game... Is it possible to create.
2018-03-28 15:20:21 <Amir1> This discussion, albeit being very useful, is outside of scope of the TAIM as it's not technical. I recommend asking this question in #wikidata, or wikidata-l or in Wikidata office hours
2018-03-28 15:20:48 <Amir1> that way, you will get better answers
2018-03-28 15:20:48 <sjoerddebruin> Indeed. ^
2018-03-28 15:21:43 <Csisc> I see. I will ask Wikidata team about that even if I thought that the creation of such a tool is a technical issue.
2018-03-28 15:22:18 <Lucas_WMDE> well, there are a lot of things to discuss before you get to the technical parts of creating the tool
2018-03-28 15:22:29 <Csisc> My next question. I have created
2018-03-28 15:23:46 <Csisc> I see Mr. Lucas. So, I will return to discuss this issue with you once I finish discussing it with Wikidata team.
2018-03-28 15:25:10 <Csisc> My next question is that I created an algorithm that generates the IPA of a one-word lexeme in Tunisian. I ask if it is possible to implement it to Lexicographic data.
2018-03-28 15:27:18 <Csisc> The algorithm can be used with minor revisions to Arabic.
2018-03-28 15:28:00 <Lucas_WMDE> in general, I think you could write a bot to add pronunciation statements to the relevant forms
2018-03-28 15:28:10 <Lucas_WMDE> (instead of integrating this into the WikibaseLexeme extension directly)
2018-03-28 15:28:25 <Lucas_WMDE> that said… I’m very sceptical whether automatically generating IPA is possible in any language
2018-03-28 15:28:37 <Lucas_WMDE> unless it’s a language with a highly regular pronunciation
2018-03-28 15:29:50 <Csisc> Mr. Lucas, you are right. For languages like French, it is not practical
2018-03-28 15:30:10 <Amir1> As a Persian language speaker, since Arabic script doesn't write vowels I'm sceptical of that too. e.g. "بست" can be either "Bast", "Best", "Bost", "Bosot", etc.
2018-03-28 15:31:17 <Csisc> You are right. That is why I will add diacritics to lexemes before using the tool
2018-03-28 15:31:47 <Csisc> This will solve the problem.
2018-03-28 15:32:19 <Csisc> Arabic diacritizers are performant nowadays...
2018-03-28 15:32:48 <Lucas_WMDE> in general, I think you would need to discuss this with the community
2018-03-28 15:32:56 <Lucas_WMDE> during the approval process for the bot, I guess
2018-03-28 15:33:18 <Lucas_WMDE> I’m definitely not the right person to decide whether this is a good idea or not :)
2018-03-28 15:33:21 <Csisc> https://en.m.wikipedia.org/wiki/Arabic_diacritics
2018-03-28 15:33:53 <Csisc> I agree with you. We should discuss that with the community.
2018-03-28 15:34:30 <Csisc> However, I have created an IPA generator for Italian and I can prove that it works well.
2018-03-28 15:35:21 <Lucas_WMDE> does anyone else have technical questions, by the way? just checking :)
2018-03-28 15:37:40 <Csisc> You have said me that I can make a bot for my algorithms. I ask if it is possible to convert my Python code into a bot.
2018-03-28 15:38:26 <Amir1> Csisc: If it's an algorithm, by definition it's possible
2018-03-28 15:38:55 <Csisc> How...
2018-03-28 15:39:40 <thedj> well most wikibots are written in python, so that should already help :)_
2018-03-28 15:40:15 <Amir1> This is the framework you can use https://www.mediawiki.org/wiki/Manual:Pywikibot
2018-03-28 15:40:32 <Amir1> specially: https://www.mediawiki.org/wiki/Manual:Pywikibot/Wikidata
2018-03-28 15:40:37 <masti> Csicshttps://github.com/wikimedia/pywikibot
2018-03-28 15:41:37 <Csisc> Thank you. Is there a query limit when using Pywikibot.
2018-03-28 15:42:22 <Csisc> I mean. Is there a number of edits by minutes that should not be exceeded.
2018-03-28 15:43:03 <Amir1> Yes, it automatically handles that
2018-03-28 15:43:17 <Amir1> it sleeps between each edit
2018-03-28 15:43:43 <Csisc> That is absolutely excellent. Thanks.
2018-03-28 15:45:03 <thedj> Csisc: not that most wiki's have bot policies however for making edits.. Not sure what the plan is exactly, but you might have to check the community of the specific wiki and get their buy-in as well.
2018-03-28 15:45:09 <Csisc> I am sorry if I disturb.
2018-03-28 15:45:29 <thedj> that's what this hour is for.
2018-03-28 15:45:48 <Csisc> Thanks. I will absolutely do that
2018-03-28 15:48:59 <Csisc> My next question is not very technical. But, I have to ask it. My question is can I use Medline APIs and Pywikibot to automatically generate Medical information in Wikidata
2018-03-28 15:50:01 <Csisc> For example, adding all PubMed articles to Wikidata, finding the risk factors of all diseases...
2018-03-28 15:50:52 <Lucas_WMDE> I think that’s a question for WikiProject Medicine
2018-03-28 15:51:06 <Csisc> I ask if this is technically practical and if it would not cause licenses matters.
2018-03-28 15:51:44 <Csisc> Mr. Lucas, I will ask Wiki Project Med about that.
2018-03-28 15:56:20 <Csisc> I just have a last question. As well, it is not technical. I am doing with other scientists from the University of Sfax some computer science researches about Wikimedia projects. We are asking if it is possible for us to host Wikimedia Hackaton in Sfax so that our WikiResearch community can improve their computing skills. Learn how to use Pywikibot, QuickStatements, APIs...
2018-03-28 15:56:43 <Csisc> I mean Wikimedia Hackathon
2018-03-28 15:58:48 <Csisc> I think that this will be an excellent push for our Wikitech research community.
2018-03-28 16:00:00 <Csisc> I mean... What is the procedure for applying to host Wikimedia Hackathon.
2018-03-28 16:00:46 <Lucas_WMDE> I have no idea, sorry
2018-03-28 16:02:53 <Lucas_WMDE> I don’t even know who the right person to ask this would be
2018-03-28 16:03:00 <Lucas_WMDE> I’m afraid that question is too non-technical for me :D
2018-03-28 16:03:11 <Csisc> I see. I will try to contact the organizers of Barcelone's Hackathon to see what they have done to host the Hackthon
2018-03-28 16:04:15 <Csisc> Thank you. I am sorry for disturbing you this evening. As you have seen, many of my questions were not technical.
2018-03-28 16:04:15 <Amir1> Csisc: You should contact Rachel Farrand from WMF and she can point you out to the right person
2018-03-28 16:04:33 <Lucas_WMDE> anyways – the time is up, that was the Technical Advice IRC Meeting :)
2018-03-28 16:04:38 <Lucas_WMDE> thanks to everyone who participated!
2018-03-28 16:04:46 <Csisc> Thank you. I will do that.
2018-03-28 16:04:53 <Csisc> Good Bye.
2018-03-28 16:14:17 <greg__> I'm wondering about suggestions for the best way to load sample content from one wiki into an empty wiki
2018-03-28 16:14:54 <greg__> Say I'm doing a migration, and I don't have access to the database directly, nor the filesystem for images, would the API be the best solution?
2018-03-28 16:14:59 <MatmaRex> greg__: probably easiest to use Special:Export and Special:Import
2018-03-28 16:15:26 <MatmaRex> the API would be easier if you need to automate this, of course
2018-03-28 16:15:27 <greg__> or, would Special:Export and Special:Import work for a limited subset of content?
2018-03-28 16:15:51 <MatmaRex> it has some options to define what you want to export. e.g. by category
2018-03-28 16:16:01 <greg__> right
2018-03-28 16:16:02 <MatmaRex> there's a help page for it somewhere
2018-03-28 16:16:05 <MatmaRex> !import
2018-03-28 16:16:37 <MatmaRex> hmph, i always forget the bot is not in this channel. :)
2018-03-28 16:16:43 <MatmaRex> [18:16] <wm-bot> To import a few pages, use Special:Import - you can also import pages directly from another wiki (see !importsources). For mass imports, use importDump.php - see <http://www.mediawiki.org/wiki/Manual:Importing_XML_dumps>; for details. NOTE: when using content from another wiki, follow the LICENSE TERMS, especially, attribute source and authors!
2018-03-28 16:16:48 <MatmaRex> [18:16] <wm-bot> To export pages from a wiki, navigate to Special:Export on the wiki, type in the names of the pages to export, and hit "export". See <http://en.wikipedia.org/wiki/Special:Export>; for an example of this form. See also: !import
2018-03-28 16:17:32 <greg__> But, what about images included in the page(s)? I don't think there's a way to get those via Special:Export
2018-03-28 16:18:04 <MatmaRex> greg__: i think there are no good export options for images, if you don't have FTP access to the server (or similar) :/
2018-03-28 16:18:41 <MatmaRex> you could probably fairly easily write a tool to use the API to list them and download them, and upload to the other wiki. but i don't think such a tool already exists
2018-03-28 16:19:12 <MatmaRex> if you do have FTP access, you can get the whole images/ directory, and use maintenance/importImages.php on the other wiki
2018-03-28 16:19:30 <greg__> That's why I was thinking of doing it programatically via the API.. at least I could query and find images and then fetch them individually if needed (or "export" from the file: namespace)
2018-03-28 16:20:25 <greg__> MatmaRex: yeah, I guess that was my real question: Is there a tool for this?
2018-03-28 16:21:49 <MatmaRex> not that i know of
2018-03-28 16:21:53 <greg__> I offer wiki hosting, and would like to be able to offer a "preview" or "demo" site and/or simple migration strategy without direct FTP/SSH access to the origin wiki
2018-03-28 16:23:00 <greg__> Of course, final migration is dependent on LocalSettings.php and so it's never possible to just clone a wiki from afar
2018-03-28 16:26:44 <greg__> MatmaRex: thanks for the help
2018-03-28 16:33:18 <greg__> Anyone ever pull images OUT of Amazon S3 into a traditional wiki /images/ layout?
2018-03-28 16:33:51 <greg__> I know that there are cli tools for interacting with S3, so I can hack something together
2018-03-28 16:46:17 <d3r1ck> Sorry for hopping in late
2018-03-28 16:46:23 <d3r1ck> I've got some question(s)
2018-03-28 16:47:06 <d3r1ck> Is it possible I edit an entry of this page via MW api?
2018-03-28 16:47:23 <d3r1ck> https://meta.wikimedia.org/wiki/Connect/Individuals
2018-03-28 16:47:46 <d3r1ck> The page has wikitext (consisting of a template per entry)
2018-03-28 16:48:02 <d3r1ck> Wish to know if it's possible to edit just a particular entry via the MW api?
2018-03-28 16:48:29 <d3r1ck> I've been trying to see how possible it is to convert a MW template to an Object in JS but no luck so far
2018-03-28 16:59:01 <d3r1ck> Ivy: ^
2018-03-28 17:06:11 <MatmaRex> d3r1ck: not really. you'll probably have to parse and edit the wikitext
2018-03-28 17:06:47 <d3r1ck> MatmaRex: Thanks and is there any library to do that, I'm aware of something that does a similar issue but on Lua tables instead
2018-03-28 17:07:00 <d3r1ck> Was looking for something to enable me do same for Wikitext
2018-03-28 17:07:46 <MatmaRex> d3r1ck: i don't know of any, but it might exist
2018-03-28 17:08:12 <MatmaRex> d3r1ck: alternatively, you could try using Parsoid to convert it to HTML document, edit that HTML document, then use Parsoid to convert it to wikitext. but that might be more complicated than you need
2018-03-28 17:08:17 <d3r1ck> Okay, I'll keep digging.
2018-03-28 17:08:40 <d3r1ck> Yes, that's pretty complicated
2018-03-28 17:08:52 <d3r1ck> But I'll keep digging, thanks so much for the heads up MatmaRex
2018-03-28 17:08:53 <MatmaRex> d3r1ck: here's a script that does that, you might be able to steal some ideas: https://en.wikipedia.org/wiki/User:Jackmcbarn/editProtectedHelper.js
2018-03-28 17:09:01 <MatmaRex> it actually also edits some template parameters, i think
2018-03-28 17:09:14 <d3r1ck> Yeha, thanks
2018-03-28 17:09:23 <d3r1ck> yeah, I'm gonna steal some ideas :D

This page is generated from SQL logs, you can also download static txt files from here