Wikidata:Bot requests

From Wikidata
Jump to: navigation, search

Bot requests
If you have a bot request, create a new section here and tell exactly what you want. You should discuss your request first and wait for the decision of the community. Please refer to previous discussion. If you want to request sitelink moves, see list of delinkers.

For botflag requests, see Wikidata:Requests for permissions.

Add a new request
On this page, old discussions are archived after 2 days, if they are marked with {{Section resolved}}. An overview of all archives can be found at this page's archive index. The current archive is located at April.


Import geographic coordinates from the French Wikipedia[edit]

Some bots import geographic coordinates from various Wikipedias. But I've seen no one which do it from the French Wikipedia. It would be useful for many French places which haven't an article in other languages. Thanks in advance. Ayack (talk) 14:51, 15 July 2013 (UTC)

Anyone? Ayack (talk) 15:19, 28 September 2013 (UTC)

Will import article coords as part of this task. --WikedKentaur (talk) 17:14, 28 September 2013 (UTC)

Great, thanks! Ayack (talk) 17:20, 28 September 2013 (UTC)

Partly done -- script has problems with utf encoding. --WikedKentaur (talk) 18:10, 24 November 2013 (UTC)

The bot seems blocked. Is anybody importing coordinates from Wikipedias?--Pere prlpz (talk) 15:33, 23 December 2013 (UTC)

 :-( — Ayack (talk) 19:19, 26 January 2014 (UTC)

Remove obsolete imported from (P143)->Wikipedia-based sources[edit]

imported from (P143) is important to show from which Wikipedia a claim has been imported from. This allows to find the source of errors (e.g., Wikidata:Database reports/Constraint violations). Importing basic claims from Wikipedia is in turn important as an empty item has no information to identify itself. However, we all agree that Wikipedia is no "real" source to be used in Wikipedia and elsewhere again. Thus, all sources imported from (P143)->Wikipedia where the claim is referenced by any other third party source should be removed. Ideally, in soon future, no imported from (P143)->Wikipedia remain. This work should be performed continuously and, thus, by a bot living at Labs. Manual example edit: [1].  — Felix Reimann (talk) 08:09, 6 August 2013 (UTC)

  • Symbol support vote.svg Support Emw (talk) 02:46, 8 August 2013 (UTC)
  • No. I think it is still useful to indicate where it came from. --  Docu  at 11:44, 10 August 2013 (UTC)
  • Symbol oppose vote.svg Oppose per Docu. --Ricordisamoa 00:11, 26 August 2013 (UTC)
  • Symbol support vote.svg Support It makes no sense to keep because it is an artifact of automated tasks - which bot was more active, such language is in front. Because the same statement can (and almost always) exist in multiple Wikipedias, emphasizing of particular language is not correct. --Infovarius (talk) 18:29, 6 November 2013 (UTC)
  • Symbol support vote.svg Support Most of these „references“ are sensless or misleading or have now a „real“ source. --Succu (talk) 21:24, 6 November 2013 (UTC)
  • Pictogram voting comment.svg Comment I am not going to vote on this but I am going to say again here what I have already said on the project chat: Wikipedians are going to measure us by how many statements are sourced to Wikipedia. There are stats for exactly this here: The trust in our data is one of the biggest issues we face and "imported from Wikipedia" is one of the biggest obstacles in gaining more trust. --Lydia Pintscher (WMDE) (talk) 21:36, 6 November 2013 (UTC)
  • Oppose. It should be clear where a fact is coming from, regardless of whether the source is reliable or is an "artifact" or otherwise (per en:WP:SAYWHEREYOUGOTIT). The only way to indicate that is the use of the "source", and in fact that's the best place for that information to reside.

    OTOH, I would certainly agree that a bot that detects multiple sources for a claim, at least one of which is not imported from, then the imported from claim should be removed. --Izno (talk) 23:24, 6 November 2013 (UTC)

    I agree. --Yair rand (talk) 08:57, 11 February 2014 (UTC)
    Switch to support, because Felix was good enough to inform me that I apparently can't read. --Izno (talk) 14:44, 11 February 2014 (UTC)
  • Pictogram voting comment.svg Comment One possibility to make everybody happy is to retain the historical information (which is useful for error correction among other things) but clearly distinguish it from real sources. Pichpich (talk) 18:50, 7 November 2013 (UTC)
  • Symbol oppose vote.svg Oppose If it is meant to remove the source AND the claim the result will be an almost blank Wikidata. Also Symbol oppose vote.svg Oppose if it is meant to remove only the source but leave the claim without source it is very difficult to find the source (if any) form that wikipedia. HenkvD (talk) 18:11, 9 November 2013 (UTC)
I misinterpreted the request. Agreed to remove P143 if an other valid source is present. HenkvD (talk) 19:20, 13 February 2014 (UTC)
  • Symbol oppose vote.svg Emot If a bot removes P143, when they add a "real" source, I agree. But just remove them en masse, would harm the traceability of discovered errors. I think the bots have to show not only which project they take the information from, but also which page and version. A url-property can be used together with P143 for that purpose. -- Lavallen (talk) 19:14, 9 November 2013 (UTC)
  • Symbol oppose vote.svg Oppose --Kizar (talk) 12:04, 23 November 2013 (UTC)
  • Symbol oppose vote.svg Oppose. --Yair rand (talk) 08:57, 11 February 2014 (UTC)
  • Pictogram voting comment.svg Comment Yair rand, Lavallen, Izno: this is exactly what I tried to propose: In case of multiple sources for one specific statement, where at least one of the sources is based on a source according to Help:Sources (i.e., with stated in (P248)), any additional imported from (P143) (i.e., pointing to where to statement came initially) should be removed. See my example: The statement had two sources, one based on P143, one based on P248 and thus, P143 can be safely removed in this case.  — Felix Reimann (talk) 09:59, 11 February 2014 (UTC)
    Oh, I'm dumb. Good call. I'll switch me-self. I'll ping HenkvD and Ricordisamoa as well, because they also may have misunderstood. --Izno (talk) 14:44, 11 February 2014 (UTC)
    I agree, that is what I tried to say above, but it was maybe Lost in Translation (Q107270). -- Lavallen (talk) 17:28, 11 February 2014 (UTC)
  • GA candidate.svg Conditional support on a case-by-case basis, provided that the bot will be explicitly designed to replace imported from (P143) with Wikipedia-independent sources of which it can reasonably ensure the verifiability. There should be specifical tasks dealing with certain property sets (e.g. SuccuBot could do that for IUCN conservation status (P141)). --Ricordisamoa 00:58, 12 February 2014 (UTC)
  • Symbol support vote.svg Support. @Yair rand, Kizar: your oppose seems to be based on a misunderstanding of this proposal. Could you either elaborate your on your opinion or strike your oppose ? --Zolo (talk) 09:43, 13 April 2014 (UTC)

Add descriptions[edit]

In this days I'm working on item of Category. Someone ask to me to add descriptions also for other languages. So if user add in this table correct description and also if is necessary replace if already exist description, I can do it. I can use descriptions in MediaWiki:Gadget-autoEdit.js, but not all are correct (Wikipedia/Wikimedia) and there aren't indication about replace or not. --ValterVB (talk) 19:50, 29 October 2013 (UTC)

Shouldn't it just be the same text as in Wikimedia category page (Q4167836), Wikimedia disambiguation page (Q4167410) and Wikimedia template (Q11266439)? --тнояsтеn 20:46, 29 October 2013 (UTC)
I guess it would ideally be the same, but by default the label of Wikimedia category page (Q4167836) is equal to the sitelink, and in many languages we still have things like "Wikipedia:Disambiguation" instead of the more elegant "Wikipedia disambiguation page".
I think the auto-edit descriptions are ok except that in the "wikipedia" has been to to "wikimedia category" at some points and some languages have not yet been updated. --Zolo (talk) 22:35, 29 October 2013 (UTC)

Shouldn't we then also update pages like Help:Description#Non-article_items and Help:Description/de?--Zuphilip (talk) 14:16, 1 November 2013 (UTC)

Maybe objects with type Wikimedia list article (Q13406463) could be handled similarly? --Zuphilip (talk) 18:37, 2 November 2013 (UTC)

The French translations are wrong: Wikimedia must NOT be translated to Wikimédia (corrected in the table below). See w:fr:Wikimedia. -- Bjung (talk) 22:02, 27 December 2013 (UTC)

OK none oppose so I start to replace "Wikimédia" with "Wikimedia" --ValterVB (talk) 10:44, 11 January 2014 (UTC)

Lang description Replace(Y/N) Sign
it categoria di un progetto Wikimedia Y ValterVB (talk)
fr page de catégorie d'un projet Wikimedia Y Zolo (talk)
de Wikimedia-Kategorie Y тнояsтеn
pt categoria de um projeto da Wikimedia Y ValterVB (talk)
pt-br categoria de um projeto da Wikimedia Y ValterVB (talk)
ru категория в проекте Викимедия Y Infovarius (talk)
sv kategorisida 20:59, 6 November 2013 (UTC)
eo Vikimedia-kategorio לערי ריינהארט (talk)
ro categorie pe paginile Wikimedia לערי ריינהארט (talk)

Next task:

Lang description Replace(Y/N) Sign
it pagina di disambiguazione Y ValterVB (talk)
fr page d'homonymie d'un projet Wikimedia Y Zolo (talk)
de Wikimedia-Begriffsklärungsseite Y тнояsтеn
pt página de desambiguação de um projeto da Wikimedia MisterSanderson (talk)
pt-br página de desambiguação de um projeto da Wikimedia MisterSanderson (talk)
sv grensida 20:59, 6 November 2013 (UTC)
eo apartigilo לערי ריינהארט (talk)
ro pagină de dezambiguizare לערי ריינהארט (talk)
Lang description Replace(Y/N) Sign
it template di un progetto Wikimedia Y ValterVB (talk)
fr modèle d'un projet Wikimedia Y Zolo (talk)
de Wikimedia-Vorlage Y тнояsтеn
pt predefinição de um projeto da Wikimedia MisterSanderson (talk)
pt-br predefinição de um projeto da Wikimedia MisterSanderson (talk)
ru шаблон проекта Викимедия Y Infovarius (talk)
eo Vikimedia-ŝablono לערי ריינהארט (talk)
ro format pe paginile Wikimedia לערי ריינהארט (talk)

disambiguation pages with additional statements[edit]

Hi! re: Wikimedia disambiguation page (Q4167410) I noticed that more and more disambiguation pages contain additional statements. Please do not add the descriptions in this case. Please let us know what pages are affected (please let us know an url). לערי ריינהארט (talk) 08:36, 26 March 2014 (UTC)

disambiguation pages causing constraint violations[edit]

Hi! Archetype (Q346973) caused a constraint violation. See . I removed the relation see .
Please verify if "special:WhatLinksHere" links to a relevant Wikidata object ("Qxxxx") and generate a report on the identified pages. Thanks in advance! לערי ריינהארט (talk) 08:55, 3 April 2014 (UTC)

Create items for Wikipedia articles[edit]

Many Wikipedia articles are currently not linked to Wikidata, so new items should be created. Some of them may be duplicates, but that seems to be a small minority, and they are easier to spot if they are in Wikidata anyway. --Zolo (talk) 08:55, 11 December 2013 (UTC)

is there a specific Wikipedia where more items are missing then in other Wikipedias? --Bene* talk 14:48, 11 December 2013 (UTC)
Some time ago many articles with sv:Template:Insjöfakta Sverige did not have any item here. -- Lavallen (talk) 15:24, 11 December 2013 (UTC)
The English Wikipedia, obviously.--Ymblanter (talk) 20:13, 11 December 2013 (UTC)
@Zolo:: You stated: „Some of them may be duplicates, but that seems to be a small minority, and they are easier to spot if they are in Wikidata anyway“ (my emphasis). Is this knowledge or only a guess? How useful are new created items without a single (useable) statement? --Succu (talk) 20:39, 11 December 2013 (UTC)
For the languages I have checked in the topics I have browsed: more missing items in Chinese and English than in French and German.
@Succu: yes, this is only a guess based on a non-random experience. To state things more rigorously, I have seen rather many Wikipedia articles that did not have any link to Wikipedia, and almost all of those I checked could not be linked to any existing item. Once it is in Wikidata, an item gets a chance to get statements. Actually, what prompted me to make this request is user:Magnus Manske's new "Widar" tool that allows to add statements based on Wikipedia category but leaves out articles that are not in Wikipedia (of course in this case, another solution would be to add an option for creating items through the tool ;). --Zolo (talk) 21:47, 11 December 2013 (UTC)
Short-term, I can offer Swedish people on en.wp without Wikidata item. --Magnus Manske (talk) 22:45, 11 December 2013 (UTC)
A lot of new items were created just after the start of Wikidata, but afterwards we didn't setup anything to keep importing new articles.
I wrote added a bot to Pywikibot to make it easy to create new items. I do a query to find articles without an item and use this as input for the bot.
The bot doesn't touch an article if it's created to recently (default: 3 weeks) and if it has been edit recently (default: 7 days). This prevents new items from being created when people might be still working on it.
I'm thinking about setting up a shared account (on Toollabs) with this bot in it. It could work on every Wikipedia, but we should probably start with a smaller subset.
What do you think? Is this a good idea? Who wants to help to setup this bot and maintain it? Multichill (talk) 15:58, 8 March 2014 (UTC)

Improve Wikidata:Country subdivision task force/United States[edit]

That is:

  1. Create items of political divisions of the United States if there're no article in Wikipedia.
  2. Add country (P17), contains administrative territorial entity (P150), is in the administrative-territorial entity (P131), coordinate location (P625), type of administrative entity (P132)/instance of (P31), and one of FIPS 55-3 (locations in the US) (P774), FIPS 6-4 (US counties) (P882), FIPS 5-2 (code for US states) (P883), FIPS 10-4 (countries and regions) (P901) to political divisions of the United States.

Currently, FIPS 55-3 (locations in the US) (P774), FIPS 6-4 (US counties) (P882), FIPS 5-2 (code for US states) (P883), FIPS 10-4 (countries and regions) (P901) are used in less than 150 items, and Q816 even have no contains administrative territorial entity (P150) claims! Please make a bot to do them. The number of political divisions of the United States is: County (6,372), City or Town (147,410), Congressional District (7,619), Census Tract (170,359), Metro/Micro Area (1,215), ZIP Code/ZCTA (166,081), Other (175,807). So it's a big work to do.--GZWDer (talk) 05:46, 12 December 2013 (UTC)

I will try to help the taskforce by manual adding/preparing as well as some automatic tasks (a bot for everything is maybe a little too much to ask for). First, I will concentrate on the county level. Some questions and comments:
--Zuphilip (talk) 14:18, 15 December 2013 (UTC)
Update: FIPS 5-2 (code for US states) (P883) ✓ Done, FIPS 6-4 (US counties) (P882) ✓ Done, FIPS 10-4 (countries and regions) (P901) ✓ Done --Zuphilip (talk) 20:27, 22 January 2014 (UTC)

Replace imported from (P143) to stated in (P248) for some value[edit]

Such as:

  1. JPL Small-Body Database (Q4026990)
  2. Virtual International Authority File (Q54919)
  3. International Standard Name Identifier (Q423048)
  4. IUCN Red List (Q32059)
  5. SIMBAD (Q654724)
  6. Minor Planet Center (Q522039)
  7. INSEE code (Q156705)
  8. OKATO (Q856636)
  9. World Odonata List (Q13561342)
  10. Jonathan's Space Report (Q6272367)
  11. National Central Library of Florence (Q460907)
  12. Istituto Nazionale di Statistica (Q214195)
  13. International Plant Names Index (Q922063)
  14. VizieR (Q1662358)

and so on. They are trusted database, so we should use stated in (P248) instead of imported from (P143).--GZWDer (talk) 15:45, 21 December 2013 (UTC)

imported from (P143) already has several uses: After importing authortiy numbers from any Wikipedia the respective Wikipedia is attributed as source. And after bulk import of VIAF numbers and freebase Id's based on external mappings the authority file or concrete dump was given as source. Personally however I use stated in (P248) on individual items when I have actively checked the validity and appropriateness of the authority number in its respective database. And especially for VIAF imho stated in (P248) does not make much sense, since (similar to wikidata Q-numbers) the numbers stand for clusters of "real" entries reshuffled by automatic processes on a regular schedule (these numbers are quite stable of course, but in a sense VIAF does not make any semantic statements of its own, it just refres to its constituents). I know here is not the proper place to discuss source qualifiers for authority control properties, I only intend to utter that things may be tricky and "trust" has many aspects). -- Gymel (talk) 12:04, 11 January 2014 (UTC)

Consistency of Commons links[edit]

Regrettably, the ill-conceived and incomplete solution of the relation between Wikidata and Commons caused doubleness of the links to Commons. Wikidata have properties Commons category (P373) and Commons gallery (P935) as well as "Wikimedia Commons page linked to this item" section in interwiki links but none of the possibilities is fully-fledged and fully functional yet. Until any definitive arrangement is applied, we need to keep consistency between both types of links.

The interwiki link in the section "Wikimedia Commons page linked to this item" should be imported, exported or compared in relation to properties linking to Commons.

  • If the iw link contains a category page of Commons, the link should be automatically rendered (copied) to P373 property (Commons category). Possible collision should be reported as possible error.
  • If the iw link contains a gallery page of Commons, the link should be automatically rendered (copied) to P935 property (Commons gallery). Possible collision should be reported as possible error.
  • P373 and P935 properties should be rendered (exported) as interwiki links linking from the Wikidata item to Commons - however, the priority must be specified for the cases that both these properties are filled and for such cases that more Wikidata items link (through these properties) to identic Commons page. --ŠJů (talk) 22:38, 22 December 2013 (UTC)
report for first item - sourcecode. --Steenth (talk) 15:52, 5 March 2014 (UTC)

Sports Statistics Update Bot[edit]

My idea is a bot that automatically updates sports statistics. Preferably, it would get statistics from the assorted [sport] sites (e.g. etc.). It would pull information from an athletes' page every so often and update that athlete's Wikipedia article.

Thanks! Newyorkadam (talk) 22:39, 30 December 2013 (UTC)Newyorkadam

By the way, I was forwarded from's Bot Request page.

I would want a hell of a lot more information about how this would work before I even considered letting Wikidata have any control over another project's content. Resolute (talk) 18:33, 1 January 2014 (UTC)
What do you mean how it would work? It would pull data from those websites and update athletes' articles. Newyorkadam (talk) 18:43, 1 January 2014 (UTC)Newyorkadam
Hi Newyorkadam, thanks for this interesting idea. I have got two questions:
  1. Are those webpages reliable? Where does the data come from?
  2. Under which license is the data published? Just noting that Wikidata's content is licensed under the CC0.
Best regards, -- Bene* talk 21:34, 2 January 2014 (UTC)
Bene - under US law, bare statistics are not copyrightable, so there shouldn't be an issue there. (England is apparently a more problematic though). I'm not sure of any issue that might relate to more advanced statistics (of which Baseball is legendary for, but also which I doubt Newyorkadam is looking to add).
Newyorkadam - Sorry, I meant to ask how Wikidata would work vis a vis the articles themselves. The point of a site like Wikidata is to centralize certain types of information across all wikis. Statistics could be one aspect, but my belief is that it would ultimately mean the loss of control over the data, styles, formatting, etc. by local wikis. That concerns me a little. Resolute (talk) 21:47, 2 January 2014 (UTC)
No, formatting is up to Wikis (with the only current restriction that a Wiki page can only import info from the Wikidata item corresponding to this page), and the wikis need to opt in anyway, so if they want to keep the statistics locally they are free to do it. The problem is of course that if some data is vandalized it shows up incorrectly at all wikis which use it, but this is a general problem for centralized storage of info.--Ymblanter (talk) 07:05, 3 January 2014 (UTC)
──────────────────────────────────────────────────────────────────────────────────────────────────── @Resolute: however, if you read [2] you can see that they don't allow this use of data. Quoting:

Please do not attempt to spider data from our web sites, as spidering violates the terms and conditions that govern your use of our web sites: Site Terms of Use
If we notice excessive activity from a particular IP address we will be forced to take appropriate measures, which will include, but not be limited to, blocking that IP address. We thank you in advance for respecting our terms of use.

Regards Bene* talk 09:58, 3 January 2014 (UTC)
Fair enough. If they don't want you to take the data from their site, that is obviously their right. The bare data itself is not copyrightable, however. And thank you for the response, Ymblanter. Resolute (talk) 22:40, 3 January 2014 (UTC)
We can always use another website like ESPN for statistics (if it isn't copyrighted too). Newyorkadam (talk) 01:03, 4 January 2014 (UTC)Newyorkadam
Probably we can't: «Any business use, "re-mailing" or high-volume or automated use of WDIG Sites is prohibited» from Terms of use --ValterVB (talk) 08:12, 4 January 2014 (UTC)
I haven't looked at the data, but [3] looks interesting. Legoktm (talk) 08:31, 4 January 2014 (UTC)
I'm not expert about license. Is compatible?: «This database is copyright 1996-2014 by Sean Lahman.» «This work is licensed under a Creative Commons Attribution-ShareAlike 3.0 Unported License. For details see: » --ValterVB (talk) 09:08, 4 January 2014 (UTC)
The alike-part seems problematic. We could copy the content to wikidata and redistribute it, but the licence must stay the same. Since the content in wikidata is CC0, I guess that this is not possible. --Zuphilip (talk) 10:48, 4 January 2014 (UTC)

Fix invalid dates[edit]

Some dates were entered in a wrong format with one leading "0" missing (for instance here). According to Lydia Pintscher (WMDE) (talkcontribslogs), it should be a fixed bug, but current values still need to be fixed, can a bot do it ? --Zolo (talk) 08:20, 10 January 2014 (UTC)

@Zolo: Pywikibot can (apparently) fix such values automatically (example) when deserializing claims and serializing them back, or it could be a WikibaseRepo feature. --Ricordisamoa 22:43, 12 April 2014 (UTC)

Replace values[edit]

Hi, I need a bot to remove any instances of member of political party (P102)  independent politician (Q327591) and replace it with member of political party (P102)  no value. This is to clean up the constraints page for member of political party (P102) and remove the classification of independent politician (Q327591) as an instance of a political party. This was discussed at User talk:SPQRobin#Independent politicians. --Wylve (talk) 15:22, 11 January 2014 (UTC)


Last year someone ran a bot to keep the statistics of tennis players up to date. At the moment, this is primarily the win/loose stats for both single and double tournaments. This data can be derived from either the WTA/ATP/ITF websites, where this is kept up to date. The data is live used on some languages (dk/es/ru, from the top of my head). Once we can use numbers on WikiData, I think there will be more properties connected with this (making auto-update even more interesting). Edoderoo (talk) 10:22, 17 January 2014 (UTC)

I would like to add that the German Wikipedia Portal:Tennis is right now also looking into possibilities of using Wikidata in infoboxes and that this is the main issue to solve before we can proceed. I think that this is true for all languages. As far as we see, the update has to be run once a week (Monday) when the new world rankings are published.

In addition, there are sometimes differences between ITF on one side and WTA/ATP on the other side and it is our understanding that ATP/WTA data should be used. It would be great if someone could set up a bot or even rerun the once existing bot.

Please ping me on German Wikipedia if there are any updates as I am not a regular visitor on Wikidata. --Mad melone (talk) 13:09, 18 March 2014 (UTC)

The main problem will be this: Terms of Use Agreement. According to this it is not allowed to copy data from the ATP website and put them in our database. --Pasleim (talk) 18:37, 18 March 2014 (UTC)
@Pasleim: Thanks a lot for your input. Not surpisingly, the WTA has a similar segment in their TUA Section 5, however the wording is different.
However, isn't this basically what we do today in the local wikipedias already? As far as i see it, we go to the two websites every monday and insert the data into the local infobox templates, thereby "storing" them as well. My understanding is that we can do that due to the fact that we cite the respective source, even though the content is copyrighted (on a side note: a huge portion of what we cite as sources is copyright protected, e.g. newspaper articles, books, etc, so that shouldn't be the issue at hand.). Only difference is that we would use a crawler bot.
That being said, I am definitely not an expert on the question of what we can do and what we can't do within the legal framework that the Foundation has set up for Wikimedia projects. In my personal opinion, we should be able to do so, but it surely isn't a bad idea to involve people more knowledgeable in this area. Is there a special place on Wikidata for that? --Mad melone (talk) 07:19, 19 March 2014 (UTC)
The key difference is that Wikipedia includes prose that should discuss the statistics, which means fair-use is a defense. However infoboxes (on English Wikipedia and others, but maybe not all Wikipedias) often contain data which isnt critically discussed in the prose, which means it isnt immune from a copyright battle, but the 'owners' of this data probably dont want to sue Wikipedia. However if Wikidata offered a high quality replica of their data, able to be easily extracted via our APIs, we need to be extremely careful about how we obtained that data to avoid violating w:database rights. John Vandenberg (talk) 08:32, 19 March 2014 (UTC)
Thanks for your input as well. Even though this isn't in our favor, after doing some reading I agree with you. Wikimedia has always been very tight on copyright issues - and rightfully so, as we have to protect our users that rely on the free use of our output. This is true even though most of the media or other people using copyrighted data use it on a "they won't sue me" basis, which we can't afford to use. I thought about contacting WTA/ATP in this regard, but I want to let sleeping dogs lie.
Nonetheless, this is an unfourtunate situation, as it dooes not only affect tennis, but more or less all infoboxes with regularly changing data. Probably it's best to get a broad concensus on this one, but I am no expert on WIkidata and would rely on someone else to initiate such discussion.--Mad melone (talk) 10:28, 19 March 2014 (UTC)
As iknow there is a big problem to crawl data automaticaly if there is a data copyright on the crawled websites. In the business there was always a problem to get data auto via an automatized process. It is no problem to take a secretary to type thousands of data published in publicy. But it is a big problem to crawl them. Imho we should use iuristical help in this case because lawyers have in most cases the sight of both parties, the right owners and the users. So maybe there is a lawyer in the WP or WM to look on this case? --DonPedro71 (talk) 11:53, 19 March 2014 (UTC)

Lydia has provided me with a link to the Foundation's legal team analysis of the situation: It basically backs up what was mentioned earlier, so at this time being I see no chance for crawling websites :( --Mad melone (talk) 16:38, 21 March 2014 (UTC)

Fix my mistakes ![edit]

I have added p31: craft (Q2207288) to several items through Widar, but actually I was misled by the French label, it seems that the correct item was profession (Q28640). Can someone correct that ? I do not think it has to be restricted may edits, as it seems that we should always use craft (Q2207288) in p31 to keep some consistency. --Zolo (talk) 09:19, 18 January 2014 (UTC)

Done.--GZWDer (talk) 14:02, 12 April 2014 (UTC)
I think that this discussion is resolved and can be archived. If you disagree, don't hesitate to replace this template with your comment. GZWDer (talk) 14:02, 12 April 2014 (UTC)

Wikisource categories link suggestion[edit]

Some Wikisource projects are not best example of inter-projects linking :-)

I'd like to automate next suggestions for Wikisource categories:

  • List of categories in Wikipedia and Wikisource with same name. Please note that year categories have different meaning: Wikisource ones are about works created in particular year.
  • Author categories. For example in Russian Wikisource author pages and categories have same name. Category item could be found via page item's Property:P910.

I'd like to highlight suggestions.

EugeneZelenko (talk) 15:01, 18 January 2014 (UTC)

Looks like these author-categories can be linked to Commons, who have many categories about authors. -- Lavallen (talk) 16:00, 18 January 2014 (UTC)
EugeneZelenko, I agree it would be nice to interlink such categories automatically. You could also enable mw:Extension:Wikibase_Client#Other_projects_sidebar on ru.s. ;) --Nemo 09:22, 14 April 2014 (UTC)

Import release dates for media entries[edit]

Of the thousands of media works (music albums, movies, etc) present on Wikidata, there are but under two hundred entires that have a supplied date of publication (P577), which I think is very valuable and relevant data.

For all instances of album and film, I propose we import the value for "release date" from infoboxes on the English Wikipedia, if available. However, per this discussion there may be concerns regarding reliability or suitable sourcing. I personally feel querying Wikipedia will give us at least some data with which can be improved if need be over time. Wikipedia data is well-monitored and hence arguably fairly accurate. Alternatively we could import the data from Rovi (they also have data on video), widely considered a reliable source, but as you could imagine, using their API is not free. I also wonder if there could be potential legal implications given how we're using their data. I cannot think of another source that is as easily accessible. That being said, I think Wikipedia is still the best option. Note you'd have to parse the infoboxes to get the actual release date, as formats may vary, or use dbpedia, which provides standardized data formats imported from Wikipedia.

Thanks! — MusikAnimal talk 23:19, 19 January 2014 (UTC)

Removing aliases[edit]

Please remove all aliases "Lepilemuridae" from genus item as (no label) (Q3830736) is family of primates now. Infovarius (talk) 15:02, 20 January 2014 (UTC)

The same task is for family Salmonidae: to remove Salmoniformes from all languages aliases, because order Salmoniformes is Q9394365. Infovarius (talk) 20:35, 25 January 2014 (UTC)

Import virus identifiers[edit]

There is a list of the codes following this link:

The third column of the table has the name of the virus. First one for example is Poxviridae Poxviridae (Q948421). For this item get the string from the second column (in this case 00.058) and set it as ICTV virus ID (P1076).

For sourcing I would use: url = author = Cornelia Büchen-Osmond title = Taxonomy and Classification of Viruses

Because the table is sometimes updated it would also be good to set "date retrieved". It might also be good to make an item for that page/table, so we can change the url in one place if it ever changes. In that case I guess we should use "stated in" and that item. --Tobias1984 (talk) 10:28, 24 January 2014 (UTC)

Add brothers and sisters[edit]

When two persons have the same father (P22) and the same mother (P25), add them respectively as brother (P7) (or sister (P9) according to their sex). Thanks. — Ayack (talk) 19:26, 26 January 2014 (UTC)

This can be deduced.. There is no need. GerardM (talk) 08:22, 1 February 2014 (UTC)
GerardM: Sorry but I don't understand your point. Are you saying that the brother (P7) and sister (P9) properties have no use as they can be deduced? If so, I don't agree because the father or the mother are not always known, which means you can't deduce the brothers and sisters. — Ayack (talk) 15:54, 1 February 2014 (UTC)
There is this property for "family relations" that one can be used I think :) GerardM (talk) 08:34, 2 February 2014 (UTC)

Global Economic Map Bot[edit]

The Global Economic Map Bot will be the primary bot to update the Global Economic Map project.

It will retrieve data from World Bank Indicators, UN Statistics, International Labor Organization, Bureau of Economic Analysis, Gapminder World, OpenCorporates and OpenSpending. The data retrieved will automatically update Wikidata with economic statistics and it will also update the Global Economic Map project.

Mcnabber091 (talk) 21:31, 26 January 2014 (UTC)

I can do that Amir (talk) 21:32, 26 January 2014 (UTC)

RFC - DBpedia import process[edit]

This conversation is now taking place at Wikidata:Requests_for_comment/DBpedia_import_process.

I have been to the DBpedia conference in Amsterdam. One frequent topic was how DBpedia and Wikidata could cooperate / coexist and how DBpedia can use its processes of its own update for Wikidata as well. What we came up with is more or less along these lines:

  • The first step is data collection
  • The second step is match the results to Wikidata and DBpedia.
    • produce a lists with differences where Wikidata and the new data differ.
    • update the new information to Wikidata / DBpedia simultaneously
  • Ask the communities to reconcile the differences between the source Wikipedia and Wikidata

The other thing that is really relevant is that many people at the conferences represented GLAM partners of our movement. There were people from Europeana, the Dutch National Library among others. They are quite interested to run pilot projects on subjects like "Dutch authors" and "Dutch monuments". These would then be run along the lines as described above.

In one meeting I mentioned the Commons/Wikidata integration. I explained why the use of the "Creator" and "Institution" templates are a good step in preparation of this integration. I urged them to keep as much meta data as possible after an initial import so that we may make use of it when the Commons/Wikidata integration is a reality.

This is an RFC. My question is to get comments on the proposed process of importing data as harvested by the DBpedia processes (including the reporting on differences). Thanks, GerardM (talk) 09:00, 1 February 2014 (UTC)

Did you discuss about the model/property mappings ? It seems important. I know there is already work done on this, but we need to precise that. TomT0m (talk) 13:36, 1 February 2014 (UTC)
For them the way we have our "model" is/will be very much a challenge on so many levels. They have a big challenge to continuously keep a mapping to what our model happens to be. We do have qualifiers and processes to add but particularly remove properties. They already have a Wikidata project within Wikidata so they are aware of such issues. Thanks, GerardM (talk)

One very important thing to mention is DBpedia's Mapping Wiki. There they maintain their ontology as well as the mappings of templates and template parameters on different Wikipedias. They already cross-reference Wikidata. This is an extremely valuable resource for anyone wanting to map between DBpedia and Wikidata, as well as anyone wanting to import data from Wikipedia templates. It might even be possible to write a generic import bot that relies solely on the mappings defined on the mappings wiki. -- Duesentrieb (talk) 13:37, 3 February 2014 (UTC)

GerardM, please create an RFC in Wikidata:Requests_for_comment. I also recommend having the RFC put into the 'Selected proposals and requests for comment' notice at the top of 'Watchlist'. Thanks, Emw (talk) 13:59, 3 February 2014 (UTC)

I created an RFC.. there is probably some additional MediaWiki magic needed for it to be visible. GerardM (talk) 15:05, 5 February 2014 (UTC)
There is no magic, it has to be manually add it to the list of open RfCs. I'll do it. TomT0m (talk) 17:16, 5 February 2014 (UTC)

High volume: NGC labels[edit]

There are a huge amount of items about astronomical objects that don't have labels in English and German and have no statements. An example would be NGC 6909 (Q1148443). Can somebody scan those infoboxes (ru and nl probably have a lot of transclusions) and add the English and German label. German with "-" please. E.g.: NGC-6909 @Paperoastro: Is there a property that a bot could set and readily source? --Tobias1984 (talk) 12:20, 6 February 2014 (UTC)

(Just curious): Why "German with "-" please""? As far as I can see in de:Kategorie:New General Catalogue, neither the German Wikipedia articles nor their Wikidata items use this format currently. --YMS (talk) 16:49, 6 February 2014 (UTC)
I am not sure why German-Wikipdia is not using the "-". It is usually considered bad form in German to separate dual names. But some people remember their grammar de:NGC-507-Gruppe. I just saw that you are native German. Do you think we should go for correct grammar or treat the catalogue number as special? --Tobias1984 (talk) 17:12, 6 February 2014 (UTC)
I don't think writing NGC-123 would be correct grammar. It's a name and a number, no need for hyphenation (cf. street numbers, movie titles, etc.). The case of NGC-507-Gruppe is a different one, it's a case for (no label) (Q1267038), as two things "NGC 507" and "Gruppe" get connected together, and so all spaces get replaced by dashes). --YMS (talk) 17:29, 6 February 2014 (UTC)
Oh that's right, my example is a case of Durchkopplung. I am not so sure about the analogy to street numbers because those are usually words and not strings like "NGC". But I can live without the dash in this case. --Tobias1984 (talk) 17:59, 6 February 2014 (UTC)
Using alias, for me there isn't any problem to use NGC<space>number and NGC<dash>number if used in a specific language. The right nomenclature is here. --Paperoastro (talk) 19:35, 6 February 2014 (UTC)

Persons and it.wikipedia[edit]

In this report there are persons with no claim sex[569]_AND_noclaim[21] many of them has it.wikipedia link. it.wikipedia is the only wiki that has got a template for all persons. It contain name, last name,sex etc.. we can inport all missing information from, at the same time check that there is instance of (P31)->human (Q5) and remove DEPRECATED main type (GND) (P107) if present. The best choice is scan [4] and use this category for alla data @ValterVB:@Ricordisamoa:--Rippitippi (talk) 00:34, 8 February 2014 (UTC)

For whoever wants to take it, the job has already been discussed previously and got consensus, all necessary properties created. The "translation" of the data to import is described at User_talk:Legobot/properties.js#Italian_person_data. --Nemo 02:03, 8 February 2014 (UTC)

UN security council resolutions[edit]

For all pages with a link to a Wikipedia article in en:Category:United Nations Security Council resolutions (depth = 1 as they are categorized by year). We could have:

  • description:
  • ca: Resolució del Consell de Seguretat de les Nacions Unides
  • de: Resolution des UN-Sicherheitsrats
  • en: United Nationa Security Council resolution
  • es: Resolución del Consejo de Seguridad de las Naciones Unidas
  • fr: résolution du Conseil de sécurité des Nations Unies
  • zh: 联合国安全理事会决议
  • The label can also be automatically created but it is a bit more difficult. In French that would be:
  • en sitelink = "United Nations Security Council Resolution XX" -=> fr label: "résolution XX du Conseil de sécurité des Nations unies" --Zolo (talk) 18:59, 8 February 2014 (UTC)
I'm working on it. --ValterVB (talk) 09:34, 9 February 2014 (UTC)
✓ Done can be useful add property preceded by (P155) and succeeded by (P156)?
Thanks. Yes, I think it would be useful as a qualifier of instance of (P31) but I have seen that some people object to adding qualifiers to p31. --Zolo (talk) 11:58, 9 February 2014 (UTC)
Qualifiers will not be queryable in the foreseeable future. Furthermore, qualifiers on P31 (i.e., rdf:type) are not well-defined and so likely will not be included in any RDF export. I recommend using 'preceded by' and 'succeeded by' as top-level properties here. Emw (talk) 15:50, 9 February 2014 (UTC)

Can someone please rerun the bot with the correct English spelling (United Nations is misspelt)? --—Wylve (talk) 10:47, 17 March 2014 (UTC)

Clean up sources for Property:P21[edit]

Will be good idea to remove from Property:P21 sources specified by Property:P143 as Italian Wikipedia or Swedish Wikipedia, if Property:P143 was also imported from Virtual International Authority File or patronymic. --EugeneZelenko (talk) 15:16, 15 February 2014 (UTC)

@EugeneZelenko: This request looks simlair to "#Remove obsolete etc" above? -- Lavallen (talk) 19:16, 15 February 2014 (UTC)

Add qualifiers[edit]


Could someone add burial plot reference (P965) as a qualifier of place of burial (P119) : Père Lachaise Cemetery (Q311) ? E.g.

You can use categories of the French Wikipedia (fr:Catégorie:Personnalité enterrée au Père-Lachaise par division) or Wikimedia Commons (commons:Category:Grave of Claude Chabrol.

After that, I'll add sources. Pyb (talk) 13:36, 16 February 2014 (UTC)

Redirects for alternate page titles of translatable project pages[edit]

It would be helpful if there was a bot that went through pages translatable through Translate that have localized titles and made redirects out of those titles where possible. For example, Wikidata:Indsatsgrupper could be created to redirect to Wikidata:Task forces/da. --Yair rand (talk) 07:29, 25 February 2014 (UTC)

Symbol oppose vote.svg Oppose – the search function should be smart enough to show the translated page first. If that doesn't happen, then there is a problem with Search. --Ricordisamoa 17:24, 13 March 2014 (UTC)

Remove claim Property:P31: Q387917[edit]

The claim instance of (P31): administrative territorial entity of Germany (Q387917) is added by user User_talk:Androoox#German_city to cities/municipalities while the item administrative territorial entity of Germany (Q387917) should only be added to subclasses of administrative entities. Please can a bot remove these claims. Michiel1972 (talk) 11:21, 26 February 2014 (UTC)

Fix coordinates[edit]

I think this request has already be made, but it appears that many invalid items remain. There were apparently errors in the handling of the globe parameter of the geocoordinate type and many items miss it. It can be mostly fixed by assuming that no globe = earth, but that is not really good (actually they show as errors in edit summaries like here). The problem can be quickly fixed by hand through doing a dummy edit like adding a white space at the end, but there are really many items, so we would need a bot to fix it. --Zolo (talk) 08:10, 28 February 2014 (UTC)

PS: a new version of module:Coordinates with a smarter use of Wikidata is ready to be deployed in fr.wikipedia but it shows a n error message when Wikidata data are invalid (I think it is a good thing, but only once this bug is fixed). --Zolo (talk) 16:30, 28 February 2014 (UTC)
@Zolo: So to fix this could simple be to null edit all items with coordinate location (P625)?  Hazard SJ  19:10, 1 March 2014 (UTC)
@Hazard-SJ:. Not really. Apparently, the missing "globe = Earth" gets automatically added whenever you edit coordinates through the UI but you have to edit the p625, it doesn't work if you edit something else in the item. And I doubt it works the same way when editing through the API. --Zolo (talk) 20:25, 1 March 2014 (UTC)
addendum: it's not just the globe. Precision is also nil on some items. Unless I am mistaken, this is also a bug. --Zolo (talk) 22:32, 4 March 2014 (UTC)
addendum2: Also, it seems that the precision parameter wzs set to the highest possible level in many items where it should not have been. For instance coordinates of the form latitude = Z.66666666667, longitude = Z.66666667 (sort of like here) should almost certainly have a precision to an arcminute. That should be fixed, otherwise, some Wikipedia display weird things like 45° 00' 00.0" while it should only be "45°". --Zolo (talk) 19:48, 6 March 2014 (UTC)
See Wikidata talk:Coordinates tracking#Comparing coordinates and Bugzilla62105. Multichill (talk) 11:32, 8 March 2014 (UTC)

Remove sources for categories, templates, portal, etc items[edit]

I think will be good idea to remove source for Property:P31 values of category, template, portal, etc. Of course will be good idea to check that all links are from particular namespace. --EugeneZelenko (talk) 15:02, 6 March 2014 (UTC)

This week end my bot will work on category to fix an old error, If no one oppose I can delete instance of (P31) source of instance of (P31) on categories. I can do it also for Template and Portal. --ValterVB (talk) 19:33, 6 March 2014 (UTC)
Please not that I asked to delete source, not property value itself. --EugeneZelenko (talk) 15:19, 7 March 2014 (UTC)
Of course I mean delete source of instance of (P31) :) --ValterVB (talk) 16:01, 7 March 2014 (UTC)
Why do you think this would be a good idea?
First I get shouted at for not including the source and now you want to remove it? No, let's not do this. @ValterVB: please stop doing this. Multichill (talk) 11:42, 8 March 2014 (UTC)
@Multichill: For now I remove only a wrong source: example (my old error). If there isn't consensus, I don't delete source. --ValterVB (talk) 12:29, 8 March 2014 (UTC)
Support deleting. We don't need sources for categories, templates and project pages etc. It is clear that they've been imported from Wikipedia, so what's the sense of adding a source to say this? --Stryn (talk) 12:23, 8 March 2014 (UTC)
Symbol oppose vote oversat.svg Strong oppose, sources are useful to detect errors. for example, If a bot add p31=disambigion page to an item which is not linked to a disambigion page in enwiki, it's useful to detect this page in which language is a disambigion. And it's also useful to detect this problem (now cleaned).--GZWDer (talk) 12:30, 8 March 2014 (UTC)
We need dedicated check for disambiguation status in projects directly. Source will not help. --EugeneZelenko (talk) 15:50, 8 March 2014 (UTC)

WD items for categories on rowiki[edit]

It's necesary to change items for categories of stadiums for rowiki, like i've done here:

  • Categorie:Stadioane în *country* → Categorie:Stadioane din *country*
  • Categorie:Stadioane de fotbal în *country* → Categorie:Stadioane de fotbal din *country*

We have an consensus somewhere on local village pump. I already recategorised all pages to correct categories. Its necessary only WD items to adjust here. Thanks you. XXN (talk) 19:46, 7 March 2014 (UTC)

@XXN: done also some unconnected ones in [5]. Not sure if it's enough. --Ricordisamoa 17:28, 8 April 2014 (UTC)
Nice job! Thank you! --XXN (talk) 21:25, 10 April 2014 (UTC)

I think that this discussion is resolved and can be archived. If you disagree, don't hesitate to replace this template with your comment. 10 April 2014

Bot request of create items and add claims of an item[edit]

I want to create some items and add claims of the item exiting in wikidata from the artcles and infobox of wikipedia  – The preceding unsigned comment was added by Qianwei0823 (talk • contribs).

@Qianwei0823: please clarify from what articles the bot should import claims. --Ricordisamoa 18:06, 13 March 2014 (UTC)

Add some information of Jiangsu Province in China  – The preceding unsigned comment was added by Qianwei0823 (talk • contribs).

Huge work: from type of astronomical object (P60) to instance of (P31)[edit]

Hi, while the decision concerning the deletion of type of astronomical object (P60) is pending (see this discussion), I open a request for a bot that copy P60 statements to P31.

More precisely, for every item with P60 property, the bot should copy the value in an instance of P31, with also the source. In some cases the statements was already copied, but not the source, that should be copied. Probably exist items with both properties but with different values! In this case I hope it is possible make a list to check manually (!!) the items. Note that the bot will check nearly 70000 items (see Database_reports/Popular_properties). Thank you very much! --Paperoastro (talk) 10:17, 17 March 2014 (UTC)

Time2wait.svg On hold per PFD. --Ricordisamoa 00:10, 20 March 2014 (UTC)

Add source to the statement instance of (P31) in items of constellations[edit]

I think that this discussion is resolved and can be archived. If you disagree, don't hesitate to replace this template with your comment. Hi, I'd like to add a source to the statement instance of (P31) in items of constellation. All these items should have already instance of (P31) -> constellation (Q8928). The source I need is:
P854 ->
P364 -> Q1860
P357 -> The Constellations
P813 -> today
P123 -> Q6867

You can see the result in the item Boötes (Q8667).

The items involved are 88: Q9256, Q10481, Q10506, Q10576, Q10586, Q9253, Q10584, Q10476, Q10488, Q8832, Q8849, Q10452, Q10538, Q9305, Q10535, Q10470, Q10464, Q8844, Q10468, Q8839, Q10457, Q10508, Q10425, Q9285, Q10413, Q10406, Q10517, Q9282, Q10542, Q8921, Q9302, Q8837, Q8675, Q10438, Q10433, Q8913, Q8923, Q10563, Q10448, Q10574, Q10578, Q10416, Q10450, Q10430, Q8853, Q10403, Q10446, Q10580, Q10571, Q10443, Q10484, Q9289, Q10492, Q10428, Q10435, Q10582, Q10503, Q8906, Q8860, Q10515, Q8864, Q10511, Q10441, Q10486, Q8679, Q10409, Q9251, Q10473, Q10498, Q10513, Q8866, Q8865, Q9286, Q10529, Q8910, Q10525, Q10570, Q10546, Q10565, Q10422, Q10567, Q8918, Q10478, Q10521, Q8842, Q10437, Q10519.

I excluded from this list Q8667, already done. Thank you very much. --Paperoastro (talk) 11:36, 27 March 2014 (UTC)

@Paperoastro: ✓ Done 87 edits. --Ricordisamoa 18:10, 28 March 2014 (UTC)
Thanks! :-) --Paperoastro (talk) 21:29, 29 March 2014 (UTC)

academic degree (P512)[edit]

Descriptive conventions of the Integrated Authority File (Q36578) includes the academic degree. Can someone write a bot to import this data? Example: Christiaan Hendrik Persoon (Q346451), GND identifier (P227): 116091509. Academic title: Dr., Sources: Online GND and DNB portal. BTW: We could also import the field "Occupation", but this would be a bit more work. --Kolja21 (talk) 17:29, 27 March 2014 (UTC)

Unfortunately, the academic degree isn't part of the rdf files, Christian Hendrik Persoon. And it is quite annoying to parse the html if there is a rdf. Maybe I will do it if Wikidata:Requests_for_permissions/Bot/PLbot_4 gets accepted. --Pasleim (talk) 13:04, 7 April 2014 (UTC)

Population of French communes[edit]

@Lavallen, ValterVB, JulesWinnfield-hu: we need to get it right, but it is a bit tricky. So a lengthy post about what might be done:

get the data

at least two solutions:

  1. download the excel file with the 2010 census data ([6]) and go to the "Communes" sheet. Population is splitted by sex and age, so you need to add a column with the sum of columns D to W (line 10 should be "784").
  2. Use the commune-specific pages ([7]). The bad thing is that it requires some parsing of the html code. The good thing is that it can be easily extended to all years. "Population municipale" is the most important figure to have but it might be interesting to have the "population totale" too.
map onto Wikidata

if you change the format in the Excel file to see more decimal after the unit you will see a suprising thing. Population of L'Abergement-Clémenciat is 784,00 but the male population less than 3 year old is "17.087179". This seems to be because total population is extrapolated and rounded, and only then distributed among sex and age group. That would entail that the precision should not be set to 0 (it should be at least 1, but probably more, I don't know exactly).--Zolo (talk) 08:10, 30 March 2014 (UTC)

@Zolo: I'm working on it with xls file. Step 1: I prepare the mapping list with INSEE municipality code (P374). Precision: I round to integer so precision must be +- 1. Qualifier:point in time (P585)= 2010 Isn't to much generic? I use criterion used (P1013) = (no label) (Q15715409). Source= is correct? --ValterVB (talk) 11:17, 30 March 2014 (UTC)
If they haven't specified more than "2010", it's enough to have precision=year. -- Lavallen (talk) 11:37, 30 March 2014 (UTC)
@ValterVB:. Thanks. Sorry, actually, I have just found a slightly better file at It can be used essentially the same way except that the total population is directly given in column I (no need to sum columns).
It is true true that having only the year seems a bit vague when population is precise to +/-1 but I as far as I can tell it is the only info provided.
For the rest, I think it is current, except that we should probably have a few additional things in the source, but I am not sure how to do it with an excel file, perhaps use the title given in and do it this way ? --Zolo (talk) 15:36, 30 March 2014 (UTC)
Regarding the diff, shouldn't it be "publisher:INSEE (Q156616)" and not "author"? -- Lavallen (talk) 15:48, 30 March 2014 (UTC)
We should really have all of the available data (including gender ratios, population by age, etc.) available in the items, but there isn't any clean way to do that yet. I recommend holding off on adding this data until Bug 62555 is resolved. --Yair rand (talk) 04:18, 31 March 2014 (UTC)
@Yair rand:: Yes, it seems a bit realy for the detailed data, but I think we can already do the total population (this is widely used in Wikipedias, and we would really benefit from having it here). Beside, in this case, the population by age is somehow estimated based on the total population rather than the other way around. --Zolo (talk) 07:13, 31 March 2014 (UTC)
Either way, I imagine there's probably some complex data we would want to import. If, in the end, the figure for total population is not going to be included as an independent separate simple statement, we would have to do a whole series of edits removing the statements added now if we don't wait until a full solution is available. --Yair rand (talk) 08:44, 31 March 2014 (UTC)
We do not know yet when (or even if) bug 62555 will be solved. -- Lavallen (talk) 09:54, 31 March 2014 (UTC)
That is correct. I assume, though, that if the bug is not going to be solved, the Wikidata team will tell us at some point. (For the record, I Symbol oppose vote.svg Oppose adding the data before there's a resolution.) --Yair rand (talk) 21:36, 31 March 2014 (UTC)
Adding all range of data for age, sex etc, does neither I support at the moment. But I support these two numbers. -- Lavallen (talk) 01:16, 1 April 2014 (UTC)
If this bug is to be solved one day, I am afraid it is not going to be anytime soon. We still do not have all the datatypes that were supposed to be released during the first year of development... --Zolo (talk) 21:24, 9 April 2014 (UTC)
@Zolo: I have finished mapping. There are 165 municipalities without data (list here). If you confirm that your diff is correct, I can start. --ValterVB (talk) 18:22, 31 March 2014 (UTC)
@ValterVB: I think it is correct but I have asked for feedback on fr.wikipedia. I have just added date of publication (P577) [8].
Missing data appear to be former communes that have been abolished. --Zolo (talk) 18:59, 31 March 2014 (UTC)
OK, ping me when ready. --ValterVB (talk) 19:58, 31 March 2014 (UTC)
@Zolo: Use the web page scheme to source your file: URL, original language, title, point in time, date of publication, author or publisher but in that case better author because INSEE did the data collection and the calculations. Snipre (talk) 15:20, 1 April 2014 (UTC)
@Snipre:. That seems to be what I have done. My main question is actually about what title we should use. It seems that the best file to use is actually . "Populations légales des communes associées ou déléguées en vigueur au 1er janvier 2014" sounds good, but it is only the name of one sheet in an unnamed workbook. --Zolo (talk) 18:25, 1 April 2014 (UTC)
I would say that is better to link a page with the file link like [9] and indicate somehow the source, rather than the file itself. --JulesWinnfield-hu (talk) 19:13, 1 April 2014 (UTC)
@Zolo: And what's about the license of the data ? Is the copyright compatible with the CC0 license ? This is always the question of the " la réutilisation est toutefois subordonnée au respect de l'intégrité de l'information et des données et à la mention précise des sources" which is not totally correct with the CC0 license as reutilization of data from WD is not following that rule. Snipre (talk) 19:44, 9 April 2014 (UTC)
I think that "subordonnée au respect de l'intégrité de l'information" means that we cannot attribute to the INSEE claims they did not make, which seems ok. "et à la mention précise des sources." is more problematic, I think the issue should rather be discussed at WD:PC#Modifying license ?. --Zolo (talk) 21:24, 9 April 2014 (UTC)

Updating category and template labels[edit]

Is there a bot operator whose bot would repeatedly process the following task?

When a category/template link is moved, the label should also be moved (i.e. deleted and recreated elsewhere). This isn't done often, though. I want the bot to (1) delete labels which are/should be placed elsewhere (regardless capitalisation of the first letter or the letter after the first colon) and (2) update labels of items which are linked with categories with changed titles. Matěj Suchánek (talk) 12:28, 30 March 2014 (UTC)

Add IMSLP ID property[edit]

Dear all,

In Wikpedia EN, there is a nice template w:Template:IMSLP

That is used

{ { IMSLP|id=Dvořák, Antonín|cname=Antonín Dvořák } }


{ { IMSLP|Dvořák, Antonín } }

in the corresponding WIKIDATA article it should become

IMSLP ID (P839) : Dvořák, Antonín

And it's done !


(by User:Xmlizer, 30. März 2014, 22:53 Uhr)

Note: IMSLP ID (P839), import "Category:Dvořák, Antonín" etc. en:Antonín Dvořák for example is outdated. (No bot import in this case possible.) --Kolja21 (talk) 00:02, 31 March 2014 (UTC)
Indeed, it was a very bad example. please use all the other Xmlizer (talk) 22:27, 1 April 2014 (UTC)

Migrating links from urwiki to wikidata[edit]

Hi all, Is there anyone who migrate the interwiki links of categories and articles to wikidata from urdu wiki? I have lot of cats and arts, Thanks. --Muhammad Shuaib (talk) 22:40, 1 April 2014 (UTC)

@محمد شعیب: It's not a bot, but you can try this tool. — Ayack (talk) 09:39, 3 April 2014 (UTC)

Openpolis work[edit]

Hi, I need a bot work in order to add a new property:

  1. parse all pages about Italian politicians on in order to find the XXX code in {{Openpolis|id=XXX}} or in {{Openpolis|XXX}} (read the manpage!)
  2. check that Openpolis ID (P1229) is not present yet
  3. add Openpolis ID (P1229) to related pages and add XXX as value
  4. check that reference URL (P854) doesn't contain link to Openpolis, otherwise remove it
  5. all pages about Italian repubblic politicians that doesn't have template Openpolis should be listed in order to check them manually (ping me in order to get the list)

Thank you very much! --★ → Airon 90 10:20, 4 April 2014 (UTC)

I didn't get point #4. Can you give an example? --Pasleim (talk) 18:26, 7 April 2014 (UTC)
✓ Done --ValterVB (talk) 09:50, 13 April 2014 (UTC)

MEP directory identifier (P1186)[edit]

Please import data from de:Vorlage:MdEP and en:Template:MEP. --Kolja21 (talk) 03:10, 7 April 2014 (UTC)

Proposal to import data for the Internet Broadway Database from English Wikipedia[edit]

Wikidata has four properties for different types of entries in the Internet Broadway Database (Q31964), but very few items link to them. Perhaps the following can be imported from templates on English Wikipedia?

--Bensin (talk) 23:43, 9 April 2014 (UTC)

Harvest defunct data from Wikipedia[edit]

Hello, over at Wikipedia Wikiproject Video games, we need to clean up our article code and remove a hell of a lot of redundant data.

Since the project started, a number of template fields have been added and removed from the main article template. We now have 15 defunct fields that still appear in the code for some article pages, and we're now in the position where all of this old data is getting in the way and making things confusing for new users (They copy over template code from existing articles only to find that some template fields aren't working after they have populated them with data.)

Initially we were just going to delete the data, but a request to save it for Wikidata was made, which is why I'm here.

In order to aid all users (especially new ones) in editing the infobox code, and at the same time preserve the data, we need a bot run by Wikidata to harvest and then remove the defunct data fields.

As we have over 11,000 articles that need this process carried out on them, a bot really is the only way of collecting this data.

We have a tracking category that lists every article that needs editing.

The discussions around this job are at the following:



I've also put the details in a table to make things easier to read.

Its not going to be easy as some of the fields contain user's own unique - and sometimes differing - formatting styles, but we know you'll find a way to cope with it. Hope you can help. - X201 (talk) 08:25, 10 April 2014 (UTC)

Sorry to press for a reply, but we need to start deleting these; probably next month. Could someone tell me if this can/will be done, or if collecting from the article edit history is just as easy for you? - X201 (talk) 13:07, 17 April 2014 (UTC)

IAAF ID (P1146)[edit]

Please import data from pl:Szablon:IAAF name: {{IAAF name|id=XXX}} or {{IAAF name|XXX}} with this category. Thanks, MATEUSZ.NS talk 07:00, 13 April 2014 (UTC)

Blocking open proxies[edit]

I checked some IPs of non-constructive edits. I found some potential open proxies: (talkcontribslogs), (talkcontribslogs), (talkcontribslogs). Maybe a lot of such IPs are open proxies, so should we make a bot to block open proxies automatically, or create a page like Open proxy detection? Note I'm not sure that any of them are really open proxies.--GZWDer (talk) 16:47, 16 April 2014 (UTC)

Such requests have been rejected by the community: Wikidata:Requests for permissions/Bot/HaroldBot and Wikidata:Requests for permissions/Administrator/HaroldBot. --Ricordisamoa 23:42, 16 April 2014 (UTC)
We don't seem to be having that big of an open proxy abuse problem, and even then I don't like how block bots tend to clutter the local block log.--Jasper Deng (talk) 23:58, 16 April 2014 (UTC)
@Ricordisamoa, Jasper Deng: Do you think something like Open proxy detection is needed?--GZWDer (talk) 04:56, 17 April 2014 (UTC)
I don't think it's bad enough to the point that we need such a centralized effort. I've blocked open proxies as I've found them, but manually and only after verification.--Jasper Deng (talk) 05:50, 17 April 2014 (UTC)

Cleanup in description[edit]

A lot of descriptions in German have a Wikipedia syntax like [[Tschechen|tschechischer]] Gelehrter, Kaplan, Dekan und Propst. This syntax has its origin in the PERSONA DATA feature of dewiki.
Some of them I've corrected already, but it's too much. Could one of the bot owners please do the corrections? The job is to strip the brackets and cut off the part before (and including) the pipe. -- Pütz M. (talk) 13:22, 17 April 2014 (UTC)

Wikidata:Requests for permissions/Bot/AkkakkBot 4 --Ricordisamoa 13:32, 17 April 2014 (UTC)