Wikidata:Bot requests

From Wikidata
Jump to: navigation, search

Bot requests
If you have a bot request, create a new section here and tell exactly what you want. You should discuss your request first and wait for the decision of the community. Please refer to previous discussion. If you want to request sitelink moves, see list of delinkers. For botflag requests, see Wikidata:Requests for permissions.
On this page, old discussions are archived after 2 days, if they are marked with {{Section resolved}}. An overview of all archives can be found at this page's archive index. The current archive is located at 2015/11.


Intel processors[edit]

Hello. Can some bot please read Intel's site and gather information from that database for the Wikidata? Its needed that it catches the name of the processors, create an item corresponding to each one and complete the item with the properties sockets supported (P1041), instruction set (P1068), manufacturer (P176) and number of processor cores (P1141).--MisterSanderson (talk) 15:58, 17 May 2014 (UTC)

Why don't you write them to ask that they release that data via a dump/machine readable API under a free license (or rather CC-0)? Even better, they could add it themselves here on Wikidata, to save us some work. --Nemo 17:16, 17 May 2014 (UTC)
I could not find an appropiate e-mail adress at, so there is no way to contact them.--MisterSanderson (talk) 18:45, 17 May 2014 (UTC)
Try any of the first four in "Intel PR Departments" [1] (calling yourself an analyst) and [2], you'll be fine. --Nemo 15:51, 23 May 2014 (UTC)
Ok, I sent them a message.--MisterSanderson (talk) 11:37, 25 May 2014 (UTC)
The contact was closed without response.--MisterSanderson (talk) 16:50, 29 May 2014 (UTC)
So the creation of the items needs to be made by Wikidata robots...--MisterSanderson (talk) 15:32, 4 July 2014 (UTC)

Today I found the link "Export Full Specifications" that generates a XML file with the data. I think this will turn easy to gather the information with bots.--MisterSanderson (talk) 15:06, 2 October 2014 (UTC)

Here, I even extracted myself manually the data and created a table: I think that now there is no excuse to not include these informations on Wikidata.--MisterSanderson (talk) 18:52, 3 October 2014 (UTC)

The table looks good. However, we can't yet add values with a dimension (e.g. Hz, MB, nm) so the only information we can now extract is the number of cores (number of processor cores (P1141). Are there already items on Wikidata about intel processors or should a new item be created for every row in the table? --Pasleim (talk) 19:15, 3 October 2014 (UTC)
Not only number of processor cores (P1141), there are other properties too: sockets supported (P1041), instruction set (P1068) and manufacturer (P176). I think that maybe there is a "release date" property too, but I could not find. And there is the subclass of (P279): all Celeron models are a subclass of the Celeron family. Some processors already have an item, but in Wikipedia is more common to create articles about a family of processors, not to individual models, so I think that each row must be an item.--MisterSanderson (talk) 22:39, 3 October 2014 (UTC)

New table: .--MisterSanderson (talk) 23:20, 6 December 2014 (UTC)

New table: .--MisterSanderson (talk) 17:02, 1 February 2015 (UTC)

New table: .--MisterSanderson (talk) 22:41, 1 February 2015 (UTC)

You can add the data by yourself using . However, it is still not possible to add columns 3-6 to Wikidata as there is no support for quantities with units. Adding sockets supported (P1041) and instruction set (P1068) can be interesting but I do not find these data on your page. --Pasleim (talk) 13:54, 3 February 2015 (UTC)
This tool only add statements to already existing items. But there are not items for all these processors. That's why I need that a robot creates them, I don't want to create manually, I already did my part of the job by creating the tables from the ARK. sockets supported (P1041) is the title of the posts, and instruction set (P1068) is not available for all the processors. There is too little information for processors released before 2006. --MisterSanderson (talk) 18:11, 2 March 2015 (UTC)

New table: .--MisterSanderson (talk) 18:11, 2 March 2015 (UTC)

New table: .--MisterSanderson (talk) 21:14, 2 April 2015 (UTC)

New table: .--MisterSanderson (talk) 00:35, 11 May 2015 (UTC)

New table: .--MisterSanderson (talk) 03:20, 15 June 2015 (UTC)

What's the point of saying that every month? Sjoerd de Bruin (talk) 07:29, 15 June 2015 (UTC)
I'm not saying the same thing each month, I'm notifying a new table every time.--MisterSanderson (talk) 00:19, 22 June 2015 (UTC)

New table: .--MisterSanderson (talk) 01:38, 12 July 2015 (UTC)

New table: .--MisterSanderson (talk) 02:27, 23 August 2015 (UTC)

New table: .--MisterSanderson (talk) 02:26, 31 October 2015 (UTC)

New table: .--MisterSanderson (talk) 01:35, 3 November 2015 (UTC)

New table: .--MisterSanderson (talk) 19:12, 3 November 2015 (UTC)

Importing Commons Categories (P373)[edit]

Dutch Wikipedia has a nl:Categorie:Wikipedia:Commonscat zonder link op Wikidata (Commonscat without equivalent in Wikidata). From there the P373 statement could be filled. -- Pütz M. (talk) 23:13, 13 June 2014 (UTC)

In en:Category:Commons category without a link on Wikidata is even more. -- Pütz M. (talk) 23:28, 13 June 2014 (UTC)
 Writing something up for this, I'll work on it as much as I can. George Edward CTalkContributions 19:25, 15 January 2015 (UTC)
@George.Edward.C: any progress in this or do you know how to do that? We would like to import commonscat links from fi-wiki to Wikidata. --Stryn (talk) 17:59, 21 January 2015 (UTC)
Doing the parser right now. Will work on it more at the weekend. If I can figure out the parser, it should be easy to complete the rest of it. When I've written it, and it works, I will run both NL and FI after that. George Edward CTalkContributions 18:05, 21 January 2015 (UTC)
Please don't run a bot on nl. There are a lot of conflicts in them, also a lot of pages with double templates. Sjoerd de Bruin (talk) 18:12, 21 January 2015 (UTC)
Noted. :-) The bot will only run on EnWiki and FiWiki (as long as there's no problems with either of them). (Edit: I will probably need a category similar to those mentioned in the reques') George Edward CTalkContributions 18:23, 21 January 2015 (UTC)
Been a while, but I've finally finished the code, and tested it with 4 edits (2 didn't work as planned, so I'm going to work on that, as it happens when Commonscat defaults to the pagename when no value is specified). Expect a request at RFP soon. --George (Talk · Contribs · CentralAuth · Log) 08:47, 20 February 2015 (UTC)
@Caliburn: Are still working on this task? --Pasleim (talk) 14:33, 26 July 2015 (UTC)
I've been working on other Internet projects recently, but I do plan on picking this back up again. I ran into a bug on User:Caliburn/, where it wouldn't properly detect Commons Categories that did not exist.  :) --George (Talk · Contribs · CentralAuth · Log) 16:21, 26 July 2015 (UTC)

Adding SOC job codes[edit]

I've noticed that SOC Occupation Code (2010) (P919) didn't have much data (, while a canonical list from the Bureau of Labor Statistics (thus PD and freely reusable, cf does exist (we're talking 7K items).

The data is tabulated as follows:

SOC Codes and Job Titles
2010 SOC Code 2010 SOC Title 2010 SOC Direct Match Title
11-1011 Chief Executives CEO
11-1011 Chief Executives Chief Executive Officer
11-1011 Chief Executives Chief Operating Officer
11-1011 Chief Executives Commissioner of Internal Revenue
11-1021 General and Operations Managers Department Store General Manager It is available at : (with more related files at

What I think could be nice are adding the SOC codes based on the 2010 SOC Direct Match Title since so many outside services and pages use the SOC codes (which seems to be a requirement for a lot of job offers in the US).

I've already added manually the French equivalent of the SOC code in Wikidata, so being able to match national codes and job titles through Wikidata would be cool.Teolemon (talk) 18:59, 29 November 2014 (UTC)

SOC is based on an UN standard, ISCO. Each country's bureau of national statistics have their own translation/adaptation of ISCO. There is a standard for historical occupations too, HISCO. In Denmark the national adaptation of this code system is called DISCO, for Norway STYRK and so on. Adding all these codes will be messy, since they may be non-overlapping. But it may also becoma a source for statisticians to get translations from one code to another. H@r@ld (talk) 23:12, 19 January 2015 (UTC)
I have created Wikidata:WikiProject_Occupations_and_professions. The idea is to have them all in Wikidata, so that we can display templates in all the Wikipedia, and get public information, both for statisticians and job seekers doing research. The end game would be to help them both be able to compare or research professions in any countries--Teolemon (talk) 21:17, 17 July 2015 (UTC)

Adding missing labels for all biological species, genus, families, etc[edit]

In a lot of languages the labels are missing, while a label can be added. This is especially easy for items with instance of (P31)  taxon (Q16521) + taxon name (P225) -> the scientific name that is added with P225, can be added to the labels of all languages with the Latin alphabet. If the label already has been given in a language, and the label differs from the Latin name, the taxon name (P225) should be added as synonym (if it is not already added). This should finding the right items make a lot easier.

This bot run should be done for all languages with Latin alphabet, and those include: en, de, fr, it, es, af, an, ast, bar, br, ca, co, cs, cy, da, de-at, de-ch, en-ca, en-gb, eo, et, eu, fi, frp, fur, fy, ga, gd, gl, gsw, hr, hu, ia, id, ie, is, io, kg, lb, li, lij, mg, min, ms, nap, nb, nds, nds-nl, nl, nn, nrm, oc, pcd, pl, pms, pt, pt-br, rm, ro, sc, scn, sco, sk, sl, sr-el, sv, sw, vec, vi, vls, vo, wa, wo, zu. Are there any languages missing where this should be done as well?

Thanks! Romaine (talk) 17:25, 13 May 2015 (UTC)

There is no consensus for this (see latest discussion). --Succu (talk) 17:48, 13 May 2015 (UTC)
What I found, is a discussion of exactly one year old, and just one person that is not supporting because of "the gadgets then need to load more data". Is that the same "no consensus" as you meant? Edoderoo (talk) 20:13, 13 May 2015 (UTC)
(after edit conflict)The discussion you link to links to another discussion in what only 1 person is objecting is because it would increase the size of items, and some tools can't handle such. This was and is silly reasoning, the adding will happen anyway, it only will take a bit longer. Now we have the problem that a lot of items can't be found because of missing labels. Is Wikidata intended to serve bots only, or can we humans get a better serving by having a label in my own language? Romaine (talk) 20:16, 13 May 2015 (UTC)
Almost all taxa have an english label. So searching for them you should have a hit. Adding these labels is not a big deal. My bot can easely do this. Simply get consensus for this. --Succu (talk) 20:33, 13 May 2015 (UTC)
In the search section on top of every page, I get to see only the items with a label in the language I have set in my preferences. So it happens a lot that I can't easily search and select. Secondly, to be able to link to an item, I need to have a label set in my language to be able to link it on other items. This is frustrating users to work easily with Wikidata. Romaine (talk) 20:45, 13 May 2015 (UTC)
Romaine: An example would be helpfull. --Succu (talk) 21:10, 13 May 2015 (UTC)
An random example is Q12488462, searching for it fails. This is not a taxon, but illustrates that we need more labels in more languages on many more items, including taxa. Romaine (talk) 22:01, 13 May 2015 (UTC)
Truly said, search for "Kahayan" gives this item for me (ru-interface). --Infovarius (talk) 16:16, 3 November 2015 (UTC)
  • I haven't changed my position: it would be best if the software could read P225 as if it were the label (this would save a whole lot of redundant content). And otherwise, Succu is the user that has the closest knowledge of the situation here. - Brya (talk) 11:06, 14 May 2015 (UTC)
Fully agree with Brya here. The Lua-modules used on WP and in other places should often prefer to use the name-properties when they exist. Introducing "wrong label" by bot, especially to this kind of items, would cause a lot of wrath on my home-wp. The search-engine-problem can be solved otherwise. Earlier, I could search for correct items by the content in the properties. I have not seen that for some time here now. -- Innocent bystander (talk) 11:40, 14 May 2015 (UTC)

Elo ratings for chess players[edit]

The FIDE database is the official source of present and past Elo ratings (e. g. Magnus Carlsen: Elo history where 1503014 is Property:P1440). Column RTNG is the Elo rating. It would be nice to have those values in the items with Fide-ID - the present value is the most interesting one, but the full history could be useful as well. That would add Elo ratings to thousands of chess players. The request was originally posted by an IP at the German bot requests page. --mfb (talk) 13:47, 7 June 2015 (UTC)

✓ Done should I repeat it every month? --Pasleim (talk) 19:53, 12 June 2015 (UTC)
Thanks, but you only added the Elo ratings of the current month. My wish is, that every Elo rating a player had in the past is added. The FIDE database contains the ratings back till 2000. Is it possible for you that your bot adds all these ratings? Note that until 2009, there were only four Elo ratings a year, from 2009 until 2012 there were six ratings a year, and since 2012, every month has its new rating. 07:59, 13 June 2015 (UTC)
Thanks. A monthly bot run to get the most recent numbers would be great. --mfb (talk) 20:53, 13 June 2015 (UTC)
I wonder if copying all historical ratings from the FIDE database would violate the FIDE's copyright (database rights). Also I wonder if that would be necessary or even very useful.
Many Wikipedias give two FIDE ratings for chess players: the most recent rating and the highest rating the player ever received. I thought that the FIDE also noted this peak ratings on their player cards, but it appears not to be so. The highest rating can not be derived from the historical rating list which the FIDE gives for players (or at least not always), because that list does not have all older ratings. For example Jan Timman has reached 2768 in March 1988, according to his Wikipedia article, which is probably correct since he was one of the best players of the world at that time, but the information the FIDE gives at present starts in 2000. If the historical information can be verified, an extra property for peak ratings would be useful. Bever (talk) 18:49, 3 August 2015 (UTC)
I don't think that Elo ratings are subject of copyright. There are lots of websites which publish elo ratings, recent and historical ones. Such a full overview of elo ratings could be used to automatically generate graphs like File:Elo rating Anand.png (e.g. with the graph-extension). FIDE has such graphs only back to 2000, so this would be an significant improvement. 12:12, 4 August 2015 (UTC)
Meanwhile, the graphs in the FIDE database are hidden behind a login page. So it would be very useful if we are able to create such graphs on our own! 20:15, 12 August 2015 (UTC)

US-American => P27:Q30 (currently 83579 items)[edit]

It might be worth doing this addition: Search descriptions for "American". --- Jura 16:29, 9 June 2015 (UTC)

I suspect that might not always be accurate, for various reasons. For example, there are 294 items on that list that have their date of death listed as before the United States was founded. --Yair rand (talk) 07:11, 10 June 2015 (UTC)
I'm afraid, after ten thousands of P27-statements have been (semi)roboticaly added based on much weaker guesses than this one, it doesn't really matter. The P27 data can't be trusted anymore. Surely not those without references.--Shlomo (talk) 11:00, 10 June 2015 (UTC)
We can probably clean it up, given enough time. That is, assuming nobody goes around continuing to make things worse. @Jura1: Should I assume you're verifying your hundreds of additions on this manually? --Yair rand (talk) 16:42, 10 June 2015 (UTC)
If I placed it here, it wasn't to do it myself. The sample could be limited by P570. --- Jura 16:46, 10 June 2015 (UTC)
It would need to be limited quite a bit further than that, I think. For example, the items with labels beginning with "American Samoa[n]", many of whom do not have American citizenship. I don't recommend doing this, even if a reasonable number of extra conditions can be figured out. Better to find some actual sources, and add those. --Yair rand (talk) 16:58, 10 June 2015 (UTC)
An alternate route is to use fairly reliable categories from Wikipedia. --- Jura 17:17, 10 June 2015 (UTC)

Change qualifier type of P1343[edit]

According to Wikidata:Project chat#Change described by source (P1343) qualificator for Wikisource articles discussion please change all stated in (P248) qualifiers of described by source (P1343) property to subject of (P805). Documentation and LUA modules will be updated after bot work complete. -- Vlsergey (talk) 20:54, 21 June 2015 (UTC)

@Vlsergey: On Wikidata:Project chat/Archive/2015/06#Change described by source (P1343) qualificator for Wikisource articles I don't see a consensus for this task. Were their any further discussions? --Pasleim (talk) 10:03, 29 July 2015 (UTC)
@Pasleim: there were some discussion before, and couple of users were against usage of stated in (P248) as qualifier. But if you see no consensus, feel free to close the request as "no consensus" and we will continue to use stated in (P248) as qualifier then. Sounds much simpler for me. -- Vlsergey (talk) 10:54, 29 July 2015 (UTC)

Import titled player from FIDE database[edit]

Is it possible that a bot imports e.g. all players with the title "International Master" frome the FIDE database? At the moment these are 3434 players, and only a few of them are already in Wikidata. Per notability criteria in, e.g., english and german wikipedia, all International Masters are notable. 22:35, 30 June 2015 (UTC)

Add place of burial (P119) based on image of grave (P1442)[edit]

Commons might have the information to add the property for the 2573 items currently listed at Wikidata:Database_reports/Constraint_violations/P1442#.22Item_place_of_burial_.28P119.29.22_violations. --- Jura 05:58, 6 July 2015 (UTC)

Can you point me to an example in which Commons stores this information? --Pasleim (talk) 12:56, 26 July 2015 (UTC)
Normally, there should be category on the image that would indicate the location, but Commons isn't that well structured. For the first five on todays constraint report:
  1. Q882#P1442 > File:Charlie_Chaplin_grave.jpg > Commons:Category:Charles Chaplin Grave > Commons:Category:Cemetery of Corsier-sur-Vevey > no QID linked, item doesn't exist yet
  2. Q2767#P1442 > File:Grave of Daniel Morgan, Winchester, Virginia - Stierch.jpg > Commons:Category:Mount Hebron Cemetery (Winchester, Virginia) > no QID linked, Q16895462 exists.
  3. Q5977#P1442 > File:HollyGrave850909.JPG > Commons:Category:Graves in the United States > no QID, Q30 might be too general
  4. Q7122#P1442 > File:Tombe Michel Audiard, Cimetière de Montrouge.jpg > Commons:Category:Cimetière de Montrouge > no QID linked, Q2972544 exists.
  5. Q24078#P1442 > File:La Jana - Waldfriedhof Dahlem.jpg > Commons:Category:Waldfriedhof Dahlem > no QID linked, Q875626] exists
Numbers 2,4,5 could work if the category was linked to the Wikidata item for the cemetery.
The easiest approach might be to query the images for categories and them attempt to search the corresponding QID for the more frequent ones and add those. --- Jura 13:15, 26 July 2015 (UTC)

100 000 localities of Mexico[edit]

Everything in,_%C4%8Dlanci_u_za%C4%8Detku that is not in should have

  • instance of = locality of Mexico (31:20202352)
  • English label = Spanish label = the part from the shWP article before the ","
  • English description = "in " + the part after ", "
  • Spanish description = "en " + the part after ", "
  • country = Mexico (17:96)

That is a preparation for placing them in the correct municipality and adding INEGI locality identifier (P1976) to these items. Eldizzino (talk) 13:11, 7 July 2015 (UTC)

Swedish (sv) label = Spanish label
Swedish (sv) description = "ort i " + as above or "ort i Mexiko" if separator is missing
-- Innocent bystander (talk) 13:39, 7 July 2015 (UTC)
@Eldizzino: The descriptions are looking odd. What about "locality in " + the part after ", " and "localidad en " + the part after ", "? --Pasleim (talk) 09:15, 29 July 2015 (UTC)
@Innocent bystander, Pasleim:
German (de) label = Spanish label
German (de) description = "in " + as above. I would not repeat the type in the description. 00:17, 31 July 2015 (UTC)
It's standard to include the type in the description or at least "location in", "place in", "settlement in". See also our guidlines about descriptions. The only descriptions of geographical features starting with "in" were done by socket puppets of Tobias Conradi, see [3], [4], [5]. --Pasleim (talk) 12:37, 31 July 2015 (UTC)

Adding Q20651139 to source section[edit]

After some discussion with ‎Innocent bystander, we came to the conclusion that it's preferable to use an item such as Q20651139 to indicate that a statement was derived from the inverse or symmetric property. Thus the following changes would need to be made.

For capital of (P1376), sample change: here

(itemA) P1376 (itemB) ref: P248 (itemB)
(itemA) P1376 (itemB) ref: P143 Q20651139

For diplomatic relation (P530), sample change: here

(itemA) P530 (itemB) ref: P248 (itemB)
(itemA) P530 (itemB) ref: P143 Q20651139

For spouse (P26), sample change: here

(itemA) P26 (itemB) ref: P143 (itemB)
(itemA) P26 (itemB) ref: P143 Q20651139

For child (P40), sample change: here

(itemA) P40 (itemB) ref: P143 (itemB)
(itemA) P40 (itemB) ref: P143 Q20651139

If there is already a bot that could do that, that would be most helpful. --- Jura 14:18, 9 July 2015 (UTC)

Sorry, I believe that using specific item is better (as a source) than single Q20651139. --Infovarius (talk) 12:06, 10 July 2015 (UTC)
"Stated in" Brazil (Q155) does not look like a good option. See my diff here. "Imported from Q20651139" "page:Q155" is then maybe an alternative? -- Innocent bystander (talk) 12:17, 10 July 2015 (UTC)
Yes, I like some variant of "imported from Q155". --Infovarius (talk) 21:29, 12 July 2015 (UTC)

Population of French communes[edit]

Can any bot migrate data about population of French communities from to Wikidata? Data for particular community is in template Modèle:Données/X/évolution population, where X is a name of a community. Mati7 (talk) 18:30, 20 July 2015 (UTC)

I actually came here today to request a bot to import information from fr-WP, including population. It is better to get the information from a reliable source. The French government agency responsible for censuses (INSEE) has published census information since 1962 in XLS format. The only restriction on reuse (in French) is to acknowledge the source, which Wikidata would do anyways in the form of a reference. Spreadsheets with population data from the 2012 census can be found on this page. In addition to population, the spreadsheets contain additional data which I think can be added easily to Wikidata, even if it is not very useful.

For reference, the hierarchy of administrative districts in France is:

  1. région
  2. département
  3. arrondissement
  4. canton - this is primarily a district for elections (ie. an electoral district)
  5. commune

Communes can be split between multiple cantons, cantons can span multiple arrondissements, but other levels cannot be split (eg. a commune cannot belong to multiple arrondissements). There are also "associated communes" ([[en:w:Associated communes of France|Wikipedia article) which are recognized districts within communes.

INSEE codes[edit]

Every administrative district in France has an INSEE code. The ISEE code is used for other purposes where a code is used. The INSEE code for départements is widely used, such as on vehicle license plates and the names of websites, even when not necessary...for example the website of en:w:Haut-Rhin (Q12722) contains numerous subpages with titles that incorporate its INSEE code (68). Since the population spreadsheets contain the codes for all administrative districts, they should be added while adding the populations.

I believe all communes already have the INSEE municipality code (Property:P374). However, the few that I viewed have the Dutch Wikipedia as the source, so the commune INSEE codes should be checked with the INSEE codes in the population spreadsheets and then change the reference to INSEE (the population spreadsheet). A property should be created for "INSEE department code" for departments (French: départements). I don't know if it's necessary to create a property for every administrative level, but the other levels should have "INSEE code" (Q156705) added. The INSEE code should be added to other levels as well (the linked INSEE municipality code is only for communes).

The INSEE codes for arrondissements, cantons, & communes all begin with the two-digit department code. The master file in the next section contains the codes for arrondissements, cantons, & communes without the department prefix. For example, the INSEE code for Colmar in the Haut-Rhin department is 68066 (68 is the INSEE code for Haut-Rhin department), but in the master file there are columns for the department (which has 68) and for the commune (066). The first column in the spreasheet in the "Older populations" section contains the complete INSEE code.

2012 Population[edit]

There are two population values:

  • Population municipale is the number of people who have their usual residence in the district, including people in penitentiaries, homeless people present in the commune at the time of the census, and people in mobile homes.
  • Population totale includes the population municipale plus people residing in the district but usually have a home elsewhere (eg. students living away from their usual home, people without a fixed residence).

The master file for the whole of France is here, using data from the 2012 census (reference date: 1 January 2012). It is produced and published by INSEE (Q156616). It contains 9 sheets:

  1. Regions
  2. Départements
  3. Arrondissements
  4. Cantons - ignore (Canton boundaries were adjusted in 2015 so this is no longer relevant)
  5. Communes
  6. Fractions cantonales - ignore (for communes that are divided between multiple cantons or for multiple cantons in one commune, this lists the population that lies in each canton. However, canton boundaries were adjusted in 2015 so this information is no longer relevant)
  7. Communes associées ou deleguées - "associated communes" (explained above), some may not have a Wikidata page
  8. Collectivitées d'outre-mer - populations of communes in overseas territories (collectivities). Unlike the rest of France, the entire area of an overseas territory is not divided into communes.
  9. Documentation

New boundaries were created for cantons effective in 2015. A spreadsheet with the 2012 population of the cantons based on the 2015 boundaries can be found here.

Older populations[edit]

I think it is most important to add the most recent population (2012 census). The populations from 1962, 1968, 1975, 1982, 1990, 1999, 2007, & 2012 for each commune is contained in this spreadsheet. It is produced by INSEE. It has three sheets:

  1. Métropole - European France
  2. DOM - Overseas departments, which have the same status as departments in the Métropole (eg. like Hawaii is a US state with the same status as a state in the continental US). Note that first three censuses were in 1961, 1967, & 1974.
  3. Arm (populations in the arrondissements of Paris)


Please add comments below, not in the above text. If I do not respond to a comment for a few days, please leave a message on my English Wikipedia talk page en:w:User talk:AHeneen. AHeneen (talk) 07:33, 24 July 2015 (UTC)

Are we allowed tu publish the census data under CC0 (Q6938433)? --Pasleim (talk) 15:33, 24 July 2015 (UTC)
The only restriction on resuse is to mention the source (similar to CC-by). AHeneen (talk) 05:06, 25 July 2015 (UTC)

Hello. Before anyone starts the job, I bring some clarifications (it is I who updates the census data in fr Wikipedia). You should know that unlike most countries where census are done on the entire territory periodically, in France since 2004 legal population is produced for each municipality every year, but the census type varies according to each municipality :

  • Municipalities with fewer than 10 000 inhabitants are recorded every 5 years by complete census
  • For those with more than 10 000, a sample of the population is counted every year. The annual collection covers a sample of addresses drawn randomly and representing about 8% of the population.

Then every year, there are three types of data:

  • complete census : municipalities less than 10,000 inhabitants that are the subject of a real census
  • estimated : municipalities less than 10,000 inhabitants that are'nt the subject of a real census in the year
  • Sampling : municipalities with more than 10,000 inhabitants

In the French WP, the choice was made not display in graphs and tables only data corresponding to those of the actual census and those of towns of over 10,000 inhabitants. Then in Wikidata, it is essential to have this qualifier characterizing the census type.

Unless I am mistaken, there is currently in Wikidata Q39825 that "census" should be added the following qualifiers

  • complete census
  • estimate census
  • sample census

Without this information, graphs and tables of the french wikipedia will never use wikidata datas. I can give these qualifier for each municipality and for each year (before 2006 all census are actual) but I want to see first these new qualifiers before, to be sure that we tell the same language.Roland45 (talk) 05:53, 21 October 2015 (UTC)

@Roland45: Do you mean that like presently for instance in Urt (Q842706) and Arles (Q48292), population (P1082) would be used with the qualifier determination method (P459) and an appropriate value? This value could be
  • before 2004, or after 2004 under pop. 10,000 every 5 years, the real (full) census: census (Q39825)
  • after 2004 under pop. 10,000 the rest of the time, the estimate (from previous years?) without a census taken that year: estimation (Q791801)
  • after 2004 over pop. 10,000, an estimate (for the full population if I understand correctly?) from a restricted sample: maybe sample (Q49906)?
Oliv0 (talk) 12:35, 21 October 2015 (UTC)
It's just that. To know the type of census for each municipality, simply look at the calendar.
You can see :
Urt is a municipality under 10 000 habitants - collection year : 2017 (and 2012, 2022, 2027, etc)
Arles is a municipality above 10000 habitants : collection year : each year by sampling
And then we have the following data
year Urt (Q842706) Arles (Q48292)
Value type Value type
1999 1702 census 50467 census
2006 1988 estim. 51970 sample
2007 2028 census 52197 sample
2008 2092 estim. 52729 sample
2009 2183 estim. 52979 sample
2010 2195 estim. 52661 sample
2011 2208 estim. 52510 sample
2012 2220 census 52439 sample

2006 is the first year of publication under the new method. I precise that even if the type of census can be different by year, all values are legal populations. For each commune and each year, I can give data from 1999 to 2012 and the correspondant type of census (by crossing this spreadsheet, this one and this other one).Roland45 (talk) 16:53, 21 October 2015 (UTC)

Sources are good to know, and here with their url; does each value use all the three of them? Oliv0 (talk) 19:01, 21 October 2015 (UTC)
Sources are :
Year Definition of the population (in french) Source 1 (data) Source 2 (calendar)
1962-1999 Population sans doubles comptes années 1962, 1968, 1975, 1982, 1990, 1999
2006 Populations légales des communes en vigueur au 1er janvier 2009 - Date de référence statistique : 1er janvier 2006 - limites territoriales en vigueur au 1er janvier 2008
2007 Populations légales des communes en vigueur au 1er janvier 2010 - Date de référence statistique : 1er janvier 2007 - limites territoriales en vigueur au 1er janvier 2009
2008 Populations légales des communes en vigueur au 1er janvier 2011 - Date de référence statistique : 1er janvier 2008 - limites territoriales en vigueur au 1er janvier 2010
2009 Populations légales des communes en vigueur au 1er janvier 2012 - Date de référence statistique : 1er janvier 2009 - limites territoriales en vigueur au 1er janvier 2011
2010 Populations légales des communes en vigueur au 1er janvier 2013 - Date de référence statistique : 1er janvier 2010 - limites territoriales en vigueur au 1er janvier 2012
2011 Populations légales des communes en vigueur au 1er janvier 2014 - Date de référence statistique : 1er janvier 2011 - limites territoriales en vigueur au 1er janvier 2013
2012 Populations légales des communes en vigueur au 1er janvier 2015 - Date de référence statistique : 1er janvier 2012 - limites territoriales en vigueur au 1er janvier 2014

From 1962 to 1999, there is no specific calendar because the whole census was done in one year (1962, 1968, etc). From 2006 to 2010, I have no longer the source of the calendar. In fact the qualifier is easy to deduce from the other sources. The only difference can come from the municipalities which population crosses the threshold of 10000 habitants.Roland45 (talk) 05:10, 22 October 2015 (UTC)

So the "calendar" means the table showing (like in the smaller Urt/Arles table above) which year in the 5-year cycle a given municipality has a "census" and not an "estimation". Oliv0 (talk) 06:29, 22 October 2015 (UTC)
That's right. In fact with these sources you can upload data and qualifiers, by doing year by year. The problem you can find is for the municipalities which have disappeared between 2006 and 2012. You won't have the qualifier for theses municipalities.Roland45 (talk) 11:31, 22 October 2015 (UTC)

missing pairs items[edit]

  • here item A (column 1) has P1344:B, but B (column 2) doesn't have P710:A .they are more than 141,000 cases!

I collected a list which shows missing pairs items queries. please also run the bot on it's queries regularly (every month). Yamaha5 (talk) 22:23, 27 July 2015 (UTC)

Ok please remove the case that are not for general case.Yamaha5 (talk) 09:29, 28 July 2015 (UTC)
If you don't mind, can we annotate it? (I hesitate editing pages in people's userspace). --- Jura 09:34, 28 July 2015 (UTC)
ِyour welcome you or others can edit this page (add or remove). Yamaha5 (talk) 10:08, 28 July 2015 (UTC)

new suggestion[edit]

we can separate case at User:Yamaha5/List of missingpairs querys in 2 or 3 groups

  1. completely possible like (*Item A:father (P22) > B existed but not B:child (P40) > A .Should add child (P40) > Query) which are completely true
  2. conditional possible like (Item A:sister (P9) > B existed but not B:brother (P7) or sister (P9) > A .Should add brother (P7) or sister (P9) > Query)
  3. needs more checks like (Item A:shares border with (P47) > B existed but not B:shares border with (P47) > A .Should add shares border with (P47) > Query). for this example the bot should the geographical level and do not link country with city!

Yamaha5 (talk) 10:19, 28 July 2015 (UTC)

I separated the list. any comments? Yamaha5 (talk) 10:57, 28 July 2015 (UTC)
Good work! --- Jura 11:00, 28 July 2015 (UTC)
BTW, the other day I added missing P22/P25 based on P40/P21. Prior to doing so, I check if all involved items did have P31:Q5. I know it can apply to others, but this approach avoids most potential problems. --- Jura 11:10, 28 July 2015 (UTC)

Import date of birth (P569)/date of death (P570) from Wikipedia[edit]

Lang 2007
→ ja 334
[en] 326 (~5%)
⇒ ru 124
⇒ uk 121
→ zh 116
pt 115
es 103
→ ar 100
fr 91
hu 83
tr 79
→ ko 56
id 52
et 49
fi 44
→ el 40
→ th 34

Wikidata:Database reports/Deaths at Wikipedia lists items with dates of death at Wikipedia (10-15% of all). Some dates in articles of most languages are likely to be suitable for import by bot. For other dates, the only formatted part may be the year of death category. --- Jura 08:06, 2 August 2015 (UTC)

@Multichill: didn't you had a script for this? It only works when there is a strict format, yes. Sjoerd de Bruin (talk) 18:24, 3 August 2015 (UTC)
Symbol oppose vote oversat.svg Strong oppose for second time imports same data from the same Wikipedia. Any kind of automatic and repeatable Wikipedia->Wikidata copy work makes all others Wikipedia vulnerable to mistakes (and vandalism) in single. -- Vlsergey (talk) 19:55, 3 August 2015 (UTC)
None of these pages currently have P570 defined, thus it's not a matter of re-import. Many articles may only exist in 1 language. --- Jura 21:11, 3 August 2015 (UTC)
1. "Reimport" is not about statements, but about project+property. Having p570 imported from any wiki, it shall not be reimported. Especially not on scheduled/automated basis. Arguments above. 2. I'm okay with single time import of P569/P570 from those projects. -- Vlsergey (talk) 15:03, 4 August 2015 (UTC)
I agree that it shouldn't be done for the current year on an automated basis. If you look at "date added" column on the lists, you will notice that most entries are fairly old. --- Jura 08:30, 5 August 2015 (UTC)
Looking at en:Patri J. Pugliese it seems that the formatted version is fairly recent (2014), en:Victoria Arellano has persondata since 2010, pt:Joaquim Raimundo Ferreira Chaves since 2011. en:Mark Abramson since February 2013, but only the DOB got imported. tr:Yasemin Esmergül has the dates in the article lead. In any case, we can validate the year for P570. Maybe someone can assess ja,zh,uk, etc. To the right, the most frequent ones on the list for 2007. --- Jura 21:11, 3 August 2015 (UTC)
Persondata in en.WP is deprecated and should not be relied on. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 10:06, 4 August 2015 (UTC)
Can you provide references for your claims? Thanks. --- Jura 10:26, 4 August 2015 (UTC)
Discussion of persondata: RfC: Should Persondata template be deprecated and methodically removed from articles? Jc3s5h (talk) 11:33, 4 August 2015 (UTC)
The conclusion mentioned in the link only supports Pigsonthewing's first claim. How about the second? --- Jura 11:37, 4 August 2015 (UTC)
Q8078 refers. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:41, 4 August 2015 (UTC)
Funny. Wasn't it depreciated because Wikidata could hold the data rather than for data quality reasons? --- Jura 08:30, 5 August 2015 (UTC)
No. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:43, 6 August 2015 (UTC)
Any progress on the missing reference? --- Jura 04:34, 8 August 2015 (UTC)
In reply to @Sjoerddebruin: yes I imported date of birth and date of death in the past. I was certainly not the only one. I'm quite confident the persondata template on the English Wikipedia got scraped to Wikidata quite some time ago. I don't think there is much data left to scrape from that corner. My focus was on items about humans with a link to the Dutch Wikipedia, but without data of birth. I used regular expression to extract the data of birth from the introduction of the article. You could do that for other languages too. You just need to start conservative and expand a bit in each iteration. I was able to import thousands of birth dates this way. Multichill (talk) 17:17, 4 August 2015 (UTC)
Thanks for your helpful feedback. enwiki might indeed be mostly done. For the sample year 2007 in the table above, it's just 5%. BTW nl is not on the reports as there are no nl categories for persons by year of death. --- Jura 08:30, 5 August 2015 (UTC)
Actually of the 326 for enwiki, 300 do have persondata. --- Jura 08:36, 5 August 2015 (UTC)

I imported today birth and death dates of people deceased in 2000 by parsing the introduction phrase of the English article. If the edits [6] are okay, I could continue with other years and other languages. I pay attention not to import dates before 1924 and I will not run the script twice on the same article. --Pasleim (talk) 18:42, 12 August 2015 (UTC)

Thanks! I checked 10 and they were all fine. All but 2 or 3 had the sames dates in infobox and/or persondata too.
I noticed that many trwiki articles have a person infobox, maybe this could be imported as well. --- Jura 11:04, 15 August 2015 (UTC)
That was quick. Good work! It did reduce the numbers a bit. It might be worth applying the same method to some of the templates mentioned for enwiki.
The infobox in trwiki doesn't seem that frequent, but for ptwiki, I found that many use pt:Template:dni/pt:Template:Nascimento and pt:Template:Morte or pt:Template:morte e idade/pt:Template:Falecimento e idade. This is done in infoboxes or the article text. --- Jura 07:11, 16 August 2015 (UTC)
I did some from pt:Template:Morte. --- Jura 07:21, 17 August 2015 (UTC)
pt:Template:morte e idade/pt:Template:Falecimento e idade done as well. --- Jura 09:21, 17 August 2015 (UTC)
  • For jawiki (4370 missing), I had a look at ja:Template:死亡年月日と没年齢, but that would give only about 160 most with just the month of death. eswiki has a couple of templates that could be parsed, but there is no single one. --- Jura 04:13, 19 August 2015 (UTC)
  • I had a look at 2009: Most frequent languages are: ar 125, uk 116, en 114, es 109, ru 99, hu 86
For ukwiki, of 10 articles, 6 had an infobox (5 different ones: the uk ones from Template:Infobox ice hockey player (Q5650114), Template:Infobox scientist (Q5624818), Template:Infobox architect (Q10973090), Template:Infobox person (Q6249834), Template:Infobox artist (Q5914426) normally in the format, the other 4 had the dates in the beginning of the text in Cyrillic. --- Jura 10:37, 31 August 2015 (UTC)
For ukwiki, I just imported the dates from uk:Template:Особа. --- Jura 13:33, 7 September 2015 (UTC)

Given that we might have exhausted the bot approach, I made a request at Topic:Spgr35wayo8zy15y. --- Jura 06:20, 24 September 2015 (UTC)

labels for people's names[edit]

I use nameGuzzler to add the names of people in all western-script languages, like this. Could we get a bot to do this, if not for all instances of humans, then at least for some types, or in certain circumstances (for example, if there of the labels using western scripts are the same then copy it to the rest)? Or we could do it for anyone with a given name of one of a long list (say, Andrew, Brenda, Christopher...). Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:42, 6 August 2015 (UTC)

There are many tricky cases. Names can be written in different ways in different languages, e.g. Christopher Columbus (Q7322), Julius Caesar (Q1048), Aristotle (Q868) and even when the given name is sounding English, it still can be a transliteration from a Cyrillic language and so the name shouldn't be copied, e.g. Andrew Pavlovsky (Q290393). --Pasleim (talk) 10:31, 15 August 2015 (UTC)
So let's try to whittle those away, and do the rest - maybe based on birthplace, or the existence of three Wikipedia articles with the same title, or whatever. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 11:54, 15 August 2015 (UTC)

MCN number import[edit]

There are 10,031 identifiers for MCN code (P1987) that can be extracted from [7] or this English version. Many (but not all) items cited are animal taxons, which can be easily machine-read. For the rest, it would be useful if the bot generated a list presenting possible meanings (by comparing the English and Portuguese versions of the xls file with Wikidata language entries). Pikolas (talk) 12:38, 14 August 2015 (UTC)

What's the copyright status of those documents? Sjoerd de Bruin (talk) 13:04, 14 August 2015 (UTC)
It's unclear. I've opened a FOIA request to know under what license those are published. For reference, the protocol number is 52750.000363/2015-51 and can be accessed at Pikolas (talk) 13:40, 14 August 2015 (UTC)
I heard back from them. They have assured me it's under the public domain. How can I prove this to Wikidata? Pikolas (talk) 01:48, 2 October 2015 (UTC)

One image for P158 and P94[edit]

I noticed there are a lot of items in which seal image (P158) has the same value as coat of arms image (P94). Browsing articles of localities in a wikipedia which imports from wikidata elements such as seals and coats of arms, i saw many articles with 2 coats of arms (and one flag). For example look at Quito (Q2900): both seal image (P158) and coat of arms image (P94) uses image Coat of Arms of Quito.svg. It's necessary to find all items in which value of seal image (P158) is the same as of coat of arms image (P94) and to remove value for seal image (P158) (usually image which is duplicated is a coat of arms). -- 23:06, 14 August 2015 (UTC)

This doesn't look as simple as you present it. Take for example New York City (Q60), with your proposed logic the wrong field would be removed. I also see cases where it's not clear if we're talking about a seal or a coat of arms. The English Wikipedia seems to be mixing up the two too, see for example en:Villanueva de San Juan. Probably better to get the seal/coat of arms/flags people involved here. Maybe through Wikidata:WikiProject Heraldry? This doesn't look like something that could be fixed with a bot. Multichill (talk) 09:13, 15 August 2015 (UTC)
Bot operator can put one more condition before removal, to be safe. For example if value of P158 is identical to value of P94, AND if value of seal image (P158) contains strings such as "CoA" / "coat of arms" / "CoatOfArms", THEN remove seal image as it is not seal. This can clean about half of problematic items. The rest can be analyzed by humans. -- 09:46, 15 August 2015 (UTC)
This reminds me that I fixed a few items where a bot had inverted flag and seal image. --- Jura 07:13, 16 August 2015 (UTC)
For an overview: Property talk:P158/list. As most seem to be COA, I tend to agree with --- Jura 15:48, 17 August 2015 (UTC)
Actually, I think it should be deleted and re-created: Wikidata:Properties_for_deletion#seal_image_.28P158.29. --- Jura 06:48, 19 August 2015 (UTC)

Propagating labels from en/de/fr/es for artists to other euro languages[edit]

Hi, I have worked on propagating labels from the ULAN and RKD matches to en/de/fr/es and now I would like "for each label that is the same in those languages to be copied to other western euro languages that have blank labels today". Any takers? The ULAN/RKD matches to use can be downloaded from Mix-n-Match. I suppose you could also run a bot for all humans whose label is the same in all of those langauges - this will of course pick up more than what is in the ULAN & RKD. Thx Jane023 (talk) 11:29, 18 August 2015 (UTC)

How do you intend to do with persons who have origin from countries with a language which uses Cyrillic script (Q8209)? -- Innocent bystander (talk) 12:33, 18 August 2015 (UTC)
Not. I only want the Western languages (Italian, Swedish, Danish, Hungarian, Luxembourg, Frisian, Catalan, etc). Those countries with special alphabets will have to use their own databases. The advantage is that most top museums interface with the ULAN so you can probably get their cross-reference and do it yourself. For example, the Dutch Rijksmuseum keeps artwork titles and creator names in English and Dutch (but by using the RKD list I know that I have those covered anyway, so I didn't need it). Jane023 (talk) 13:53, 18 August 2015 (UTC)
The problem is if the original name was Russian and it was transliterated to English. In that case you can't copy the English name to other western euro languages. An example: The name of Yury Annenkov (Q1362480) is written in all languages different because all languages use a different transliteration system. Also the ULAN record [8] doesn't help. It lists many different spellings but without the corresponding language. --Pasleim (talk) 14:05, 18 August 2015 (UTC)
No that is not a problem. I only selected the names that were the same in both RKD and ULAN and which had blank labels in Dutch and in English. Then I selected all the names in both RKD and ULAN that had the same labels in Dutch and in English (so more than the first pass, including many handwritten names). This already excludes lots of names that for example have the Dutch patronymic in Dutch and don't in English. Next I added all in that subset to German labels only if the German label was missing, ditto French, and ditto Spanish. Now I want the same exercise for the other languages. Please note you first have to exclude all non-null labels per language. There are no Cyrillic characters in any names at all in the subset of the dataset I am talking about. Jane023 (talk) 17:00, 18 August 2015 (UTC)
Perhaps the bot could ignore anything with a label in a Cyrillic script; or where given name (P735) doesn't match the existing western-script label; or where two western-script labels are different?? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:45, 18 August 2015 (UTC)
I know that Lithuanian often transcribe from a latin to latin scripts. See Brūss Springstīns as an example. -- Innocent bystander (talk) 17:27, 18 August 2015 (UTC)
See #labels for people's names, above, for a similar request. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:41, 18 August 2015 (UTC)

To be clear, this work has already been done, and I just want to have it repeated for other euro languages. Check the history of my edits or check the history of specific items. I did not touch Yury Annenkov (Q1362480) for example. Jane023 (talk) 17:05, 18 August 2015 (UTC)

Here are some links illustrating how this is useful by adding names over Q numbers to female artists indexed in the RKD: de:Benutzerin:Jane023/Frauen_in_RKDartsts and en:User:Jane023/Female_RKDartists. If I set up this listeria bot in one of the languages that hasn't been done yet you will see mostly Q numbers for names. Jane023 (talk) 17:58, 18 August 2015 (UTC)

Taxon names[edit]

Where we have a taxon name (P225) and no label in a western-alphabet language, could we use a bot to apply the taxon name as a label in that language? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 18:25, 22 August 2015 (UTC)

Isn't User:Succu working on that? --Pasleim (talk) 08:01, 25 August 2015 (UTC)
See #Adding missing labels for all biological species, genus, families, etc --Succu (talk) 08:08, 25 August 2015 (UTC)
That seems to require wider discussion. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 19:46, 26 August 2015 (UTC)

New properties[edit]

The adoption of new properties is limited by the fact that neither here on Wikidata nor elsewhere on other projects are there specific actions designed to monitor them carefully. I suggest that we start tagging all existing properties with their creation date using inception (P571) and that we then keep going when new ones are created after that. Can a bot do both things? The monitoring of new ones could be based on a daily check, I would say. Thierry Caro (talk) 04:07, 24 August 2015 (UTC)

What do you plan to do if all properties have a inception (P571) claim? A list of new properties you get by --Pasleim (talk) 08:00, 25 August 2015 (UTC)
I know, but this is not usable on Wikipedias, except through a straight link. If we want to generate lists of useful properties on projects there, we should fill the properties of properties better here, starting with a creation date for instance. Thierry Caro (talk) 13:05, 25 August 2015 (UTC)

Add links to Arabic sister projects[edit]

Hey, Is there a way to add articles of sister projects in Arabic (wikiquote, books, source, .. ) to wikidata by a bot ? For example this article in Arabic wikiquote is linked to the Arabic article in wikipedia by a template. But if we take a look to the wikidata article there is no link to the Arabic wikiquote article. Or if you can at least extract a list of same name articles in Arabic wikiquote/wikisource/wikibooks and wikipedia but not linked in wikidata. --Helmoony (talk) 20:15, 25 August 2015 (UTC)

US places with names in the format "<locality>, <US state name>"[edit]

These could converted to "<locality>" and the label added as alias. Sample: diff.

For places with labels in the format "<locality>" and enwiki articles in the format "<locality>, <US state name>", the enwiki title could be added as alias. --- Jura 12:30, 26 August 2015 (UTC)

Better to add a description because aliases aren't seen in the search drop-down. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 22:11, 28 August 2015 (UTC)
Adding description is another option, but not part of this request. Please keep it focused and provide a reference for the reminder of your comment. --- Jura 06:57, 29 August 2015 (UTC)
I'm focussed on making the best improvements to Wikidata. Your proposal would remove visibility of the state-name from the search drop-down, which would be a negative effect. As such, I would object to it, unless modified. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 10:23, 29 August 2015 (UTC)
But you don't have any reference to support your POV? --- Jura 10:35, 29 August 2015 (UTC)
Many place names in the US are repeated in the US in several different states, so in any environment where one is choosing a US place name from a menu it is essential to include the state in the menu. Example: Manchester (Q48370), Manchester (Q1905654), Manchester (Q753909), Manchester (Q9027296), and all the others listed at w:Manchester (disambiguation). Jc3s5h (talk) 12:02, 29 August 2015 (UTC)
It doesn't actually support Pig's POV. --- Jura 12:06, 29 August 2015 (UTC)

ALL CAPS[edit]

A series of items for people have labels in capital letters .. if these could be converted into a more standard format .. --- Jura 08:00, 28 August 2015 (UTC)

To start off with it, I created a quarry list: Most of these labels can be converted but there are also some excpetions, e.g. RUBEN XYZ (Q7277963) --Pasleim (talk) 16:19, 16 September 2015 (UTC)
I think the Japanese ones in that list could be left out. Latin isn't the usual script for Japanese, so a Latin script name is likely be deliberately written the way it is. The ones I checked all had a jawiki sitelink which is all caps like the label. - Nikki (talk) 18:06, 16 September 2015 (UTC)
I just did the en ones for items with P31:Q5. I don't think there was any CamelCasing in it. Thanks for the list! --- Jura 09:49, 30 September 2015 (UTC)
It seems that some didn't get included, sample: Q20734549 (has two spaces). --- Jura 09:27, 1 October 2015 (UTC)

Check links between article and categories items[edit]

Analyze Wikipedia and sister project pages from main space and categories for linking through Property:P301 and Property:P910. Categories should have w:en:Template:Catmore, pages in main space - special sort keys (space, *, etc). Will be also good idea to check if category item for Commons link with page Property:P373 (same, absent, etc). Probably report should be enough to avoid false positives. --EugeneZelenko (talk) 14:13, 6 September 2015 (UTC)

Непонятно, в чём вопрос. Вы предлагаете рассоединить категории и статьи из одного элемента? Про Викиновости знаете? --Infovarius (talk) 21:26, 6 September 2015 (UTC)
Я предлагаю проверить, соединены ли элементы для статей и категорий. --EugeneZelenko (talk) 02:13, 7 September 2015 (UTC)
Some reports you can find at violation constraints: Wikidata:Database reports/Constraint violations/P301#Inverse and Wikidata:Database reports/Constraint violations/P910#Inverse. I suppose that categories with Template:Catmain were already imported into Wikidata, at least once. --Infovarius (talk) 08:43, 10 September 2015 (UTC)

Harvest elevation above sea level (P2044) from Swedish Wikipedia[edit]

Could someone harvest the value for elevation above sea level (P2044) from Swedish Wikipedia (svwiki) where it is available in articles that use the template sv:Mall:Insjöfakta Sverige in the parameter höjd. Size of harvest would be approximately above 55000 items affected. Ainali (talk) 22:32, 10 September 2015 (UTC)

@Ainali: Take a look at Galaktos' comment about this property at WD:PC. Do we (or at least Nasko) know which level this is relative to? -- Innocent bystander (talk) 06:36, 11 September 2015 (UTC)
I checked with SMHI, and they have different tables for height, which uses different relative levels. So unless Nasko can tell us which one was used, I guess this request should be put on hold. Ainali (talk) 11:06, 11 September 2015 (UTC)
I am not sure I really care about the exact level used for the moment. Thierry Caro (talk) 11:21, 11 September 2015 (UTC)
Post-glacial rebound (Q1161410) is a very big issue in parts of this set of articles, so I think this really is an important question. no label (Q18182959) is for example located 0 meters above sea level according to this source. If that was calculated compared to the de facto sea level, this lake would have been flooded half of the time, but it isn't. In fact the stream that runs from this lake to the sea have some tiny waterfalls in it and the lake has not been flooded for maybe 200 years. -- Innocent bystander (talk) 13:40, 11 September 2015 (UTC)
All the small lakes user the geodetic system RT38 and RH70 for altitude. Larger lakes uses RT90 and RH00 or SWEREF99 and RH2000. Now what is the difference between these geodetic systems and WGS84? Then it comes to altitude two factor are important. The first one is the distance from the earths center of mass (the "radius"). The second is flattening of the poles. The northern hemisphere is actually flatter than the southern because of an old impact.
RT90 uses geodetic system=Bessel 1841, Semi-major axis=6377397.155 m, Flattening of the poles=1/298.1528128
WGS84 uses geodetic system=WGS 84, Semi-major axis=6378137 m, Flattening of the poles= 1/298.257223563
When comparing a traditional leveling and satellite based altitude information, the altitudes may differ by many meters.[9] [10] [11] (swedish) Other factors: Different seas has different sea-levels (10-20 cm). The gravity from under-water mountains attracts water and the sea rises several meters. The moon causes tidal waves. Isostasy is also a factor, but it needs hundreds of years to shift the land several meters. And so on...[12] (english)
Conclution: Traditional leveling and satellite based altitude information will give different results especially when local geodetic systems and leveling is compared to modern satellite based surveying techniques. I would harvest the altitudes. A simple version of the Swedish national altitude database was released under a free license this summer, so it is possible to improve the data later if it is needed. Nasko () 17:23, 11 September 2015 (UTC)

Date for birth for recently deceased[edit]

For items appearing on Wikidata:Database reports/Recently deceased at Wikipedia, it might be worth attempting to import some data automatically (date of birth, place of birth, gender, nationality, identifiers, occupation, BUT NOT date of death) or on-demand/subject to review (this could include date of death/place of death).

@Hsarrazin:: as you worked on them quite a lot: what do you think? --- Jura 10:19, 13 September 2015 (UTC)

the problem with auto-imported data, for now, is that they are never sourced and it is really hard to find from which language the data comes, and finding the source in the language article is often tricky.
wp projects (at least fr and de) accuse wikidata of not properly sourcing the data, but I could often see that a lot of wp articles are poorly or not sourced at all :( - retrieving source whenever possible is very important.
personally, as I work on death date, I manually systematically (or nearly) import date/place of birth, gender, nationality, identifiers, occupation, using User:Magnus Manske/wikidata_useful.js. I also add given name (P735) whenever I can sort it out (hungarians are tricky).
bot-importing those data would be risky ; an on-demand retrieving tool (like User:Magnus Manske/import statements.js but working) or a reviewing tool (like Primary source tool) could be more efficient, but for most recent deaths, the articles are very young, often unstructured, and data is hard to auto-retrieve, and often harder to eye-retrieve.
it would be important to have data imported from all linked languages, when different, so that different data on different languages could be reviewed. (en is the most common, but not necessarily the most accurate).
  • auto-importing date of death for recent death is NOT good, as errors and spam are quite high :)
  • info could be retrieved from Infobox (when existing), categories or plain text (like WD Useful).
  • data to retrieve : date of birth (P569), date of death (P570), place of birth (P19)/place of death (P20), country of citizenship (P27), occupation (P106), image (P18) would be nice too :)
  • sex or gender (P21) is often tricky in many languages. I often use picture to determin it - maybe in a reviewing tool.
  • VIAF identifier (P214) would be nice too, but I often found those wrong or outdated on various wp. I prefer to add it manually from personal search (it's my day-job), and import other IDs from VIAF with User:Tpt/viaf.js.
  • when different data exist in different wikilinks, each should be retrieved and at least marked as imported from (P143)+language link on each auto-imported data.
  • such a tool should also try to retrieve sourcing info (when available) from each existing language link (something like Sourcerer that don't work for me).
  • and whenever possible, add date of input (like you do on the "recent death" report).
  • an on demand (button) tool, or a better, a reviewing tool would be very useful.
such a tool (similar to Harry Potter's magic wand) that would retrieve such data could also be useful on any human (Q5) item, so I would like it very much ;) --Hsarrazin (talk) 13:42, 13 September 2015 (UTC)
To some extent "Primary Sources" (indirectly) and "Suggestor" (with lots of noise) end up doing that, but a dedicated tool would indeed help.
I suppose its overall usefulness depends on how close (or how far) we are from importing all ±formatted data from Wikipedia. --- Jura 14:04, 13 September 2015 (UTC)
I don't know Suggestor, but existing tools I know mainly work from enwiki. There is a lot of work to autofetch info from other languages, I think. :) --Hsarrazin (talk) 20:58, 13 September 2015 (UTC)
There is now to add the date of death. --- Jura 13:37, 23 September 2015 (UTC)

Commons Category <> Sitelinks?[edit]

Hello, Is it possible/feasible for a bot to sync the Commons Category property with the sisterlinks? There were some cases where I added a sisterlink but forgot to add the property, e.g. Q6226954. A bot could easily read/write these back and forth. I'm not sure if this has been proposed before or not. I'd do it myself, but admittedly, I don't understand enough about how Wikidata works. Avicennasis (talk) 05:03, 18 September 2015 (UTC)

I think this would be an interesting task. I added hundreds of commons category properties using QuickStatements but as far as I know there is no equivalent tool to add sitelinks. Using commons category property to add sitelinks would be great.--Pere prlpz (talk) 22:28, 9 October 2015 (UTC)

Cyrillic merges[edit]

This included pairs of items with articles at ruwiki and ukwiki each (Sample: Q15061198 / Q12171178). Maybe it's possible to find similar items merely based on labels in these languages and merge them. --- Jura 03:33, 19 September 2015 (UTC)

I cannot find any ru-uk pairs. Are they all done? --Infovarius (talk) 16:27, 3 November 2015 (UTC)
The ones on that list are identified based on dates of birth/death and we regularly go through them. The occasional findings there (also with ru/be) suggest that there are more (without dates). A query would need to be done to find them. --- Jura 16:33, 3 November 2015 (UTC)

Harvest en:Template:BCGNIS for BC Geographical Names ID (P2099)[edit]

Could someone please harvert the line id= from the template en:Template:BCGNIS in enwiki and put the result in BC Geographical Names ID (P2099)? --Fralambert (talk) 00:21, 24 September 2015 (UTC)

imported with harvest templates all values which are not used as reference in the article. --Pasleim (talk) 10:45, 5 November 2015 (UTC)

Footballer senior career and national team[edit]

Hi. I'm working on using wikidata data in french footballer infobox. I realize that many information of "member of sports team" (P:54) are missing, whereas the information is described in English infobox (and italian and french, also). I've just verified in Bot requests archive and I've found this one, dating from April 2013 (ping @Sven Manguard, Legoktm:). It is exactly the point : "It seems like it'd be an easy task for a bot to read the "Infobox football biography" template over on Wikipedia for the "Senior career" and "National team" fields and then use those to populate "member of sports team" (P:54). Once we get qualifiers, the bot can then copy over the years field attached to each entry as well.".

So, do you think a bot could do (finish) the job ? Thanks in advance. --H4stings (talk) 09:00, 28 September 2015 (UTC)

My code for this is [13], and I think I ran it at some point. I unfortunately don't have much time to pick up any new tasks :( Legoktm (talk) 22:02, 1 October 2015 (UTC)
Ok, thanks. I hope some other football fan bot runner could adopt this cause. Face-smile.svg --H4stings (talk) 13:21, 2 October 2015 (UTC)

Fix redirects from pagemoves at Wikipedia (bug T92789)[edit]

Per Wikidata:Project chat#How_to_get_rid_of_redirects_.28in_sitelinks.29, some of the redirects are due to a bug in the software. These should be fixed: the redirect in sitelink replaced by target article.

If there is a way to identify that the other page moves at Wikipedia that haven't be updated at Wikidata at the same time, this could help. --- Jura 07:17, 29 September 2015 (UTC)

As a first step and following a request, here is a query that can help for enwiki:
It filters on pagemove edit summaries (excluding many others). Some of the data is stale and needs purging of the page at enwiki. --- Jura 10:52, 29 September 2015 (UTC)

Remove the description "Wikimedia disambiguation page" from items that have a value in P31 other than Wikimedia disambiguation page (Q4167410)[edit]

Currently there are about 30 items with that [14].

There are more in other languages and variations of the description text. --- Jura 11:54, 4 October 2015 (UTC)

For "Wikipedia disambiguation page", there are 3259 items (check). --- Jura 12:02, 4 October 2015 (UTC)
Good catch. Though, those also contains items that are marked as intance of surname but should also include P31:disambiguation page, for example Yule (Q19857423). --Stryn (talk) 16:08, 4 October 2015 (UTC)
Items like Q5313389 shouldn't have (and don't have it). If eventually someone adds it to some of the others, another a bot will add the descriptions.
In the meantime, removing it avoids a lot of confusion. ✓ Done for some of the en/fr descriptions. --- Jura 11:50, 10 October 2015 (UTC)

Hierarchic data for German places[edit]

Hi folks,

extract from Wikidata IRC:

[09:29] <wikiwilli> I use geographic data from wikidata for a web project and noticed, that in several cases the hierarchic information about places is inconsistent.
[09:31] <wikiwilli> For example in Germany some places are marked as contained in the approbiate 'Landkreis', but some are only marked in the 'Bundesland'
[09:31] <wikiwilli> In Wikipedia there is always the full hierarchic information.
[09:32] <wikiwilli> Is it necessary to update all thesse Wikidata entries manually or is there a bot active, which extract this information from Wikipedia?
[09:39] <wikiwilli> examples:
[09:39] <wikiwilli> Full hierarchic info: Damp Q662717
[09:39] <hoo> I doubt there's a bot active, although it would probably not to hard to get that right
[09:41] <wikiwilli> No hierarchic info: Eckernförde Q490453
[09:45] <wikiwilli> But there is a general used template in German WP for places (with the 'Landkreis' data). Wouldn't it be quite easy to extract the info by an bot?
[09:45] <hoo> Possibly, yes
[09:45] <hoo> But someone would need to do it
[09:46] <hoo> also data from Wikipedia often isn't properly sourced, so people might not want it. No idea about that data, though
[09:48] <wikiwilli> The basic data about German places in the German WP is nearly 100% correct (and I think, there is no need for a proper source to state that a village is in a certain county)
[09:49] <hoo> If you don't want to/ can't do that yourself, you can go and ask nicely on
[09:49] <hoo> Someone might pick it up

Is there any bot out there, which could do this job? -- 07:53, 8 October 2015 (UTC)

I've edited the IRC bit you quoted to make it readable and also removed a couple of unrelated lines and made the item IDs link to the items. Hope you don't mind. :)
It's not quite clear to me which places you mean, but looking at the examples, it seems to be the German municipalities (Gemeinden), is that right?
If so, it seems like a bot needs to go through all articles using de:Vorlage:Infobox Gemeinde in Deutschland and make sure that there is a located in the administrative territorial entity (P131) statement for the smallest administrative division. It would probably also make sense to remove any located in the administrative territorial entity (P131) statements matching the larger administrative divisions, as those should be inferable (then it would also make sense to go through de:Vorlage:Infobox Gemeindeverband in Deutschland and de:Vorlage:Infobox Landkreis to make sure those have the right located in the administrative territorial entity (P131) statements too).
- Nikki (talk) 11:33, 8 October 2015 (UTC)
Hi Nikki, this is exactly what I mean! Thanks for editing my request and making things clear. Can You set up a bot to do this?
If this is possible, maybe in a next step the bot can be adjusted to do similar jobs of setting up/enhancing located in the administrative territorial entity (P131) in other countries and for other types of places (German cities,...) -- 11:45, 8 October 2015 (UTC)
The problem with adding intermediary layers and removing the main layers is that they are more likely to get obsolete and people end up we end up missing the key layers. --- Jura 16:33, 13 October 2015 (UTC)
I don't see how we would end up "missing the key layers" unless someone incorrectly removes a statement when the structure changes instead of adding an end date and if they do, their change should be undone and an end date added instead. Adding all the parents to every item is not a good solution, it creates huge amounts of work when things do change, because you have to copy the same changes to everything within the item which has changed, which could be hundreds or even thousands of items. - Nikki (talk) 18:23, 13 October 2015 (UTC)
I suppose it depends on the country: some layers keep changing, other remain stable; some layers are unimportant, others are key. If the key layers are there, the structure is usable and stable. If one relies on having each individual intermediary layer, this tends to break fast. And Tobias keeps adding new intermediary layers to hundreds or even thousands of items. --- Jura 18:47, 13 October 2015 (UTC)
Sorry, but this opinion is pretty strange. Should we add all locations P131 to Germany (Q183) or better Europe (Q46), because there could be some administrative changes (or wars, annexions, ...) in future? --Arch2all (talk) 11:53, 14 October 2015 (UTC)
Good point: all locations in Germany should have Germany (Q183), but not in P131. We have country (P17) for this. You might want to read both definitions. --- Jura 12:09, 14 October 2015 (UTC)
There are quite a lot of redundant definitions in Wikidata right now. In my opinion minimal redundancy together with a wellformed structure are the big advantages of wikidata (and similar thesauri). Instead of trying to keep the repository "idiotensicher" I would prefer to engage in better structures (with even more intermediary layers) and avoid/remove redundant definitions! --Arch2all (talk) 12:29, 14 October 2015 (UTC)

Import P569 and P570 from trwiki (text)[edit]

Wiki trwiki
Items without P569 count (all)
As of Oct 12 12696 (22 %)

Quite a lot of articles at trwiki have well formed dates in the format "(d. DD MMM YYYY[, place]) - (ö. YY MMM YYYY[, place])," (sample Q17431004). These could be imported by bot. --- Jura 07:00, 11 October 2015 (UTC)

Alternate format:
  • Item: Q6055352, Ercüment Batanay
  • Sample text: Ercüment Batanay (d. 9 Nisan 1927, İstanbul - ö. 4 Mayıs 2004, İstanbul),
--- Jura 12:29, 12 October 2015 (UTC)

Import P569 and P570 from azwiki (template)[edit]

Wiki azwiki
Items without P569 count (all)
As of Oct 12 9379 (47 %)

Dates in azwiki are frequently available in az:Template:Vəfat tarixi və yaşı. These could be imported by bot. --- Jura 07:02, 11 October 2015 (UTC)

Transfer of asteroids labels[edit]

Help need to copy-paste label content in pages, that link to Q3863 from Russian to Armenian. Please, include only labels with titles, that consist a complex of letters and figures like Q8604, but not names like Q2356051, if possible. Transfer from English is prefered, if brackets () added to asteroid number. No limits to contents in that case. Please, don’t touch labels, that almost exist in Armenian. Thank you - Kareyac (talk) 07:37, 11 October 2015 (UTC)

Import P569/P570 dates from slwiki (text)[edit]

Wiki slwiki
Items without P569 count (all)
As of Oct 12 11116 (31 %)

slwiki has dates in the formats "* YYYY" and "† YYYY." or more precise. These could be imported by bot. --- Jura 07:50, 11 October 2015 (UTC)

Import values from and[edit]

Hi, Openpolis ID (P1229), Finnish MP ID (P2181) and Kansallisbiografia ID (P2180) have been created. A bot should import values for these properties:

--★ → Airon 90 08:05, 11 October 2015 (UTC)

done with harvest templates --Pasleim (talk) 14:12, 13 November 2015 (UTC)

Import P569/P570 dates from kowiki (text)[edit]

Wiki kowiki
Items without P569 count (all)
As of Oct 12 21386 (31 %)

Dates in kowiki are frequently well format and placed:

  • Sample: Q19940305, 심봉섭
  • Sample text: 심봉섭(沈鳳燮, 1930년 1월 16일 ~ 2000년 9월 15일)
  • Format: (.., YYYY년 MM월 DD일 ~ YYYY년 MM월 DD일)

These could be imported by bot. --- Jura 08:07, 11 October 2015 (UTC)

By the way, I'm now working on this. - Nikki (talk) 10:07, 9 November 2015 (UTC)
Thanks. I think it is already showing up on sitelinks and P569. --- Jura 16:36, 10 November 2015 (UTC)
And occasionally, they pop up on Wikidata:Database reports/identical birth and death dates/1. --- Jura 07:11, 13 November 2015 (UTC)

Import P569/P570 dates from jawiki (text)[edit]

Wiki jawiki
Items without P569 count (all)
As of Oct 12 96921 (37 %)

Some of the dates in jawiki are in the following format:

  • Sample: Q11478231, 川島喜代詩
  • Sample text: 川島 喜代詩(かわしま きよし、1926年10月29日‐2007年4月24日
  • Format: YYYY年MM月DD日‐YYYY年MM月DD日

These could be imported by bot. --- Jura 08:11, 11 October 2015 (UTC)

If someone decides to do this, please be careful when extracting the dates. The Japanese Wikipedia often includes dates using the era system, e.g. ja:木村作次郎, and if the dates are extracted too liberally, it could end up saying the date of birth is the 5th of July, 5 AD instead of the 8th of August 1872 or something similarly wrong.
People appear to be put into Category:YYYY年生 (Category:YYYY births) and Category:YYYY年没 (Category:YYYY deaths) which could be used to verify the years of the extracted dates.
The usual format (in my experience) has spaces around the dash between the begin and end dates (i.e. YYYY年MM月DD日 - YYYY年MM月DD日), unlike Jura's example.
The bracketed bit usually includes more information, most commonly name in kana (P1814). If the text is being parsed anyway, it would be nice if we could import that too. If someone has a way to generate a list of extracts (the first paragraph or couple of sentences or whatever) for articles on jawiki which are humans and have no begin date, I would be really interested in looking at the data (ideally including the Wikidata ID and page name on jawiki)... I might even see if I can generate a list myself if nobody has anything already.
- Nikki (talk) 09:52, 11 October 2015 (UTC)
The other format I found is:
  • Sample: Q11595189, 秋玲二
  • Sample text: 秋 玲二(あき れいじ、1910年(明治43年)11月1日 - 2006年(平成18年)2月24日 は、
It's slightly more complicated to parse. If you look at the lists at on Wikidata:Database reports/Deaths at Wikipedia, you can find colons with "ja". These would be items with missing P570, but a category at jawiki. PLTools to browse extracts is linked there as well. --- Jura 10:06, 11 October 2015 (UTC)
After some investigation, I think we could add a lot of dates even if we completely ignore the ones which include the era system forms. I fetched extracts of 2500 jawiki items. 1883 of them contain 19YY年 or 20YY年 and I was able to extract the date section in 1625 of those using a regex which checks for:
  • A comma (、 or ,) or opening parenthesis (( or ()
  • A date (in the form YYYY年MM月DD日, with month and day optional), or "生年不詳", "生年不明" or "不明" (which mean the birth date is unknown), or a question mark (? or ?)
  • A character that looks like a dash (usually the ASCII one)
  • A date (in the form YYYY年MM月DD日, with month and day optional), or "没年不明", "没年不明" or "不明" (which mean the death date is unknown), or a question mark (? or ?)
  • A closing parenthesis () or ))
with optional space (ASCII or fullwidth) between each part.
(The exact regex I was using, if that's useful to someone, is /(、|,|(|\()( | )?(([0-9]{4}年([0-9]{1,2}月([0-9]{1,2}日)?)?|生年不詳|生年不明|不明|?|\?)( | )?(-|ー|—|−|-|–|‐|~)( | )?([0-9]{4}年([0-9]{1,2}月([0-9]{1,2}日)?)?|没年不明|没年不詳|不明|?|\?)?)( | )?()|\))/)
Of the 258 that weren't matched by it, 142 included the era system form of the date and the rest were either less clearly formatted or included extra words or symbols (e.g. the word for "about" or a question mark after the date) or the date in the extract wasn't a birth/death date.
- Nikki (talk) 17:56, 12 October 2015 (UTC)
Oh, and the death date part is optional. Not everyone is dead yet. :P - Nikki (talk) 18:04, 12 October 2015 (UTC)
jawiki is one of the top candidates on User:Jura1/sitelinks and P569. --- Jura 18:13, 12 October 2015 (UTC)
I've started doing this one myself, although only dates from the 1900s or 2000s for now. - Nikki (talk) 15:27, 16 October 2015 (UTC)
I've more or less finished with the ones I can confidently extract, which has halved the number without dates (yay \o/). Of the ones which are left, ~15000 have nothing resembling a date, ~6400 say the date is unknown, ~4600 have the day and month but no year (phab:T104749), ~10000 have dates from before 1873 (which is when Japan switched to the Gregorian calendar) and ~8500 look like they contain a date but a date of birth couldn't be confidently extracted (usually either vague or unusual formatting or a date which isn't a date of birth).
I'm not sure what to do with the ones from before 1873. Japan never used the Julian calendar (they switched straight from their lunisolar calendar to Gregorian), so I could assume that all western-style dates are using the proleptic Gregorian calendar unless explicitly marked as Julian (like on ja:五龍局). Looking at Wikidata:Project chat#How_to_add_dates_with_undefined_calendar_.3F, we're not going to get a better solution any time soon and we're expected to guess the calendar if we don't know it, so I think going ahead with that assumption is the best we can do. If something does change in the future, finding the dates from before 1873 is easy enough. Any objections to that?
- Nikki (talk) 14:30, 2 November 2015 (UTC)
Great work. Thanks. jawiki is now at 17% (down from 37%), better than enwiki (20%). For items that have either, it's 15.6%(ja) compared to 16.3%(en).
I'm mostly done with Database reports/Deaths at Wikipedia (≥2000) and I don't think there is much from jawiki left there either.
I hope the pre-1873 question gets sorted out this month. Maybe input from jawiki could confirm your assumption. 井戸端 might help. --- Jura 15:12, 2 November 2015 (UTC)
Mentioned it here --- Jura 15:28, 2 November 2015 (UTC)
I've asked on Wikidata:井戸端. - Nikki (talk) 17:07, 2 November 2015 (UTC)
All this focuses on when Japan adopted the Gregorian calendar, depending on the convention for writing historical dates in Japanese. In English, it is customary to use either Gregorian or Julian, depending on which was in force at the time and place of the event (various authors take various approaches when writing in English about places where neither calendar was in force). I see the Japanese Wikipedia contains articles about people from all over the world, so it is necessary to find out what date would be given for a birth or death that occurred where the Julian calendar was in force. It appears in the case of Kings George II and George III of the UK, Gregorian dates are given. I wonder if this practice is consistent. Jc3s5h (talk) 17:45, 2 November 2015 (UTC)
The pages in question are almost all about East Asian people (i.e. from countries which never used the Julian calendar) and typically include an East Asian-style date plus a western-style date. The main question is then not what they do for countries which did use Julian, but which western calendar do they convert East Asian-style dates to, when neither Julian nor Gregorian were in use in the region at the time. I got an answer from Wikidata:井戸端 anyway, dates after 1582-10-15 should use Gregorian, dates before should use Julian. - Nikki (talk) 20:24, 3 November 2015 (UTC)

Import P569/P570 dates from ruwiki (text)[edit]

Wiki ruwiki
Items without P569 count (all)
As of Oct 12 37361 (12 %)

Some ruwiki articles still have dates in the article text:

  • Sample: Q20742501, Худалов, Харитон Алексеевич
  • Sample text: Харитон Алексеевич Худалов (9 января 1905, с. Махческ, Терская область - 7 июля 2000, Владикавказ)
  • Format: (D MMM YYYY, ... - D MMM YYYY, ...)

To avoid calendar issues, skip pre-1919 dates. --- Jura 08:53, 11 October 2015 (UTC)

Import P569/P570 dates from dewiki (text)[edit]

Wiki dewiki
Items without P569 count (all)
As of Oct 12 19612 (3 %)

dewiki articles are generally consistently formatted:

  • Sample: Q20752215, Marianne Mogler
  • Sample text: Marianne Mogler (* 7. Januar 1933; † 8. November 2000)
  • Format: (* DD. MMM YYYY; † DD. MMM YYYY)

Some dates could be imported from the above format. --- Jura 08:59, 11 October 2015 (UTC)

It shouldn't be a problem since the dates are clearly marked with symbols, but note that the place of birth/death is often included too, e.g. de:Margarete Aburumieh and de:August von Ahlefeld. - Nikki (talk) 13:30, 11 October 2015 (UTC)
yes, all requests made here are fairly straightforward. If I made separate requests, it's mainly to make it easier for users of those wikis to work on them. --- Jura 16:19, 11 October 2015 (UTC)

Import P569/P570 dates from nlwiki (template)[edit]

Wiki nlwiki
Items without P569 count (all)
As of Oct 12 12294 (7 %)

nlwiki has no tracking categories, so no articles are included in Wikidata:Database_reports/Deaths_at_Wikipedia.

Some of the articles contain infoboxes with dates. A tracking category could be added to those with dates there, but no data here at Wikidata. Sample category for frwiki: Q20113586. The dates could then be imported to Wikidata. --- Jura 09:07, 11 October 2015 (UTC)

A lot of dates are already imported. The leftovers are the articles without infobox or syntax errors. Trying to lower that every day. Sjoerd de Bruin (talk) 17:02, 12 October 2015 (UTC)
I tried to do some stats: User:Jura1/sitelinks and P569 --- Jura 18:11, 12 October 2015 (UTC)

Correct claims[edit]

The elements supervillain (Q6498903) and superhero (Q188784) are instances of fictional profession (Q17305127) and they are used in a lot of elements as instance of (P31), can anybody change those from instance of (P31) to occupation (P106)? At least as I understand I think that it is more correct putting that information in occupation (P106). -- Agabi10 (talk) 12:57, 11 October 2015 (UTC)

Of course, not. They are fictional, not real professions. --Infovarius (talk) 05:47, 12 October 2015 (UTC)
@Infovarius:: Does it really mind to put them in the property occupation (P106)? Even if they are fictional they are still professions... -- Agabi10 (talk) 09:46, 12 October 2015 (UTC)
Oh, sorry. I've answered wrong question (I thought that you propose to make them kind of subclass of (P279) of occupation (Q13516667)...). As for property I don't know. May be it is possible. --Infovarius (talk) 13:45, 12 October 2015 (UTC)

Move qualifiers with "Source" constraint to the references block[edit]

The properties that are supposed to have to be used as references are used as qualifiers in some cases and there are too many cases to correct them manually. The properties are reference URL (P854), quote (P1683), retrieved (P813), imported from (P143) and stated in (P248). The elements that are using it in any way different than the sources block are listed in their respective constraint violations report. If it is not used as a qualifier it's better to use it like it is. Thanks. -- Agabi10 (talk) 15:56, 11 October 2015 (UTC)

Import P569/570 from etwiki (text)[edit]

Wiki etwiki
Items without P569 count (all)
As of Oct 12 8983 (25 %)

The following formats are used:

  • Sample item: Q16404152, Aleksander-Henn Schipai
  • Sample text: Aleksander - Henn Schipai (21. november 1929 Tallinn - 5. märts 2004 Tallinn)
  • Format: (DD. MMM YYYY [place] - DD. MM YYYY [place])

A bot could import them. --- Jura 16:21, 11 October 2015 (UTC)

Import P569/570 from svwiki (text)[edit]

Wiki svwiki
Items without P569 count (all)
As of Oct 12 7315 (4 %)

Text frequently follows this format:

  • Sample item: Q6010919, Knut Nilsson
  • Sample text: Knut Johan Evert Nilsson, född 23 september 1923 i Stehags församling, Skåne, död 2 oktober 2004 i Handen, Österhaninge församling
  • Format: född DD MMM YYY ... död DD MMM YYY ..

A bot could import them. --- Jura 12:25, 12 October 2015 (UTC)

Import P569/570 from nowiki (text)[edit]

Wiki nowiki
Items without P569 count (all)
As of Oct 12 4899 (4 %)

Text frequently follows this format:

  • Sample item: Q17110549, Miling Svalastog
  • Sample text: Miling Svalastog (født 29. april 1919 i Trondheim, død 8. juni 2004)
  • Format: (født DD. MMM YYY [i <place>], død DD. MMM YYY [i <place>])

A bot could import them. --- Jura 12:25, 12 October 2015 (UTC)

Import P569/570 from huwiki (text)[edit]

Wiki huwiki
Items without P569 count (all)
As of Oct 12 18402 (24 %)

Text frequently follows this format:

  • Sample item: Q12816373, Székely Erzsébet
  • Sample text: Székely Erzsébet (Kolozsvár, 1922. június 15. - Kolozsvár, 2001. január 6.)
  • Format: ([<place>,] YYYY. MMM DD. - [<place>,] YYYY. MMM DD.)

This could be imported. --- Jura 13:55, 12 October 2015 (UTC)

Import P569/P570 from eowiki (text)[edit]

Wiki eowiki
Items without P569 count (all)
As of Oct 12 7198 (20 %)

The following format seems quite frequent:

  • Sample item: Q12350072, Juozas Kazlauskas
  • Sample text: Juozas KAZLAUSKAS (naskiĝis la 28 - an de januaro 1909 en vilaĝo Nuoriškiai, distrikto Biržai; mortis la 29 - an de julio 2001 en Panevėžys)
  • Format: (naskiĝis la DD - an de MMM YYYY [..]; mortis la DD - an de MMM YYYY [..])

--- Jura 16:31, 12 October 2015 (UTC)

I disagree. User:Jura1 has not pointed out any ability to distinguish Julian and Gregorian calendars, so these dates are not suitable for dates before 1923 or so. Jc3s5h (talk) 15:32, 23 October 2015 (UTC)

Import P569/P570 from plwiki (text)[edit]

Wiki plwiki
Items without P569 count (all)
As of Oct 12 59354 (22 %)

The following format can be read easily:

  • Sample item: Q21014908, Karol Świetlik
  • Sample text: Karol Świetlik (ur. 4 listopada 1924, zm. 14 września 2006)
  • Format (ur. DD MMM YYYY, zm. DD MMM YYYY)

--- Jura 16:28, 13 October 2015 (UTC)

I disagree. User:Jura1 has not pointed out any ability to distinguish Julian and Gregorian calendars, so these dates are not suitable for dates before 1923 or so. Jc3s5h (talk) 15:33, 23 October 2015 (UTC)
Have a look at the sample .. --- Jura 15:35, 23 October 2015 (UTC)
The dates should be Gregorian, but mistakes happen. Sometimes there are both dates Julian and Gregorian like in George Washington (Q23) and Johann Sebastian Bach (Q1339). Paweł Ziemian (talk) 21:15, 25 October 2015 (UTC)
The example is useless because the subject of the article is Polish and Poland had already adopted the Gregorian calendar by 1924. Jc3s5h (talk) 21:27, 25 October 2015 (UTC)

VIAF Identifiers for Dutch streets[edit]

There are about 900 items like Q19302580 for individual streets in the Netherlands. They all are instance of (P31)  street (Q79007), country (P17)  Netherlands (Q55), have dutch label "straat in ...", no sitelinks at all and additionally properties P276, P969, P625 and P281.

VIAF has got notice of these items and wrongly assigned them to (mostly geographic) entities. Recently these assignments had been imported here, resulting in P214 with reference stated in (P248)  Virtual International Authority File (Q54919) and retrieved (P813) some time this year. To remedy that situation on both sides P214 should be deleted on these items and then set to "novalue" (VIAF will take this as a hint to dissassociate the WD item from its clusters). -- Gymel (talk) 19:31, 16 October 2015 (UTC)

Creating family name items[edit]

To allow sorting players in squad navigational boxes according to their family name (as done in de:Vorlage:Navigationsleiste Kader von Montreal Impact), all player items contained in the squad items (e.g. roster of Montreal Impact (Q21011427)) need to have the family name (P734) specified. In many cases, the corresponding family name items do not exist and need to be created manually. This is, however, an annoying task.
Therefore, I ask to have family name items of football players created by a bot. --Leyo 23:15, 16 October 2015 (UTC)

@Leyo: I have an idea how to do this and I'm currently thinking about it. I will get back to you soon. Yellowcard (talk) 16:48, 28 October 2015 (UTC)
If you compile a list, I could do that for you. --- Jura 23:42, 30 October 2015 (UTC)
Thank you. Let's wait for Yellowcard's proposition. --Leyo 21:26, 2 November 2015 (UTC)

Importing data from template Bio to Wikidata[edit]

Wiki itwiki
Items without P569 count (all)
As of Oct 12 38009 (13 %)

Template:Bio (Q19610561) on itwiki contains many data that can be imported into Wikidata. It has the following parameters

--★ → Airon 90 12:00, 17 October 2015 (UTC)

I partly disagree. User:Airon90 has not pointed out any ability to distinguish Julian and Gregorian calendars, so these dates are not suitable for dates before 1923 or so. Jc3s5h (talk) 15:34, 23 October 2015 (UTC)
  • Pictogram voting comment.svg Comment I tried P19 through LuogoNascitaLink with harvesttemplates. For 1700 (of 50000) items, it took 30 minutes, adding only 8 (Maybe the ones that come first have already been checked). --- Jura 13:57, 6 November 2015 (UTC)

Connecting pages[edit]

On and are several thousands of homonymous categories not connected in wikidata. It's necessary to check cross-wiki if pagename exists on +, and also on — then connect these pages in one item - in 99% of cases they are about taxons. --XXN, 12:34, 22 October 2015 (UTC)

Example of pages to be interconnected

--XXN, 00:44, 3 November 2015 (UTC)

Wikispecies sitelinks[edit]

We can add sitelinks to Wikispecies now. There are hundreds of thousands of pages, so a bot would definitely be useful. I've listed a few ways I think a bot could add sitelinks below, feedback would be welcome.

Pinging @Magnus Manske: (who asked about it on the Wikidata mailing list) and @Dan Koehl, Koavf, MPF, Pigsonthewing: (from Wikidata:Wikispecies/Get involved) Josve05a (talk)
FelixReimann (talk)
Infovarius (talk)
Daniel Mietchen (talk)
Soulkeeper (talk)
Brya (talk)
Klortho (talk)
Delusion23 (talk)
Andy Mabbett (talk)
Dan Koehl (talk)
Achim Raschka (talk)
Rod Page
Joel Sachs

Pictogram voting comment.svg Notified participants of Wikiproject Taxonomy

- Nikki (talk) 14:23, 22 October 2015 (UTC)


There are about 500 pages linked to Wikidata via species:Template:Wikidata. A bot could extract the item ID from the template and then add the sitelink to that item.

If no one else is already doing this one, I would like to try it. --Tobias1984 (talk) 18:11, 22 October 2015 (UTC)
Slowly adding the links: Special:Contributions/Checkallthestrings_bot --Tobias1984 (talk) 19:06, 22 October 2015 (UTC)
✓ Done 138 items according to my bot. --Tobias1984 (talk) 20:57, 22 October 2015 (UTC)

Disambiguation pages[edit]

There are about 3000 disambiguation pages in species:Category:Disambiguation pages. A bot could find a matching Wikidata item by looking for an item marked as instance of (P31) Wikimedia disambiguation page (Q4167410) which has the same English(?) label.

Interwiki links[edit]

Some pages (no idea how many) have interwiki links in the page text, e.g. species:Nicolaus Joseph von Jacquin. A bot could find the Wikidata item the linked pages are connected to and add the sitelink to that item.

It looks like User:Reinheitsgebot is doing this now. - Nikki (talk) 03:21, 23 October 2015 (UTC)
Yup. I don't remove the interwiki links from the wikitext, though. If anyone has a bot for that, please clean up after me ;-) --Magnus Manske (talk) 12:11, 23 October 2015 (UTC)
Thanks! :) Can this subsection be marked as done now? Regarding the interwiki links in the wikitext, I know of one global bot (meta:User:YiFeiBot) and I asked about it last night on meta:User_talk:Zhuyifei1999#Wikispecies.2C_Meta-Wiki_and_Mediawiki.org_interwiki_links, it seems it will need approval first. I don't know if there are any others. - Nikki (talk) 15:54, 23 October 2015 (UTC)

Page names matching taxon names[edit]

The taxon pages have page names matching the taxon name. A bot could go through all pages using species:Template:Taxonav (about 459,000) and try to find a Wikidata item which has a taxon name (P225) statement matching the page name.  – The preceding unsigned comment was added by ? (talk • contribs).

A list of non-unique values is at Property talk:P225.  – The preceding unsigned comment was added by Jura1 (talk • contribs) at 13:30, 23 October 2015‎ (UTC).
Doing that ATM. --Magnus Manske (talk) 15:00, 23 October 2015 (UTC)


Some tens of Wikipedias use en:Template:Wikispecies and en:Template:Sister project links linking to Wikispecies. It should be easy for a bot to find the Wikidata item linked to the Wikipedia page using the template and the Wikispecies page linked by the template.


Wikispecies has pages about journals, with titles in the format species:ISSN_0035-418X. We can match the ISSNs. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:57, 23 October 2015 (UTC)

✓ Done There are 1569 pages in species:Category:ISSN. I managed to find matches for 630 of them using ISSN (P236), but it seems sometimes items have multiple ISSNs (see Wikidata_talk:Wikispecies#Multiple_ISSNs for that). There are also a few pairs where the sitelink is on one item and a matching ISSN was found on another (so possibly duplicates, possibly mistakes):
- Nikki (talk) 15:43, 23 October 2015 (UTC)
AIUI, journals that change title are supposed to use a different ISSN. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:51, 23 October 2015 (UTC)


Wikispecies has pages about taxon authors. We can match on author abbreviations, and/or authority control values in species:Template:Authority control. Magnus has done a number of these already. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:57, 23 October 2015 (UTC)

species:Template:Authority control doesn't look like it would be useful for a bot to find sitelinks. It's only used 33 times and there are only 2 pages which aren't already linked, species:Donald L.J. Quicke and species:John Richard Spence. - Nikki (talk) 14:15, 23 October 2015 (UTC)
I'm surprised it's so low; but we now have a full set, as I've manually created the two you mention. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 17:36, 23 October 2015 (UTC)
I think you might have missed the 24K or so that are not associated with a Wikidata item at the moment. --Magnus Manske (talk) 19:15, 23 October 2015 (UTC)
I understood Andy's reply to mean that all the pages using the template we were talking about are now linked to Wikidata. There are still lots of pages in species:Category:Taxon Authorities which aren't using that template, but the template obviously can't help us there. - Nikki (talk) 20:08, 23 October 2015 (UTC)
Indeed so. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 20:33, 23 October 2015 (UTC)

Swedish districts[edit]

All pages using sv:Template:Infobox distrikt Sverige should have statements like:

instance of (P31):civil registry district of Sweden (Q18333556) (qualifier:startdate 2016-01-01) [source:stated in (P248):SFS 2015:493 (Q21199810)]
country (P17):Sweden (Q34)

Not all articles have items yet.

N.B. Do NOT add any statements about district code yet from that template. The division is already revisoned and the articles are not corrected yet. The number of districts have decreased with 1 and therefor many districts north of Scania in the articles have a number that is to high. -- Innocent bystander (talk) 08:44, 26 October 2015 (UTC)

Convert Tinyurl to template[edit]

Currently a few queries are stored at tinyurl (samples). Could these be retrieved and stored directly on the pages instead? --- Jura 23:41, 30 October 2015 (UTC)

Welsh books[edit]

In using the Wikidata game to find possible humans I'm finding that there are a lot of items for books on the Welsh Wikipedia that have no instance of or other information here, but with very complete infoboxes. I think it would therefore be useful (if possible) for a bot to import this data.

If an article at has cy:Nodyn:Gwybodlen llyfr then the following can be added to it's wikidata item:

then from the infobox

I'm not sure anything else (e.g. original language of work (P364), publisher (P123), author (P50)) can be reliably imported. Thryduulf (talk: local | en.wp | en.wikt) 20:23, 2 November 2015 (UTC)

Religious buildings by year[edit]

A couple of hundred category-items like Category:Religious buildings completed in 1778 (Q8649491) are missing Commons category (P373) and their site-link to Commons. It should be fairly easy to iterate over them all from 1528 (which is the earliest I've found) to 2015 and add "Category:Religious buildings built in $YEAR". /ℇsquilo 08:26, 6 November 2015 (UTC)

Remove "was a", "is a" etc at the beginning of English descriptions[edit]

A few description start with "is a", samples:

etc. That part of most descriptions can be removed. --- Jura 10:15, 7 November 2015 (UTC)

If you have a text editor with an advanced enough find and replace function, I think this would be pretty easy to do by selecting the IDs and descriptions using Quarry, removing the "is a" (etc) with a find/replace that only matches the start of the description and then use QuickStatements to do the edits. There might be a way to do the find/replace step in SQL instead of in a text editor, but I'm not very familiar with MySQL and a quick search suggests MySQL doesn't have an equivalent of PostgreSQL's regexp_replace. - Nikki (talk) 10:01, 9 November 2015 (UTC)
Something like could do. I had fixed once a few of these.
As it's probably a re-occurring task, maybe someone wants to do some as well. --- Jura 10:34, 9 November 2015 (UTC)


Hello.Is it possible to connect 7٬574 Categories with wiki data by a bot?Thanks --ديفيد عادل وهبة خليل 2 (talk) 10:16, 7 November 2015 (UTC)

ديفيد عادل وهبة خليل 2:
If you want to do it yourself, you could use QuickStatements (Q20084080) for it ("Scommonswiki").
List of items is here.--- Jura 10:29, 7 November 2015 (UTC)

@Jura1: I do not want to add a property,but linking commons categorie pages with items.Thank you --ديفيد عادل وهبة خليل 2 (talk) 11:14, 7 November 2015 (UTC)

ديفيد عادل وهبة خليل 2: You can do that with QuickStatements, just use "Scommonswiki" instead of a property id. Category name needs to be in double quotes (").--- Jura 11:18, 7 November 2015 (UTC)
@Jura1: Please link pages by yourself because I do not understand this Page.Thank you --ديفيد عادل وهبة خليل 2 (talk) 11:28, 7 November 2015 (UTC)
Be nice to beginners, Jura1. QuickStatements is damn far from intuitive to use. /ℇsquilo 19:47, 7 November 2015 (UTC)
ℇsquilo: It seems like a good task to get started with QuickStatements .. it occurred to me afterwards, that I could have explained it in more simple way. Anyways, maybe you want to have a go? --- Jura 12:15, 8 November 2015 (UTC)
I'd love to use it for #Religious buildings by year above, but I have no idea on how to get started. /ℇsquilo 10:31, 10 November 2015 (UTC)
To deal with QuickStatements you need either an advanced text editor (e.g. Notepad++, Atom) or a spreadsheet application (e.g. Excel) to format you input. If you have excel, open a blank workbook, and add fill in the following data
1 Category:Religious buildings completed in 1528 P373 "Category:Religious buildings built in 1528"
2 Category:Religious buildings completed in 1529 P373 "Category:Religious buildings built in 1529"
Then select cells A1-C2 and drag it to row 488. Excel will automatically ajust the year so you see in row 488:
488 Category:Religious buildings completed in 2015 P373 "Category:Religious buildings built in 2015"
Select now all cells from A1-C488 and copy it to QuickStatements. In the input field "First column are articles from" add enwiki and click Do it. --Pasleim (talk) 11:10, 10 November 2015 (UTC)
I think your sample should read "Scommonswiki" instead of "P373". --- Jura 11:54, 10 November 2015 (UTC)

enwiki Persondata[edit]

FYI Wikidata_talk:Primary_sources_tool#Migration_of_enwiki_Persondata. --- Jura 10:43, 7 November 2015 (UTC)

Refresh pages[edit]

Could we refresh some of the older pages?

However, a count on page_props shows that the information is available for only 16,846,670 items today. At least 1.8 million items are missing it.

With a refresh, the information should get populated.

This would help identify items without claims (more likely to not have been edited recently). --- Jura 01:38, 9 November 2015 (UTC) --- Jura 10:47, 9 November 2015 (UTC)

I will try to do this myself. --- Jura 07:12, 13 November 2015 (UTC)

USA population[edit]

I would like USA populations data for USA communes. Source:

Thank You! --B.Zsolt (talk) 10:46, 10 November 2015 (UTC)

Data from English Wikipedia chemboxes, drugboxes and relevant infoboxes[edit]

Hi everyone

While collecting data for our data platform I scraped English Wikipedia's chemboxes, drugboxes and mineral/drug/element infoboxes to include this data on our platform. These are an excellent source of information and we link back to the corresponding English Wikipedia page--see here.

I thought it might be useful to contribute this data back to Wikidata so that no one else will have to go through this same process. Would the Chemistry or Physics communities here be interested? Also, would anyone be willing to write the bot to do this? I have formatted json files that should be easy to work with.

Thanks! Jo (talk) 22:03, 11 November 2015 (UTC)

By Wikipedia you certainly refer to all language versions of Wikipedia. Note that en.wikipedia is not the Wikipedia. --Leyo 00:46, 12 November 2015 (UTC)
My apologies, the data I have is from the English Wikipedia only. I have updated my post to reflect this. --Jo (talk) 00:31, 13 November 2015 (UTC)
Please first get agreement from the Wikidata: Wikiproject Chemistry. I don't want to see data added again in the wrong items. Thanks Snipre (talk) 12:26, 20 November 2015 (UTC)

Move nobility titles from occupation to royal or noble rank[edit]

There are 475 entities at the moment that have a subclasses of royal or noble rank (Q355567) as occupation (P106) which should have been moved to noble title (P97) maintaining all of their current qualifiers (if any). The tools I know to do this allow me to move the elements from one property to the other, but I can't keep the qualifiers, which is not an option in this situation. -- Agabi10 (talk) 10:56, 13 November 2015 (UTC)

Do you have a list of these 475 entities? With the WDQ claim[106:(tree[355567][][279])] I only find 20 items. --Pasleim (talk) 11:00, 13 November 2015 (UTC)
Pasleim It is the number that appears in the Constraint report -- Agabi10 (talk) 11:34, 13 November 2015 (UTC)

Import names in Latin script from kowiki[edit]

There are a few items for persons that link only to kowiki and don't have labels in English. Samples:

The articles for these in kowiki have names in Latin script (or other scripts) defined in the introduction.

This could be imported to Wikidata as label or alias.

The two samples were already merged as they appeared on the report for identical birth and death dates. --- Jura 10:54, 14 November 2015 (UTC)

Transfer images from SvWP[edit]

The category sv:Kategori:Ingen bild på Wikidata men det finns lokalt (No image on Wikidata but exists locally) contains ~3000 articles with images where there are none on Wikidata. Could those be transferred to Wikidata with a bot? Preferably the one in any infobox (as those usually are the most representative images). Parameter names in infoboxes for images are image or bild. For extra security you could skip those that only has images in a gallery and has the template robotskapad as it is a big chance those images has been added by bot without any manual check. -- 09:20, 19 November 2015 (UTC)

Symbol oppose vote.svg Oppose Lsjbot added a lot of wrong pictures to articles with taxa. --Succu (talk) 09:35, 19 November 2015 (UTC)
Those will be skipped if you skip those with the template robotskapad and that only has images in galleries. --Averater (talk) 10:02, 19 November 2015 (UTC)
I did check ~50 of the ~700 bot generated articles with images now and found just one error where there was a mixup of a plant and a worm with the same name. However, that one is of course fixed now. So there will not be many "wrong" images even if the bot-added images are included. --Averater (talk) 11:18, 19 November 2015 (UTC)
For general image additions, is most helpful and fairly quick despite not being a bot. --- Jura 13:45, 19 November 2015 (UTC)
Thank you! That was a wonderful little tool. --Averater (talk) 15:35, 19 November 2015 (UTC)

The category has updated itself somehow and now contains ~14000 articles.I have checked ~1000 articles and it seems that around 2 % if the bot-added images are wrong. Some are due to bad iw-links or due to wrong images on other languages but most are due to multiple genera with the same name. If those should be skipped it is not enough to skip only the image in galleries but those articles should be skipped completely. --Averater (talk) 14:39, 21 November 2015 (UTC)

Transform references to Dictionary of Art Historians[edit]

Thanks to mix'n'match, we already have items for all entries in the Dictionary of Art Historians (Q17166797). They are currently referred to with described by source (P1343) > Dictionary of Art Historians (Q17166797) with section, verse, or paragraph (P958) > [Dictionary of Art Historians ID].

Could somebody please make their bot remove these statements and replace them with Dictionary of Art Historians (P2332) > [Dictionary of Art Historians ID]? Thanks! Jonathan Groß (talk) 10:04, 23 November 2015 (UTC)

@Magnus Manske: Jonathan Groß (talk) 10:10, 23 November 2015 (UTC)

On it. --Magnus Manske (talk) 14:09, 23 November 2015 (UTC)
Done. There are ~15 items where the mix'n'match updater refused to set the value, because they are duplicates in DAH. If someone could do the honors... --Magnus Manske (talk) 14:47, 23 November 2015 (UTC)