Wikidata:Bot requests

From Wikidata
Jump to: navigation, search





for permissions


for deletions


for deletion

for comment

and imports

a query


Bot requests
If you have a bot request, create a new section here and tell exactly what you want. You should discuss your request first and wait for the decision of the community. Please refer to previous discussion. If you want to request sitelink moves, see list of delinkers.

For botflag requests, see Wikidata:Requests for permissions.

Tools available to all users which can be used to accomplish the work without the need for a bot:

  1. PetScan for creating items from Wikimedia pages and/or adding same statements to items
  2. QuickStatements for creating items and/or adding different statements to items
  3. Harvest Templates for importing statements from Wikimedia projects
On this page, old discussions are archived. An overview of all archives can be found at this page's archive index. The current archive is located at 2016/10.
Filing cabinet icon.svg
SpBot archives all sections tagged with {{Section resolved|1=~~~~}} after 2 days.


Cyrillic merges[edit]

This included pairs of items with articles at ruwiki and ukwiki each (Sample: Q15061198 / Q12171178). Maybe it's possible to find similar items merely based on labels in these languages and merge them. --- Jura 03:33, 19 September 2015 (UTC)

I cannot find any ru-uk pairs. Are they all done? --Infovarius (talk) 16:27, 3 November 2015 (UTC)
The ones on that list are identified based on dates of birth/death and we regularly go through them. The occasional findings there (also with ru/be) suggest that there are more (without dates). A query would need to be done to find them. --- Jura 16:33, 3 November 2015 (UTC)
Today the list includes quite a few, thanks to new dates of birth/death being added. --- Jura 16:43, 2 December 2015 (UTC)
A step could involve reviewing suggestions for missing labels in one language based on labels in another languages with Add Names as labels (Q21640602): sample be/ru. --- Jura 11:44, 6 December 2015 (UTC)
I came across a few items that had interwikis in ukwiki to ruwiki, but as they were on separate items, these weren't used to link the articles to existing items (sample, merged since). --- Jura 10:17, 15 December 2015 (UTC)
SELECT DISTINCT ?item ?Spanishlabel ?item2 ?Italianlabel
  	VALUES ?item { wd:Q19909894 }
  	?item wdt:P31 wd:Q5 .

    VALUES ?item2 { wd:Q16704775 }
  	?item2 wdt:P31 wd:Q5 .

    ?item rdfs:label ?Spanishlabel . FILTER(lang(?Spanishlabel)="ru")
	BIND(REPLACE(?Spanishlabel, ",", "") as ?Spanishlabel2)

    ?item2 rdfs:label ?Italianlabel . FILTER(lang(?Italianlabel)="uk")

    FILTER(str(?Spanishlabel2) = str(?Italianlabel))
  	FILTER(str(?Spanishlabel) != str(?Italianlabel))

#added by Jura1

SPARQL query

The above currently finds one pair. It times out when not limited to specific items ;) Maybe there is a better way to find these.
--- Jura 14:19, 3 April 2016 (UTC)

In the meantime the two items were merged, so it doesn't work anymore.
--- Jura 16:54, 4 April 2016 (UTC)
See also User:Pasleim/projectmerge/ruwiki-ukwiki. XXN, 08:22, 8 September 2016 (UTC)

Sorting flags by level of government[edit]

Hello. I'm trying to control constraint violations for applies to territorial jurisdiction (P1001). Could someone please:

  1. For items in w:en:Category:National flags, change instance of flag (Q14660) to national flag (Q186516).
  2. For items in w:en:Category:United States state flags could you please change instance of flag (Q14660) or national flag (Q186516) to flag of a country subdivision (Q22807280).
  3. For items in subcategories of w:en:Category:Flags of cities by country, change instance of flag (Q14660) to flag of a municipality (Q22807298).

Thank you! --Arctic.gnome (talk) 20:59, 15 February 2016 (UTC)

@Arctic.gnome: Sorry for the delay, I'm ready to do this. Could please just in case provide a rationale why it is okay to do this task? Matěj Suchánek (talk) 13:32, 16 April 2016 (UTC)
@Arctic.gnome: please respond to this section, otherwise it will be archived. Sjoerd de Bruin (talk) 08:51, 12 October 2016 (UTC)
@Matěj Suchánek: Since making this request I've changed the constrants of applies to territorial jurisdiction (P1001), so this is no longer strictly necessary. It might still be a good idea for pages to diferentiate between national flags, sub-national flags, and non-government flags. But I don't know whether or not "instance of" is the correct place to do that. On some types of pages, the "instance of" property uses a very detailed value, while in other cases (like "human" and "position") we only use a general term for the instance property. --Arctic.gnome (talk) 15:43, 12 October 2016 (UTC)

Taxon labels[edit]

For items where instance of (P31)=taxon (Q16521), and where there is already a label one one or more languages, which is the same as the value of taxon name (P225), the label should be copied to all other empty, western alphabet, labels. For example, this edit. Please can someone attend to this? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 15:11, 10 March 2016 (UTC)

Do you mean label or alias? I would support the latter where there is already a label and that label is not already the taxon name. --Izno (talk) 17:03, 10 March 2016 (UTC)
No, I mean label; as per the example edit I gave. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 20:16, 10 March 2016 (UTC)
See your last request: Wikidata:Bot_requests/Archive/2015/08#Taxon_names. --Succu (talk) 18:57, 10 March 2016 (UTC)
Which was archived unresolved. We still have many thousands of missing labels. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 20:16, 10 March 2016 (UTC)
Nope. There is no consensus doing this. Reach one. --Succu (talk) 20:22, 10 March 2016 (UTC)
You saying "there is no consensus" does not mean that there is none. Do you have a reasoned objection to the proposal? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 20:56, 10 March 2016 (UTC)
Go back and read the linked discussions. In the nursery of wikidata some communities had strong objections. If they changed their mind my bot can easily execute this job. --Succu (talk) 21:19, 10 March 2016 (UTC)
So that's a "no" to my question, then. I read the linked discussions, and mostly I see people not discussing the proposal, and you claiming "there is no consensus", to which another poster responded "What I found, is a discussion of exactly one year old, and just one person that is not supporting because of 'the gadgets then need to load more data'. Is that the same 'no consensus' as you meant?". There are no reasoned objections there, either. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 21:24, 10 March 2016 (UTC)
For the lazy ones:
--Succu (talk) 21:53, 10 March 2016 (UTC)
I already done for Italian label in past. Here other two propose: May 2014 and March 2015 --ValterVB (talk) 09:54, 11 March 2016 (UTC)
@ValterVB: Thank you. Can you help across any other, or all, western-alphabet languages, please? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 13:18, 16 March 2016 (UTC)
Yes I can do it, but before to modify 2,098,749 items I think is necessary to have a strong consensus. --ValterVB (talk) 18:14, 16 March 2016 (UTC)
@ValterVB: Thank you. Could you do a small batch, say 100, as an example, so we can then ask on, say, Project Chat? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 15:03, 18 March 2016 (UTC)
Simply ask with the example given by you. --Succu (talk) 15:16, 18 March 2016 (UTC)

──────────────────────────────────────────────────────────────────────────────────────────────────── @Pigsonthewing:

  • Test edit: Q14945671, Q21444273, Q2508347, Q25247.
  • Languge: "en","de","fr","it","es","af","an","ast","bar","br","ca","co","cs","cy","da","de-at","de-ch","en-ca","en-gb","eo","et","eu","fi","frp","fur","ga","gd","gl","gsw","hr","ia","id","ie","is","io","kg","lb","li","lij","mg","min","ms","nap","nb","nds","nds-nl","nl","nn","nrm","oc","pcd","pl","pms","pt","pt-br","rm","ro","sc","scn","sco","sk","sl","sr-el","sv","sw","vec","vi","vls","vo","wa","wo","zu"
  • Rule:

Very important: is necessary verify if the list of languages is complete. Is the same that I use for disambiguation item. --ValterVB (talk) 09:42, 19 March 2016 (UTC)

    • I really don't like the idea of this. The label, according to Help:Label, should be the most common name. I doubt that most people are familiar with the latin names. Inserting the latin name everywhere prevents language fallback from working and stops people from being shown the common name in another language they speak. A very simple example, Special:Diff/313676163 added latin names for the de-at and de-ch labels which now stops the common name from the de label from being shown. - Nikki (talk) 10:29, 19 March 2016 (UTC)
      • @Nikki: The vast majority of taxons have no common name; and certainly no common name in every language. And of course edits can subsequently be overwritten if a common name does exist. As for fallback, we could limit this to "top level" languages. Would that satisfy? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:02, 19 March 2016 (UTC)
        • As far as I'm aware most tools rely on the absence of certain information. Adding #10,000 csv file of Latin / Welsh (cy) species of birds. would be rendered to handcraft. --Succu (talk) 23:11, 19 March 2016 (UTC)
          • Perhaps this issue could be resolved by excluding certain groups? Or the script used in your example could overwrite the label if it matches the taxon name? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 11:14, 23 March 2016 (UTC)
        • It may be the case that most taxon items won't have a common name in any language, but I don't see anything here which is only trying to target the taxon items which have no common names. Adding the same string to lots of labels isn't adding any new information and as Succu pointed out, doing that can get in the way (e.g. it makes it more difficult to find items with missing labels, it can get in the way when merging (moving common names to the aliases because the target already has the latin name as a label) and IIRC the bot which adds labels for items where a sitelink has been recently added will only do so if there is no existing label). To me, these requests seem like people are trying to fill in gaps in other languages for the sake of filling in the gaps with something (despite that being the aim of the language fallback support), not because the speakers of those languages think it would be useful for them and want it to happen (if I understand this correctly, @Innocent bystander: is objecting to it for their language). - Nikki (talk) 22:40, 22 March 2016 (UTC)
          • Yes, the tolerance against bot-mistakes is limited on svwiki. Mistakes initiated by errors in the source is no big issue, but mistakes initiated by "guesses" done by a bot is not tolerated at all. The modules we have on svwiki have no problem handling items without Swedish labels. We have a fallback-system which can use any label in any language. -- Innocent bystander (talk) 06:39, 23 March 2016 (UTC)
            • @Innocent bystander: This would not involve an "guesses". Your Wikipedia's modules may handle items without labels, but what about third-party reusers? Have you identified any issues with the test edits provided above? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 11:14, 23 March 2016 (UTC)
              • No, I have not found any issue in the examples. But this is not my subject, I would not see an issue even if it was directly under my nose. Adding correct statements for Scientific names and Common names looks more important here for the third party users than labels, which cannot be sourced. NB, the work of Lsjbot have done that Swedish and Cebuano probably have more labels than any other language in the taxon set. You will not miss much by excluding 'sv' in this botrun. -- Innocent bystander (talk) 07:00, 24 March 2016 (UTC)
                • If a taxon name can be sourced, then by definition so can the label. If you have identified no errors, then your reference to "guesses" is not substantiated. true, adding for Scientific names and Common names is important, but the two tasks are not mutually exclusive, and their relative importance is subjective. To pick one example at random, from the many possible, Dayus (Q18107066) currently has no label in Swedish, and so would benefit from the suggested bot run. indeed, it currently has only 7 labels, all the same, and all using the scientific name. Indeed, what are the various European language's common name for this mainly Chinese genus? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:34, 25 March 2016 (UTC)
          • No, this is not "trying to fill in gaps in other languages for the sake of filling in the gaps". Nor are most of the languages affected served by fallback. If this task is completed, then "find items with missing labels" will not be an issue for the items concerned, because they will have valid labels. Meanwhile, what is the likelihood of these labels being provided manually? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 11:14, 23 March 2016 (UTC)
            • If this is not trying to fill in the gaps for the sake of filling in the gaps, what problem is it solving and why does language fallback not help? (I'm sure the development team would be like to know that language fallback is not working properly). The taxonomic names are not the preferred labels and valid is not the same as useful (adding "human" as the description for humans with no description was valid, yet users found it annoying and useless and they were all removed again), the labels for a specific language in that language are still missing even if we make it seem like they're not by filling in all the gaps with taxonomic names, it's just masking the problem. I can't predict the future so I don't see any point in speculating how likely it is that someone will come along and add common names. They might, they might not. - Nikki (talk) 23:02, 24 March 2016 (UTC)
              • It solves the problem of an external user, making a query (say for "all species in genus X") being returned the Q items with no labels, in their language. This could break third party applications, also. In some cases, there is currently no label in any language - how does language fallback work then? How does it work if the external user's language is Indonesian, and there is only an English label saying, say, "Lesser Spotted Woodpecker"? And, again, taxonomic names are the preferred labels for the many thousands of species - the vast majority - with no common name - or with no common name in a given language. The "human" examples compares apples with pears. This is a proposal to add specific labels, not vague descriptions (the equivalent would be adding "taxon" as a description). Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:26, 25 March 2016 (UTC)
                • Why should an external user query a Wikidata internal called label and not rely on a query of taxon name (P225)? --Succu (talk) 22:04, 25 March 2016 (UTC)
                  • For any of a number of reasons; not least that they may be querying things which are not all taxons. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 18:32, 26 March 2016 (UTC)
                    • Grand answer. Maybe they are searching the labels for aliens, gods, fairy tales or something else? A better solution would be if the Wikibase could be configured to take certain properties like as taxon name (P225) or title (P1476) as a default value as a language independent label. --Succu (talk) 21:09, 27 March 2016 (UTC)
                      • Maybe it could. But it is not. That was suggested a year or two ago, in the discussions you cited above, and I see no move to make it so, no any significant support for doing so. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 21:19, 27 March 2016 (UTC)
                        • So what? Did you reached an agreement with svwiwki, cebwiki, warwiki, viwiki or nlwiki we should go along your proposed way? --Succu (talk) 21:43, 27 March 2016 (UTC)
    • @ValterVB: Thank you. I think your rules are correct. I converted the Ps &Qs in your comment to templates, for clarity. Hope that's OK. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:02, 19 March 2016 (UTC)
  • Symbol oppose vote.svg Oppose That majority of taxons does not have a common name, does not mean that all western languages should automatically use the scientific name as label. Matěj Suchánek (talk) 13:23, 16 April 2016 (UTC)
    • Nobody is saying "all western languages should automatically use the scientific name as label"; if the items already have label, it won't be changed. If a scientific label is added as a label, where none existed previously, and then that label is changed to some other valid string, the latter will not be overwritten. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:31, 20 April 2016 (UTC)

We seem to have reached as stalemate, with the most recent objections being straw men, or based on historic and inconclusive discussions. How may we move forward? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 18:28, 16 May 2016 (UTC)

That's simple: drop your request. --Succu (talk) 18:33, 16 May 2016 (UTC)
Were there a cogent reason to, I would. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 15:57, 17 May 2016 (UTC)
Anyone? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:04, 10 September 2016 (UTC)
@Pigsonthewing: I'll support the proposal if it is limited to major languages that don't have other fallbacks. For most taxons, the scientific name is the only name, and even for taxons with a common name, having the scientific name as the label is better than having no label at all. I'm reluctant to enact this for a huge number of languages though, as it might make merges (which are commonly needed for taxons) a pain to complete. Kaldari (talk) 23:02, 28 September 2016 (UTC)
@Kaldari: Thank you. Please can you be more specific as to what you mean by "major languages that don't have other fallbacks"? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 10:35, 29 September 2016 (UTC)
@Pigsonthewing: Maybe just the biggest Latin languages: English, German, Spanish, French, Portuguese, Italian, Polish, Dutch. Kaldari (talk) 18:29, 29 September 2016 (UTC)
I'm not sure why we'd limit ourselves to them, but if we can agree they should be done, let's do so. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 18:40, 29 September 2016 (UTC)
@Kaldari: Did you see my reply? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:48, 10 October 2016 (UTC)

Symbol oppose vote oversat.svg Strong oppose As said before...--Succu (talk) 22:02, 10 October 2016 (UTC)

What you actually said was "There is no consensus doing this. Reach one.". My reply was "You saying 'there is no consensus' does not mean that there is none. Do you have a reasoned objection to the proposal?", and you provided none then, nor since. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 13:11, 11 October 2016 (UTC)

Add labels from sitelinks[edit]

There used to be a bot that added labels based on sitelinks (enwiki sitelink => en label). I think it stopped running at some point. Maybe some alternative should be found.
--- Jura 08:32, 8 April 2016 (UTC)

I have seen, that Pasleim's bot is doing some job in this area, at least for German and French. --Edgars2007 (talk) 16:20, 9 April 2016 (UTC)
I do it for all the languages, but only for item that have one of these values in instance of (P31):

There is the problem with uppercase/lowercase --ValterVB (talk) 16:30, 9 April 2016 (UTC)

Another rule that I use: add label if the first letter of sitelink is one of this list:
  • (
  • !
  • ?
  • "
  • $
  • '
  • ,
  • .
  • /
  • 0
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9

If you have other suggestion I can add it --ValterVB (talk) 16:41, 9 April 2016 (UTC)

  • Pictogram voting comment.svg Comment Just to make sure this is clear: this is mainly for items that exist and where someone added manually a sitelink to, e.g., enwiki, but the items doesn't have a label in the corresponding language yet. It does not concern items that don't have an English label, but no sitelink to English. I don't think search finds such items if they have no label defined at all. It's key that at least a basic label is defined for such items.
    If you are looking for rules to implement, then try the ones used by PetScan (Q23665536). It mainly removes disambiguators in round brackets. I think this works fine for Wikipedia. A large amount of pages are created that way. It might not work well for Wikisource.
    --- Jura 10:50, 10 April 2016 (UTC)
Jura, these rules are applied only on item that have a sitelink but don't have a label in language of the sitelinki. I check for all the sitelink that end with "wiki", excep "commonswiki", "wikidatawiki", "specieswiki", "metawiki" and "mediawikiwiki" and I delete disambiguation with parenthesis. --ValterVB (talk) 12:13, 10 April 2016 (UTC)

{{Section resolved|Sjoerd de Bruin (talk) 07:45, 19 September 2016 (UTC)}}

I think we still should try to do something about this. If we want to be purist and not add them as labels, I think they should be added at least as alias.
--- Jura 13:32, 20 September 2016 (UTC)

Maybe we should move the discussion to WD:PC or WD:RFC to get community consensus on one or the other solution. --Pasleim (talk) 13:59, 20 September 2016 (UTC)

Import GNIS ID (P590) from en:Template:Infobox settlement[edit]

We have started to use that property on frwiki, through Template:Infobox settlement (Q23779748), and as you can see in fr:Catégorie:Page utilisant P590. Thank you to any bot who would do this! Thierry Caro (talk) 04:33, 11 April 2016 (UTC)

Do you have some examples where it hasn't been imported? I already added thousands of those a couple of months ago. - Nikki (talk) 18:45, 12 April 2016 (UTC)
I was about to mention some place like Cheraw (Q1070214), but apparently you've found this and have already added the data. Thank you! Thierry Caro (talk) 17:16, 13 April 2016 (UTC)
The same thing with FIPS 55-3 (locations in the US) (P774) would be awesome, by the way. Thierry Caro (talk) 17:17, 13 April 2016 (UTC)

Oh! I found an example of missing GNIS ID (P590). Chincoteague (Q1073686) for instance. Thierry Caro (talk) 17:19, 13 April 2016 (UTC)

@Thierry Caro: Thanks! :) It turns out I hadn't actually finished importing the ones I'd started to import before and I'd missed some. Most of them should be there now, we now have twice as many as before. :D There's still a few hundred left that I'm going to look at once the GNIS site is back up (they're quite inconsistently represented in Wikipedia so there could be more I missed, so feel free to continue linking me to any that are missing if you want to).
Regarding FIPS codes, how urgent is it? I'm currently trying to write a bot to add information from GNIS and the US census data (and hopefully also look for information that's wrong or add missing references if I get that far). It looks like the data includes FIPS codes, so I should be able to add them from there once I'm far enough with the bot that I can add data. That would be easier than trying to extract data from templates (and I could add references too).
- Nikki (talk) 13:07, 16 April 2016 (UTC)
OK. Perfect! i'll wait for the bot, don't worry! Thierry Caro (talk) 13:50, 16 April 2016 (UTC)
@Nikki:, what is the status of this? Sjoerd de Bruin (talk) 07:27, 27 September 2016 (UTC)
Not yet done. :( I've been quite busy with other things and it got neglected. I'm still interested in doing it though and I'm going to try and continue working on it soon. - Nikki (talk) 10:47, 9 October 2016 (UTC)

Add P1082 (population) and P585 (point in time) from PLwiki to Wikidata[edit]

Looks like PLwiki has lots of population information other Wiki does not have. It will be useful to have it for all of us. בורה בורה (talk) 18:23, 12 April 2016 (UTC)

It might be helpful to give some supporting links here, to be sure to get the right information from the right place into the right fields. Can you list one pl-article and one corresponding wikidata-item that is manually filled with the desired information? Than I can see if I can get the information filled by a script in the same way. Edoderoo (talk) 18:26, 16 April 2016 (UTC)
Edoderoo sorry for the late reply. I was on vacation. Take for example the article "Żołynia" in PLwiki. It has a population of 5188 as of 2013. However this information does not exist on Wikidata item (Q2363612). There are thousands of examples like this, but you got the idea... PLwiki is really great on population. Share it with us all. בורה בורה (talk) 10:19, 4 May 2016 (UTC)
It would be better to find a reliable source instead. Sjoerd de Bruin (talk) 07:44, 19 September 2016 (UTC)

Take care of disambiguation items[edit]

Points to cover

Somehow it should be possible to create a bot that handles disambiguation items entirely. Not sure what are all the functions needed, but I started a list on the right side. Please add more. Eventually a Wikibase function might even do that.
--- Jura 13:36, 18 April 2016 (UTC)

Empty disambiguation: Probably @Pasleim: can create User:Pasleim/Items for deletion/Disambiguation . Rules: Item without sitelink, with P31 that have only 1 value: Wikimedia disambiguation page (Q4167410). For the other point my bot alredy do something, (for my bot a disambiguation is an item with P31 that have only 1 value: Wikimedia disambiguation page (Q4167410)). Descriptions I use description used in autoEdit Label: I add the same label for all the latin language only if all the sitelink without disambiguation are the same. With these 2 operation I detect a lot of duplicate: same label+description. For now the list is very long (maybe >10K item) but isn't possible to merge automatically too much errors. Another thing to do is normalize the descriptions, there are a lot of item with not standard description. --ValterVB (talk) 18:02, 18 April 2016 (UTC)
  • Personally, I'm not that much worried about duplicate disambiguation items. Mixes between content and disambiguations are much more problematic. It seems they keep appearing through problems with page moves. BTW, I added static numbers to the points.
    --- Jura 10:06, 19 April 2016 (UTC)
    You will always have duplicate disambiguation items, since svwiki has duplicate disambiguation-pages. Some of these duplicates exists because they cover different topics and some of them exists since the pages otherwise becomes to long. A third category are the bot-generated duplicates. They should be treated as temporary, until a carbon based user has merged them.
    And how are un-normalized descriptions a problem? -- Innocent bystander (talk) 10:58, 19 April 2016 (UTC)
About "un-normalized descriptions": ex I have a disambiguation item with label "XXXX" and description "Wikipedia disambiguation", if I create a new item with label "XXXX" and description "Wikimedia disambiguation" I don't see that already exist an disambiguation item "XXXX", if the description is "normalized" I see immediately the the disambiguation already exist so I can merge it. --ValterVB (talk) 11:10, 19 April 2016 (UTC)
For some fields, this proved quite efficient. If there are several items that can't be merged, as some point, there will be something like "Wikimedia disambiguation page (2)", etc.
--- Jura 12:10, 19 April 2016 (UTC)

Lazy start for point (4): 47 links to add instance of (P31)=Wikimedia disambiguation page (Q4167410) to items without statements in categories of sitelinks on Category:Disambiguation pages (Q1982926): en, sl, simple, lt, bg, es, sv, da, ja, ka, fi, la, bs, ba, ca, eo, ru, hy, et, uk, pl, it, mk, kk, pt, sh, nl, id, el, fr, sk, az, be_x_old, hr, eu, sr, de, tr, be, hu, sq, nn, ro, no, lv, cs, zh,
--- Jura 12:07, 23 April 2016 (UTC)

The biggest problem is to define what pages are disambiguation pages, given names and surnames. For example Backman (Q183341) and Backman (Q23773321). I don't see what is the difference between enwiki and fiwiki links. Enwiki page is in category "surnames" and fiwiki page in categories "disambiguation pages" and "list of people by surname", but the page in fiwiki only contains surnames, so basically it could be in the same item as the enwiki link. --Stryn (talk) 13:10, 23 April 2016 (UTC)

I think people at Wikidata could be tempted to make editorial decisions for Wikipedia, but I don't think it's up to Wikidata to determine what Wikipedia has to consider a disambiguation page. If a language version considers a page to be a disambiguation page, then it should go on a disambiguation item. If it's an article about a city that also lists similarly named cities, it should be on an item about that city. Even if some users at Wikidata attempted to set "capital" to a disambiguation page as Wikipedia did the same, such a solution can't be sustained. The situation for given names and family names isn't much different. In the meantime, at least it's clear which items at Wikidata have what purpose.
--- Jura 14:20, 23 April 2016 (UTC)
You then have to love Category:Surname-disambigs (Q19121541)! -- Innocent bystander (talk) 14:35, 23 April 2016 (UTC)
IMHO: In Wikipedia disambiguation page are page that listing page or possible page that have the same spelling, no assumption should be made about the meaning. If we limit the content to partial sets whith some specific criterion we haven't a disambiguation page but a list (ex. list of person with the same surname List of people with surname Williams (Q6633281). These pages must use tag __DISAMBIG__ to permit bot and human to recognize without doubts a disambiguation from a different item. In Wikidata disambiguation item are item the connect disambiguations page with the same spelling. --ValterVB (talk) 20:02, 23 April 2016 (UTC)

Disambiguation item without sitelink --ValterVB (talk) 21:30, 23 April 2016 (UTC)

I'd delete all of them.
--- Jura 06:13, 24 April 2016 (UTC)

Some queries for point (7):

A better way needs to be found for (7a).
--- Jura 08:07, 25 April 2016 (UTC)

I brought up the question of the empty items at Wikidata:Project_chat#Wikidata.2C_a_stable_source_for_disambiguation_items.3F.
--- Jura 09:39, 27 April 2016 (UTC)

As this is related: Wikidata:Project chat/Archive/2016/04#Deleting descriptions. Note, that other languages could be checked. --Edgars2007 (talk) 10:30, 27 April 2016 (UTC)

I don't mind debating if we should keep or redirect empty disambiguation items (if admins want to check them first ..), but I think we should avoid recycling them for anything else. --- Jura 10:34, 27 April 2016 (UTC)
As it can't be avoided entirely, I added a point 10.
--- Jura 08:32, 30 April 2016 (UTC)
Point (3) and (10) are done. For point (2) I created User:Pasleim/disambiguationmerge. --Pasleim (talk) 19:22, 2 July 2016 (UTC)
Thanks, Pasleim.
--- Jura 05:02, 11 July 2016 (UTC)
  • Matěj Suchánek made User:MatSuBot/Disambig errors which covers some of 7b.
    Some things it finds:
    • Articles that are linked from disambiguation items
    • Disambiguation items that were merged with items for concepts relevant to these articles (maybe we should check items for disambiguation with more than a P31-statement or attempt to block such merges)
    • Pages in languages were the disambiguation category isn't correctly set-up or recognized by the bot (some pages even have "(disambiguation)" in the page title). e.g. Q27721 (36 sitelinks) – ig:1 (disambiguation)
    • Pages in categories close to disambiguation categories. (e.g. w:Category:Set indices on ships)
    • Redirects to non-disambiguations. (e.g. Q37817 (27 sitelinks) idwiki – id:Montreuil – redirects to id:Komune di departemen Pas-de-Calais (Q243036, not a disambiguation)

Seems like an iceberg. It might be easier to check these by language and once the various problems are identified, attempt to sort out some automatically.
--- Jura 05:02, 11 July 2016 (UTC)

Note that my bot only recognizes pages with the __DISAMBIG__ magic word as disambiguations. If you want a wiki-specific approach, I can write a new script which will work only for chosen wikis. Matěj Suchánek (talk) 09:12, 12 July 2016 (UTC)
  • Step #4 should be done for now. The above list now includes links for 160+ sites.
    --- Jura 22:02, 5 August 2016 (UTC)
  • For step #3a, there is now Phab:T141845
    --- Jura 22:30, 5 August 2016 (UTC)
List of disambiguation item with conflict on Label/description --ValterVB (talk) 13:57, 6 August 2016 (UTC)
  • Add #11.
    --- Jura 02:05, 21 September 2016 (UTC)

Exploitation visa number[edit]


Can we add the exploitation visa number (P2755) (number of the exploitation visa of a movie in France) to all movie avaible in the website of the CNC? Maybe the bot can compare the label in french with the title of the movie or the year, the duration, the country, etc.

Then, can it add the CNC film rating (P2758) with :

It's written in the legal notice:

Sauf mention particulière, toute reproduction partielle ou totale des informations diffusées sur le site internet du CNC est autorisée sous réserve d’indication de la source.
Unless otherwise stipulated, any total or partial reproduction of the information published on the CNC website is authorized subject to indication of the source.

--Tubezlob (🙋) 16:58, 26 April 2016 (UTC)

Integrate data about the relationships from the Foundational Model of Anatomy into Wikidata[edit]

Most Anatomical concepts on Wikidata already have information about the Foundational Model of Anatomy ID. On the other hand the lack the information about hypernyms, holonyms and meronym that are found in the Foundational Model of Anatomy ID ontology.

On their website they describe the availability of the database:

The Foundational Model of Anatomy ontology is available under a

Creative Commons Attribution 3.0 Unported License (link is external). It can be accessed through several mechanisms:

1. The latest OWL2 files are available at These can be viewed in the latest version of Protege.

Furthermore I think that valuable infoboxes could be created based on the data from the FMA-ontology within Wikipedia.

ChristianKl (talk) 20:47, 26 April 2016 (UTC)

  • Our database is CC0, not CC3.0 Unported. This would be a copyright problem. --Izno (talk) 14:03, 27 April 2016 (UTC)
    • CC3.0 Unported doesn't require derivitive works to use CC3.0 Unported. It's not a share-alike license. What's required is attribution. If every entry in Wikidata would cite the FMA as the source, Wikidata would fulfill the attribution requirement and thus the terms of the license.ChristianKl (talk) 09:47, 18 May 2016 (UTC)

Mother/father/brother etc.[edit]

Can some bot (regularly) update these statements, putting information in all relevant items? --Edgars2007 (talk) 14:57, 12 May 2016 (UTC)

There is a list at User:Yamaha5/List_of_missingpairs_querys.
--- Jura 07:59, 20 May 2016 (UTC)
The Bot LandesfilmsammlungBot is currently under test for this request --Landesfilmsammlung (talk) 13:01, 2 June 2016 (UTC)
User:Landesfilmsammlung: Please check the P31 values as suggested on Yamah5's list. This to avoid edits like on Q629347.
--- Jura 05:26, 3 June 2016 (UTC)
Oh thanks... I will fix it... can someone correct the Node Samuel (Q629347). Cause the children-Property seems very unusual. --Landesfilmsammlung (talk) 11:55, 3 June 2016 (UTC)
Landesfilmsammlung: the general idea is that that someone should be you. You might want to keep an idea on the constraint violation reports for properties you added a day or two earlier.
--- Jura 09:42, 4 June 2016 (UTC)

labels from name properties[edit]

For people for which we know family name (P734) See with SQID and given name (P735) See with SQID, it could be possible to autogenerate a label in several relevant languages following certain rules. Could someone set up a robot to do this ? author  TomT0m / talk page 10:13, 15 May 2016 (UTC) Ash Crow
Harmonia Amanda
Чаховіч Уладзіслаў
Place Clichy

Pictogram voting comment.svg Notified participants of Wikiproject Names

@TomT0m: I can imagine working on this but I feel it can be controversial (therefore I want more comments on this). Query for this: Matěj Suchánek (talk) 09:45, 12 July 2016 (UTC)
I have seen people insist on one item for each spelling of a name, which means an approach like this would be unreliable (at best) when languages don't copy the original spelling. I think something based on this idea could work though if it takes into account things like where the person is from and the target language (and it would be better if people who speak the target language can confirm that they would expect the original spelling to be used for all people from those countries, because there might be things which are different that we're not aware of).
For example, I can't think of many examples of British people whose names are written differently in German so if a person is British and the names match the English label, using the same label for the German label sounds like it would be very unlikely to cause a problem. At the other extreme, Japanese writes almost all foreign names in katakana based on the pronunciation, so Michael Jackson (Q2831) (American), Michael Mittermeier (Q45083) (German) and Michael Laudrup (Q188720) (Danish) are all written differently in Japanese despite all having the same given name (P735) Michael (Q4927524) statement.
Generating the expected name from the statements and comparing it to the most appropriate label seems like a good sanity check. If the name expected from the statements doesn't match the actual label, there must be a reason for it. Some of the labels or statements could be wrong and need fixing, or perhaps the person is most commonly known by a different name.
Looking at that query, a few things already stand out to me: It says "Kirsten Johnson" for Czech for Kirsten Johnson (Q6416089), but the Czech sitelink for Betsey Johnson (Q467665) is "Betsey Johnsonová". For Azeri it also says "Kirsten Johnson", but the Azeri sitelink for Boris Johnson (Q180589) is "Boris Conson". It says "Bert Jansen (příjmení)" for Czech for Bert Jansen (Q1988186).
- Nikki (talk) 10:36, 15 July 2016 (UTC)
I support doing this, per the proposal, for some languages, but not for others. I'd be happy to collaborate on drawing up a "safe list" of languages. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 17:41, 12 July 2016 (UTC)
P.S. See also #Taxon labels, above. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 18:00, 12 July 2016 (UTC)
Just be careful. The transcriptions of names with origin in Cyrillic script is transcribed differently depending on the original and transcripted language. In Swedish we also have different transcriptions of Middle East Arabic and North African Arabic names. -- Innocent bystander (talk) 19:14, 12 July 2016 (UTC)
To be honest @TomT0m: I'm really really wary about this. I saw too many P735/P734 errors to believe we can accurately expand on it. I saw too many P735/P734 correct uses which don't permit to deduce label (for example, second or third given name, with the pseudonym-family name, noble people, etc.). It seems to me like every label should be checked manually and that's not possible with a bot. @Ash_Crow: had started working on something a little different: if people had the exact same label in English, French and German, he expanded the labels to all languages in Latin script and same naming usages. --Harmonia Amanda (talk) 09:32, 20 July 2016 (UTC)
@Harmonia Amanda: What's the worst, having an item really without a label so possibly very difficult to identify or an item with a close but non accurate label ? Considering the fact that the labels are probably missing in a very big number of languages in most cases, I think a non perfect information is better that no information at all. Plus ... a good way to improve quality of datas is actually use them to start spotting and correcting errors. I guess to be really useful both as easy to maintain, the bot should check if a name property has been modified since the last time he put the label. Hence the correction would propagate in each languages in a minimal numbers of edits and it would be clear that we should focus on the naming properties to optimize the interproject effort. author  TomT0m / talk page 09:41, 20 July 2016 (UTC)
Given the number of errors already present, especially for ancient/medieval people (Romans for a start) when names used to be translated, I'd be *very* careful.
I suggest you limit your action to people who:
  1. lived in the last century or so
  2. already have at last a label in one language matching the First+last name combination
  3. don't have a pseudonym (P742) or nickname (P1449).
Ash Crow (talk) 10:53, 20 July 2016 (UTC)
Strong reticence. This task would lead to too many false positive. Furthermore, there is an active community around the question of names, available to work on these topics. The bot would impede their work. --Dereckson (talk) 12:12, 20 July 2016 (UTC)
@Dereckson, Ash Crow, Harmonia Amanda: Can you be (publicly) more specific on how the work would be impeded ? Maybe solutions can be found to make everyone happy.
I've already have a suggestion to put only the computed label as alias. author  TomT0m / talk page 12:28, 20 July 2016 (UTC)

@Matěj Suchánek: are you still interested in working on this? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:51, 10 October 2016 (UTC)

I don't usually do potentially controversial tasks whereas many concerns and objections were raised. So not for the time being. Matěj Suchánek (talk) 12:58, 10 October 2016 (UTC)

date of birth (P569) with century precision (7)[edit]

For date of birth (P569) with century precision (7), values could be changed

Change Sample (displays as "20. century") WQS
From +(century)00-00-00T00:00:00Z/7 +2000-00-00T00:00:00Z/7 2000-01-01
To +(century-1)01-00-00T00:00:00Z/7 +1901-00-00T00:00:00Z/7 1901-01-01

For dates of birth, it seems that for century precision "dates", it would be better to specify the first year in the century rather than the last one.

When queried at WQS these appear as January 1.
--- Jura 07:38, 16 May 2016 (UTC)

Symbol oppose vote.svg Oppose With the current implementation of the time datatype lower order elements can be omitted for reduced precision without the need of doing any further calculations. --Pasleim (talk) 09:14, 16 May 2016 (UTC)
That actual leads you to mix-up 20th century people (born maybe 1910) with people from the 21st century (born 2005).
--- Jura 09:59, 16 May 2016 (UTC)
I don't unterstand your example. A person born in 1910 has the value +1910-00-00T00:00:00Z/9, born in 2005 the value +2005-00-00T00:00:00Z/9, born in 20th century the value +2000-00-00T00:00:00Z/7 and born in the 21th century the value +2100-00-00T00:00:00Z/7. If precision 9 is given, you have to omit everything except the first 4 digits, with precision 7 you have to omit everything except the first 2 digits. --Pasleim (talk) 10:34, 16 May 2016 (UTC)
The sample I have in mind would be a person born (maybe) in 1910 using +2000-00-00T00:00:00Z/7 compared to a person born in 2005 using +2005-00-00T00:00:00Z/9 . If you use just wdt:P569 rounded to the century digits you would get "20" for both.
--- Jura 15:30, 16 May 2016 (UTC)

Bloomberg Privat Company Search[edit]

Crawl all ~300.000 companies and add them to wikidata.  – The preceding unsigned comment was added by (talk • contribs) at 16:09, 23 May 2016 (UTC).

(related) I've research and spoken to Bloomberg employees previously on importing their symbols (BBGID). I've tried quickly proposing clear cut properties with some taking nearly a year to be approved (What you'd need). Disappointingly we've imported notability from Wikipedia with people worrying about too many items. There's also significant structural problems with Wikidata because its a crappy mirror of Wikipedia (and the smaller ones at that). Movie soundtracks can't be linked to the article's Soundtrack section (many items => 1 article). Multi-platform video games are currently a mess (1 article => many items).

To start you'll need to propose a new property Dispenser (talk) 20:09, 23 May 2016 (UTC)

@Dispenser: I added a property proposal: ChristianKl (talk) 11:20, 24 September 2016 (UTC)

MCN number import[edit]

There are 10,031 identifiers for MCN code (P1987) that can be extracted from [1] or this English version. Many (but not all) items cited are animal taxons, which can be easily machine-read. For the rest, it would be useful if the bot generated a list presenting possible meanings (by comparing the English and Portuguese versions of the xls file with Wikidata language entries). Pikolas (talk) 12:38, 14 August 2015 (UTC)

What's the copyright status of those documents? Sjoerd de Bruin (talk) 13:04, 14 August 2015 (UTC)
It's unclear. I've opened a FOIA request to know under what license those are published. For reference, the protocol number is 52750.000363/2015-51 and can be accessed at Pikolas (talk) 13:40, 14 August 2015 (UTC)
I heard back from them. They have assured me it's under the public domain. How can I prove this to Wikidata? Pikolas (talk) 01:48, 2 October 2015 (UTC)
@Pikolas: I have only just noticed that you haven't had the courtesy of a reply. The best method would be to get them to put a statement to that effect on their website. Failing that, you could get them to email OTRS. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:55, 10 October 2016 (UTC)
@Sjoerddebruin: Reopening this thread since I forgot to ping you. NMaia (talk) 15:45, 1 June 2016 (UTC)
Updated links: Portuguese version, English version. NMaia (talk) 19:35, 2 June 2016 (UTC)

Mark items being used in lists as "used" on Wikidata:Requests for deletion[edit]

Currently Benebot marks some items as "in use" when there are links on a given item.

As sites such as Wikipedia start using arbitrary access to retrieve information from Wikidata, the above approach doesn't capture what may be key uses for some items.
--- Jura 16:01, 5 June 2016 (UTC)

Lowercase adjectives[edit]

It might be worth doing another conversion of lowercase adjectives in descriptions of people, "italian" → "Italian", "british" → "British", etc.
--- Jura 11:51, 22 June 2016 (UTC)

Get GeoNames ID from the Cebuano or Swedish Wikipedia[edit]

Currently there are many concepts such as no label (Q22564260) that refer to geographical features that have articles in the Cebuano and Swedish Wikipedia. For most of them there's an Infobox with information at the respective Wikipedia but not all of the information is available in Wikidata. I would propose that the information get's copied over by a bot. There are to many articles to copy information manually. Especially the GeoNames ID should be easy to copy automatically. ChristianKl (talk) 15:52, 6 July 2016 (UTC)

Be very very careful! The GeoNamesID's that has been added here before, based on the Wikipedia-links in the GeoNames database are very very often very very wrong! Starting with copying the geonames-ID's from the sv/ceb-articles is a good start! We then can detect mismatching in Wikidata and GeoNames. Other kind of information can thereafter be directly be collected from GeoNames. But even that data is often wrong. An example, large parts of the Faroe Islands (Q4628) in GeoNames is located on the bottom of the Atlantic. -- Innocent bystander (talk) 16:26, 12 July 2016 (UTC)
@Innocent bystander: Note: I did import few thousands of Geonames IDs some few weeks ago. Can't say, how many are left there. If svwiki had some tracking category, that would be helpful :) --Edgars2007 (talk) 17:18, 31 July 2016 (UTC)
@Edgars2007: I'll see what I can do (tomorrow). One issue here is that a tracking-category cannot separate the Lsjbot-articles from the others. -- Innocent bystander (talk) 18:56, 31 July 2016 (UTC)
@Innocent bystander: To clarify, I'm only asking about category for Geonames parameter, not about others. I don't see any reason why this fact (who created article) is relevant in this situation. If needed, that can be get with database query. --Edgars2007 (talk) 19:43, 31 July 2016 (UTC)
@Edgars2007: I intend to create (at least) two categories. One for when P1556 is missing here and one for when WD and WP do not agree about the geonames-id. A third potential category could be used to detect when there is a geonames-parameter in WP and it matches P1556. In such cases, the parameter could be removed from WP. -- Innocent bystander (talk) 05:25, 1 August 2016 (UTC)
@Edgars2007: ✓ Done Category:Wikipedia:Articles with a geonames-parameter but without P1566 at Wikidata (Q26205593)! It will take some time until the category is completely filled with related articles. It will also take some time after you have added the property here, until the category is removed on svwiki. -- Innocent bystander (talk) 07:01, 1 August 2016 (UTC)
The category is now filled with almost 250000 pages. A category for the cases when WD and svwp contradicts each other have ~4000 members. -- Innocent bystander (talk) 07:10, 2 August 2016 (UTC)
Yesterday evening that was some 300 pages (for the first category) :D --Edgars2007 (talk) 07:17, 2 August 2016 (UTC)
@Edgars2007: Any progress? Lsjbot is halted for some more time, so there is a possibility to catch up with hir! I am daily sorting out some of the more complicated constraints-problems and other problems reported on svwiki. -- Innocent bystander (talk) 06:37, 21 August 2016 (UTC)
@Innocent bystander: I haven't forgot about you. Yes, I haven't had (much) time to do this yet, but will try to clean-up the category. --Edgars2007 (talk) 07:38, 21 August 2016 (UTC)

Import vernacular names from Wikispecies[edit]

Wikispecies stores vernacular names for taxons, using a template, species:Template:VN. For example, on species:Passer_domesticus, the template markup includes:

{{VN |af=Huismossie |als=Spatz, Schbads |ar=عصفور دوري |de=Haussperling |en=House Sparrow

(and many more entries besides). Note that Spatz, Schbads represents two names, separated by a comma.

We need to have these names imported to taxon common name (P1843), avoiding duplication of existing values, and with the language codes. Assuming the latter is technically possible, can someone do this, please? We can then work on converting the template to pull its data from Wikidata. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 11:49, 7 July 2016 (UTC)

Pictogram voting info.svg Info There is a related discussion at Wikispecies: Moving vernacular names to Wikidata. --Succu (talk) 11:32, 18 July 2016 (UTC)

Can anyone help with this, please? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:06, 10 September 2016 (UTC)

I'll look into it. --Magnus Manske (talk) 13:04, 10 October 2016 (UTC)
OK Magnus, more details? Josve05a (talk)

FelixReimann (talk)
Infovarius (talk)
Daniel Mietchen (talk)
Soulkeeper (talk)
Brya (talk)
Klortho (talk)
Delusion23 (talk)
Andy Mabbett (talk)
Dan Koehl (talk)
Achim Raschka (talk)
Rod Page
Joel Sachs
Prot D
Michael Goodyear
Lymantria (talk)
Pictogram voting comment.svg Notified participants of Wikiproject Taxonomy --Succu (talk) 21:57, 10 October 2016 (UTC)

It's >66K pages, scan is still running. I am creating a list of QuickStatement commands, which I can run by bot once it's complete. Edits will look like this. So far, ~95K statements prepared. I'll post here once it's done. --Magnus Manske (talk) 22:02, 10 October 2016 (UTC)
@Magnus Manske: Is there any agrement to run this? --Succu (talk) 22:09, 10 October 2016 (UTC)
Yes, Andy and I agree. --Magnus Manske (talk) 08:57, 11 October 2016 (UTC)
I informed Wikispecies. --Succu (talk) 14:54, 11 October 2016 (UTC)

Update The batch is now complete, ~263K statements altogether. The proposed statements will all have "Source:Wikispecies", in case we want to mass-check/clean/remove them later. Some people on Wikispecies think these is some "crud" in there; I checked a few manually, and at least the English names seem to be OK. I will start the import later today, unless there are (good) objections. --Magnus Manske (talk) 08:35, 12 October 2016 (UTC)

Did you notice that wikispecies is using some "wrong" language codes? e.g. als instead of gsw for Allemanic or no instead of nb for norsk bokmål? --Pasleim (talk) 11:35, 12 October 2016 (UTC)
Is there a mapping somewhere, or a list to exclude? --Magnus Manske (talk) 13:53, 12 October 2016 (UTC)
You could use mw:Manual:$wgDummyLanguageCodes but I don't know if Wikispecies users even invented more language codes. --Pasleim (talk) 14:25, 12 October 2016 (UTC)
Well, Wikidata would not accept made-up language codes, right? So worst case scenario, names in those languages won't be added. I consider that "acceptable risk".--Magnus Manske (talk) 15:58, 12 October 2016 (UTC)
Sometimes taxon name (P225) is used for vernacular name. --Succu (talk) 14:35, 12 October 2016 (UTC)
Not sure I follow. I am parsing the VN template from the wikitext on Wikispecies, and add those to taxon common name (P1843). Where does taxon name (P225) come into play? --Magnus Manske (talk) 15:58, 12 October 2016 (UTC)
Sorry. The vernacular name for a language given in the template is sometimes the scientific name (=P225) and not really a vernacular name. --Succu (talk) 16:05, 12 October 2016 (UTC)
What about a common names that differ slightly? E.g. Amazonetta brasiliensis (Q456423): IOC says "kachnička amazonská" - Wikispecies says "Kachnička amazonská". How do you handle this? There are a lots of hyphens or apostrophes involved in this kind of name but essentially refering to the same vernacular name. --Succu (talk) 22:27, 12 October 2016 (UTC)
I have not applied any filters for this. Wikispecies being a wiki, there is always the chance of getting some instances wrong. But having all entries in Wikidata will help Wikidata (e.g. with search), will get more eyes on the issues, and allow to replace the Wikispecies text with Wikidata-derived templates. I would call that a win.
I'll write something to remove the taxon names, though scanning through them, there seem to be few. --Magnus Manske (talk) 08:08, 13 October 2016 (UTC)

Update Filtered out some taxa names, corrected some language codes. Uploading now. 256,187 new statements in total. Edits here. --Magnus Manske (talk) 10:54, 13 October 2016 (UTC)

Updating population of US towns[edit]

Hello, I was wondering if a bot can be used to update population estimates in the U.S for 2015. I think a good source of information is here. It is a government website. Is this feasable?MechQuester (talk) 06:09, 26 July 2016 (UTC)

If there is any desire to do this, it should be as additional information, and should not replace official census information from 2010. This is because many laws have different provisions depending on the population of a town or city, and such laws always reference official census results which are done once every 10 years. Interim results from the Census Bureau are not recognized by law. Jc3s5h (talk) 14:22, 10 September 2016 (UTC)

Import Template:Bio from itwiki[edit]

To avoid that the gap gets to big, it might be worth doing another import. There are a series of steps outlined in Help:Import Template:Bio from itwiki.
--- Jura 17:08, 29 July 2016 (UTC)

Labels in English for items whose names begin with "The"[edit]

Items with a label in English beginning with "The " (note space; case insensitive), like "The Dark Side of the Moon", should have an alias in the form "Dark Side of the Moon" and/or "Dark Side of the Moon, The", if one does not already exist. Can someone do this, please? I'm ambivalent as to which of the varieties, or both, is used. This also applies to labels in types of English, such as en-GB. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 09:52, 4 August 2016 (UTC)

Category:Commons category without a link on Wikidata[edit]

Hello.Please connect the sub-pages of this category with items (AWB should be used).Thank you --ديفيد عادل وهبة خليل 2 (talk) 13:01, 6 August 2016 (UTC)

I would like to see, how it could be done with AWB ;) --Edgars2007 (talk) 23:57, 9 August 2016 (UTC)
AWB should be used if possible.It is generally useful Face-glasses.svg --ديفيد عادل وهبة خليل 2 (talk) 08:50, 10 August 2016 (UTC)
That isn't an answer to his question. Sjoerd de Bruin (talk) 09:05, 10 August 2016 (UTC)
Personally, I don't understand this request. What is meant with a "sub-page"? AFAIK, AWB doesn't work on Wikidata, so if the bot should work on another wiki, this request does not belong here. Matěj Suchánek (talk) 14:36, 10 August 2016 (UTC)
For clarity: Special:Diff/362948011. The request is to simply clean-up Category:Commons category without a link on Wikidata (Q11925744). Did import few ten thousands weeks or months ago from enwiki, but it wasn't enough - there are a lot more. --Edgars2007 (talk) 14:47, 10 August 2016 (UTC)
Maybe the time has come to convert this to normal sitelinks.
--- Jura 14:56, 10 August 2016 (UTC)

Zerozero footballer Ids & others in refs[edit]

It seems that many of en.Wikipedia's use of en:Template:Zerozero is in <ref></ref> tags. Does anyone have a bot that can compare the subject of the target page, and add matches using footballzz ID (P3047)? And similar cases?Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 19:14, 9 August 2016 (UTC)


-- Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 13:30, 11 August 2016 (UTC)


-- Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 19:46, 4 October 2016 (UTC)

Remove incorrect svwiki sitelinks[edit]

Many svwiki articles about places of China created by Lsjbot have incorrect zhwiki sitelinks, which are imported to Wikidata. Some examples I have fixed manually are:

  1. sv:Gaohu (köpinghuvudort i Kina, Jiangxi Sheng, lat 28,93, long 115,24), incorrectly links to no label (Q11671865) (a person), correct item is Gaohu (Q2656384) (a town)
  2. sv:Yushan (köpinghuvudort i Kina, Chongqing Shi, lat 29,53, long 108,43), incorrectly links to no label (Q22079207) (a person), correct item is no label (Q13714765) (a town)
  3. sv:Hongfenghu, incorrectly links to no label (Q15928338) (a lake), correct item is no label (Q14143028) (a town). I have added the correct svwiki link to the former article
  4. sv:Bianhe (köping i Kina, Anhui), incorrectly links to no label (Q11137293) (a disambiguation page), correct item is no label (Q11137300) (a subdistrict, formerly a town)
  5. sv:Chenyaohu, incorrectly links to no label (Q16935572) (a lake), correct item is no label (Q14343855) (a town)

Request to:

  1. Remove all svwiki sitelinks and Swedish labels in (all are errors)
  2. Remove all svwiki sitelinks and Swedish labels in (all are errors)
  3. Remove all svwiki sitelinks and Swedish labels in (all are errors)
  4. Remove all svwiki sitelinks and Swedish labels in sv:Kategori:Robotskapade Kinaartiklar if the distance between the Wikidata P625 and coordinate in svwiki>100km

--GZWDer (talk) 15:43, 11 August 2016 (UTC)

Note: Lsjbot have not completed all articles about places of China. This should be done again when it's completed.--GZWDer (talk) 15:47, 11 August 2016 (UTC)
Looks like Chinese to me. Maybe @Innocent bystander: can explain it in Swedish to Lsj.
--- Jura 16:01, 11 August 2016 (UTC)
@Lsj: most likely speaks/writes better English than me.
I think Lsjbot has finished with Peoples republic of China. (The nations has been edited in alphabetic order after the ISO-codes, with a few exceptions for nations who have been requested (ex: Syria) or was a part of a benchmark (ex: South Sudan)) This is a known bug, but it has been hard to undo all the mistakes. It happend when places in China was interwiki-linked to articles on zh-wiki with the same "label" as the Chinese places in GeoNames. So the Swedish labels are not necessarily wrong in these cases, even if the sitelinks are. -- Innocent bystander (talk) 16:46, 11 August 2016 (UTC)
There're still a red link in sv:Bianhe. For labels: zhwiki articles always use full name, not short names; in addition, the full name may not be unique, which disambiguation pages are needed.--GZWDer (talk) 17:18, 11 August 2016 (UTC)
I ping @Bothnia: who is skilled in both East Asian and Swedish. My knowledge in Chinese is extremely limited! -- Innocent bystander (talk) 18:29, 11 August 2016 (UTC)

Requests #2 and #3 ✓ Done. Not done #4 - is a bit more difficult, and #1 – not sure if it's 100% sure (all articles in sv:Kategori:Robotskapade Kinaartiklar that doesn't have official name (P1448) in their WD items). --XXN, 20:07, 8 September 2016 (UTC)

Done #4. If somebody is interested, removed sitelink and label from 396 items, the record holder is Rock Creek (Q2352739) - the distance is 12435.39 km. I have data also for other articles for that category, if somebody wants to analyze further. Many coords are really close (<5 km), but few hundrets are in that 5-100 km part. --Edgars2007 (talk) 07:16, 13 October 2016 (UTC)


SIMBAD ID (P3083) cannot be collected from en.Wikipedia using HarvestTemplates, as most of the values are in subtemplates, such as en:Template:Planetbox reference, which builds them by concatenating two other values. Can anyone fetch them, please? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:34, 13 August 2016 (UTC)

Redundant P1343 for DNB00[edit]

A lot of items contain two described by source (P1343) statements for the exact same article. One links the article directly, and one uses Dictionary of National Biography (1885-1900) (Q15987216) as value with the article as qualifier. I think the latter is the correct way of linking these, thus the redundant statements should be removed. See this item for a example. Sjoerd de Bruin (talk) 19:17, 17 August 2016 (UTC)

Wikidata:WikiProject DNB recommends the other. Using Q15987216 would be redundant as it's present in the linked item.
--- Jura 05:47, 18 August 2016 (UTC)
@Sjoerddebruin: I also prefer to link directly to the article, as recommended in Wikidata:WikiProject DNB#Examples. Is it okay for you if I remove the statement with the qualifier? --Pasleim (talk) 12:47, 24 August 2016 (UTC)
Please do, everything is better than this duplicated stuff. Sjoerd de Bruin (talk) 13:33, 24 August 2016 (UTC)
Definitely not ok. -- Sergey kudryavtsev (talk) 04:28, 26 August 2016 (UTC)
Just realized that there are more than 20,000 of such duplicated claims. @Sergey kudryavtsev: Are these claims used in a LUA module? --Pasleim (talk) 21:42, 25 August 2016 (UTC)
@Pasleim: Yes. The claims described by source (P1343) = Dictionary of National Biography (1885-1900) (Q15987216), Dictionary of National Biography, first supplement (Q16014700) or Dictionary of National Biography, second supplement (Q16014697) with qualifier stated in (P248) are widelly used in ruwp and ruws. At ruws such pages collected in categories: s:ru:Категория:Викитека:Ссылка из Викиданных:DNB, s:ru:Категория:Викитека:Ссылка из Викиданных:DNB01 and s:ru:Категория:Викитека:Ссылка из Викиданных:DNB12, altogether about 1000 pages.
Using described by source (P1343) with a qualifier stated in (P248) greatly optimize access in LUA modules — а described by source (P1343)'s values acts as some flag to decide to load or no to load an article's item. (I said all this to Jura1 several months ago, but he do not heed the voice of reason...) In opposed case LUA module should load and inspect all described by source (P1343)'s values until a required article would be found! -- Sergey kudryavtsev (talk) 04:24, 26 August 2016 (UTC)
I see. However, in most cases items only have one described by source (P1343) value. Then no additional LUA calls should be needed. In cases where there are two or three P1343 values, one resp. two more LUA calls are needed but this should still be okay performancwise. Items with more than three P1343 values are rather rare, see [2] for the distribution. --Pasleim (talk) 12:33, 26 August 2016 (UTC)
But average P1343-per-item value will actively grow because most enciclopedic article still not linked for now. The perfomance critical is a entity loading operation, which you calls as "LUA calls". I not afraids exceed the limit, but ruws sometime gets unstable timeout errors on wikidata-linked pages. So every additional loading operation makes timeout more probable. -- Sergey kudryavtsev (talk) 06:37, 27 August 2016 (UTC)
It's kind of cool being confused with Joe F., but still. Last time I discussed this with Sergev, the Russian module needed fixing and didn't make use of DNB at all. A fix was offered. The argument advanced here seems to be the same as the one we got for adding country qualifiers to every place of birth/death statements. Incidentally, I think both modules using that were designed by the same former contributor.
--- Jura 11:40, 27 August 2016 (UTC)

For what it's worth, I find the construction with the DNB article item as qualifier to be acceptable. It is a serious issue, clearly; but given that Wikisource only gradually proofreads articles (e.g. Britannica 1911 is only slowly being completed), it makes a lot of sense to add the main work, and then qualify with the article when that is possible/available. I realise this a Wikisource argument rather than a Wikidata argument. But I'm not going to apologise too much for that. Charles Matthews (talk) 11:48, 30 August 2016 (UTC)

VIAF import[edit]

Please see Property_talk:P214#Import_.3F.
--- Jura 05:09, 25 August 2016 (UTC)

Revert label additions by Edoderoobot in the beginning of May[edit]

In the beginning of May, Edoderoobot (talkcontribslogs) copied a lot of labels from other languages like here and here. You can clearly see that these aren't acceptable labels in Dutch. I've asked the bot operator multiple times to clean this up, but they are still there. Can someone help me? Sjoerd de Bruin (talk) 07:17, 25 August 2016 (UTC)

The following query uses these:

  	?item wdt:P31 wd:Q13406463 .
	?item rdfs:label ?labelnl FILTER(lang(?labelnl)="nl")
  	?item rdfs:label ?labelen FILTER(lang(?labelen)="en" && str(?labelnl) = str(?labelen) )

SPARQL query

Above a list of (all) NL labels that are identical with EN (4698). You could use QuickStatements to delete the label for some or all (or replace it).
--- Jura 07:36, 25 August 2016 (UTC)

I would not be bothered if they were cleared all. I have now filtered for what items/instance of (P31) it makes sense to take over the English description (right now items like human (Q5)), so if any are deleted in excess I can re-do them with my (repaired) bot script. But i will have a look myself if this SPARQL-script can automate a repair action. This might be the opening I needed to get it fixed myself. Edoderoo (talk) 08:21, 25 August 2016 (UTC)
I created a repair script based on the above SPARQL-query... will run it tomorrow, as right now another script isn't finished yet. Please be adviced that there might be more P31-types, but those can be fixed with the same script. Most likely Sjoerd will keep contact with me about those, but feel free to contact me in case someone finds another case. Once more thanks to Jura for this helpful SPARQL-script! Edoderoo (talk) 13:24, 25 August 2016 (UTC)

Also a lot of errors in January, see here for a example. Sjoerd de Bruin (talk) 11:37, 31 August 2016 (UTC)

Also a broad selection of subjects, see Special:Diff/330646941. Can't we mass-revert? Sjoerd de Bruin (talk) 14:30, 8 September 2016 (UTC)

Idea: Multi-wiki KML bot[edit]

Are there any bot operators willing to work on a multi-work bot task? If so, please see meta:Talk:KML files - Evad37 [talk] 04:10, 27 August 2016 (UTC)

NSW Flora IDs[edit]

Values for NSW Flora ID (P3130) are held in English Wikipedia's en:Template:NSW Flora Online, but split over multiple parameters, preventing the use of HarvestTemplates. Please can someone import them? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:23, 1 September 2016 (UTC)

My bot will import this from the original source. --Succu (talk) 16:44, 1 September 2016 (UTC)

@Pigsonthewing, YULdigitalpreservation, ChristianKl: At the moment this property works only for species (formatter: href=/cgi-bin/, but the website has

too. This is simmilar to GRIN URL (P1421) or AlgaeBase URL (P1348). So the datatype of this property should be changed to URL. --Succu (talk) 08:13, 2 September 2016 (UTC)

@Succu: one of possible work-arounds is to have property value "nswfl" and a new qualifier with value "fm". Of course, not perfect... But this probably has to be discussed somewhere else, not on BOTREQ page.--Edgars2007 (talk) 08:18, 2 September 2016 (UTC)
@Edgars2007 Do you have a working example for this „workarond“? --Succu (talk) 21:16, 2 September 2016 (UTC)
@Succu: No, I don't have. --Edgars2007 (talk) 04:01, 3 September 2016 (UTC)


The simplest fix will be to rename this property, and have another for other ranks. Otherwise, change the formatter URL and use IDs like:

  • lvl=sp&name=Avicennia~marina
  • lvl=in&name=Avicennia~marina+subsp.~australasica
  • lvl=gn&name=Avicennia

-- Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 13:54, 2 September 2016 (UTC)

I doubt this is a reasonable option. --Succu (talk) 18:21, 2 September 2016 (UTC)
Two, mutually-exclusive, options were suggested. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:58, 5 September 2016 (UTC)
I count three:
  1. recreate with datatype URL - straightforward
  2. create two additional taxon properties for the same dataset - complex (we don't use 3 properties for GRIN)
  3. reuse the current property with an URL fragment - a strange mixup between datatypes external ID and URL
--Succu (talk) 20:36, 8 September 2016 (UTC)
I was referring to my post, to which you replied in the singular. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:02, 10 September 2016 (UTC)

NLSZ authority name (HuBpOSK) import from VIAF[edit]

Hey, I was wondering if someone could import NSZL name authority ID (P3133) from VIAF. The trick is that this ID is not the primary ID on the NLSZ record (that would be NSZL ID (P951)), but the one under HuBpOSK in NLSZ records (so for it is "114", see Antal Szerb (Q570810)). Thanks! – Máté (talk) 05:33, 4 September 2016 (UTC)

Calendar date[edit]

For every instance of calendar date (Q205892), please can someone's bot add calculated values like in these edits. It may also be possible to calculate labels in other languages; and values for other properties. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:51, 5 September 2016 (UTC)

P.S. query. --Edgars2007 (talk) 13:39, 10 September 2016 (UTC)

Delete redirects[edit]

Lists like Josve05a's User:Josve05a/dupes tend to be full of redirects with sitelinks. Samples: Q5145525 and Q23812706. Would you delete all sitelinks on these items that are redirects?

Obviously, a solution that would solve this for even more item would be better.
--- Jura 23:41, 12 September 2016 (UTC)

For seeking/finding/implementing a more general solution, I created phab:T145522.
--- Jura 07:38, 14 September 2016 (UTC)

add locative as (en) alias[edit]

For places with a native label (P1705) in a language that has a locative case, it would be helpful if the form would be added as an alias.
Sample: Q1799#P1705 "Wrocław" > alias "Wrocławiu".
--- Jura 09:47, 18 September 2016 (UTC)
Question: Why? How can this be useful, in English particularly?
Concern: I have never heard of a bot which would successfully work with a Fusional language (Q318917). Matěj Suchánek (talk) 07:50, 24 September 2016 (UTC)
The problem is that sometimes locations in Wikipedia (and elsewhere) are given in this form and don't necessarily link to corresponding articles.
Currently, there is no way for people to find this locations on Wikidata.
The bot would just add an alias. Obviously, the input lists would first need to be compiled.
--- Jura 12:56, 24 September 2016 (UTC)

Species acronyms[edit]

Items for pages in species:Category:Repositories should probably have the full name listed on species:repositories as label and the acronym as alias.

If no better P31 value can be found, maybe P31=organization can do. Not sure what to suggest for the related categories. These are for taxa whose type specimen is held by that institution.
--- Jura 14:39, 20 September 2016 (UTC)

Dropped a note at Wikispecies. --Succu (talk) 19:48, 20 September 2016 (UTC)
The category is a mess replete with duplicates that I've long given up trying to deal with, but generally these pages should be treated as cross-wiki links for the equivalent institution (cf. species:AMNH, species:DNMNH, the latter of which I just merged into the proper institution).
In and of itself this is straightforward. The problem comes where many, if not most of these don't necessarily have straightforward matching pages on other wikis (e.g. because often a given institution correspond to several "collections" that have no been merged in Wikispecies), and that's not counting renamed institutions, collections that were moved/merged years ago, or the occasional outright ambiguous or incorrect name. Circeus (talk) 01:55, 21 September 2016 (UTC)
@Circeus: As long as the acronyms in Wikispecies article titles match the entry in the list, it should be fairly straightforward. Adding the full name to item labels would make it easier to find duplicates/merge them with other items. That some institutions have changed their name since or that collections were absorbed by others shouldn't be much of an issue. Wikidata is a good place to hold historic data as well.
--- Jura 13:01, 24 September 2016 (UTC)
I think we need a property to map an institution to a code. Index Herbariorum (Q11712089) (website) is an example for a register of herbarium (Q181916). --Succu (talk) 19:18, 24 September 2016 (UTC)
This might help, but isn't necessarily needed for this request.
BTW short name (P1813) could also be used.
--- Jura 10:48, 26 September 2016 (UTC)

Automatically creating a human subclass for anatomical features that don't already have subclasses[edit]

Some statements are true for the fingers of every species but others are human specific. Currently we often don't have separate items for the concept in humans. I think it would be valuable to have a bot that automatically creates human subclasses. ChristianKl (talk) 09:45, 22 September 2016 (UTC)

Can you please provide a list with all anatomical features? --Pasleim (talk) 11:56, 26 September 2016 (UTC)
We have animal structure (Q25570959). That then get's subclassed in different ways. That should produce a long list of anatomical features where most exist in humans. ChristianKl (talk) 14:43, 1 October 2016 (UTC)

Official tourist website[edit]

As discussed here and as summarized here I would request to transfer the information (relevant to website and toponym) stored in it:voy: into the agreed structure.

The information are stored in the following templates divided per toponym type:

When the associated Wikidata instance of a toponym do not have the property tourist office (P2872), it must be created, adding as a value the instance that contains the official tourist website. If this instance does not exist, it must be created with the following properties and values:

  • official website (P856) with the official tourist website extracted from the Quickbar template
  • country (P17) with the same value of the main instance. Obviously if the main instance is a Country the value it that instance itself
  • instance of (P31) with official tourism agency (Q26989327) ... this value is to declare its official status
  • If exist a "city property" can be stored the instance associated to the QuickbarCity
  • If exist a "territory property" can be stored the instance associated to the QuickbarRegion

I hope it is enough clear, if not, feel free to ask. Thanks, --Andyrom75 (talk) 20:11, 24 September 2016 (UTC)

Guardian data about US police killings[edit] provides data about individuals in the US who were killed by the police. Should we import the data? If we import the data it might also be interesting to make a public statement that invites other people to contribute data about those people. ChristianKl (talk) 14:38, 26 September 2016 (UTC)

Import it to what item? Jc3s5h (talk) 15:38, 26 September 2016 (UTC)
Items for the people who are killed. The Guardian lists their names and data about them. It would be possible to automatically create lists in Wikipedia that show all police killings in month X. ChristianKl (talk) 16:26, 26 September 2016 (UTC)
I object to another bot that will create lots of items without making an effort to see if there is already an item for the person. Of course, if there were an existing item it would be necessary to rigorously investigate whether the person who was killed was the same person named in an existing item. I realize that occasionally duplicate items will be created accidentally, but doing it en mass with a bot doesn't seem like a good idea to me. Jc3s5h (talk) 17:07, 26 September 2016 (UTC)
Why? Merging items is easy. Especially with the merging game. ChristianKl (talk) 17:38, 26 September 2016 (UTC)
What data? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 18:19, 26 September 2016 (UTC)
There seems to be something of an API:, although it isn't documented. Kaldari (talk) 22:45, 28 September 2016 (UTC)
Names of the people who are killed. manner of death (P1196) "killed by police gunshot" (and other classes for deaths that aren't gunshots). date of death (P570). ethnic group (P172). I think there's interest in having Wikipedia lists of people killed by police by race. ChristianKl (talk) 23:18, 28 September 2016 (UTC)
  • enwiki has fairly detailed lists. If you are interested in the topic, you could import these.
    --- Jura 09:33, 1 October 2016 (UTC)

French cities not marked as such[edit]

Ajaccio (Q40104) is a municipality (Q15284) with a population (P1082) of 58,000 inhabitants, yet it was not instance of (P31) city (Q515).

I guess their might be work for a bot to correct similar cases?

In France a "ville" (city) is defined by having more than 2000 inhabitants.

Thanks! Syced (talk) 07:41, 28 September 2016 (UTC)

  • Maybe you'd want to make a dedicated item for such "ville", but then, what's the advantage compared to the current situation.
    --- Jura 09:35, 1 October 2016 (UTC)

film budget[edit]

The following query uses these:

SELECT ?item
  ?item wdt:P2769 [];
        wdt:P31/wdt:P279* wd:Q11424

SPARQL query

The query above returns over a thousand items, whereas the entertainment media jargon 'budget' refers to estimated cost (P2130) and not budget (P2769). These published numbers are estimated after production, and are not actually the planned budget. I would like someone to move these statements to the correct property while keeping the qualifiers. – Máté (talk) 12:04, 30 September 2016 (UTC)

Copy data from the property documentation to property statements[edit]

The syntax in which the formatter url and other statements are stored in the property documentation should be easy to understand for a bot, that can automatically create statements from them. ChristianKl (talk) 11:34, 1 October 2016 (UTC)

This should be "move", not "copy". Some statements are suitable for this; others not. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 18:46, 1 October 2016 (UTC)
Wikidata:Requests for permissions/Bot/MatSuBot 2. Matěj Suchánek (talk) 18:58, 1 October 2016 (UTC)

{{Section resolved|1=Sjoerd de Bruin (talk) 08:55, 12 October 2016 (UTC)}}

Postponing archiving per Andy's comment and Topic:Td698cyh1l7depz2. My bot's RfP only allows copying, not (re)moving data from talk pages. So this can remain an open task, although I can imagine someone would like to keep it more detailed than what property statements can provide. Matěj Suchánek (talk) 17:33, 14 October 2016 (UTC)

MySpace IDs[edit]

Unfortunately, en.Wikipedia deleted its MySpace template, with each transclusions substituted. This left over 3,000 MySpace URLs in articles using ordinary wikitext links. Can someone import the values to MySpace ID (P3265), please? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 13:33, 6 October 2016 (UTC)

Item located in Rome, with coordinates not in Rome[edit]

I just discovered that Embassy of Sweden, Rome (Q5369877) had wrong coordinates putting it in Milano, hundreds of kilometers from Rome.

Since this item is located in the administrative territorial entity (P131) Rome (Q220), would it not be possible to detect such inconsistencies?

A bot could use an OpenStreetMap shapefile (good opportunity to make sure all OpenStreetMap countries and major cities are linked to Wikidata), or any other reverse geocoding solution, and generate a list of potential errors for human editors to check.

Any volunteer? :-) Syced (talk) 03:35, 12 October 2016 (UTC)

@Syced: maybe this will be enough? Query:
select * where {
  ?item wdt:P131/wdt:P279* wd:Q220 .
  ?item wdt:P625 ?coords .
SPARQL query
--Edgars2007 (talk) 06:16, 12 October 2016 (UTC)
The human who investigates whether the items really are in Rome would have to be quite familiar with Italian laws about municipalities; sometimes cities have unexpected jurisdiction over places that seem remote from the city. Jc3s5h (talk) 09:08, 12 October 2016 (UTC)
Hello Edgars2007! Rome was only an example :-) Checking using shapefile is needed here, because location borders sometimes include remote places: For instance Tokyo includes a few tiny islands very far in the Pacific Ocean. Cheers! Syced (talk) 09:12, 12 October 2016 (UTC)
@Syced: This query lists all places in Italy (Q38) which are further than 50km from their city (Q515):
SELECT ?place ?placeLabel ?placeDescription ?location ?locationLabel ?dist WHERE {
  ?place wdt:P17 wd:Q38;
         wdt:P625 ?coord;
         wdt:P131 ?location .
  ?location p:P31/ps:P31/wdt:P279* wd:Q515;
            wdt:P625 ?locationCoord .
  BIND( geof:distance(?coord, ?locationCoord) AS ?dist ) .
  FILTER( ?dist > 50 ) .
  SERVICE wikibase:label {
    bd:serviceParam wikibase:language "en" . 
  } .
} ORDER BY DESC(?dist)
SPARQL query
On the other hand, what you suggest would propably help in 100% cases whereas queries are really limited to simple circles.
Matěj Suchánek (talk) 14:14, 12 October 2016 (UTC)

Statements using P2044 with unit, precision and multivalued references[edit]

I would like to add the following statements, but using:

I think this is currently not possible with QuickStatements, ping Magnus.

Thanks in advance. --abián 16:21, 16 October 2016 (UTC)

@Abián: is Alborge (Q1650353) what you want? P.S. If Great Aragonese Encyclopedia ID (P1807) is used in all reference URL (P854) fields, then the statement should look a little bit different (like this). --Edgars2007 (talk) 16:27, 17 October 2016 (UTC)
@Edgars2007: That's it. :D Indeed, Great Aragonese Encyclopedia ID (P1807) is used in all reference URL (P854) fields, so you can use the first property instead of the second. You can also use ±30 instead of ±50, I've been investigating a little more. --abián 17:20, 18 October 2016 (UTC)
@Abián: I have written the script for this, but we/you have to decide what to do with precision. See Wikidata:Project chat#+- workaround. --Edgars2007 (talk) 15:48, 24 October 2016 (UTC)

Add a band/musician database to Wikidata.[edit]

Hello I am wondering if I need a bot or to use QuickStatements for uploading data (as opposed to inserting "items" one by one).

I am currently manually collected a dataset of around 400 Latvian rockbands, their participant names (more than 1000 musicians) and the role in the band, (musical instrument they are playing). I would like to upload that data set to Wikidata, so an embedded network graph could be made trough Wikidata query, and users could add and contribute to that graph trough Wikidata.

The data are in a spreadsheet: first column is the bandnames (the name of each band is repeated as many times as there is different bandmembers), the second column is the musicians (some of them have instruments and recording labels as extra columns)

Is there a way by using my dataset to:

  • -upload a list of bandnames as "items", being "instance of: bands" with "country of origin: Latvia"
  • -upload a list of musician names as "items", with "instance of: human" and for example "instance of: bassist", "instance of: female"
  • -upload a list of band-musicion pairs, creating for each of the bands "has part: (the musician name)"

Thank you if someone has the time to answer.
--- LinardsLinardsLinards 04:00, 20 October 2016 (UTC)

Are all the band notable? Without reference I don't think they are --ValterVB (talk) 06:32, 20 October 2016 (UTC)
Hi ValterVB, thanks for the comment, most of the bands in the data-set are notable even by Wikipedia standards (recorded albums, national radio rotation, awards, press publications, coverage's and interviews) but here on Wikidata, the purpose of uploading the bands serves the third Wikidata notability criterion " It fulfils some structural need, for example: it is needed to make statements made in other items more useful." The bands in Latvia are highly interconnected, with a lot of musicians playing in more than one band. (See the press coverage of network graph of the same data, of connecting the bands by having at least one band member common therefore every of the bands serves as structural connection between other bands. Adding the dataset also allows to query such things as: male/female distribution in Latvian rockmusic, bands belonging to a specific label, distance between two bands by their players, bands coming from a specific city, the bass players in Latvia etc. and also query network graphs of band genealogy again based on specific parameters. --LinardsLinardsLinards (talk) 13:59, 21 October 2016 (UTC)
If we made a "loop" of not notable item, we haven't notable items. The third criterion relates to notable item. If item "A" is notable then I can create item "B" for structural need. But If I create item "A" that isn't notable also if I create item "B" not notabale and link it to "A" and viceversa, I haven't 2 notable item. --ValterVB (talk) 17:22, 21 October 2016 (UTC)
No problem ValterVB, I am aware of that. The data-set consists of most known Latvian bands. The data are collected manually from press interviews and other data sets. It sounds that I should ad these sources. I will do that. And what about the actual uploading, is a bot necessarily to do that or there exists already solutions that are able to do such uploading?--LinardsLinardsLinards (talk) 02:52, 22 October 2016 (UTC)
It's a problem because you know thet they are « are most known Latvian bands », but we can't know it. So, if the band aren't in wikipedia, wich kind of source you add to prove that are notable? --ValterVB (talk) 08:26, 22 October 2016 (UTC)
After I got the data manually from interviews with latvian music journalists, social media, and and presented them in a network graph form, I reached out to "Latvian performer and producer union" (, they collect information about the usage of music in radio, films, advertisements, collects the money and divides across the musicians. And since they are a public institution, they shared their database of registered bands and their participants. I could use that as a reference, since these are bands that gets to be played on radio. P.S. Unfortunately Latvian Wikipedia is very small and it currently does not include a lot of bands, what it should. So I want to show the capabilities of open databases trough data visualisations, to attract more editors in this specific field.--LinardsLinardsLinards (talk) 15:35, 22 October 2016 (UTC)
Can you add an id or link for every band in site? Maybe we can create a new property, something like "id on laipa.prg" --ValterVB (talk) 16:48, 22 October 2016 (UTC)
The data I got from only consists of unique integer IDs for Bands and musicians (There are several musicians with the same names). Would it maybe be appropriate to add ID as reference to the bands that are in data-set provided by them, and then I could find the interviews and mentions for the rest of the bands in the set?--LinardsLinardsLinards (talk) 17:47, 22 October 2016 (UTC)
Who should I ask about adding the ID property? I also have the information of who of the musicians have died, but no specific date or year. Does it makes sense to add property "date of death" as unknown? --LinardsLinardsLinards (talk) 22:45, 25 October 2016 (UTC)

BOT for programming project[edit]

Hello Wikidata! I want to create a Wikidata Bot for my university course, it's my first time programming bots and I'm a bit lost... can someone help me or show me a starting point or something? Thanks a lot!

You can start from here or here --ValterVB (talk) 17:25, 21 October 2016 (UTC)
You can also find some links and example scripts here. Edoderoo (talk) 14:48, 23 October 2016 (UTC)

Sitelink removal[edit]

This lists pages at Wikipedia that are not disambiguations, but are linked from items that have P31=Q4167410. Could you remove them? I will add them to appropriate items by QuickStatements afterwards.
--- Jura 04:25, 25 October 2016 (UTC)

I strongly oppose the use of a bot before cleaning out all the items on the list which aren't given names, because the categories concerned are not only "given name" and 'disambiguation" but also 'surnames', for example. I'm working on a equivalent list since early September and I don't see why we should do it badly with a bot when we'll still have to pass individually on each article to clean it correctly. The list was nearly twice as long when I started and it's going down steadily. --Harmonia Amanda (talk) 05:27, 25 October 2016 (UTC)
I agree that some should be given names, other "name" item. Don't worry about that.
--- Jura 05:45, 25 October 2016 (UTC)
Uh yes, I worry! How exactly do you intend to treat it? When I see your query happily mixing names, given names and disambiguation pages (because the sitelinks other than the English one can have always been genuinely a disambiguation page) and you only say "I'll treat it", I worry. --Harmonia Amanda (talk) 06:44, 25 October 2016 (UTC)
Refers to Wikimedia disambiguation page (Q4167410). Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 13:06, 25 October 2016 (UTC)

Remove redundant human settlement (Q486972)[edit]

Bjankuloski06 added thousands of instance of (P31)  human settlement (Q486972) to items that already contain a subclass of that. Can someone mass-revert? Sjoerd de Bruin (talk) 07:11, 25 October 2016 (UTC)

Suggest reverting only on items where there was no other instance of (P31) claim. no label (Q12911027) is fine, for example. @Bjankuloski06: perhaps you can revert yourself. --Izno (talk) 11:48, 25 October 2016 (UTC)

@Izno: Since I lack the experience, I would appreciate if you can do it for me, as it needs selecting those that are doubled-up or unnecessary. Sorry about the mistake. --B. Jankuloski (talk) 14:43, 25 October 2016 (UTC)

I'm sure someone can make a nice sparql query to get items that have both instance of human settlement (Q486972) and instance of some subclass of human settlement (Q486972). Next step would be to remove it with quickstatements. Anyone? Multichill (talk) 18:01, 25 October 2016 (UTC)
Pictogram voting comment.svg Comment This rises some questions I have had regarding P31:småort (Q14839548) and P31:urban area of Sweden (Q12813115)-claims. I am currently adding such claims into items who already have P31:human settlement (Q486972)-claims. I am doing this intentionally without removing the P31:Q486972-claims since P31:Q14839548 and P31:12813115 always should have at least a start date and sometimes also an end date. But that something has started or ended being a Q14839548, does not mean that it has started or ended being a Q486972. P31:Q486972 probably have to stay as "backup" in items with P31:Q14839548 and P31:12813115-claims. Maybe such statements even have to be added to such items, instead of being removed.
To make it more obvious: Solsidan (Q1800816) became a småort (Q14839548) in 1990, and ended that in 2005, when it became a urban area of Sweden (Q12813115). But there have been living people here, probably at least since Iron age. Removing a :P31:human settlement (Q486972) in such cases, would say that nobody lived here before december 31, 1990. What the P31:Q14839548 (start date 1990)-claim says, is that Statistics Sweden started to recognise the place in 1990. -- Innocent bystander (talk) 18:39, 25 October 2016 (UTC)
If you can make a more-specific claim than human settlement, that's probably best. But if you can't, then you should use "human settlement" with an end date of 1990 IMO. --Izno (talk) 18:50, 25 October 2016 (UTC)
If I had access to the National archives and unlimited measures of time, I could look into older official records, but I am afraid I have some limits. -- Innocent bystander (talk) 18:59, 25 October 2016 (UTC)


There are some updates related to items with Swedish urban area code (P775) in the pipeline! A number of items should have this change (the add of P813 in the reference is probably optional). All of them should have a P31:Q12813115-claim already, but if they are missing it, add it! This relates to all items who have P775 with any of these values. If there is no item with any of these PP775-values, please let me know! -- Innocent bystander (talk) 06:02, 26 October 2016 (UTC)