User talk:Ivan A. Krestinin/Archive

From Wikidata
Jump to navigation Jump to search
This page is an archive. Please do not modify it. Use the current page, even to continue an old discussion.

&bot=1

Hi, your bot is not sending &bot=1 so it's changes are showing up in recentchanges. Please fix this. Thanks, Legoktm (talk) 01:30, 30 March 2013 (UTC)

  • Hi, I did this, but bot parameter does not work for wbcreateclaim action. For example Q4518574 was edited using request: wikidata.org/w/api.php?format=txt&token=xxxxxxxxxxxxxxxxxxxx%2B%5C&baserevid=6407630&bot&action=wbcreateclaim&entity=q4518574&property=p31&snaktype=value&&value={"entity-type":"item","numeric-id":4022}. I think this is bug https://bugzilla.wikimedia.org/show_bug.cgi?id=44857. It is fixed, but looks like that the fix was not appied to main server yet. — Ivan A. Krestinin (talk) 02:57, 30 March 2013 (UTC)
  • Hm, I try to replace ...&bot&... to ...&bot=1&... and it works! ...&bot&... form works for another actions (action=wbsetsitelink, action=edit), but not for wbcreateclaim. Looks like this is another bug. — Ivan A. Krestinin (talk) 03:11, 30 March 2013 (UTC)
I've blocked the bot per the below talk page section. Your input is appreciated at WD:AN#KrBot. --Izno (talk) 01:47, 1 April 2013 (UTC)

Creation of new properties without proposal

Hi Ivan. I requested both of your properties you created (P375 & P377) to delete, see here. The reason is you didn't propose this properties as required. You also choice the wrong data type for P377. --Nightwish62 (talk) 01:36, 1 April 2013 (UTC)

Hi. Please stop creating properties without a discussion at WD:Property proposal. Thanks. --Izno (talk) 01:43, 1 April 2013 (UTC)
Hello, is discussion needed in plain cases? — Ivan A. Krestinin (talk) 04:29, 1 April 2013 (UTC)
Yes, for sure. Only because you (or someone else) THINK it's a plain case, it doesn't mean it's really a plain case. E.g. if the property already exist under another name or another data type is more appropriate. Or if the property is unwanted by the community. Last but not least: It's only fair to the other who propose her property. I hope you see the point. If you promise also to propose new property beforehand, I will withdrawn my request for deletion, okay? --Nightwish62 (talk) 10:14, 1 April 2013 (UTC)
I will use WD:PP for creating properties. And possible I try to change project rules to remove PP page as required phase for property creation process. I think post-discussion (same as for items or for pages in Wikipedia) is more effective. It is more friendly because bureaucracy level is less than in case of pre-discussion case. Another problem of PPP page is language - the page is English written. Many editors (include me) have problems with this language. Invalid datatype is not problem: it can be corrected using bot. Duplicate property is not problem too it can be corrected post-factum too. — Ivan A. Krestinin (talk) 10:45, 1 April 2013 (UTC) P. S. Could you please place link to WD:PP to Special:NewProperty page to prevent possible conflicts?
No need to do that, there was already a discussion and the majority supports a restriction of property creation, see here (attention: this site is already archived, so don't reply to this anymore). --Nightwish62 (talk) 12:08, 1 April 2013 (UTC)
What do you mean with you PS? There is already a note Please discuss new properties at Wikidata:Property proposal and wait 2–3 days for consensus before creating any new property when creating a new property. --Nightwish62 (talk) 12:09, 1 April 2013 (UTC)
Russian comment does not contain link to the PP: "Убедитесь, что проверили, не существует ли уже такое свойство! Вы должны указать для всех новых свойств метку (название) и описание, и вдобавок — задать допустимый тип свойства." — Ivan A. Krestinin (talk) 17:59, 1 April 2013 (UTC)


Property:P448 "launch site"

Hi,

The above property is now available and can be used on items. I noticed you first proposed its creation. --  Docu  at 15:05, 21 April 2013 (UTC)

Property:P450 "mission"

Hi,

The above property is now available and can be used on items. I noticed you first proposed its creation. --  Docu  at 19:43, 21 April 2013 (UTC)


Other property requests

Hi Ivan,

I made a request to enable you to create them directly yourself. Please see Wikidata:Requests_for_permissions#Requests_for_the_property_creator_right. If you agree, please comment there. --  Docu  at 08:54, 25 April 2013 (UTC)

По правилам, я могу присвоить Вам, если Вы не возражаете. Выразите, пожалуйста, своё согласие.--Ymblanter (talk) 10:56, 25 April 2013 (UTC)

about adding P373

Hello, per User talk:Dexbot#Invalid Commons links I come to say the latter bug is solved now and one of my friends is checking to find old bugs. but about the former, I thought about that and I think it's not something needs solving. why? let me give you an example:

  • In near future using of commonscat template will be like this:{{commons cat}} so If you want to add commons:Category:Lunokhod program to the articles of w:Lunokhod 2 or w:Lunokhod programme, you should simply add the text I said in the articles of your wiki (we assume it's ru)
  • you don't want to add this cat to the Lunokhod 2 but you want to add to the Lunokhod programme, because of your wiki's policies or anything else
  • you should simply add {{commons cat}} to the Lunokhod programme but not to the Lunokhod 2, because there is not any related "Lunokhod 2" category in the commons that you want to use in your article. As simple as that

I hope you can understand what I mean. I'm not very good at talking :( Amir (talk) 13:12, 25 April 2013 (UTC)

  • {{commons cat}} is not so problematic template. This template can be added or removed simple. Or it can be used in two parameters form: {{commons cat|Lunokhod program|Lunokhod programme}} to indicate difference between article's theme and Commons category's theme. More problematic templates are cards (for example ru:Template:Река), often it contains Commons link at the bottom part (for example see ru:Малый Енисей, after map). May be this is only ruwiki tradition to place Commons link to cards, but this method is widely used in ruwiki. It can be solved using additional card parameter with "suppress Commons link showing" mean, but this is bad way. — Ivan A. Krestinin (talk) 17:47, 25 April 2013 (UTC) P. S. Offtopic: are you know some tool, dump or something another way to extract all property values from wikidata database? (for example P373: "Q1: Universe; Q3: Life; Q4: Death; ...")
So. about that I think you can use {{#ifeq:}} to check the river's name is okay or not. about that other you must use "What links here" at the toolbox or backlinks for the API (for bots or tools) Amir (talk) 10:13, 26 April 2013 (UTC)
ifeq to what? Commons cat (Little Yenisei River) = page name (Малый Енисей)? They are not equal in most cases. backlinks is not good idea, number of requests = N/5000 + N, there N = 182193 for P373. — Ivan A. Krestinin (talk) 18:35, 26 April 2013 (UTC)

Invalid commons category

Thank you. It's fixed now. --Kizar (talk) 10:19, 1 May 2013 (UTC)

Invalid IMDb identifiers

Hi!

I've noticed that you've added some automatically-generated lists at Wikidata:Database reports. I'm curious, would it be possible to get a list of all items with a P:P345 that doesn't match this regexp: ^(tt|nm|ch|co)[0-9]{7}$

The list doesn't need to be continuously updated, since Special:AbuseFilter/17 now reacts when somebody adds an invalid identifier. Gabbe (talk) 08:40, 2 May 2013 (UTC)

Please prevent your bot to update this, since Special:WhatLinksHere/Q123509 shows that no items link to "extinction". --Ricordisamoa 13:47, 4 May 2013 (UTC)

Re: Template:Constraint/doc

There's a bug in Extension:Translate, so I reverted my edit and merged the translations made in the meantime. Now it should be works fine with {{LangSwitch}}. --β16 - (talk) 08:26, 7 May 2013 (UTC)

Bot request

Hi! Thanks for your amazing bot! Is it possible to improve it to remove from the list of constraint violations the violations that doesn't append any more? I would be very useful for pages like Wikidata:Database reports/Constraint violations/P213 where the list of duplicates contains a lot of rows where the issue is fixed. Thanks in advance! Tpt (talk) 06:00, 16 May 2013 (UTC)

They get removed unless an item is deleted. I think Ivan noted somewhere that deletions are not reported in the daily incremental dumps, thus you'd have to wait till the next full dump gets processed. You could also ask our devs to change the daily dumps. --  Docu  at 06:12, 16 May 2013 (UTC)
Ok. Thanks! I'll wait. Tpt (talk) 12:22, 16 May 2013 (UTC)
Hi, I find workaround for this problem. Now reports must be fine. — Ivan A. Krestinin (talk) 17:35, 17 May 2013 (UTC)

Property:P522: Type of orbit

A property you proposed is now available: --Tobias1984 (talk) 09:14, 16 May 2013 (UTC)

type of orbitP522Itemorbit: orbit a satellite has around its central bodyInternational Space Station <type of orbit> low Earth orbit-

Constraint violations/P227

Indeed: great work! A minor correction: Property talk:P227 has a format constraint, that needs to be improved. GND numbers like:

listed on Wikidata:Database reports/Constraint violations/P227 are correct. --Kolja21 (talk) 15:31, 18 May 2013 (UTC)

PS: A helpful rule would be: "Only numbers starting with 1-9 (not 0) and dashes are allowed." --Kolja21 (talk) 16:12, 18 May 2013 (UTC)
You can edit the formatting constraint directly on Property talk:P227. --  Docu  at 16:14, 18 May 2013 (UTC)
Sorry, I don't speak Tech. I only know that "(1|10)\d{7}[0-9X]|[47]\d{6}-\d|[1-9]\d{0,7}-[0-9X]|3\d{7}[0-9X]" seems to be not correct. We can continue the discussion at Property talk:P227#STICKY: Explanation of format constraints. --Kolja21 (talk) 16:57, 18 May 2013 (UTC)

Hi Ivan, maybe you can setup your bot to start making reports for this constraint? You can see example usage at Property talk:P74. Multichill (talk) 11:04, 18 May 2013 (UTC)

Thanks for setting up. I see the first reports coming in. They will become useful when we harvested more taxoboxes. Did you take into account that the template can be on the page twice? See for example Property talk:P74. Multichill (talk) 12:07, 19 May 2013 (UTC)
Hi, than you for message, already fixed, please wait for the report update. — Ivan A. Krestinin (talk) 12:11, 19 May 2013 (UTC)
Seems to work just fine. Cool. I'd reverse the output though. Instead of A1: B1, A2: B1, I'd list: B1 <= A1, A2, A3. --  Docu  at 15:53, 19 May 2013 (UTC)

Names of items

Hi, great work with the constraint violation reports! I was wondering if it would be possible to have the label names of the items also appear in tables such as Wikidata:Database_reports/Constraint_violations/P27#Values_statistics. It would mean users won't have to clikc the item ID number to work out what is being linked to. Thanks again. Delsion23 (talk) 11:59, 18 May 2013 (UTC)

Hi Ivan,

I added the template to a few properties. Most are aren't much used yet, so I think they are unlikely to generate many results. "dentai code" is already the popular list. Curious how it will work. --  Docu  at 13:23, 28 April 2013 (UTC)

Somehow the results on P359 disappeared without the items being fixed. --  Docu  at 19:00, 5 May 2013 (UTC)

The reports seem to run fine now. Excellent work! I added a reference to them on WD:Bots#Statement_adding_bots and suggested it for the weekly news update on WD:Status_updates/Next. On the later, I also mentioned property summary. BTW to make the summary sortable, instead of using "—" the cell should just be empty. Possibly the merged cells in the header row need to go as well. --  Docu  at 05:50, 12 May 2013 (UTC)

If you feel like adding additional checks: Symmetric properties might be easy to do. --  Docu  at 11:35, 12 May 2013 (UTC)

Existing file violations shows "Q4504: Komodo_dragon_distribution.gif" on P181. This despite that the fact that the file exists. --  Docu  at 23:48, 15 May 2013 (UTC)

Some thinks:

  1. {{Constraint:Target required claim}} - too long name, maybe rename it to something short and clean;
  2. {{Constraint:Item}} - too short :-) and unclear, another name needed;
  3. {{Constraint:Item}} - its parameters are not well determined, for example: {{Constraint:Item|Q215627}}, what this mean? Is publisher (P123) = person (Q215627) good for this constraint? Maybe better make it`s params same as in {{Constraint:Target required claim}}?
  4. {{Constraint:Reciprocal}} - conditionproperty and conditionvalue currently are not processed by bot. This is very specific case, maybe remove this functionality? — Ivan A. Krestinin (talk) 19:04, 19 May 2013 (UTC)
I agree about the names for "target required claim" and "item".
The parameters for {{Constraint:Item}} are due to the way Magnus' tool works. Try one of the URLs on Property talk:P131. As it's based on toolserver replication, results are currently somewhat outdated. When replication works, they are live. If you want to include them in your reports, that might be fine. We could still keep Magnus' as a backup.
For {{Constraint:Reciprocal}}, I can remove it if you prefer. --  Docu  at 20:48, 19 May 2013 (UTC)
For Constraint:Item, I updated the syntax and added Magnus' newer site. --  Docu  at 06:25, 20 May 2013 (UTC)
Good. I implement check for it in bot. Some ideas about template naming: target required claim => Value type, Item => Property required. — Ivan A. Krestinin (talk) 06:38, 20 May 2013 (UTC)
I rephrased the English text of "target required claim". BTW I like the term "target" in the name. "Property required" doesn't make it clear if it is about the target or the source item. --  Docu  at 07:15, 25 May 2013 (UTC)

Some details about [1]: there is a bug in my bot`s wiki-parser. I have planes to fix it. To avoid reports generation break lets stay incorrect layout as temporary solution. — Ivan A. Krestinin (talk) 19:09, 20 May 2013 (UTC)

P297 isn't easy to read. Would you add the label and the number of violations to the headers for constraint violation reports for items?

Sample instead of
"Item P107 = Q618123" violations
something like:
item "P107 (P107)" = "geographical feature (Q618123)" violations (7)

For stable anchors, <span id="Item P107 = Q618123 violations"></span> could be added just below or above the header.--  Docu  at 07:20, 25 May 2013 (UTC)


"please do not move logic to lang-specific subtemplates, it is hard to modify all subtemplates if I need change template logic"

На самом деле, менять надо только один подшаблон, для английского языка. После того, как администратор перевода помечает изменение к переводу, это изменение автоматически распространяется на все языковые подшаблоны. --Michgrig (talk) 10:24, 30 May 2013 (UTC)

Кроме того, в случае использования новой схемы все переводы собраны в одном месте, а при использовании LangSwitch'ей еще надо найти, где этот шаблон переводится... --Michgrig (talk) 10:30, 30 May 2013 (UTC)

  • Здесь несколько проблем: изменения обычно носят связанный характер, т. е. я поменял шаблон, затем поменял код бота под это изменение, бот уже поменял страницы, а пока администратор перевода отметит шаблон, всё будет находиться в разваленном состоянии. Потом когда ты начинаешь изменять логику шаблона, то ты не видишь то, как шаблон используется на других языках, для этого тебе нужно пройтись по всем подшаблонам, что затруднительно. Ещё один момент: мне не удобно работать на английском языке, я могу это делать, но мне это тяжело. Также непонятно, если мне нужно внести какое-нибудь изменение в тексты сразу на всех языках, то как это сделать (хотя здесь я возможно просто пока не разобрался с инструментарием). В целом я не против нового механизма, если он будет использоваться только для перевода текстов и внутрь него не будет затягиваться логика работы шаблонов, насколько я вижу это пытались сделать выделив его часть в подшаблон /text, но не довели дело до конца и бросив всё-таки затянули туда и логику тоже. — Ivan A. Krestinin (talk) 17:33, 30 May 2013 (UTC)
  • Попробую ответить по порядку.
    Для начала встречный вопрос: А что, изменения в этих шаблонах часты и глобальны, что в их результате страницы будут находиться в разваленном состоянии? Если часты, то, возможно, пока не стоит переводить их на новые рельсы, а уже потом, когда все стабилизируется.
    "Потом когда ты начинаешь изменять логику шаблона, то ты не видишь то, как шаблон используется на других языках" - а при системе LangSwitch'ей вы проходитесь по всем языкам и смотрите, как ведет себя шаблон?
    "Ещё один момент: мне не удобно работать на английском языке, я могу это делать, но мне это тяжело." - т.е. вы делаете русские тексты, а потом кто-нибудь добавляет английские? Это серьезный аргумент. Могу предложить свою помощь с английским - если надо, стучитесь в скайп (см. мою ЛС ру-вп).
    Насчет системы перевода. Если вы посмотрите любую переводимую страницу (хотя бы подстраницу /text этого шаблона), то увидите, что там используются теги <translate></translate>, которые обрамляют текст, который надо переводить. После простановки этих тегов традмин помечает страницу к переводу, а софт проставляет внутри этих тегов другие, <!-- Т:ХХ -->, формируя т.наз. транслейшн юниты - переводимые блоки. При изменении исходной страницы и пометки к переводу специальный бот разносит новые версии страницы по уже имеющимся переводам. При этом те изменения, которые не относятся ни к одному транслейшен юниту, применяются сразу, а если в исходном тр.юните что-то поменялось, то его перевод устаревает, и переводчики должны поправить соответствующий перевод. Т.е. если вам нужно внести какое-то изменение на все языки, то после правки надо подождать/дернуть традмина. После этого бот быстро сработает (в большинстве случаев, в пределах нескольких секунд).
    "В целом я не против нового механизма, если он будет использоваться только для перевода текстов и внутрь него не будет затягиваться логика работы шаблонов" - Одним из основных отличий от системы LangSwitch'ей является необходимость выносить в подстраницу всего блока, содержащего требующие перевода фрагменты. Иногда туда приходится вносить и некоторую часть кода, вызывая потом в основном шаблоне также с переменными.
    Есть, правда, еще одна модель, которая применяется в основном в старой модели перевода (см., например, Template:Anonblock) - подстраница /layout с разметкой шаблона и вызов этой подстраницы в основном шаблоне. Такую же модель использует шаблон Template:Policies с новой системой перевода.
    "насколько я вижу это пытались сделать выделив его часть в подшаблон /text, но не довели дело до конца и бросив всё-таки затянули туда и логику тоже." - я не так давно являюсь традмином, поэтому не всегда сразу знаю, как что делается. Хотя в случае с этим шаблоном я не вижу в истории каких-то метаний туда-сюда. Это было где-то на другой странице. --Michgrig (talk) 21:13, 30 May 2013 (UTC)
  • Иван, ответьте, пожалуйста, а то я не пойму - убедил я вас в своей правоте или нет, могу ли я снова выносить переводимые фрагменты в подшаблоны или нет... --Michgrig (talk) 11:48, 3 June 2013 (UTC)
    • Прошу прощения, был занят другими задачами. Относительно количества изменений, можете посмотреть историю этих шаблонов. Сейчас назревает потребность в расширении функционала, с соответствующей модификацией шаблонов. С LangSwitch-ами я действительно прохожусь по всем языкам и либо меняю там тексты, либо модифицирую новую логику с учётом потребностей отдельных языков. По поводу русского языка я обычно поступаю так: пишу шаблон на-русском, отлаживаю его с помощью предпросмотра, когда всё готово, то делаю английский перевод, после этого записываю изменения. Когда совсем надоедает возиться с английским, то оставляю только русский. Мне было бы очень неудобно, если бы в окне предпросмотра во время модификации и отладки шаблона отображался бы английский текст. По поводу метаний, я уже не помню с каким конкретно шаблоном они были, кажется участник Beta16 с этим возился, но это не важно, основной момент в том, что в подшаблоне /text так и осталась часть логики, а не только тексты. — Ivan A. Krestinin (talk) 12:14, 3 June 2013 (UTC) P. S. Вообще система перевода не самая простая в устройстве, боюсь, когда её сложность умножается на нетривиальность системы шаблонов, то получается уж совсем неудобовариваемый зверь, с которым разобраться уже практически никто не может, хотя бы из-за невозможности его обозреть ввиду размазанности по десяткам страниц.
      • ОК. В принципе, нет никакой кровь-из-носной необходимости переводить шаблоны на новую систему переводов во что бы то ни стало, поэтому настаивать не буду. Может, впоследствии получится сделать так, как я описывал для {{Policies}} --Michgrig (talk) 12:50, 3 June 2013 (UTC)

Constraint: is subclass of

Hello, I was wondering if it would be possible for you to create a constraint for the P168 (P168) property, that would say that the value must be a subclass of Q811979, that is to say that the value should be an item that has the statement P:P279: Q811979 or is itself of subclass of Q811979.--Zolo (talk) 16:11, 19 May 2013 (UTC)

You could use this. The "is itself of subclass of Q811979"-part can't currently be done, but is probably not really needed. --  Docu  at 16:17, 19 May 2013 (UTC)
It needs to be recursive, because the value itself does not need to be a first-degree subclass of Q811979 (actually, it should probably never be, because it would it would mean that it uses a much too imprecise item). --Zolo (talk) 16:55, 19 May 2013 (UTC)
I think "One of" is better for P168 (P168) because it is more strong. Negative cite of it is manual work for subclass checking. I will think about it, maybe find some general solution for this constraints of this type. — Ivan A. Krestinin (talk) 19:19, 19 May 2013 (UTC)
I see only two cases, where using subclass of (P279) would not work:
  1. the value is incorrect
  2. the value does not have the correct chain of subclass of (P279)
Case 2 is likely to produce many errors, but I do not think it is a bad thing. It would make the manual work would become more interesting (adding relevant P279 values to items instead of making flat lists of accepted values). --Zolo (talk) 19:55, 19 May 2013 (UTC)
You could find another item to track them and use Template:Constraint:Target required claim instead. --  Docu  at 20:41, 19 May 2013 (UTC)
✓ Done, {{Constraint:Value type}} is created for these cases. — Ivan A. Krestinin (talk) 17:33, 24 June 2013 (UTC)

Property:P620 "landing date", Property:P621 "launch date"

Hi,

The above property is now available and can be used on items. I noticed you first proposed its creation. --  Docu  at 20:40, 9 June 2013 (UTC)

The list IS the problem

Hoi, you indicate on my talk page that I should solve the problem. As far as I am concerned there is no problem. Consequently when it is not clear WHY this list exists and why I should care, the list IS the problem. GerardM (talk) 16:30, 17 June 2013 (UTC)

I think GerardM's question was: is it useful to have constraint violations ? Personnally, I think that is some of them are useeful, for sanity checks, or to help find missing statements. But I I think that constraint violations implying P:P107 are useless because P107 itself is useless. --Zolo (talk) 19:43, 19 June 2013 (UTC)

  • Could you explain a bit how you use it ? I think I will try to make a proposal to replace it by a better system, so that would help me. --Zolo (talk) 20:14, 19 June 2013 (UTC)

Empty second parameter in Template:Label

Hello, Ivan A. Krestinin. You have new messages at Template talk:Label#Empty second parameter.
You can remove this notice at any time by removing the {{Talkback}} or {{Tb}} template.

--Ricordisamoa 00:43, 30 June 2013 (UTC)

RfC on Wikidata's primary sorting property

You recently participated in a deletion discussion for P107 - main type (GND). The discussion has been closed, as it is clear that a resolution won't come from PfD, and an RfC has been opened on the matter at Wikidata:Requests for comment/Primary sorting property. You are invited to participate there. Please note that this is a mass delivered message, and that I will not see any replies you leave on this page.

Yours, Sven Manguard Wha? 18:24, 30 June 2013 (UTC)

The property spacecraft bus (P707) that you proposed is available now. --Tobias1984 (talk) 09:21, 24 July 2013 (UTC)

The property JPL Small-Body Database SPK-ID (P716) that you supported is available now. Please help out with other science identifiers at Wikidata:Property_proposal/Term#Chemistry_.2F_Chemie_.2F_Chimie and Wikidata:Property_proposal/Term#Stratigraphy. --Tobias1984 (talk) 09:09, 26 July 2013 (UTC)

The property Minor Planet Center observatory code (P717) that you supported is available now. --Tobias1984 (talk) 09:14, 26 July 2013 (UTC)

The property OKATO ID (P721) that you supported is available now. --Tobias1984 (talk) 09:34, 28 July 2013 (UTC)

The property asteroid family (P744) that you supported is available now. --Tobias1984 (talk) 22:24, 1 August 2013 (UTC)

Mass-merger

Hello, Ivan A. Krestinin. You have new messages at Wikidata:Bot requests#Merge multiple items.
You can remove this notice at any time by removing the {{Talkback}} or {{Tb}} template.

--Ricordisamoa 01:49, 3 August 2013 (UTC)


The property Lake ID (Sweden) (P761) that you supported is available now. --  Docu  at 14:36, 4 August 2013 (UTC)


The property OKTMO ID (P764) that you supported is available now. --  Docu  at 16:05, 4 August 2013 (UTC)

There seem to be issue about me creating properties like the above, please see Wikidata:Administrators'_noticeboard#User:Docu_and_the_property_creator_permission. --  Docu  at 18:03, 4 August 2013 (UTC)

The property INE municipality code (P772) that you supported is available now. --Tobias1984 (talk) 08:04, 7 August 2013 (UTC)

Объединение

Приветствую, Иван! Подскажите, как объединить эту страницу с этой? Лукас Фокс (talk) 11:59, 9 August 2013 (UTC)

Не то удалили. М-дя.... ))) Лукас Фокс (talk) 03:41, 10 August 2013 (UTC)

Daily update

Hi Ivan,

In the past, the daily updates took place somewhen in the evening. This had the advantage that items fixed that day disappeared in the next update. Now the update takes place just about when the next incremental dumps start. So it takes 2 days to see fixed items disappear. If you have the possibility, it would be nice if we could have the updates a bit earlier.

In general, I think the reports work well and help us improve the data. --  Docu  at 18:51, 14 August 2013 (UTC)

Constraint: Required claims for items in this category

Hi Ivan, these last days we were having a discussion about how to add statements to items contained in certain categories. Instead of doing it with properties it would be better to do it with a template in the talk page of the category items. What do you think about doing this as a Constraint template "Required claims for items in this category"?--Micru (talk) 14:29, 20 August 2013 (UTC)

Interesting idea, but there are technical problems: the bot need information both from Wikidata and Wikipedias. Bot needs a lot of dumps (every dump size is 10-20 Gb) or it must use API access to servers, but API is slow for large number requests. I think that toolserver/wikilabs interactive tool is more perspective for this task. Additionally report-based system is not good because the report will contain many false-positives. — Ivan A. Krestinin (talk) 17:44, 20 August 2013 (UTC)
Thanks for your comments! Do you know of anyone who has experience with the toolserver/wikilabs interactive tool?--Micru (talk) 20:37, 20 August 2013 (UTC)
Maybe these tools author? — Ivan A. Krestinin (talk) 20:56, 20 August 2013 (UTC)

The property you suported is now available. --Fralambert (talk) 22:58, 20 August 2013 (UTC)

Constraints on locations

Hi Ivan,

You might want to look into Wikidata:Property_proposal/Place#Locatable_on_earth. --  Docu  at 07:43, 24 August 2013 (UTC)

Hi Ivan,

Somehow the above report didn't get updated with these changes. --  Docu  at 08:01, 24 August 2013 (UTC)

This template combines the two constraints of P107 (P107) -> person (Q215627) and has sex or gender (P21). But this means that the constraints are no longer being reported in the constraint reports. It would be really good if this could be fixed. Harryboyles (talk) 07:22, 26 August 2013 (UTC)

Пожалуйста, добавьте поддержку этого шаблона и ещё {{Constraint:Taxon}}. Вы ведь можете загружать шаблон, извлекать из него актуальные ограничения и дальше действовать по накатанной. Infovarius (talk) 18:27, 9 February 2014 (UTC)

Good news

Today's incremental dump was created! Hopefully it didn't skip three days. --  Docu  at 21:31, 27 August 2013 (UTC)

File size suggests it might include several days of changes. --  Docu  at 21:34, 27 August 2013 (UTC)
Good. I run bot already. — Ivan A. Krestinin (talk) 21:43, 27 August 2013 (UTC)
Thanks. Seems we might be missing edits from Saturday. --  Docu  at 22:26, 27 August 2013 (UTC)
Apparently the missing days have been re-run one-by-one. --  Docu  at 01:54, 29 August 2013 (UTC)
Sorry, I do not understand your mind. Missed dumps cause errors in reports. The errors will be fixed with the next full dump. Part of the errors will be fixed earlier because every item`s change causes full item save to incremental dump. You can make some minor edit of critical item to force the process. — Ivan A. Krestinin (talk) 14:58, 29 August 2013 (UTC)

It might be fairly complete if you load the following in this sequence:

  1. http://dumps.wikimedia.org/other/incr/wikidatawiki/20130824/
  2. http://dumps.wikimedia.org/other/incr/wikidatawiki/20130825/
  3. http://dumps.wikimedia.org/other/incr/wikidatawiki/20130826/
  4. http://dumps.wikimedia.org/other/incr/wikidatawiki/20130827/
  5. http://dumps.wikimedia.org/other/incr/wikidatawiki/20130828/

This even if the 28 was already loaded. I guess we could also wait till the next full dump. --  Docu  at 18:57, 29 August 2013 (UTC)

"Single value" constraint: allow repeated but qualified values

Hi Ivan, thanks a lot for your good work on constraint violation reports!

Looking at false positives in Wikidata:Database reports/Constraint violations/P268#"Single value" violations, I think we should not report violations of template:Constraint:Single value when a value is repeated but with different qualifiers:

Any different qualifier or any different qualifier value should be enough to make them non-single.

Thanks - LaddΩ chat ;) 13:15, 30 August 2013 (UTC)

  • Hi, interesting case. I think we use non-optimal data model. Pseudonym is not simple synonym of the author name. Pseudonym can have such properties as "the first/last usage", different identifiers (as in your sample) and etc. Looks like we need separate Q-entities for every pseudonym. This solves problem that you describe too. — Ivan A. Krestinin (talk) 15:39, 30 August 2013 (UTC)
Mmmm yeah, we could do it that way... but I still think that the uniqueness should be on (value+[qualifier+qualValue]*) - LaddΩ chat ;) 16:40, 30 August 2013 (UTC)
For Jan and Joël Martel (Q1312123), I made two additional items. I think the VIAF can be removed from Jan and Joël Martel (Q1312123). --  Docu  at 17:06, 30 August 2013 (UTC)
Thanks Docu. I removed both VIAF & BnF identifiers from Jan and Joël Martel (Q1312123) - though I suspect that we just made the life more difficult for the 3 wikipedias that were linking to that item :) - LaddΩ chat ;) 17:44, 30 August 2013 (UTC)

Scope of P766 (P766)

Ivan

As we don't seem to be getting any closer to agreement on P766 (P766) I have asked on Project chat for other people to have a look at this and say what they think. Filceolaire (talk) 23:44, 31 August 2013 (UTC)

Please comment Wikidata:Project_chat#Problem_with_2_properties about this problem. Snipre (talk) 09:47, 1 September 2013 (UTC)
This is another problem as my English knowledge allows to see. My common position is: do not mix different relation types in one property either if these relations have same name on some language. Every property must accurately define relation between linked objects. — Ivan A. Krestinin (talk) 10:26, 1 September 2013 (UTC)

New RFC

I have started an RFC on 'Place' related properties including P766 (P766). Filceolaire (talk) 17:43, 2 September 2013 (UTC)

Unique value with qualifiers

Hello, do you think you could upgrade the "unique" constraint" so that there is an option to restrict it to values using a particular qualifier. That makes it a bit more complicated, but I think it will be very useful. See Property talk:P528#Extension of the property. --Zolo (talk) 05:53, 5 September 2013 (UTC)

  • I think we need no extend this property, but divide it to multiple properties, one property per notable catalog. Mixing different catalogs in one property make data usage too complex without needless. — Ivan A. Krestinin (talk) 16:57, 5 September 2013 (UTC)
    • In a sense, it is a bit strange to have specific properties for some catalogues, but not for others is a bit strange, but there are really many possible catalogues, that would make a lot of properties. And in some ways, specific properties have problems too. If we use a catalogue property using a catalogue item as a value, we can look at the item about the catalogue to know what it is about, while we cannot do that when we useproperties speficific to one catalogue (there is no machine-readable documentation of properties). Anyway, qualifiers are part of Wikidata, so if your tools can take them into account, it will certainly be useful at some point. --Zolo (talk) 20:28, 5 September 2013 (UTC)

Property

Hi Ivan! Can you look at this proposal? I think if it is renamed to "carries instruments it could be useful for the space task force. --Tobias1984 (talk) 19:33, 6 September 2013 (UTC)

New property

Hi Ivan,

One of the properties you proposed is now available: State Water Register Code (Russia) (P884). --  Docu  at 20:11, 11 September 2013 (UTC)

New property

Hi,

One or several properties you proposed are now available: United States Navy aircraft designation (P847), United States Army and Air Force aircraft designation (P897), Japanese military aircraft designation (P849), --  Docu  at 21:22, 11 September 2013 (UTC)

BTW, would you have a look at this proposal? --  Docu  at 11:29, 15 September 2013 (UTC)

Babel tool

Hi Ivan, I much appreciate your contributions to Wikidata. I noticed that you do not use the Babel tool, very useful to simultaneously display labels & descriptions on more than one language, for any item that you view. You can use it by adding a template of this form on your user page User:Ivan A. Krestinin : {{#babel:ru-N|en-3|uk-2|de-2}}. Thanks a lot for all your work - LaddΩ chat ;) 23:17, 19 September 2013 (UTC)

Bots adding sitelinks to Commons

Hello, your bots is currently adding many links from normal items to common categories. I do not think we have a consensus to do things this way (heterogeneous namespaces may prove troublesome). --Zolo (talk) 13:08, 27 September 2013 (UTC)

Hi, as I see most part existing interwiki links are between categories and articles, see commons:Category:William Gullick, commons:Category:Windsor (Siřejovice) as examples. The number of category-category links is less. So consensus exists, bot only copy existing interwiki to Wikidata to make its synchronization more simple. I tested this linkage type, it works fine, so I do not think that some troubles exists with it. — Ivan A. Krestinin (talk) 18:13, 27 September 2013 (UTC)
Yes, but current interwikis on Commons are messy, and sometimes the gallery has the same interwikis as the category. There is a discussion at Commons:Commons:Village Pump#Wikidata is here!. --Zolo (talk) 07:07, 29 September 2013 (UTC)
Galleries is not major type of entities on Commons. Galleries exists to very small number of themes (~ 4%). Galleries are often outdated and there are no resources to update its. We avoid to link articles to galleries and prefer categories at least in ruwiki. Sometimes I head comments like "Let`s delete galleries at all". We can link galleries using separate property, if its really needed. So, galleries are not problem. — Ivan A. Krestinin (talk) 08:47, 29 September 2013 (UTC)
It's not the importance of the galeries (I also beleive they are useless), but the nature imself of the galleries, who are “wikimedia category”. Your robot linked the commons category to Woodlands County (Q8032936), but the category, Category:Woodlands County (Q6987959), already exist. I think the best would be to create a RfC on the subjet before making massive import. --Fralambert (talk) 15:26, 29 September 2013 (UTC)
Please see commons:Category:Woodlands County, Alberta, it is linked with en:Woodlands County, Alberta, not with en:Category:Woodlands County, Alberta. This is not only this page`s case. This is commonly used practice. Bot does not change it. If we want to change commons interwiki system we need some global discussion, not RfC on Wikidata. — Ivan A. Krestinin (talk) 16:25, 29 September 2013 (UTC)
I am prepared to stop this bot if you do not stop it yourself, because it is very unclear to me that your edits have consensus.
Thanks. --Izno (talk) 15:28, 29 September 2013 (UTC)
Ok, I stoped the bot. Please clarify why you think that existing interwiki system on commons has no consensus? — Ivan A. Krestinin (talk) 16:25, 29 September 2013 (UTC)
I do not think that it has consensus here, however accurately it reflects the previous arrangement (and that's pretty clear from the above two users commenting, as well as the description at WD:Wikimedia Commons). I personally don't think it's a smart decision (or correct) to have links from page to category directly via the interwiki system, but would rather leave my rationale to a discussion at an RFC (whether hosted here or at Commons and with the Commons users taking part of course). There are also Commons users who disagree with the practice....
I'll go unblock the bot now in case you want to run other tasks. --Izno (talk) 16:50, 29 September 2013 (UTC)
Direct linking is more rationale because commons category is additional materials for Wikipedia article. So link between category and article is more important then category-category. There are another reasons. But you are right, full discussion is needed, please stop another import bots until it complete. — Ivan A. Krestinin (talk) 17:44, 29 September 2013 (UTC)
When you will start the RFC, just leave me a note. --Fralambert (talk) 18:22, 29 September 2013 (UTC)
I start it Wikidata:Requests for comment/Commons links, but it needs to be translated before publishing. And as I told above it is needed to be moved to some another project/place because its affects to Commons, Wikidata, Wikipedia, not only Wikidata. — Ivan A. Krestinin (talk) 19:12, 29 September 2013 (UTC)
Translation is finished, you can review. I plan to make notification on forum and another places tomorrow. — Ivan A. Krestinin (talk) 21:28, 1 October 2013 (UTC)
I put the notification on the RFC main page... And I see that you will make it the next day. sorry. --Fralambert (talk) 21:46, 1 October 2013 (UTC)
In Russian their saying is "No need to run in front of railway engine". — Ivan A. Krestinin (talk) 22:03, 1 October 2013 (UTC)
It's most dangerous that the french equivalent who is to "put the plough before the oxen". (Mettre la charrue devant les boeufs). --Fralambert (talk) 02:51, 2 October 2013 (UTC)
Is it used when something is made too early? I started the RfC. — Ivan A. Krestinin (talk) 17:45, 2 October 2013 (UTC)

Thank you Ivan, for putting the RFC together. --Izno (talk) 00:14, 3 October 2013 (UTC)

NUTS

Please, can you remove P605 from all items in "Format" violations" in Wikidata:Database reports/Constraint violations/P605? These were replaced by another property. JAn Dudík (talk) 05:34, 9 October 2013 (UTC)

"Format" violations at Format constraints for properties

Hi! Please see the note at Property talk:P947#Format constraints - length_check. This is a general issue for all similar sections. Thanks! לערי ריינהארט (talk) 06:22, 12 October 2013 (UTC)

Is it possible to add the full timestamp with hours and UTC as in 06:22, 12 October 2013 (UTC)? לערי ריינהארט (talk) 09:41, 12 October 2013 (UTC)

Property talk:P268 BnF identifier - constraint violation format - broken links

Please see Property talk:P268#BnF identifier - constraint violation format - broken links. Can you help please? Thanks in advance! לערי ריינהארט (talk) 09:34, 12 October 2013 (UTC)

Thanks a lot! לערי ריינהארט (talk)

Constraint violations report does not update properly

Please see Property talk:P268:Constraint violations report does not update properly . Regards לערי ריינהארט (talk) 07:03, 16 October 2013 (UTC)

Please help at a regex and UTF-8 showstopper

Dear Ivan, since last night I am puzzled about some javascript lines in my clone version of Magnus Manske's authority_control.js tool.
I tried to expand it to detect AC datas at WMF projects where the template name is using UTF-8 characters. Please look at my clone version and please serach for the following lines:

// else if ( lang == 'ro' ) st = t.match ( /{{Informații bibliotecare.+?)}}/i ) ;
// else if ( lang == 'ru' ) st = t.match ( /{{Библиоинформация.+?)}}/i ) ;
// else if ( lang == 'yi' ) st = t.match ( /{{ביכער(.+?)}}/i ) ;

As far as I know regex supports only some (ASCII) characters. Reffering to control characters and or UTF-8 requires another coding.

I am not able to enter the developers irc channel. If you can not answer this UTF-8 question you may ask the developers in the irc channel. Thanks for any help! Greetings from Munich Germany (gangleri aka 11:15, 12 October 2013 (UTC)) לערי ריינהארט (talk) 11:15, 12 October 2013 (UTC)

instance of human - my edits: instance of (P31): human

Hi! I could see that you edit instance human by adding sources. I was wondering what information can be retreived from articles at eo.Wikipedia and yi.Wikipedia. Do you need a template? Regards לערי ריינהארט (talk) 14:26, 13 October 2013 (UTC)

please help

Dear friend! Please see Wikidata:Requests for comment/links to wikidata from a template at an arbitrary WMF project page. Can you help please? Maybe you have done something similar or you heard about similar ongoing works. Thanks in advance! לערי ריינהארט (talk) 15:24, 16 October 2013 (UTC)

ISBN

Hi Ivan, can you please take a look at ISBN-10 (P957)? It needs a constraint template similar to ISBN-13 (P212) but I have no idea how to write it. --Kolja21 (talk) 17:48, 17 October 2013 (UTC)

Thanks! --Kolja21 (talk) 21:29, 17 October 2013 (UTC)

Hi Ivan, could you implement this in your bot? It's like {{Constraint:Single value}}, but instead of count being 1 it should be more than 1. Multichill (talk) 12:49, 19 October 2013 (UTC)

Instance of person

Please stop your bot importing these values.

Thanks. --Izno (talk) 14:51, 19 October 2013 (UTC)

Stopped, please clarify the problem. — Ivan A. Krestinin (talk) 16:43, 19 October 2013 (UTC)
For humans (items with Q5), there should be no duplication (and there are several cases that I saw where there was already a P31 claim of Q5 and the use of "person" was introduced for what I could see was no reason). This is per the consensus previously established at the P107 RFC. In general, the use of "person" should not be used with P31 unless there is not already a P31 claim, and certainly not reintroduced. --Izno (talk) 16:56, 19 October 2013 (UTC)
Hm, strange. Maybe I do not understand English meaning of these terms. In Russian there are two terms/classes: "человек" (human?) and "личность" (person?). "человек" is not subclass of "личность". "личность" is not subclass of "человек". For example Ilias Venezis (Q5360627) is instance of "человек" and Ilias Venezis (Q5360627) is instance of "личность". So I add two claims. Situation in English is different? — Ivan A. Krestinin (talk) 17:10, 19 October 2013 (UTC)
Yes, human (Q5) is subclass of person (Q215627) in English (and in Wikidata, as it were, as human (Q5) subclass of natural person (Q154954) subclass of person (Q215627)). --Izno (talk) 17:22, 19 October 2013 (UTC)
In Russian this sounds very strange. human (Q5) is about human as biological organism (homo sapiens). natural person (Q154954) is person as role in jurisprudence. person (Q215627) is person as social member. So these subclass sequence is wrong in Russian. Looks like old problem then iwiki links similar, but not exact the same terms. — Ivan A. Krestinin (talk) 17:41, 19 October 2013 (UTC)
I'm not sure that's unintentional. The point of using Q5 for English at least is that "person" is not well defined, perhaps in any language. "Human" aka "Homo sapiens" (sapiens) is. --Izno (talk) 18:13, 19 October 2013 (UTC)

Whitelist

Hello Ivan, is it possible to write a kind of whitelist for constraint violations? For example GND ID (P227):

"Unique value" violations

In this (rare) case it's not an error, but I don't know where I can note this. --Kolja21 (talk) 22:12, 20 October 2013 (UTC)

  • Currently this is not implemented. I will try to implement this. Possible additional parameter for template, something like: {{Constraint:Unique value|exceptions= {{Q|13261}}, {{Q|208008}} }}. — Ivan A. Krestinin (talk) 06:38, 21 October 2013 (UTC)

That would be great. Thank! --Kolja21 (talk) 19:54, 21 October 2013 (UTC)

New functionality testing: It worked. Congratulations!
"Violations count: 12 (+ 1 exceptions)"
Two questions: Would it be useful to list the exceptions on the database report? And/or: Could the parameter "exceptions" be linked to a separate page or list? (This might be useful for properties with more than ten or twenty exections.) --Kolja21 (talk) 21:14, 21 October 2013 (UTC)
1. I make 'exceptions' parameter visible, so exceptions are displayed on property talk page. 2. currently discussion about property metadata is in progress, lets wait until its completed, maybe more good storage will be available. — Ivan A. Krestinin (talk) 20:10, 22 October 2013 (UTC)
Thanks again. --Kolja21 (talk) 09:30, 23 October 2013 (UTC)
We have the same problem for P225. There are currently >900 [[Wikidata:Database_reports/Constraint_violations/P225#.22Unique_value.22_violations|unique value violations], out of which 1/3 are false negatives (for example, there is a plant and an animal with the same taxon name), 1/3 which have already been checked (some cannot be corrected if there is a single Wikipedia having two articles for the same taxon), and 1/3 true negatives, which we want to fix but are hard to find in the list. A whitelist would be great. It could also work, if KrBot could preserve comments after a violation: example which could be preserved as long as the violation exists.  — Felix Reimann (talk) 12:44, 25 November 2013 (UTC)
Hi, preserving comments is difficult task: bot need not only generate report, but parse it. You can create white list on property talk page, comments are possible in it: example. — Ivan A. Krestinin (talk) 20:19, 25 November 2013 (UTC)

possible mistake

Hello, can you check this out? thank you Amir (talk) 12:54, 21 October 2013 (UTC)

Bot doing property deletions for constraint violations

I see that your bot is doing deletions of properties of constraint violations. Is it documented somewhere what it is doing? I checked the RfP and it is not mentioned there, nor on KrBot's user page. It would be great if that could be remedied. Thanks.  — billinghurst sDrewth 23:41, 23 October 2013 (UTC)

URLENCODEbug

Dear friend, Please see MediaWiki talk:Gadget-AuthorityControl.js#URLENCODEbug. I am not sure if you get the point. If you do not know what to do please let me know who can come to the #mediawiki-wikidata irc channel / chat. Thanks in advance! לערי ריינהארט (talk) 06:45, 1 November 2013 (UTC)

somevalue и novalue

Здравствуйте! Хотел спросить, можно ли как-нибудь для шаблона {{Constraint:One of}} задать значения "novalue" и "somevalue". Например, свойство P60 (P60) вполне может принимать такие значения согласно источникам, например IC 880 (Q3711469) в источнике описывается как "Object of unknown nature"; также в том же источнике попадались записи с прочерком вместо типа (данные из этого источника добавлял в основном мой бот, а у него был баг, который не позволял записывать "novalue", поэтому конкретный пример сейчас затрудняюсь привести). --Art-top (talk) 15:49, 1 November 2013 (UTC)

  • Добрый день, пока нельзя, можете в шаблоне указать {{Constraint:One of|values = ..., {{Q|2088753}}, somevalue, novalue }}, на этих выходных постараюсь сделать, чтобы это заработало. — Ivan A. Krestinin (talk) 18:10, 1 November 2013 (UTC)
  • Ivan, спасибо, это не срочно, как получится, так получится. Можно вопрос немного не в тему? Может быть вы знаете, какое ограничение на длину поля data в запросе wbeditentity? --Art-top (talk) 22:40, 2 November 2013 (UTC)
    • Не знаю, не сталкивался, единственное: когда-то я на проектах не относящихся к Википедии сталкивался с ограничением на общую длину URL, которая настраивалась в файле конфигурации веб-сервера. Но чему равно значение этой настройки здесь - не знаю. — Ivan A. Krestinin (talk) 05:25, 3 November 2013 (UTC)

blacklistVIAF blacklistLCCN blacklistNLR

Hi! Please see

Property talk:P214#blacklistVIAF Property talk:P214#new constraint violations about deprecated VIAF identifiers
Property talk:P244#blacklistLCCN
Property_talk:P1003#blacklistNLR

Can you make such a constraint? Thanks! Regards לערי ריינהארט (talk)

Name    Property VIAF values  WD values    Bad codes    Error rate
ISNI    P213     3498188      190616       54619        28,65%
VIAF    P214     24498513     437784       4110          0,94%
GND     P227     9714886      225272       7627          3,39%
LCCN    P244     8786429      160741       350           0,22%
ULAN    P245     224809       443          1             0,23%
BNF     P268     1914209      17836        229           1,28%
SUDOC   P269     2303361      16034        106           0,66%
NDL     P349     1000804      35608        10044        28,21%
NLA     P409     1008519      983          1             0,10%
NKC     P691     681021       13809        449           3,25%
SELIBR  P906     174934       959          1             0,10%
RSL     P947     997          58           21           36,21%
NLI     P949     520131       744          0             0,00%
BNE     P950     538993       515          0             0,00%
NLR     P1003    146095       82           59           71,95%
PTBNP   P1005    321064       272          0             0,00%
NTA     P1006    2445197      964          0             0,00%
BIBSYS  P1015    63718        0            0             -
BAV     P1017    326876       0            0             -

Ivan A. Krestinin (talk) 08:56, 25 November 2013 (UTC)

Commons/cat

Hello, could you generate constarint reports for P935 in the same format like P373? I mean checking of existence and formatting. JAn Dudík (talk) 08:09, 12 November 2013 (UTC)

мольба о помощи :)

Джерард попросил меня сделать одно дело, и я поняла, что его масштабы огромны. Можно ли для того, чтобы оно стало посильным, попросить пройтись ботом по этому списку и проверить, что там везде указан женский пол, профессия=художник, попробовать добавить гражданство, а если не выйдет - дать мне список тех записей, по которым останется пройтись вручную?.. Пожалуйста, пожалуйста, пожалуйста. Список взят из категорий художников, пол проверяла вручную (по имени). Lvova (talk) 10:06, 12 November 2013 (UTC)

  • Пол всем персонам чьё имя оканчивается на "вна", "вич" бот уже должен был расставить. С профессией похоже не всё однозначно, например, ru:Авакова, Светлана Ивановна — скульптор. С гражданством постараюсь сделать алгоритм берущий его из карточки. — Ivan A. Krestinin (talk) 17:25, 12 November 2013 (UTC)
  • Сделал следующее: где была Скульпторы по алфавиту, добавил профессия = скульптор, где Художники по алфавиту - художник. Гражданство добавил там, где была одна страна, чтобы добавить несколько стран нужно переделывать бота. Пол, как уже говорилось был расставлен ранее. — Ivan A. Krestinin (talk) 07:17, 18 November 2013 (UTC)

Dear friend, can you please update Wikidata:Database reports/Constraint violations/P268?
Ladsgroup · talk fixed the Bibliothèque nationale de France ID (P268) values adding the "modulo 29 checksum" (and removing the cb prefix) search for "Set a claim value: BnF identifier (P268):" in your browser at special:Contributions/Dexbot|offset=2013111120&target=Dexbot. Thanks @Ladsgroup for your work!
Best regards לערי ריינהארט (talk) 15:28, 12 November 2013 (UTC)

KrBot - adding "reference" without importing

‎Hi, I spotted multiple times your KrBot addeding "imported from" reference to the claims, that were not imported by your bot (and also not from stated source). I guess that is not intentional behaviour of your bot. Please improve your bot to avoid that behaviour in future. Thanks. --Jklamo (talk) 16:26, 13 November 2013 (UTC)

  • Hi, this behavior is needed to avoid situations when request that set value passed but request that adds source failed. In this case bot add source on the second pass, but it is difficult to recognize values added by my bot and added by another bots. Bot does not add source if any source already specified or if the value does not equals to specified in "imported from" source. Is it really needed to avoid adding source in case then the value have no sources at all? — Ivan A. Krestinin (talk) 19:05, 13 November 2013 (UTC)
    • I do not see this behavior in actions of other bots importing from wikis, so I guess it is not technical problem (just do not use second pass). I think "imported from" is just info about source, but not a reference, so better none value than false value. --Jklamo (talk) 00:38, 20 November 2013 (UTC)
      • Second pass is needed because servers fail to save changes sometimes. "imported from" is important information, it allows to trace invalid values sources and fix its if needed. I do not think that "imported from" + non original site is false value. For example if I do 1, 2 and 3, is "imported from" correct now? Bot`s edit is equals to this actions sequence, only actions 1 and 2 are not done as do nothing. — Ivan A. Krestinin (talk) 20:02, 20 November 2013 (UTC)

a problem in the database report

Hi, see this edit [6] the time is "14 November 2013 9:09 UTC" and Antonius de Butrio (Q3620350) exists in the list but the item (and almost all other listed items) has been fixed in Nov 12 [7], can you check it? Amir (talk) 06:59, 16 November 2013 (UTC)

Image removal

Hi there Ivan. I don't understand this edit of yours. What happened? Cheers. Pikolas (talk) 13:25, 18 November 2013 (UTC)

bot edit summaries

Dear friend, please add the English property name at the edit summaries made by special:Contributions/KrBot in the project namespace. Example:

22:46, 17 November 2013 (diff | hist) . . (0)‎ . . Wikidata:Database reports/Constraint violations/P569 ‎ (update)

could be

22:46, 17 November 2013 (diff | hist) . . (0)‎ . . Wikidata:Database reports/Constraint violations/P569 ‎ (date of birth update)

oer eaven linkified as

... (date of birth update)

Thanks in advance! I18n (talk) 16:12, 19 November 2013 (UTC)

  • Hi, good idea, but I do not find a way to display property label depend on current language, something like {{P|31}}. Such mechanism exists, but my attempts to use it fails. Displaying always English label is not good practice because the project is international. As for me, I prefer to display Russian labels :-). — Ivan A. Krestinin (talk) 18:35, 20 November 2013 (UTC)

Hi! For reports like this, is it possible to allow two types instead of only one? --Rschen7754 21:53, 22 November 2013 (UTC)

Upper stage

Hi Ivan, I'm trying to figure out if this proposal is necessary. What is the difference between booster (Q741745) and space tug (Q1958838)?--Micru (talk) 17:22, 23 November 2013 (UTC)

  • booster (Q741745) usually is the first rocket stage or something like this. It works in atmosphere during short time (~ 1 minute). space tug (Q1958838) is separate device that transfer payload to target orbit after separation from rocket. It works outside atmosphere during a number of hours. Usually one space tug (Q1958838) can be used with different rockets, for example Briz can be used with Proton, Angara, Rockot. Rockets can be used with different space tug (Q1958838) or without it. For example Sojuz can be used with Fregat, Ikar or without it (for low orbit launches). — Ivan A. Krestinin (talk) 18:16, 23 November 2013 (UTC)
  • Thanks for the explanation, now I understand better what it refers to. Please check my comment on the property proposal and let me know if that could be an acceptable solution.--Micru (talk) 11:51, 24 November 2013 (UTC)

Commonscat and namespaces

in Wikidata:Database_reports/Constraint_violations/P373 there are on the end of Unique value" violations two items, which one of them is in main namespace and second in category namespace. MAybe the only problem is, that arwiki have articles about years with prefix (maybe separate namespace?) e.g.

JAn Dudík (talk) 09:00, 28 November 2013 (UTC)

Gadget-Merge.js

Hello Ivan, I don't use the merge gadget so far. So I have two questions:

  1. Why Abraham ben Samuel ibn Hasdai ha-Levi (Q330775) is merged into Abraham ben Samuel ibn Hasdai ha-Levi (Q15229395)? The other way around (saving the old item) seems more obvious.
  2. Why Abraham ben Samuel ibn Hasdai ha-Levi (Q330775) is not deleted? Does the merge gadget not include RFD?

Cheers --05:24, 29 November 2013 (UTC)

@Kolja21: Both merging always to the lower number QID and submitting the merged item to RFD are options that you can select. Most who use the merge tool do so, and I would advise Ivan to do so as well. --Izno (talk) 18:09, 29 November 2013 (UTC)
Hi, options of merge tools was reseted to default unexpectedly and I miss this fact. Thanks for the message. I re-megred the items. — Ivan A. Krestinin (talk) 18:18, 29 November 2013 (UTC) P. S. Interesting question to merge tool authors: why default options of the tool are non-consensual?
Yeah, I think I need to go submit that as a request to change on the merge gadget talk page. I've requested it before... --Izno (talk) 18:20, 29 November 2013 (UTC)
quick information: Michael North (Q109744) and Michael North (Q6833098) are two different persons. VIAF and after that VIAFbot made error identifying person from en.wikipedia. I fixed that tonight. – at this occasion: Thank you for all the work you have done for constraint checking. Cheers -- Make (talk) 19:30, 29 November 2013 (UTC)
Thanks for my mistake correction. — Ivan A. Krestinin (talk) 19:36, 29 November 2013 (UTC)
Thanks for your fast reply. Good work! --Kolja21 (talk) 22:25, 29 November 2013 (UTC)

Constraint violations

Hi Ivan! I have two questions about constraint violation:

Thanks in advance. — Ayack (talk) 10:35, 9 December 2013 (UTC)

Hi, bot does not touch report if no changes (except date and items count). {{Constraint:Person}} has a number of problems, so it is not supported. Better use individual templates, basically {{Constraint:Type|class=Q215627|relation=instance}} (or with Q5). — Ivan A. Krestinin (talk) 16:30, 9 December 2013 (UTC)
OK. Thanks for the answers! — Ayack (talk) 17:05, 9 December 2013 (UTC)

Hi Ivan. Your bot omitted an update of Wikidata:Database reports/Constraint violations/P225 for three days now. Is there a special problem? --Succu (talk) 22:40, 9 December 2013 (UTC)

Hi, error in regular expression, parser says: "Invalid regexp (missing ))". — Ivan A. Krestinin (talk) 03:25, 10 December 2013 (UTC)
I restored the last working regex. Is it possible to start the report manually? --Succu (talk) 07:28, 10 December 2013 (UTC)
This is possible, but this is long and resource-intensive procedure. Lets wait for the next daily dump. In appears in 20:30 UTC usually. — Ivan A. Krestinin (talk) 17:28, 10 December 2013 (UTC)
OK, thanks. My fear is, that I have to clean up a lot of duplicate taxon names. :( BTW: Thanks a lot for these constraint violation reports. --Succu (talk) 17:50, 10 December 2013 (UTC)

Wikidata:Database_reports/Constraint_violations/P771

Dear Ivan, I have restructered the constraints for Swiss municipality code (P771) and added several exceptions. Moreover, I deleted the properties on district level for about 150 objects. The reports seems to ignore these stuff. Is there a problem with the constraints? How do you treat stuff that is commented out, i.e. <!-- ... -->? --Zuphilip (talk) 18:20, 13 December 2013 (UTC)

  • Hi, currently bot has a small bug: it find all templates on talk page, both usual and commented. I changed templates name as workaround. About reports updating: bot uses incremental dumps to generate reports. These dumps have some time lag, see [8] for details. — Ivan A. Krestinin (talk) 22:13, 13 December 2013 (UTC)
Thank you for the work around. The exception inside the constraints seems not to work, e.g. Wikidata:Database_reports/Constraint_violations/P771#.22Single_value.22_violations lists exactly the 11 exceptions. Is there also some problem with the comments or format? --Zuphilip (talk) 23:13, 13 December 2013 (UTC)
Fixed. — Ivan A. Krestinin (talk) 08:12, 14 December 2013 (UTC)
Thank you. It looks almost perfect now. Do you have any idea why Constance (Q3834) is still in the constraint violation list (two times) also it is mentioned as exception? --Zuphilip (talk) 23:32, 15 December 2013 (UTC)
Fixed in bot`s code. Comma in comment was recognized as list divider. — Ivan A. Krestinin (talk) 17:09, 16 December 2013 (UTC)
Perfect! Thank you again very much! --Zuphilip (talk) 22:07, 16 December 2013 (UTC)

Adminship

You have made a lot deletion requests. Why not request an adminship?--GZWDer (talk) 08:16, 29 December 2013 (UTC)

Copy of en label to es label

Solved. When the bot finish the work I will check all Wikidata entries for errors like that. According to the rules there should not be elements with the disambiguation in the name, so the bot does not cover it. --Kizar (talk) 00:31, 8 January 2014 (UTC)

Lost links: Removing talk page links

Hi Ivan. Thanks again for this great list of yours. It's very helpful to remain the sanity of Wikidata. But there's another tiny issue with it: Currently it contains the removal of hr:Razgovor:Elvis Grbac from Q1334309. While this is a rather strange case (the Croatian article hr:Elvis Grbac has been moved to its talk page, so now in NS 0 we only find a template telling us the page is protected, while the real article is in NS 1), it was completely legit that the talk page has been removed from the item, as we don't allow talk page sitelinks. So cases like this should not appear on that list in general. --YMS (talk) 14:04, 10 January 2014 (UTC)

Great, thanks. --YMS (talk) 21:00, 10 January 2014 (UTC)
There still are pages in user namespace, e.g. pt:Usuário:Yleite/Poluição genética (Q3141144). --YMS (talk) 10:14, 11 January 2014 (UTC)
✓ Done, namespace 2 is excluded too. — Ivan A. Krestinin (talk) 10:30, 11 January 2014 (UTC)

novalue

Can your bot support P105=novalue and P225=novalue?--GZWDer (talk) 05:05, 11 January 2014 (UTC)

Исключения

Провёл небольшой тест: добавил в исключения для ISO 639-3 code (P220) два элемента из трёх, содержащих значение «sgl», в итоге из отчёта пропали все три. Мне кажется, что лучше выводить значение как ошибочное, пока абсолютно все элементы, его содержащие, не добавлены в исключения. --putnik 02:20, 17 January 2014 (UTC)

  • Сейчас работает по принципу: элементы в списке исключений не попадают в обработку совсем. Соответственно единственный оставшийся элемент стал уникальным. Поменять поведение конечно можно, но кода править много придётся, а ситуация вроде бы не очень критичная. — Ivan A. Krestinin (talk) 08:48, 17 January 2014 (UTC)
    • Мне как раз кажется, что довольно критичная. Для свойства исправляются ошибки и прописываются исключения, в итоге получается небольшая страница, на которой удобно отслеживать новые ошибки. А тут как раз ошибки не появляются (и не исправляются), хотя должны бы. --putnik 22:13, 18 January 2014 (UTC)

Format constraint

Can you have a look at the format constraint in Property_talk:P946? It looks okay for me, but everything is listed as a violation... --Zuphilip (talk) 14:52, 17 January 2014 (UTC)

(talk page stalker) I fixed the constraint. It previously required the ISIN to consist of two letters and ten combinations of a letter and a digit instead of ten letters or digits. --YMS (talk) 16:11, 17 January 2014 (UTC)
Thanks YMS. To Zuphilip: There is great tool: http://regexpal.com/ It helps me create and debug expressions. — Ivan A. Krestinin (talk) 17:04, 17 January 2014 (UTC)
Thanks both of you! And thank you for the link. --Zuphilip (talk) 17:26, 17 January 2014 (UTC)

GND type

Hi Ivan, thanks for your comment. Unfortunately it seems impossible to correct this false decision, since so much ego and emotions are involved. Cheers --Kolja21 (talk) 16:36, 18 January 2014 (UTC)

I understand. This is not singe high conflicting RfC result. I hope we take some lessons from this situation. And I think we need to treat more respectful to the needs of each other. — Ivan A. Krestinin (talk) 17:03, 18 January 2014 (UTC)

BTW: What happened to Constraint violations/P227? The curly brackets (= names) disappeared with the last update:

16. January: 4370933-3: Central Labour Camp in Potulice (Q190415), Potulice concentration camp (Q7235433)
17. January: 4370933-3: Q190415, Q7235433

Can you change it back with the next update? These lists are really helpful. Thanks --Kolja21 (talk) 17:13, 18 January 2014 (UTC)

Your bot removed images ...

Why ? GerardM (talk) 10:47, 20 January 2014 (UTC)

Косметические правки

Таких много; м.б. лучше попросить ботовода сразу убирать подчёркивания, чем потом гоняться за ним на другом боте? NBS (talk) 13:09, 22 January 2014 (UTC)

  • Беда в одном: таких ботоводов много и каждый ходит по одним и тем же граблям: тянет всякие %20%39_%48 в именах, тянет имена локальных файлов, удалённых файлов, редиректов, placeholder-ов и т. д. Я уже довольно давно повесил задачу в очередь и с тех пор она каждый день подчищает мусор за другими ботами, я уже и не слежу за какими именно и в каких объёмах. Ещё одна неприятность в том, что ботоводы здесь, по моим ощущениям, менее отзывчивы, нежели в рувики, хотя возможно виной этому языковой барьер. Потому, чем объяснять каждому ботоводу, как проверять существование файлов и как нормализовывать имена, мне проще держать задачу включенной. — Ivan A. Krestinin (talk) 18:19, 22 January 2014 (UTC)

Exception list for huge lists

e.g. in Wikidata:Database reports/Constraint violations/P373 there are thousands of constraints, but big part of them should be counted as exception. It would be very useful to take part of this list as list of exceptions and make e.g Property talk:P373/Exceptions#Unique_value instead of typing them to Constraint template. There should be only link or parameter

{{Constraint:Unique value|separate category processing=yes|exceptions=[[./Exceptions]]}}

JAn Dudík (talk) 08:54, 27 January 2014 (UTC)

[9] beginning:
JAn Dudík (talk) 20:30, 27 January 2014 (UTC)
Thanks for the sample. I reviewed some items, looks like links in Q7019790, Q8963584 are invalid. Cycling categories are more interesting case. Links are invalid in both items in pairs. Another side: currently direct linking with Commons is available. I think we need think about migration from P373 to sitelinks. — Ivan A. Krestinin (talk) 19:19, 28 January 2014 (UTC)

Target items

Hello Ivan, can you help me with a constraint question? I would like to add the following claim: Every instance of (P31) = book (Q571) should have an author (P50). Is this possible? --Mburbo (talk) 04:30, 29 January 2014 (UTC)

Hi Mburbo, currently no. But you can create constraint: every item with ISBN-13 (P212) should have author (P50) and be subclassed from work (Q386724). — Ivan A. Krestinin (talk) 17:00, 29 January 2014 (UTC)
Thanx. I'll give it a try: new constraints P212 & P747. --Mburbo (talk) 02:22, 30 January 2014 (UTC)

КОАТУУ

Эта правка ошибочная. Код КОАТУУ состоит из 10 цифр. Просто в том источнике, откуда его брали, данные в таблице Excel, а он начальные нули обрезает, поэтому получилось, что для трех областей (Крым, Винницкая и Волынская) коды неправильные.--Ahonc (talk) 00:15, 4 February 2014 (UTC)

Добрый день, я ориентировался на описание свойства на Property talk:P1077, где сказано "9 or 10 digits". Давайте так: вы на странице Property talk:P1077 поправите описание и отпишетесь почему правильным является именно 10-значный формат, если в течении пары дней возражений не будет, то я ботом расставлю недостающие нули. — Ivan A. Krestinin (talk) 04:03, 4 February 2014 (UTC)
Вон даже в статье Классификатор объектов административно-территориального устройства Украины написано, что состоит из 10 разрядов.--Ahonc (talk) 11:23, 4 February 2014 (UTC)
Возражений на странице обсуждения свойства не возникло, запустил бота, через пол-часа должен закончить. Не знаете кстати какого-нибудь сайта для скрипта MediaWiki:Gadget-AuthorityControl.js, где бы по КОАТУУ можно было посмотреть хоть что-нибудь по населённому пункту, хотя бы название? — Ivan A. Krestinin (talk) 18:14, 5 February 2014 (UTC)

Constraint for quantity datatype

does not work, see Property talk:P1082. By the way, we need a {{Constraint:Integer}} for this and so on.--GZWDer (talk) 15:32, 4 February 2014 (UTC)

{{Constraint:Range}} was created for dates. I try to extend it to quantity. Are you mean "without fraction" under {{Constraint:Integer}}? — Ivan A. Krestinin (talk) 18:54, 4 February 2014 (UTC)
{{Constraint:Range}} fixed. — Ivan A. Krestinin (talk) 18:08, 5 February 2014 (UTC)

Некорректные координаты

Наткнулся на давнее, но не исправленное. NBS (talk) 16:47, 4 February 2014 (UTC)

Координаты там корректные, на момент расстановки они и отображались нормально, потом был апдейт движка и они начали показываться красным текстом. Я баг по этому поводу написал, но он так и висит в статусе New. Ботами, движком Википедии и т. п. эти координаты получаются нормально, проблема только в отображении их страницами Викиданных. — Ivan A. Krestinin (talk) 18:47, 4 February 2014 (UTC)

Just created floors below ground (P1139). --Tobias1984 (talk) 18:14, 4 February 2014 (UTC)

Just created EHAK id (P1140). --Tobias1984 (talk) 18:19, 4 February 2014 (UTC)

is not up to date. Why?--GZWDer (talk) 18:10, 6 February 2014 (UTC)

Hi, bot does not recognize indirectly included templates, fixed. — Ivan A. Krestinin (talk) 19:29, 6 February 2014 (UTC)
It should learn to recognize :) Inclusion is very convenient when using the same stack of constraints on many properties. --Infovarius (talk) 18:35, 9 February 2014 (UTC)
I understand, but currently it does not supported by bot. There are technical problems with it. — Ivan A. Krestinin (talk) 19:01, 9 February 2014 (UTC)

Добрый день, Иван. Я еще раз изучил формат КОАТУУ и я согласен с User:Ahonc, что формат должен быть \d10. Если заглянуть в файл введения в КОАТУУ, то там есть пример в котором фигурирует ноль в первой позиции. Запустите, пожалуйста, Вашего бота еще раз с форматом \d10. Спасибо. --Voll (talk) 19:12, 6 February 2014 (UTC)

Добрый день, бот ещё вчера поменял всё на 10-тизначный формат. Отчёты Constraint violations обновляются с задержкой примерно на сутки, записи о 9-тизначном формате должны исчезнуть с него часа через три. — Ivan A. Krestinin (talk) 19:24, 6 February 2014 (UTC)
Ясно. Большое спасибо. --Voll (talk) 19:31, 6 February 2014 (UTC)

Почему отчёт не обновляется уже давно, хотя было добавлено третье ограничение? --Infovarius (talk) 18:34, 9 February 2014 (UTC)

Вроде всё чётко, три ограничения на СО, три секции в отчёте. Не совсем понял в чём проблема, может кэш браузера? — Ivan A. Krestinin (talk) 20:29, 9 February 2014 (UTC)
Да, не обновил после исправления 6 февраля. А что, {{Constraint:Value type}} не воспринимается? Infovarius (talk) 07:22, 10 February 2014 (UTC)
Как раз 6 февраля и обновил. {{Constraint:Value type}} предназначен только для свойств типа элемент. — Ivan A. Krestinin (talk) 07:38, 10 February 2014 (UTC)

Types statistics

Hello Ivan, is there a possibility to improve the types statistics? Example:

--Kolja21 (talk) 19:49, 9 February 2014 (UTC)

Hi, 1: could you create some sample of this idea? 2: I added type information to violation list, it appears during next 3 hours. — Ivan A. Krestinin (talk) 20:21, 9 February 2014 (UTC)
Hello Ivan, 1: A note like: "Types statistics: 'Creative work' (Q386724) or its subclasses required" would be enough. 2: Thanks. It works! (I've just corrected some errors with the help of the list.) --Kolja21 (talk) 00:45, 10 February 2014 (UTC)

Wikidata:Database reports/Constraint violations/P6

Removing part of this update from your KrBot does not seem to be useful to me, maybe there is some error in his code (or at least he has a bad day). --Jklamo (talk) 01:53, 10 February 2014 (UTC)

Bots having bad days… oi vey. :) --Izno (talk) 01:56, 10 February 2014 (UTC)
Item type was added to violation list. See "Types statistics" discussion above. — Ivan A. Krestinin (talk) 05:14, 10 February 2014 (UTC)

Just created World Athletics athlete ID (P1146). --Tobias1984 (talk) 14:48, 11 February 2014 (UTC)

P935/P373

Hello, I noticed, there are some items with Commons gallery (P935) = Category:Foo. Could you change them to Commons category (P373) = Foo and (if existing) Commons gallery (P935) = Foo ? JAn Dudík (talk) 13:56, 14 February 2014 (UTC)

Once more, please. --JAn Dudík (talk) 15:56, 23 March 2014 (UTC)
Bot checks this every day. One item caused problem, but you fix it already. — Ivan A. Krestinin (talk) 16:59, 23 March 2014 (UTC)

Hello, [10] says that "22374 items link to this item". Can you maybe add an English label and description? Holger1959 (talk) 19:04, 16 February 2014 (UTC)

  • Hi, my English is not enough for correct translation. Full the document name is "Сводная таблица соответствия кодов OKATO кодам OKTMO Российской Федерации на 31.12.2013", rough translation is "Summary table of correspondence OKATO codes to OKTMO codes of Russian Federation on 31 December 2013". — Ivan A. Krestinin (talk) 19:15, 16 February 2014 (UTC)
thank you! Holger1959 (talk) 00:11, 18 February 2014 (UTC)

Types statistics

Hi Ivan, can you modify the types statistics for Wikidata:Database reports/Constraint violations/P212? It would be helpful if all types, except the three major types: work (Q386724), book (Q571) and version, edition or translation (Q3331189), are listed behind the items. (I would like to make some corrections concerning work / genre. Also, for example, a film item with an ISBN most probably need to be corrected; often films and books are mixed in one item.) Thanx --Kolja21 (talk) 04:14, 19 February 2014 (UTC)

Химические свойства малины

Можно ли из элемента raspberry убрать химические свойства или бот внесёт повторно? Виновата японская статья, которую явно надо делить. WBR, BattlePeasant (talk) 12:29, 21 February 2014 (UTC)

Конкретно эту статью добавлю в список исключений бота. Но если смотреть глобально, то что делать с этой проблемой не знаю, это очень частая ситуация, когда из-за особенностей статьи в каком-нибудь разделе, боты из раза в раз добавляют некорректные свойства. Кое-где удаётся остановить этот процесс заменой значения свойства на novalue (например, Commons category (P373) на Luna 18 (Q668942)), но здесь слишком много свойств, где-то исправлением проблемной статьи, но здесь слишком далёкий язык... У меня у самого в списке наблюдения коллекция статей, с похожими ситуациями, которые я периодически откатываю после прохода чьего-нибудь бота... В общем не знаю... — Ivan A. Krestinin (talk) 18:39, 21 February 2014 (UTC)
В конце концов, надеюсь, карточки полностью перейдут на викиданные и бота можно будет отключить. А пока значит не стоит беспокоиться, пусть лишняя информация висит, пить - есть не просит. WBR, BattlePeasant (talk) 13:48, 22 February 2014 (UTC)

Hi! There seems to be something wrong with this update... Regards, SPQRobin (talk) 21:11, 23 February 2014 (UTC)

Обновления сейчас только по запросу происходят? :) Infovarius (talk)

bot error report

Your bot made a mistake: [11]. Please see what the problem was so it won't be repeated. Thanks, DGtal (talk) 08:34, 5 March 2014 (UTC)

Fregat versions

I've seen that your bot has been adding space tug (P1201), great! I have created Fregat-MT (Q15896716) which was the one used for Gaia (Q767805), do you think we should create items for the other versions too (M, SB)?--Micru (talk) 22:44, 5 March 2014 (UTC)

Hi, there is no this information in datasource that my bot uses unfortunately. Some another datasource is need to be found. — Ivan A. Krestinin (talk) 17:52, 6 March 2014 (UTC)
I found this page but it is not structured data.--Micru (talk) 21:09, 6 March 2014 (UTC)

Радуга

Hi, could you please check which ruwiki page should be linked with Raduga satellite constellation (Q15915658)? That is supposed to be the base item for the satellite constellation with each Raduga model linked to it. Thanks!--Micru (talk) 15:29, 9 March 2014 (UTC)

Maybe I should be using Q4173940 as base item instead?--Micru (talk) 15:33, 9 March 2014 (UTC)
Hi, Q4173940 describes whole system, include ground segment, non-Raduga satellites and etc. I do not find separate article about satellite constellation. — Ivan A. Krestinin (talk)
Ok, then I will leave it as it is now. --Micru (talk) 16:20, 9 March 2014 (UTC)

Hi Ivan! The report is in error: "ERROR: Invalid item key: PA00086780". I think that it's because I added this string (PA00086780) in the template as an exception: {{Constraint:Unique value|exceptions= PA00086780}}. What's the right way to do it please? Thanks. — Ayack (talk) 11:19, 15 March 2014 (UTC)

Ok, thanks, I've changed it. — Ayack (talk) 12:34, 15 March 2014 (UTC)

Krbot symbol images

Hi, Krbot has added symbol images as "official" images to items. Example, full list for that image. --Magnus Manske (talk) 22:09, 31 March 2014 (UTC)

Wow, another symbol picture (example), added by your bot over 850 times. I suggest you write a revert bot... --Magnus Manske (talk) 22:23, 31 March 2014 (UTC)
Thank you for the message. I revert vandalism and add File:Mitre (plain).svg to placeholders list as false negative. Bot will remove these images from WD items regular. — Ivan A. Krestinin (talk)

Dear Ivan,
Recently I've started to take care of this report. Would it be possible to add in section Value type human (Q5) violations a link to the related item, that is actually causing the violation? Especially in a list of mayors (like in Zurich Q72) the item, that has to get corrected, is hard to find. Thank you very much in advance. -- Pütz M. (talk) 08:48, 5 April 2014 (UTC)

From previously more than 8000 constraint violations we are now down to around 400. Tomorrow it will be even less.
To do a next step, I suggest for the report a constraint on sex/gender ... The constraints should look like {{Constraint:Value type|class=Q6581097|relation=instance}} and {{Constraint:Value type|class=Q6581072|relation=instance}}. But I could not find anywhere how to combine them in an OR-constraint. -- Pütz M. (talk) 09:21, 10 April 2014 (UTC)
You can add {{Constraint:Target required claim|property=P21}}. Allowed values of P21 are controlled by Wikidata:Database reports/Constraint violations/P21 report. — Ivan A. Krestinin (talk) 21:15, 10 April 2014 (UTC)
@Pütz M.: But there are ≈7000 items like these. Are you have an idea how to add P21 for its maybe? Some external database, or gender from personal name, or something else... — Ivan A. Krestinin (talk) 18:59, 11 April 2014 (UTC)
Most of the items are about mayors in Spain. There we could indeed try to detect the gender by the first name automatically. The number of different first names in Spain is limited. -- Pütz M. (talk) 19:08, 11 April 2014 (UTC)
I know too small about Spanish personal name traditions. Could you find list of man/women names? Is Spanish first names strongly gender dependent or there are notable exceptions count? Or your knowledge in this area is limited too? — Ivan A. Krestinin (talk) 19:49, 11 April 2014 (UTC)
I've started to work on the first items. These Spanish first names I have collected:
Male: Joan Josep José Jose Luis Lluis LLuís Alex Enric Martí Estanislau Carlos Carles Pablo Pedro Juan Jaume Ferran Narcís Pere Ángel Jesús Jordi David Bernardí Isidre
Female: Maria Mónica Pilar Modesta Ana Isabel
-- Pütz M. (talk) 10:57, 12 April 2014 (UTC)
Thanks. I try to create algorithm and I am confused by names like "Jose Maria Saiz Rodriguez", "D. Jose Maria Luengo Zapirain", "Maria Jose Fernandez Barranco", "Jose Maria Chaves Gomez-Orihuela". Could you help me with parsing these names? How to extract given name from Spanish full name? — Ivan A. Krestinin (talk) 17:35, 12 April 2014 (UTC)
Always take the first name "José Maria" ist male, "Maria José" is female. Maybe it is better, if we do first a list of all the names of the mayors, in which we collect suggestions for the sex/gender. I can have a look on this and will correct it, if necessary. After that your bot can edit the items. -- Pütz M. (talk) 18:36, 12 April 2014 (UTC)
And more ... "D." means "Don" or "Doña" or "Doctor", so you can leave it away. Spanish people usually take the name of the father and the name of the mother. The last two words are usually the full name, the part before the given name: "Jose Maria" (given name) "Luengo Zapirain" (family name), Maria Jose (given name) Fernandez Barranco (family name). -- Pütz M. (talk) 18:42, 12 April 2014 (UTC)
Thanks for your description. I modified the algorithm. Please review its results: User:Ivan A. Krestinin/Temp. I save bot`s log as is, but I can reformat it if needed. — Ivan A. Krestinin (talk) 18:57, 12 April 2014 (UTC)
Thank you very much for ths list. A very nice basis to work on. I've checked the upper part of the list and added the gender, where missing. Maybe your script can "learn" from this.
For today, I can not do more, as I have to leave now. But I think we came already far today. -- Pütz M. (talk) 19:38, 12 April 2014 (UTC)

Hi Ivan, this report is noting only 3 exceptions for 'Unique value', however there are many more listed at Property talk:P1058. I suspect that it was caused by '}}' being on the following line, which I have now fixed. Would it be possible for you to support this formatting so the exception list is more readable. John Vandenberg (talk) 05:21, 6 April 2014 (UTC)

Hallo Ivan A. Krestinin, I do not see why your bot made the revert? Regards--Oursana (talk) 07:52, 11 April 2014 (UTC)

Hi, no file by this name exists on Commons: COA_Ruffo_Calabria.PNG. Please check that file name is correct. — Ivan A. Krestinin (talk) 18:29, 11 April 2014 (UTC)
OK, thank you--Oursana (talk) 21:25, 11 April 2014 (UTC)

GND ID (P227) / KrBot fixing "Invalid GND-ID"

Hi Ivan, these cases can usually be fixed manually: The wrong Id usually stems from VIAF and strongly indicates that a GND entry exists, the GND record can be visited by a trick (remove "gnd/" from the URL), and the correct GND-ID can then be recorded. Since usually only one or two cases per day arise I would propose that the bot does not remove them. -- Gymel (talk) 08:10, 12 April 2014 (UTC)

Is there a problem with correcting the values after the bot removed them? That looks like a safer and faster solution to me. Regards, --Marsupium (talk) 04:17, 15 April 2014 (UTC)
Yep: When I don't see them in Wikidata:Database reports/Constraint violations/P227#Format I can't fix them. -- Gymel (talk) 08:37, 16 April 2014 (UTC)
I can disable this bot`s behavior. But this code removes not only badly formatted codes, but non-existing in GND codes too. The second codes are not displayed in Wikidata:Database reports/Constraint violations/P227#Format. So its will not be fixed. I can create separate report of cause, but this need some work. So is it really needed? — Ivan A. Krestinin (talk) 20:12, 15 April 2014 (UTC)
By what method do you determine "non-existing" codes? Checksum tests (those wrong Ids do have a valid checksum) or do you retrieve the actual content under http://d-nb.info/gnd/{ID} ? In that case you would have a couple of additional options:
  • Not found: Check for trailing spaces or LTR marks, repair and try again
  • Not found (and the ID does not contain a dash): Try http://d-nb.info/{ID} (without "/gnd") and
    1. remove or report as an error if still not found
    2. scrape the correct URL from the page and substitute it if it contains "/gnd" (might be duplicate to another P227 in the same entry though),
    3. remove if it does not contain "/gnd" (journals, items, other non-authority records),
  • Found: Check for differences between the URL requested and the URL scraped from the page (DNB does not issue HTTP redirect messages but silently displays the currently valid record in cases where records have been superseeded e.g. after merging two duplicate records). Substitute if different (might be duplicate to another P227 in the same entry though).
Of course this is much automation and other users might prefer more manual control (DNB sometimes has a glitch for some records, http://d-nb.info/gnd/14272095X/about/html is a current example: The record exists but has some internal issue preventing retrieval and we probably should not delete the id from Wikidata but report it to DNB: Compare with ("5" deliberately changed to "6" to produce a formally wrong id) http://d-nb.info/gnd/14272096X/about/html ). -- Gymel (talk) 08:37, 16 April 2014 (UTC)
Bot checks codes using dumps of GND database and on-line access to this DB. I implement part of logic that you describe. But I do not think that this is real problem. Bot removed ~50 invalid codes during this year only. The most errors are like this (internal VIAF identifier was added instead of GND ID). Primary bot`s goal are edits like this e. g. fixing "redirect" IDs. — Ivan A. Krestinin (talk) 19:51, 16 April 2014 (UTC)

Here date of death (P570) always was incorrectly 1440 stated by your bot: [12] and next, while ru:Эйк, Ян ван says 1441; [13] (contradictuous shows 1441 on view history) and next--Oursana (talk) 12:03, 18 April 2014 (UTC)

Hi, looks like bug in new version of Wikidata editor:
time="+00000001441-01-10T00:00:00Z" precision="9" calendarmodel="proleptic Julian calendar (Q1985786)" is displayed as 1441
time="+00000001441-01-01T00:00:00Z" precision="9" calendarmodel="proleptic Gregorian calendar (Q1985727)" is displayed as 1441
time="+00000001441-01-01T00:00:00Z" precision="9" calendarmodel="proleptic Julian calendar (Q1985786)" is displayed as 1440
But precision="9" means that month and day are insignificant numbers. For example diff tool display dates more correctly: [14]. There were no such bug then my bot imported these dates. — Ivan A. Krestinin (talk) 19:00, 18 April 2014 (UTC)
https://www.wikidata.org/w/index.php?title=Q266309&diff=80576414&oldid=80576409 same problem, must be more. Are you announcing the problem on Project chat or elsewhere? Regards--Oursana (talk) 22:28, 21 April 2014 (UTC)
No, currently not. It would be great if you report bug to https://bugzilla.wikimedia.org/ or to Wikidata:Contact the development team. — Ivan A. Krestinin (talk) 17:16, 22 April 2014 (UTC)
Please comment here Wikidata:Contact the development team#dates show wrong value - bug--Oursana (talk) 20:29, 22 April 2014 (UTC)

I reverted fatal error of your bot. Please take care. Regards --Oursana (talk) 10:23, 1 May 2014 (UTC)

template used by constraint violation reports

Hey Ivan,

I am sitting here with User:Jaredzimmerman (WMF) at the MediaWiki hackathon in Zurich. He is using the constraint violation reports your bot creates. Recently the bot no longer creates the list using the Q template. Therefor the lists are not very readable anymore. Can you use the template again so the label for each item in the list is shown instead of just the Q number? --Lydia Pintscher (WMDE) (talk) 20:15, 10 May 2014 (UTC)

removing P625

Hi. as here. We wrote a code to preserve the data in another way (see this) please undo these edits Amir (talk) 13:36, 13 May 2014 (UTC)

Hi Amir, it is bad idea use existing P625 claims as burial place coordinated. Bots copy to Wikidata any coordinates that was found in Wikipedia articles. Sometimes it was burial place coordinates, sometimes - not (for example see Nicholas van Hoogstraten (Q15262296)). "resting_place" and "resting_place_coordinates" parameters of en:Template:Infobox person are better datasource for your task. We have too many invalid claims in Wikidata already. — Ivan A. Krestinin (talk) 17:38, 13 May 2014 (UTC)
More samples: Paul Brook (Q7149533) (living people), Qaboos bin Said Al Said (Q57331) (living people), Pietro and Maria Botto House (Q7193052) (subject conflict), Tintin (Q52986) (fictional character), Powel Crosley, Jr. (Q7236048), Reginald Whinney (Q7308909). — Ivan A. Krestinin (talk) 20:11, 13 May 2014 (UTC)
Ok, I trust you Amir (talk) 11:52, 14 May 2014 (UTC)
coordinate location (P625) is still present ( and conflicting ) on hundreds of item https://www.wikidata.org/wiki/Wikidata:Database_reports/Constraint_violations/P21#One_of Ivan are you planning on removing these? Jared Zimmerman (talk) 07:22, 20 May 2014 (UTC)
Hi, I removed its for a number of times, but another bot were import its again :-). The situation is resolved now (as I think). ~10 violations of P625 will live in the report tomorrow or the day after tomorrow as I hope. — Ivan A. Krestinin (talk) 18:45, 20 May 2014 (UTC)
Ivan A. Krestinin, there are less than 100 Conflicts on https://www.wikidata.org/wiki/Wikidata:Database_reports/Constraint_violations/P21#One_of but its still shows Q items rather than the item names. you know what's up? Jared Zimmerman (talk) 06:59, 25 May 2014 (UTC)
Now the report is looked fine. I make more accurate items calculation. — Ivan A. Krestinin (talk) 09:38, 25 May 2014 (UTC)

Articles/Categories in P373 violations

Finally there are not only categories but articles too in Constraint_violations/P373. But there are some incorrect items: See Wikidata:Database_reports/Constraint_violations/P373#.22Format.22_violations few lines above there starts articles. But some of them are incorrectly there:

One of them is always category and one is article. JAn Dudík (talk) 06:59, 21 May 2014 (UTC)

Thanks, fixed, please check tomorrow report. — Ivan A. Krestinin (talk) 17:56, 22 May 2014 (UTC)

2nd NLA identifier

Hi Ivan, your bot is adding Libraries Australia ID (P409). (Good work!) Since a few days we have also the newer identifier NLA Trove people ID (P1315). Is it possible that your bot is adding both properties? (I also left a note here.) Cheers --Kolja21 (talk) 08:29, 22 May 2014 (UTC)

Hi, last time I used Template:Authority control (Q3907614) as datasource. But this template does not contain NLA Trove people ID (P1315) identifier. — Ivan A. Krestinin (talk) 18:06, 22 May 2014 (UTC)

Merging

Are you using your bot to merge items in these lists? If so, please stop, because there are a lot of false positive. For instance, here your bot replaced one item after it was wrongly merged. Regards, Andreasm háblame / just talk to me 23:10, 23 May 2014 (UTC)

Hi, bot does not merge items. Bot fixes lost links after merged item was deleted. For Insult (legal) (Q3086119) tree links was fixed. You revert these edits already. — Ivan A. Krestinin (talk) 04:21, 24 May 2014 (UTC)
In that case, could you please hold those changes at least two days after an item is deleted? Because that is usually the time needed to detect a wrong merge and revert it. It took me a lot of time to find those three links, but this would have been prevented with a longer wait between the deletion and the removal of lost links. Also, there is no rush. Andreasm háblame / just talk to me 04:45, 24 May 2014 (UTC)
This is possible, but as I know my bot is not the only tool that do this task. Usually links are fixed before deletion. Another side: issue is caused by invalid merging. There were many invalid merges with Widar label last time. Looks like this tool does not provide to user enough information to make correct decision. Discussion: Wikidata:Project chat#STOP „The Game“ (Merge). — Ivan A. Krestinin (talk) 05:49, 24 May 2014 (UTC)

Use of unknown value should not trigger constraint violation

on https://www.wikidata.org/wiki/Wikidata:Database_reports/Constraint_violations/P21#One_of the use of unknown value on item https://www.wikidata.org/wiki/Q483728 curiously displayed as "somevalue" rather than "unknown value" triggers a constraint violation, which seems incorrect. Jared Zimmerman (talk) 23:58, 23 May 2014 (UTC)

[16] this should help. — Ivan A. Krestinin (talk) 04:29, 24 May 2014 (UTC)

Things to merge

Hi. Please take a look at w: numerical deaths-by-year categories starting at w:Category:34th-century BC deaths and onwards (containter category at w:Category:Deaths by year), plus existing deaths-in-decade subcategories, starting at w:Category:2680s BC deaths. There are a lot of s: pages with Wikidata entries that could be merged with those, starting with s:Category:21st century BCE deaths, plus subcategories (container category: s:Category:Deaths by century). Could you please (a) either add the relevant Wikidata entries to your list of things to merge, or else (b) have a Bot just go ahead and merge them? Thanks! It Is Me Here t / c 12:18, 24 May 2014 (UTC)

Edit: it would probably also be worth checking the Commons container category c:Category:Deaths by year and subcategories in the same way. It Is Me Here t / c 12:22, 24 May 2014 (UTC)
Commons checking is simple task, I run bot, it completes for a hour. Processing numbers with "th", "st", "s" postfixes is more complex, I hope to complete this today. Automatic merge is complex task, I do not plan to do this. — Ivan A. Krestinin (talk) 12:37, 24 May 2014 (UTC)
Bot completes Commons processing. But it find too many false positives. I revert this. Merge pairs for "XXX0s":

Ivan A. Krestinin (talk) 17:07, 24 May 2014 (UTC)

Hi. Thank you for doing this. However, Q7235221 and Q6046038 are not the same thing; "700s" is supposed to refer to the decade, not the century. It Is Me Here t / c 20:54, 24 May 2014 (UTC)
fiwiki and svwiki have instable naming convension for these articles. I exclude these wiki from processing and update the list above. — Ivan A. Krestinin (talk) 08:37, 25 May 2014 (UTC)

Database reports

I realized that some of the database reports uses links to the elements or properties instead of {{Q}} and {{P}}. Can you update them to use those templates? At least for me it is more confortable to check the reports when the templates are used. Thanks. -- Agabi10 (talk) 09:49, 28 May 2014 (UTC)

Hi, {{Q}} can not be used more than 500 times per page unfortunately. See Template:Label for details. So bot uses {{Q}} for small reports and direct links for large reports. Development team can help to remove or increase this limit maybe. — Ivan A. Krestinin (talk) 18:39, 28 May 2014 (UTC)
Oh, really. I completely forgot the expensive parser function limit. And when a report is going to reach the limit wouldn't be possible to use the {{Q}} for at least the first 450 elements and direct links for the other links? -- Agabi10 (talk) 19:08, 28 May 2014 (UTC)
This is possible, but large reports have large load time now. This change improves its only a little, but load time will increased severely. Sometimes to avoid this problem I use custom button in wiki-editor. It replaces all [[Qxxx]] to {{Q|xxxx}} in selected text. See QButtonClick function in User:Ivan A. Krestinin/monobook.js. Or text replace function can be used from some external text editor for the same purpose. — Ivan A. Krestinin (talk) 19:26, 28 May 2014 (UTC)

More s: pages to merge

Hi again. This time I've found some U.S. Supreme Court decision pages that could do with added to Wikidata. Please cross-check members of w:Category:United States Supreme Court cases, members of s:Category:United States Supreme Court decisions and sub-categories, and members of sub-categories of s:Category:Volumes of the United States Reports. The Wikisource pages typically don't have Wikidata entries; they just need to be added to the Wikipedia pages' Wikidata entries. I've done a few manually (in my Contribs) as an example. Thanks a lot! It Is Me Here t / c 14:10, 28 May 2014 (UTC)

I add enwikisource to reports. Bot completes the report generation for a hour. — Ivan A. Krestinin (talk) 19:00, 28 May 2014 (UTC)
✓ Done, see User:Ivan A. Krestinin/To merge/enwikisource. Not many results. Bot uses existing relations to find missing links. Number of existing relations is low maybe. — Ivan A. Krestinin (talk) 19:53, 28 May 2014 (UTC)

Interesting constraint

Hi Ivan, I'm working on the Rijksmonumenten. Each Rijksmonument uses Rijksmonument ID (P359) and for this I get a nice constraint report. I want to expand this report, but not sure how and if this is possible:

  1. Every item should be
    1. instance of (P31) -> Rijksmonument (Q916333))) and
    2. instance of (P31) -> house (Q3947)/farmhouse (Q489357)/church building (Q16970) (about 50)
  2. Every item should include located in the administrative territorial entity (P131) twice
    1. located in the administrative territorial entity (P131) -> target instance of (P31) -> province of the Netherlands (Q134390)
    2. located in the administrative territorial entity (P131) -> target instance of (P31) -> municipality of the Netherlands (Q2039348)

Hope you can help. Multichill (talk) 21:08, 30 May 2014 (UTC)

  • Hi, I am apologize for delayed response, I missed notification about your message.
      1. {{Constraint:Type|class=Q916333|relation=instance}} - already present
      2. {{Constraint:Type|class=Q811979|relation=instance}} - not exact your requested, but similar
    1. {{Constraint:Item|property=P131}} is available currently, but it checks located in the administrative territorial entity (P131) is exist only.
Ivan A. Krestinin (talk) 07:31, 2 June 2014 (UTC)

Bot removes logo without providing rationale

Hi, it will be nice to know why the bot removes the logo from Q20728 without providing rationale for this. The irritating thing is that it does not remove the logo from other similar pages, so I have no clue what is going on. So I believe there are two issues. The first specific to the very page I mentioned, the second that now edit summary is provided. Cheers --[[kgh]] (talk) 07:17, 5 June 2014 (UTC)

Hi, the image was in commons:Category:Image placeholders. This edit fix the issue. — Ivan A. Krestinin (talk) 19:14, 5 June 2014 (UTC)
Yeah, great, thanks for rectifying this! However it will still be very nice if your bot could provide an edit summary so one gets at least some kind of clue about why the edit was done. --[[kgh]] (talk) 20:57, 5 June 2014 (UTC)
Large question of edit summaries is language. Wikidata is multilingual project. What language I must use in summary?... — Ivan A. Krestinin (talk) 21:09, 5 June 2014 (UTC)
This wiki instance was set up in English, so the bot will have to do its edit summary in English, too. Cheers --[[kgh]] (talk) 07:07, 6 June 2014 (UTC)
This wiki interface is multilingual. For example I see all UI elements in Russian. — Ivan A. Krestinin (talk) 17:13, 6 June 2014 (UTC)
Yeah, but what's the point? --[[kgh]] (talk) 17:15, 6 June 2014 (UTC)
I prefer default comments. Its are less informative, but multilingual. I think custom multilingual comments is good item for developer`s backlog. — Ivan A. Krestinin (talk) 17:21, 6 June 2014 (UTC)

Bogus image cleanup

FYI. --Magnus Manske (talk) 12:39, 12 June 2014 (UTC)

Good test! But only part of these properties are created by my bot. My bot present in history because it normalizes file names added by another bots. For example see [17]. I think that P18=*.svg can be cleaned up if P31=Q5. — Ivan A. Krestinin (talk) 14:01, 12 June 2014 (UTC)
✓ Done, I removed SVG from P31=Q5. Some another items are false positives (like coat of arms of the Solomon Islands (Q461197)) and was not created by my bot. Also see Property talk:P18#Notes for botmasters. — Ivan A. Krestinin (talk) 14:40, 14 June 2014 (UTC)

constraint violation report on P:P1273

Hello, I just added a constraint violation for CANTIC ID (former scheme) (P1273). Can you report it? Somehow I need it for some important fixations. Thank you Amir (talk) 15:50, 22 June 2014 (UTC)

Thank you Amir (talk) 04:39, 23 June 2014 (UTC)

Wikidata:Database reports/Constraint violations/P484

Hello
Wikidata:Database reports/Constraint violations/P484#"Single value" violations
It would be nice if 'IMA numbers' with an 'end date' didn't count. For example: chlorapatite (Q3680919)
Regards --Chris.urs-o (talk) 06:50, 24 June 2014 (UTC)
Hi @Chris.urs-o:, bot does not process qualifiers at all. But you can the items to exceptions: {{Constraint:Single value|exceptions={{Q|3680919}} }}. — Ivan A. Krestinin (talk) 18:56, 24 June 2014 (UTC)

Lost values Reports

Dear Ivan,
In dewiki we have now implemented first templates, that are directly linked and filled with information from Wikidata. Some users have expressed strong concerns, as they see Wikidatas basically unprotected against vandalism. These concerns we should take serious. They are a key factor for the acceptance of Wikidata in the Wikipedia community.
My idea now: What do you think of "Lost values Reports" similar to User:KrBot/Lost_links? Daily reports for properties, that will document all changes and removals of values.
If you think this is a good idea, could we start with something for P345 (IMDb identification)? This is the template in dewiki I'm currently working on. I could think of a report that lists all IMDb identifiers, that were replaced within the last 24 or 48 hrs from our items. -- Pütz M. (talk) 20:05, 24 June 2014 (UTC)

Just read, that User:Pasleim has already a similiar script almost working. So nothing to do for you. -- Pütz M. (talk) 22:30, 24 June 2014 (UTC)

Syntax

Hello Ivan
Is the syntax correct here: {{Constraint:Unique value|exceptions=IMA 1962 s.p., IMA 1966 s.p., IMA 1967 s.p., IMA 1968 s.p., IMA 1971 s.p., IMA 1973 s.p., IMA 1980 s.p., IMA 1982 s.p., IMA 1997 s.p., IMA 2008 s.p., IMA 2010 s.p., IMA 2012 s.p.}}
Regards --Chris.urs-o (talk) 13:09, 27 June 2014 (UTC)
Hi Chris, no, item identifiers must be used in exceptions list. {{Constraint:Unique value|exceptions= {{Q|111}}, {{Q|222}} }}, {{Constraint:Unique value|exceptions=[[Q111]], [[Q222]]}} or {{Constraint:Unique value|exceptions=Q111, Q222}}
Thx --Chris.urs-o (talk) 03:26, 28 June 2014 (UTC)

Question about a property proposal

Hi Ivan, I have doubts about this property proposal. Could you take a look and see if it is really needed? Thanks --Micru (talk) 09:21, 1 July 2014 (UTC)

Hi, our government starts general geographic names catalog. This proposal is about the catalog ID. Catalog contains verified object name, geocoords, administrative unit. This data is not copyrighted (as I understand). So it can be imported, used for verification and referencing. This proposal is similar to already existing OKATO ID (P721), OKTMO ID (P764) and State Water Register Code (Russia) (P884). One user says that he does not see any reason to display this property in ruwiki infoboxes. But Wikipedia is a bit different project. So I think the property is needed and there is no some negative points. — Ivan A. Krestinin (talk) 19:42, 1 July 2014 (UTC)
Thanks for your comment! I have created it as State Catalogue of Geographical Names (Russia) ID (P1397).--Micru (talk) 21:41, 1 July 2014 (UTC)

Und du meinst

die russischsprachige Wikipedia ist eine Quelle für einen österreichischen Autor? siehe hier. Falls du nicht deutsch verstehts, dann bitte bearbeite auch keine deutschsprachigen Themen, du richtest damit nur Schaden an. Wie viele deiner 5 Millionen Änderungen sind falsch? 50% Sollte man sicherheitshalber nicht gleich alle löschen?--Hubertl (talk) 18:51, 5 July 2014 (UTC)

Никакой из разделов Википедии не является хорошим источником для утверждений. Все разделы Википедии содержат ошибки и одинаково неавторитетны, и русский и австрийский и английский. Именно поэтому бот указал лишь imported from Wikimedia project (P143), а не stated in (P248). Если у вас есть авторитетный источник, то пожалуйста укажите его используя stated in (P248). Указание imported from Wikimedia project (P143) после этого потеряет смысл и его можно будет удалить. Свойство imported from Wikimedia project (P143) имеет скорее технический смысл. Оно нужно для выяснения источника ошибки. — Ivan A. Krestinin (talk) 19:41, 5 July 2014 (UTC)
Warum sollt ich deine Fehler ausbessern und nach Quellen suchen? Es sollte wohl umgekehrt gehen. Kein Eintrag ohne Quelle.Ich kann nichts dafür, dass in der russischsprachigen Wikipedia Blödsinn steht.--Hubertl (talk) 18:20, 6 July 2014 (UTC)
Hubertl, could you please tone down a bit? Thank you, Multichill (talk) 18:29, 6 July 2014 (UTC)
See also User_talk:Zuphilip#Wie_viele_deiner_80.000_Beitr.C3.A4ge. "Kein Eintrag ohne Quelle." Where did User:Hubertl give his Quelle for [18]? Tamawashi (talk) 03:50, 9 July 2014 (UTC

What a bullshit "Falls du nicht deutsch verstehts, dann bitte bearbeite auch keine deutschsprachigen Themen, du richtest damit nur Schaden an." Tamawashi (talk) 03:55, 9 July 2014 (UTC)

Tamawashi: Gänzlich falsche Eintragungen sind für dich ok? Wenn in der Russischsprachigen WP etwas falsches über einen deutschsprachigen Autor steht (Geburtsort), dann ist die russische Wikipedia als Quelle vorzuziehen? Das ist echt krank! --Hubertl (talk) 21:40, 14 July 2014 (UTC)
Multichill Wer bist du, dass du mir sagst, wie ich auf fahrlässigen Schwachsinn reagieren soll? Soll ich Ja und Amen dazu sagen, wenn von ganz offensichtlich völlig realitätsfremden Personen das möglicherweise wunderbarste Projekt seit Wikipedia von Anfang an zerstört wird? Unterstützt du das? Ich kann das jedenfalls nicht. Aber vielleicht äußerst Du dich dann, wenn die verschiedenen Wikipedia aus genau diesen Gründen es verhindern müssen (sic!), dass DAten aus Wikidata verwendet werden sollen. Kommentar, Multichill? --Hubertl (talk) 21:40, 14 July 2014 (UTC)
@Hubertl:: fahrlässigen Schwachsinn, realitätsfremden Personen, zerstört wird - ???? Solche Formulierungen sind wahrlich 1A projektdienlich. --Succu (talk) 21:48, 14 July 2014 (UTC)
@Succu: Nicht meine Einlassungen sind nicht projektdienlich, der Datenmüll, der hier bewegt wird, ist es!--Hubertl (talk) 22:18, 14 July 2014 (UTC)
@Hubertl: Mind to answer at Russian or English, respecting our host? --Succu (talk)

Removing deleted image claims

Hi! No idea whether you knew about my request, but thanks for removing image claims of deleted images with your bot! Trijnstel (talk) 17:55, 6 July 2014 (UTC)

Hello Ivan, it seems that the constraint report for this property is no longer been updated. Could you have a look at it please? Thanks. — Ayack (talk) 11:55, 9 July 2014 (UTC)

Exceptions to constraint violations

Hello Ivan, is there a place where I can add pages so that they are not listed as constraint violations ? For instance some people have two birth dates because two sources give different birth dates, and some people have two VIAF ID, because there is an error in the VIAF database. --Zolo (talk) 08:31, 19 July 2014 (UTC)

Bot error

This bot edit appears to be an error (the reference is to work not a person). I have reverted it. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 11:40, 22 July 2014 (UTC)

IMDB request

Hello, My bot was adding IMDb ID (P345) from cs.wiki, but I didnt noticed, that there is only number without prefix. Can you repair all IMDB values added by JAnDbot with adding nm before numbers? (maybe it will be tommorrow in Constraint violatons)? Thanks, JAn Dudík (talk) 09:04, 24 July 2014 (UTC)

I delete these values currently. You can add its again or I can do this some time latter. Additionally please see Q17403602, Q17403877, Q17403940. — Ivan A. Krestinin (talk) 19:09, 24 July 2014 (UTC) P. S. Do not forger about leading zeros.

Bot error: "invalid GND ID"

Hi Ivan, have you changed your bot configuration? Today, for the first time, KrBot removed valid GND IDs:

  1. Rudolf Reichling (Q3433775) [19]
  2. Ronald W. Jones (Q4160425) [20]

Cheers --Kolja21 (talk) 14:55, 26 July 2014 (UTC)

I think I found the error. The template in the article en:Ronald W. Jones:
{{Authority control |VIAF=108640452 |LCCN=n81052578 |GND=119128020‏ }}
produces the following link:
http://d-nb.info/gnd/119128020%E2%80%8F
The appendix "%E2%80%8F" makes the link invalid. --Kolja21 (talk) 15:09, 26 July 2014 (UTC)

"To merge"-project updates

Hi Ivan A. Krestinin!

Is there any pattern in the times of the updates of the "to merge"-pages? I've fixed quite some of them and an new update would be much appreciated. Of course the red links are obvious, but there are also items that have been switched within blue links etc.

P.S. Wouldn't it be an idea to put a Babel-box on your user page with the languages you speak or understand, or is there a specific reason you haven't done that yet? - FakirNL (talk) 23:23, 26 July 2014 (UTC)

Merge

Privet tovarish Ivan A. Krestinin! Can you merge these items by your bot: User:Calak/To merge/kuwiki? Spasibo.--Çalak talk 08:57, 28 July 2014 (UTC)

Hello, I have limited internet connection now. Please ask another users. — Ivan A. Krestinin (talk) 10:19, 28 July 2014 (UTC)

P373 update

Hello, https://www.wikidata.org/wiki/Wikidata:Database_reports/Constraint_violations/P373#.22Existing_category.22_violations was not solved (repaired and deleted in items) for many days. Have you any problems with it? JAn Dudík (talk) 09:12, 28 July 2014 (UTC)

cswiki ExternalUse

Hi Ivan. I have found that your bot had added many ExternalUse templates from ruwiki. I am now working on next templates to include some Wikidata data. Would it be possible to run the bot through cswiki same way? Could you please also run the bot periodically since I am going to work on more templates in the future? And just a little note: I very appreciate your work on Wikidata! The work you do here is great! Also I hope you will soon find the way how to do this suggestion. Matěj Suchánek (talk) 11:35, 28 July 2014 (UTC)

Давно не обновлялся. -- Vlsergey (talk) 10:58, 1 August 2014 (UTC)

Don't merge

Hi Ivan! I plan to move the page User:Pasleim/whitelist to Wikidata:Don't merge. In addition, I'd like to remove all sitelinks to reduce the size of the page. Is this okay for you? --Pasleim (talk) 11:27, 1 August 2014 (UTC)

Removing site links would make it harder to spot false false positives, though the list is surely getting very long and slow now. - FakirNL (talk) 12:28, 1 August 2014 (UTC)
Hello, good idea! Q-keys are enough for my bot, but another users can have another vision. — Ivan A. Krestinin (talk) 14:41, 1 August 2014 (UTC)
Maybe a compromise would be a listing in the format with the Q-template? So from:
*Q14852938 (nl:Locum) and Q2621243 (en:Locum)
We go to:
Locum (Q14852938) and locum (Q2621243)
because:
Q14852938 and Q2621243
would make it harder to find items. Just suggesting... - FakirNL (talk) 15:05, 1 August 2014 (UTC)

Ok, I moved the page to Wikidata:Do not merge. The format is still the same but we have to act soon. Using the Q-template is not possible because only 500 templates can be included in one page. --Pasleim (talk) 21:16, 2 August 2014 (UTC)

Maybe split the list? Would that work in some way or another? - FakirNL (talk) 21:58, 2 August 2014 (UTC)
Why would someone need to search the list for false false positives? They should have been manually checked by competent editors. Sure, there might be a few errors, but they would be incredibly time-consuming to find from the list as it is. The only way to look for them would be some shorter list generated by an algorithm that used more information than just the names, but writing that would probably require only the Q-keys. Rigadoun (talk) 18:30, 7 August 2014 (UTC)
Well, all of us make mistakes and if you have a clue what it is about (containing a certain word or letter combination) you can now search for it in the list. I have to admit, it's probably only a minor consideration and the concerns about the list becoming to long are founded better. Too bad it's not possible to have the list use the Q-keys, with 500 templates being the maximum and over 8000 items at the Do Not Merge it would require splitting the list in 33 items (currently, more to come). I don't have the definitive solution so keep the suggestions coming! - FakirNL (talk) 19:26, 7 August 2014 (UTC)
OK, how about having a bot-generated shortening that creates a list for use, but the users add items to pages that get archived somewhere so they would remain searchable? For instance, let's move the current contents of Wikidata:Do not merge to Wikidata:Do not merge/1, add new contents to Wikidata:Do not merge, but have the bot create a shortened version of archived lists at Wikidata:Do not merge/Short, moving items to archived pages (probably Wikidata:Do not merge/2, since the first is quite long)? Something like that? Rigadoun (talk) 01:31, 15 August 2014 (UTC)

Bot stopped?

Hi Ivan, your bot is no longer updating constraint violations. Could you start it again please? Thanks. — Ayack (talk) 12:30, 1 August 2014 (UTC)

OK, so I'll wait :). Thanks. — Ayack (talk) 14:56, 1 August 2014 (UTC)

To merge: possible error

Hi Ivan, thanks for the current new run of your bot regarding the To Merge-project. I do have a question though regarding possible false positives. A lot of items are flagged as conflicting with old English (User:Ivan A. Krestinin/To merge/angwiki#ruwiki) for 'Wikipǣdia:Gecorene gemynddagas/# Solmōnaþ' and 'Wikipedia:Selected anniversaries/February #', with w:ang:Solmōnaþ being February/Февраль and w:ang:Hrēþmōnaþ being March/Март. As you can see in the histories of Q5673838 and Q5832707 those items have been fixed but in your current run they are still flagged. Can you explain or, even better, fix that? - FakirNL (talk) 14:19, 5 August 2014 (UTC)

And others, such as all three at User:Ivan A. Krestinin/To merge/liwiki. - FakirNL (talk) 16:49, 5 August 2014 (UTC)
New dump will be available soon. These items must disappear from lists after the dump will be ready. — Ivan A. Krestinin (talk) 20:32, 5 August 2014 (UTC)
Maybe time for an update? It's been 20 days since the last run. Спасибо! - FakirNL (talk) 11:05, 6 October 2014 (UTC)
Started, but 2014-09-12 dump is the latest. — Ivan A. Krestinin (talk) 14:30, 6 October 2014 (UTC)
Well, a new dump would have been very useful. Any idea when a new dump will be provided or where I can ask for one? - FakirNL (talk) 18:36, 6 October 2014 (UTC)
Some progress page. This list can be used to calculate typical time between dumps. This newsgroup can be useful as I hope. — Ivan A. Krestinin (talk) 18:47, 6 October 2014 (UTC)
The intervals between dumps are getting increasingly irregular, if I'm correct the list goes 16, 17, 16, 20, 18, 18, 17, 17, 23, 26, 4 (?), 19, 20, but it's been 24 days now and counting. - FakirNL (talk) 19:14, 6 October 2014 (UTC)
I have no information about this delay reasons. I can use incremental dumps theoretically. But this is not simple task... — Ivan A. Krestinin (talk) 19:21, 6 October 2014 (UTC)
People did point me to another set of dumps which could possibly be used. They are found at dumps.wikimedia.org/other/wikidata/. - FakirNL (talk) 12:07, 8 October 2014 (UTC)
It is interesting. I need a time to investigate these dumps. Thank you. — Ivan A. Krestinin (talk) 17:01, 8 October 2014 (UTC)
✓ Done, pages update is in progress. — Ivan A. Krestinin (talk) 18:16, 11 October 2014 (UTC)

Constraint violations P570

Hi, sometime i look at Wikidata:Database reports/Constraint violations/P570. There is a Diff within range section based on Template:Constraint:Diff within range. But there is a problem with unknown dates (of birth or death), that are usually indicated not as unknown value but as (at least) century. But in these case script is not able to handle these correctly and result are false alarms. Is there any workaround (maybe exclude the items with century values)? --Jklamo (talk) 09:24, 7 August 2014 (UTC)

Hello, this is similar to bug in Wikidata value editor/visualizer. See [21], bot add value "+00000001300-01-01T00:00:00Z", e. g. 1300s years. This is 14 century, but visualizer treads it as 13 century. My bot calculates 1279-1300=-21. Another related discussion: User talk:Magnus Manske#Adding date of birth "20 century". — Ivan A. Krestinin (talk) 12:45, 7 August 2014 (UTC)

Bulk deletion requests

When you make a bulk deletion request, could you please use the button at Wikidata:Requests_for_deletions? That way, I as an admin will have direct deletion links on the requests for deletion page, which reduces my work by one click per deletion. Also, it will cause the deletion comment to be a reference to your request rather than the non-informative 'content before blanking was: ""'. Thanks in advance, Andre Engels (talk) 18:00, 9 August 2014 (UTC)

Hello, the button does not allow to add multiple items. But I can use some another list format. For example:
  1. Q16742992 (delete | history | links | logs): merged with Q8162227
  2. Q16742993 (delete | history | links | logs): merged with Q8161574
Is this format ok for you? — Ivan A. Krestinin (talk) 20:28, 9 August 2014 (UTC)

Не ясно, чем ему "12170" не нравится, ведь попадает и под "\d+", и под "\d{1-5}". -- Vlsergey (talk) 18:07, 14 August 2014 (UTC)

List of externally used properties

Hi! Is it possible for your bot to output a list of externally used properties to e.g. Wikidata:Externally used Properties? Would be nice to be able to text-search them. Best format would probably be {{P|xxx}} so that people see the list in their language. Thank you for considering. Tobias1984 (talk) 19:13, 26 August 2014 (UTC)

  • Hello, current bot`s algorithm is resource expensive and miss many cases. So it can not be used for creating and updating full list. And I think that property talk is better place for usage information than long-long single page. — Ivan A. Krestinin (talk) 19:32, 26 August 2014 (UTC)

Привет. Не знаю, как на других страницах, а эту я всегда визуально обрабатывал по совпадению/несовпадению имён и названий. Теперь без шаблона {{Q}} вообще непонятно, что делать. Нельзя ли его вернуть? --putnik 06:40, 28 August 2014 (UTC)

  • Приветствую, вернуть то можно, только он вместо названий будет показывать "Script error". Этот шаблон нельзя использовать больше 500 раз на странице. В качестве не слишком удобного варианта можно брать кусочек отчёта, в каком-нибудь текстовом редакторе делать замены "[[Q" -> "{{Q|" и смотреть результат в окошке предпросмотра. А когда отчёт станет поменьше, то бот снова начнёт использовать {{Q}}. — Ivan A. Krestinin (talk) 17:07, 28 August 2014 (UTC)
    • А можно тогда сделать Wikidata:Database reports/Constraint violations/P345/ru, в котором будут русские названия (и так со всеми)? -- Vlsergey (talk) 19:20, 28 August 2014 (UTC)
      • Беда в том, что бот уже сейчас работает около 3 часов и съедает порядка 10 Гб оперативки, а генерация нескольких страниц и подтягивание информации о метках ещё раздуют эти параметры. Лучше уж тогда бить отчёты на несколько страниц, примерно по 500 элементов на каждой. А ещё лучше было бы уговорить девелоперов поднять цифру 500 до скажем 5000. В конце концов эти отчёты - не единственное место, где в эту цифру упираются. — Ivan A. Krestinin (talk) 19:35, 28 August 2014 (UTC)
    • Там даже сейчас в запущенном состоянии где-то 850 включений, а если исправлять (что сейчас делать сложно), то будет держаться меньше 500. Можно попросить вернуть конкретно для P345? --putnik 14:59, 29 August 2014 (UTC)

Неверный импорт дат юлианского календаря

Доброго дня. В процессе повторного пеереноса дат из рувики в викиданные обнаружил, что Ваш бот некорректно перенёс даты по Юлинскому календарю. Например, для статьи https://ru.wikipedia.org/wiki/%D0%90%D0%B1%D1%83_%D0%9C%D0%B0%D1%88%D0%B0%D1%80 он скопировал дату как "10 августа 787" по григоринаскому календарю, установив флаг отображения по юлианскому. В результате корректным отображением данной даты является 3 августа (а не 10-е, как в григорианском). Напоминаю, что все даты сохраняются в пролептическом григорианском календаре. Пожалуйста, укажите, к какой дате Вы сможете поправить данное недоразумение, либо стоит ли просто удалить все даты, добавленные Вашим ботом из ру-вики и перенести их ещё раз. Ещё одним вариантом будет установить флаг отображения по григорианскому календарю для всех дат, импортиролванных Вашим ботом. Это хотя и некорректно с точки зрения дат, но достаточно распространённый "хак" для дат в Викиданных. -- Vlsergey (talk) 12:06, 31 August 2014 (UTC)

  • Приветствую, не понял где ошибка, смотрю на Abu Ma'shar al-Balkhi (Q11373), написано "дата рождения 10 августа 787", если открыть редактор, то там сказано, что это юлианский календарь. Если посмотреть в истории, то там тоже сказано "Календарь http://www.wikidata.org/entity/Q1985786". — Ivan A. Krestinin (talk) 16:52, 31 August 2014 (UTC)
    • дата 10 августа, которая отображается в редакторе Викиданных, это дата по пролептическому григорианскому календарю. А то, что там отображается "юлианский календарь", это означает, что дату нужно показать в юлианском, предварительно сместив на оппеделённое число дней. См. Также ссылку на вывод API -- http://www.wikidata.org/w/api.php?action=wbgetclaims&entity=Q11373&property=P569 -- видно, что 08-10 это именно сохранённое значение, т.е. григорианское. -- Vlsergey (talk) 17:51, 31 August 2014 (UTC)
      • Похоже мы видим одно и тоже, но интерпретируем увиденное по-разному. Почему вы думаете, что то, что сейчас показывает редактор Викиданных - это дата по григорианскому календарю? Рядом с этой датой не написано, что это григорианский календарь, в тоже время в 787 году никакого другого, кроме юлианского не было. С API аналогично, там написано 08-10 и calendarmodel="...Q1985786", т. е. 10 августа по юлианскому календарю. — Ivan A. Krestinin (talk) 18:10, 31 August 2014 (UTC)
        • Поэтому и указывается, что используется proleptic Gregorian calendar (Q1985727), который определён до бесконечности в оде стороны (ср. с Gregorian calendar (Q12138)). Что касается "почему я так думаю", я специально задавл вопрос на форуме и вот ответ от Лидии. Также это отражено в https://www.mediawiki.org/wiki/Wikibase/DataModel#Dates_and_times : "calendarmodel: URI identifying the calendar model that should be used to display this time value. Note that time is always saved in proleptic Gregorian, this URI states how the value should be displayed". Вы также можете осуществить довольно простую проверку: добавить в песочнице в качестве даты рождения 29 февраля 100-го года. Такой даты нет в григорианском, но есть в юлианском. -- Vlsergey (talk) 20:29, 31 August 2014 (UTC)
          • Поигрался, эксперимент с 29 февраля мало что значит, там и 31 февраля указать можно, несмотря на то, что его ни в том ни в другом календаре нет. Вообще если девелоперы действительно рассчитывают на такую модель, то у них большой баг в редакторе дат потому что любой неподготовленный к подвоху человек текст "4 марта 1850 Юлианский" воспримет как 4 марта 1850 по юлианскому календарю, а отнюдь не 4 марта 1850 по григорианскому календарю, которое следует отображать по юлианскому. Ожидать от пользователей, что они перед вводом даты в редактор в голове пересчитают из юлианской какого-нибудь 100-го года в григорианскую - тоже крайне странно (я даже примерно не представляю какое там смещение). В общем до приведения работы редактора в соответствие с моделью описанной в приведённом вами документе я массовые правки проводить не буду, т. к. люди всё равно будут вводить даты в соответствии с поведением редактора. В целом я склоняюсь к мысли, что девелоперы реализовали в редакторе одну логику, а в документации осталась другая. Лидия же свой ответ взяла из документации, а не из реального состояния кода. — Ivan A. Krestinin (talk) 21:35, 31 August 2014 (UTC)
            • Вы, конечно, можете ожидать бесконечно, но для интеграции Викиданных с рувики необходима корректная работа как скриптов импорта, так и модулей на Lua. Окей, тогда я самсделаю бота, который удалит заимпортированные из рувики даты рождения и смерти. То, что редактор недостаточно ясно показывает, что именно происходит, не означает, что он работает неправильно. Он меняет метку, но не меняет сохранённую дату. То, что он при этом ещё и не показывает дату в другом календаре -- это, конечно, плохо, но тоже достаточно понятно. -- Vlsergey (talk) 22:56, 31 August 2014 (UTC)
              • Я смотрю другие боты импортировали даты способом, который не похож ни на моего бота, ни на предлагаемый вами: [22], [23]. Главное пока вопрос окончательно не прояснится, не удаляйте юлианские даты из статей рувики, а то снова будет волна недовольства "опять понатащили какого-то мусора из Викиданных". — Ivan A. Krestinin (talk) 06:19, 1 September 2014 (UTC)
                • Это я и называю "стандартным хаком" -- даты, для которых нужно показывать только юлианский (без доп. новым стилем), помещались в Викиданные как даты по как-будто по григорианскому календарю. Модуль их так воспринимает и отображает без изменений, никак не изменяя. Хак, но пока работающий. Хотя на форуме об этом ругались. -- Vlsergey (talk) 12:18, 1 September 2014 (UTC)
                  • Даааа, похоже нам пора писать подробный документ о том как оно должно быть в ISO-формате, в отображении редактором, в карточке ВП, писать с конкретными примерами. А потом согласовывать этот документ с девелоперами, ботоводами и после этого добиваться того, чтобы движок ВД, боты и инструменты работали в соответствии с этим документом. С текущим бардаком мы далеко не уедем. — Ivan A. Krestinin (talk) 12:28, 1 September 2014 (UTC)

Wikidata:Database reports/Constraint violations/P625

Hi, can ask for update of that page (last updated 22 August)? Maybe there is no need to update it every day, but at least weekly will be nice. Thanks. --Jklamo (talk) 23:50, 9 September 2014 (UTC)--Jklamo (talk) 23:50, 9 September 2014 (UTC)

  • Hello, thank you for the notification. The report update procedure fails with error "The content you supplied exceeds the article size limit of 2000 kilobytes". I update the report manually for now. And I will think about ways to reduce the report size. — Ivan A. Krestinin (talk) 03:36, 10 September 2014 (UTC)
Thanks for feedback. Now i am working mainly on "conflict with" violations (and it is "nice" to have some bot creating more violations). If size is a problem, maybe we can consider to drop single value violations, as now many of them are false positives ("linear" items with multiple coors like rivers or roads). For that case ideal will be to filter out from "single value violations" those with coors and qualifier, but i am afraid that will require even more computing power. --Jklamo (talk) 09:18, 10 September 2014 (UTC)
Bad news about "Conflicts with": Wikidata:Administrators' noticeboard#User:AyackBot and coordinates of movable objects. — Ivan A. Krestinin (talk) 18:48, 10 September 2014 (UTC)
Thanks, that your bot cleaned-up some of violations! --Jklamo (talk) 08:11, 14 September 2014 (UTC)

Constraint for Property:P735 (and P734)

For this property, it would be helpful if there was a way to check if the labels of the target item are identical in languages with Roman script. A sample item that would pass the test is Q16281827. See Wikidata:WikiProject_Names for more information. --- Jura 06:17, 10 September 2014 (UTC)

An item I created (Q18012517) with the label "Prisca" appears as having two different labels. I can't figure out how they differ. Maybe some unprintable character when pasting? "Jean-Charles" and "Patrick" and "Szeréna" might have the same issue. --- Jura 20:01, 16 September 2014 (UTC)

Hi Ivan, could you please omit Wikidata Sandbox (Q4115189) in your constraint reports. Thanks. --Succu (talk) 07:27, 11 September 2014 (UTC)

Great! --Succu (talk) 20:37, 11 September 2014 (UTC)

Hello,

on Wikidata:Database reports/Constraint violations/P106 is an Error i think.

For example at the first Q354: Q937857 it is an instance of q2066131 and that is an q13516667, so it shouldn't be on the List.

And thats on the most of the 250.000 entries.

I hope you understand what i mean...and sorry for the bad english.

Greets --McSearch (talk) 07:08, 13 September 2014 (UTC)

Обновляется не полностью. Не хватает мощностей? --Infovarius (talk) 14:31, 15 September 2014 (UTC)

  • Приветствую, в смысле не полностью? — Ivan A. Krestinin (talk) 17:14, 15 September 2014 (UTC)
    • Ну там вообще-то 5 ограничений. А, они ж в шаблоне. Может всё-таки в алгоритме добавить распознавание сборных ограничений? --Infovarius (talk) 18:12, 26 September 2014 (UTC)
      • А, вы про это, старое ограничение бота. История началась с шаблона {{Constraint:Person}}, помимо технических сложностей, с ним связана ещё одна неприятность: он требует от миллионов элементов наличия в них свойств вроде "дата/место рождения" при этом понятно, что выполнить это ограничение или совсем нереально или по крайней мере оно будет выполнено весьма не скоро. В результате, например, отчёт Wikidata:Database reports/Constraint violations/P21 из отчёта с которым можно довольно продуктивно работать превратится в огромный список, который очень вряд-ли скоро исправят. Как дело обстоит с тасконами — не знаю. Давайте поэкспериментируем, воспользовавшись одним небольшим багом бота (он "видит" даже шаблоны внутри комментариев). — Ivan A. Krestinin (talk) 18:39, 26 September 2014 (UTC) P. S. Вообще нужно наверное разделить ограничения, которые "нужно" выполнять, т. е. собственно ограничения и "пожелания", которые выполнять можно по мере сил.

Hello Ivan,

At Q16281827, the property P460 lists a series of items with related given names. Normally, each of these given names should have all other given names listed with P460.

The constraint currently only allows checking if Q16281827 is listed on the other items, not if all other items have the ones listed on Q16281827. Would it be possible create a new constraint that does that? --- Jura 05:46, 19 September 2014 (UTC)

BTW, when adding missing P460 based on the constraint report, I checked if it had one of the P31 types for given names and no disambiguation item. Maybe the constraint could work as "transitive" for items with specific P31. --- Jura 19:35, 19 September 2014 (UTC)
I can create this report. But now it is too early time for this as I think. There are 2871 symmetric violations. Transitive relation is harder that symmetric. So transitive report will be too large too. — Ivan A. Krestinin (talk) 07:06, 21 September 2014 (UTC)
To avoid too many results, I think the transitive relation should be fairly restrictive. It would only work with items that had the same single specified P31 statement. BTW I just fixed the 100 violations linked to given names. --- Jura 10:00, 21 September 2014 (UTC)
Thanks! The new report looks good. --- Jura 21:10, 22 September 2014 (UTC)

Didn't understood ths change

Hi,

Could you explain to me what is this change made by KrBot on Rennes Opera house (Q3354560) ? I maybe missing something obvious but I can't see any difference…

Cdlt, VIGNERON (talk) 16:01, 21 September 2014 (UTC)

Hello, bot replaced "_" symbol to space. — Ivan A. Krestinin (talk) 17:06, 21 September 2014 (UTC)

Exceptions for Constraint:Target required claim

Hello Ivan,

For the above constraint, somehow the exceptions added at PT:P1321 don't get taken in account. --- Jura 09:00, 27 September 2014 (UTC)

Hi, Jura, fixed. Source items must be listed for all constraint templates, not target. — Ivan A. Krestinin (talk) 19:56, 27 September 2014 (UTC)

Constraints and redirects

Hi, please follow redirects when checking constraints. Example: Q3893116. --JulesWinnfield-hu (talk) 13:48, 9 October 2014 (UTC)

Hi, Jules, I am apologize for delayed response. ✓ Done, please wait for 1-2 days for data update. — Ivan A. Krestinin (talk) 19:23, 10 October 2014 (UTC)

Could you please follow redirects when checking constraints? Q12725297 --JulesWinnfield-hu (talk) 10:04, 14 October 2014 (UTC)

This is not simple task. So I used another way: I started task that fixes links to redirects. — Ivan A. Krestinin (talk) 18:02, 14 October 2014 (UTC)

Richard James Cheeswright (Q18200401) returned no value VIAF statement

I have just readded the blank statement, as I need a means to record where there is a negative result to a search. Not sure of your concern about a "no value" pointer.  — billinghurst sDrewth 00:05, 13 October 2014 (UTC)

Hi, I removed it because no value is not listed as allowed value on the property documentation (Constraint:Format). — Ivan A. Krestinin (talk) 04:53, 13 October 2014 (UTC)
@billinghurst: "no value" makes only sence in connection with the qualifier retrieved (P813). (R. J. Cheeswright for example has written a book. If he has no authority file yet, he will have one in future.) --Kolja21 (talk) 19:30, 13 October 2014 (UTC)
Thanks, I have taken the conversation to Property talk:P214 about how we should best check and track, as the future is a long time, and an absence causes an issue now. @Kolja21: Yes, I know that the person is an author, I have just completed one of his works at Wikisource, which was why I was adding the detail here.  — billinghurst sDrewth 23:33, 13 October 2014 (UTC)

Birth date probable error

this edit is likely to be wrong. In 1869 Russia observed the Julian calendar, so the birth date of persons born in Russia in that year would usually be stated in the Julian calendar. But when dates are entered in the Wikidata database, they must always be entered in the Gregorian calendar. Jc3s5h (talk) 00:16, 25 October 2014 (UTC)

The issue is caused by Wikidata editor. It marks its as Julian, but display as Gregorian. — Ivan A. Krestinin (talk) 04:29, 25 October 2014 (UTC)
Based on what I see in the diff for your next edit and other discussions I have had about the editing interface, you entered 1869-12-14 as the Gregorian birth date. That is still the date in the article today. But the English Wikipedia gives his birthdate as December 26, 1869. The Russian Wikipedia gives 14 (26) декабря 1869 (144 года). As best I can tell, he was actually born December 26, 1869, and that is the date that should be stored in the Wikidata. The "Julian" superscript that appears when viewing the item with the interface you get when you use the Wikidata search box in the upper right of the window is merely an instruction that the date should be converted to the Julian calendar before presenting it to a reader. Jc3s5h (talk) 15:27, 25 October 2014 (UTC)
Yes, but Wikidata editor ignores "Julian" superscript and displays date as Gregorian always. User needs to convert Julian dates mentally. This behavior causes many errors and misunderstandings like this. I do not recommend use dates before 1917 from Wikidata before Wikidata editor will be fixed. — Ivan A. Krestinin (talk) 16:00, 25 October 2014 (UTC)
I agree the editing interface is (insert your favorite obscenity), but the data for this property seems to be wrong. If you agree the Gregorian birth date is December 26, 1869, could you please fix the property? Jc3s5h (talk) 16:17, 25 October 2014 (UTC)
I agree, the Gregorian birth date is December 26, 1869. You can fix it. But this is only one of many this type errors. I do not see reason to fix items until the wdeditor produces these errors. Also these is alternative approach: store dates "as is", without double conversion (to Gregorian during saving and back during displaying). Current wdeditor implementation uses this approach as I see. — Ivan A. Krestinin (talk) 16:48, 25 October 2014 (UTC)
Done. Jc3s5h (talk) 17:19, 25 October 2014 (UTC)


P1566 report

Wikidata:Database reports/Constraint violations/P1566 seems to be broken "ERROR: Invalid Q-key: QQ212" (Ukraine (Q212)?) Multichill (talk) 07:57, 30 October 2014 (UTC)

Constrain reports

Hello Ivan Your constraint reports: Do I have the option to stop a lenghy report section after the first 150 items? Do I have an item for 'unknown' and 'n/a' (not applicable)? Thx. Regards --Chris.urs-o (talk) 18:31, 30 October 2014 (UTC)

Known datasources

  • Hello Chris, I view these datasources. Its are not optimized for automatic usage. Import is possible, but not simple. Also licensing question is not clear. Now I am working primary on different area. Please move the request to WD:Bot requests, maybe another botmaster has more free time. — Ivan A. Krestinin (talk) 09:37, 16 November 2014 (UTC)

Error on P723 report

Hi, report P723 states "ERROR: No constraints templates found." despite {{Constraint:Person}} on Property talk:P723. -- LaddΩ chat ;) 11:55, 1 November 2014 (UTC)

OK! I thought it was one of your templates. Thanks for adding the message on {{Constraint:Person}}. -- LaddΩ chat ;) 14:38, 2 November 2014 (UTC)

STOP KRbot please

Hi Ivan,

the DNB catalogue http://portal.d-nb.de is experiencing problems since yesterday, any query is answered with "not found" or "general error". Your bot is removing many GND IDs GND ID (P227), all of them should be still valid. Can you stop your bot and revert the deletions beginning with [24]? Thank you very much. -- Gymel (talk) 07:08, 2 November 2014 (UTC)

Thank you very much. The weekend is over now and DNB obviously has fixed their catalogues so you could unleash your bot again. -- Gymel (talk) 16:19, 3 November 2014 (UTC)
Yes, DNB is operating now. But I need to make some changes in my bot to avoid same situation in the future. — Ivan A. Krestinin (talk) 16:29, 3 November 2014 (UTC)

Dash in Career record

Hi, is any reason to have hyphen "330-215", instead of correct dash "330–215"? Best regards.--Kacir (talk) 09:32, 6 November 2014 (UTC)

Sorry, but what does it mean in property documentation:
  • Example item and value: 151 – 218
, there is example of dash (although with incorrect "space" (corr. 151–218, e.g. right record in article's infobox). Eventually, where can I inform about that problem, there on property documentation as a new section? --Kacir (talk) 17:29, 13 November 2014 (UTC)
The property talk page is good place for discussing property related issues. — Ivan A. Krestinin (talk) 19:51, 13 November 2014 (UTC)

Please explain edit

Your bot did this. Why? DGtal (talk) 07:52, 9 November 2014 (UTC)

MacTutor ID

Здравствуйте, спасибо большое за их загрузку из немецуой Википедии! Сейчас в списке [25] около 2200 биографий, а в Викиданных около 1900. Не могли бы вы составить ботом табличку отсутствующих? Взаимно-однозначного соответствия, видимо, добиться не получится (так как на мактьюторе есть несколько статей практически в жанре некролога), но я думаю, что около 250 ссылок я бы мог добавить. Danneks (talk) 18:47, 9 November 2014 (UTC)

Regular update of Mandatory constraint violations report

Hi Ivan, your new report Wikidata:Database reports/Constraint violations/Mandatory constraints/Violations is so popular and effective that it's often difficult to find in it an error that is not yet corrected. Would it be possible to generate it more regularly, say, every 4 hours? thanks -- LaddΩ chat ;) 15:38, 14 November 2014 (UTC)

Hello, now I am working on tool for cleaning the report from fixed items. I plan do this every 2 hours. Real update interval (e. g. finding new violations) is limited by incremental dumps generation period. — Ivan A. Krestinin (talk) 18:29, 14 November 2014 (UTC)
Great! thank you so much - -- LaddΩ chat ;) 18:42, 14 November 2014 (UTC)

IMDb constraint

Maybe this constraints isn't useful. There are a lot of sister/brother duo etc.. on IMDb. --ValterVB (talk) 08:14, 15 November 2014 (UTC)

Usually IMDb registers sisters/brothers as different records. I took a look to found items. The most contains errors. I will review this today in details. — Ivan A. Krestinin (talk) 08:22, 15 November 2014 (UTC)
I review the items, only 4 exceptions found: Kelley Roos (Q16028224), Garfunkel and Oates (Q4039264), The Boswell Sisters (Q894612), Siegfried & Roy (Q60806). Some of these four IMDb items are looked strange. For example nm4424651 contains text: "Siegfried & Roy is an actor, known for...". So I think such items are exceptions and possible errors, not rule. — Ivan A. Krestinin (talk) 21:40, 15 November 2014 (UTC)

P17

Hello Ivan,

The report is currently hard to open. Not sure what causes it.


BTW Would you de-activate the self-link section. For this constraint, those can actually be useful. Otherwise many "target required claims" with P17 would be hard to make. --- Jura 07:02, 16 November 2014 (UTC)

Hello, my browser opens the report without some issues. I think the issue is caused by Values statistics section. You can try to delete the section and see result. Self-link section is interesting. Botwork is needed maybe. — Ivan A. Krestinin (talk) 07:50, 16 November 2014 (UTC)
No, I just completed the missing self-links ;)
I will try to delete some sections to see if it becomes visible. --- Jura 08:01, 16 November 2014 (UTC)

Names

Hello Ivan,

If you have a moment, would you update the Person Names list? --- Jura 19:37, 17 November 2014 (UTC)

Links in mandatory constraint reports

Hello Ivan, it would be nice if in Mandatory constraints and Mandatory constraints/Candidates your bot could link the detailed reports from the violation counts like this. Petr Matas 13:30, 18 November 2014 (UTC)

False positives in constraint violation report?

I cannot find any reason for the 2650 violations apearing at Wikidata:Database reports/Constraint violations/P376#Value_type_Q18551696. Maybe I overlooked something or they are false positives. May it be related to Wikidata:Contact the development team#Some statements incorectly show their value is a deleted item? Petr Matas 23:18, 18 November 2014 (UTC)

Detecting constraints

Just a theoretical question: there's a property I proposed that currently has constraints set wrong, so your bot reports many violations that aren't really informative. I have to think about how to define them better, but when it's fixed, will your bot pick up the changes from the property talk page automatically or do I have to contact you to change bot settings? The report I'm talking about is P1587, where not all items with that ID are instance of cultural monument of Slovenia (Q12794452) (in fact, most of them aren't). — Yerpo Eh? 21:46, 21 November 2014 (UTC)

Ok, thanks. — Yerpo Eh? 08:20, 22 November 2014 (UTC)

Иван, ваш бот исправил ошибку в дате рождения, но не исправил её в квалификаторе "не ранее" в дате смерти, там ошибка того же рода [26]. Запустите его ещё раз на квалификаторы "не ранее" и "не позже". Из-за этого в Викитеке не правильно отображается дата. -- Sergey kudryavtsev (talk) 16:37, 23 November 2014 (UTC)

  • Приветствую, Сергей, раньше редактор ВД сохранял даты в формате XXXX-01-01, если пользователь указал только год. Некоторые скрипты, в отличии от редактора сохраняли в формате XXXX-00-00, мой бот приводил всё к единому виду XXXX-01-01. Однако несколько дней назад случился апдейт движка и редактор ВД начал сохранять даты в формате XXXX-00-00. В этих условиях продолжать исправлять на формат XXXX-01-01 несколько странновато, скорее нужно уж тогда везде на XXXX-00-00 переделывать, но это в любом случае долгий процесс. Потому я рекомендую исправить скрипт отображающий даты, чтобы он формат XXXX-00-00 тоже воспринимал как нормальный. — Ivan A. Krestinin (talk) 16:47, 23 November 2014 (UTC)

P189

Please take a look at Constraint violations/P189. There are some violations in the list which should be OK now, For example the 4. (Q374591) and the 6. (Q510876). Thanks, --Molarus 11:59, 28 November 2014 (UTC)

If that is the problem, I can fix the other items. thanks. --Molarus 20:31, 28 November 2014 (UTC)

Иван, ваш бот исправил ошибку в дате рождения, но не исправил её в квалификаторе "не ранее" в дате смерти, там ошибка того же рода [27]. Запустите его ещё раз на квалификаторы "не ранее" и "не позже". Из-за этого в Викитеке не правильно отображается дата. -- Sergey kudryavtsev (talk) 16:37, 23 November 2014 (UTC)

  • Приветствую, Сергей, раньше редактор ВД сохранял даты в формате XXXX-01-01, если пользователь указал только год. Некоторые скрипты, в отличии от редактора сохраняли в формате XXXX-00-00, мой бот приводил всё к единому виду XXXX-01-01. Однако несколько дней назад случился апдейт движка и редактор ВД начал сохранять даты в формате XXXX-00-00. В этих условиях продолжать исправлять на формат XXXX-01-01 несколько странновато, скорее нужно уж тогда везде на XXXX-00-00 переделывать, но это в любом случае долгий процесс. Потому я рекомендую исправить скрипт отображающий даты, чтобы он формат XXXX-00-00 тоже воспринимал как нормальный. — Ivan A. Krestinin (talk) 16:47, 23 November 2014 (UTC)
А вы не знаете, сохранение в формате XXXX-00-00 редактором ВД - это осознанное решение или очередной ляп программистов? -- Sergey kudryavtsev (talk) 06:52, 1 December 2014 (UTC)
PS: Сегодня добавил даты в Q4240550 - та же самая история. -- Sergey kudryavtsev (talk) 06:54, 1 December 2014 (UTC)
PS2: Гаджет w:ВП:WE-F сохраняет даты в формате XXXX-01-01. При открытии дату в формате XXXX-00-00 он показывает строку ISO-формате и позволяет исправить дату, так я и сделал для Q4240550. -- Sergey kudryavtsev (talk) 07:03, 1 December 2014 (UTC)
Не знаю, в принципе эти числа могут быть любыми, хоть -77-77. Почему они сейчас поменяли с -01-01 на -00-00 - не знаю, могли и неосознанно, можно попробовать поискать в багтрекере, возможно были какие-нибудь баги на этот счёт. — Ivan A. Krestinin (talk) 07:07, 1 December 2014 (UTC)


Mandatory consraint violations report

Hello Ivan,

Just to let you know: I undid a change at Property talk:P213‎ which might have crashed the generation of the report. Hope this helps. --- Jura 11:06, 7 December 2014 (UTC)

Thank you. I need to add something to avoid bot fails in such situations. — Ivan A. Krestinin (talk) 11:48, 7 December 2014 (UTC)

Qualifier values

Hi Ivan, Wikidata:Database reports/Constraint violations/P548 says that the qualifier property version type (P548) is used nowhere but that's not true. Because of this there are no statistics (tables). Do you have any idea how to fix this? Matěj Suchánek (talk) 14:48, 7 December 2014 (UTC)

Resolving redirects

Hi! Do you have comments on Resolving redirects? https://lists.wikimedia.org/pipermail/wikidata-tech/2014-December/thread.html --JulesWinnfield-hu (talk) 17:35, 10 December 2014 (UTC)

Hello, comments... yes I have a bit. All redirects are need to be resolved in WD. Not only double. Why? Many algorithms need to compare Q-keys. For example simple instance of (P31) import bot: it needs to compare adding value with existing values. Bot can simple compare Q-keys in redirect free world. World with redirects is more difficult. Bot needs to request Wikidata for item compare operations. Every data processing algorithm author must keep this in mind. Every author needs to spend his time to code and debug this. This increases complexity of every data processing algorithm. I had a number of different data processing algorithms then redirects appears in WD. It was too hard to fix comparing operation in every algorithm. So I find more simple way: fix the redirects. Negative side is invalid edits in case of merge errors. To avoid this my bot waits potential merge revert for 24 hours. — Ivan A. Krestinin (talk) 20:19, 10 December 2014 (UTC)
Thanks. The question was “Is there a bot resolving those?” If yes, please write it on the list. --JulesWinnfield-hu (talk) 20:54, 10 December 2014 (UTC)
Yes. But I am not subscribed to this list. I prefer to discuss WD on WD pages. — Ivan A. Krestinin (talk) 20:59, 10 December 2014 (UTC)

Strange edit

I have reverted --ValterVB (talk) 22:07, 10 December 2014 (UTC)

Q17142442 is redirect https://www.wikidata.org/w/index.php?title=Q17142442&redirect=no --JulesWinnfield-hu (talk) 22:13, 10 December 2014 (UTC)
Not only “said to be the same”, but now it is the same. --JulesWinnfield-hu (talk) 22:21, 10 December 2014 (UTC)
The question is: is this edit correct? — Ivan A. Krestinin (talk) 05:02, 11 December 2014 (UTC)

IMA status and/or rank (P579)

Hello Ivan
Is there a way to see which items with property 'IMA status and/or rank (P579)' don't have an English label and/or an English description?
Thx. Regards --Chris.urs-o (talk) 04:16, 21 December 2014 (UTC)
Hello Chris, ✓ Done, please see User:Chris.urs-o/IMA. — Ivan A. Krestinin (talk) 08:12, 21 December 2014 (UTC)
Thx Ivan --Chris.urs-o (talk) 14:37, 21 December 2014 (UTC)

Names report

Hello Ivan,

If you have moment, would you update them? --- Jura 08:53, 21 December 2014 (UTC)

Top offenders

Hi Ivan, would it be possible to make a report of items that show up in most reports? Take for example Wallace D. Hayes (Q73264), it showed up in 9 reports. By fixing the date of birth that should be reduced. Multichill (talk) 17:53, 21 December 2014 (UTC)

KrBot P373 revert

Do you realize that KrBot (talkcontribslogs) is removing Commons category (P373) claims instead of reverting them when a user changes Commons category (P373) from a valid to invalid link? See the history for cyst (Q193211). --Closeapple (talk) 06:37, 22 December 2014 (UTC)

Why is not URL enough as source?

I saw that your bot added imported from Wikimedia project (P143) to Zarina Diyas (Q2466357) although there was already "URL" used as source. I think that the URL is enough. --Stryn (talk) 20:58, 26 December 2014 (UTC)

Reports

Why are the constraint violations reports not updated? Sjoerd de Bruin (talk) 12:23, 28 December 2014 (UTC)

Symmetric

Hello Ivan,

Any idea why "symmetric" came up on this. The pages I check seem fine and the constraint hasn't changed recently. --- Jura 06:23, 29 December 2014 (UTC)

In any case, I edited all of them, maybe this refreshes the data. --- Jura 12:22, 29 December 2014 (UTC)
Thank you, it was bug in code, fixed. — Ivan A. Krestinin (talk) 21:06, 29 December 2014 (UTC)

Please update references when updating value

Hello Ivan, I've added Bibliothèque nationale de France ID (P268)=13898981 to Dick Richards (Q1209782) with stated in (P248)=Virtual International Authority File (Q54919), retrieved (P813)=2014-12-27 and reference URL (P854)=https://viaf.org/viaf/19301482. Shortly after your Bot User:KrBot has changed the value to Bibliothèque nationale de France ID (P268)=138989812 which no longer matches the information on https://viaf.org/viaf/19301482. In my opinion, your bot should either remove the references in such a case, or add the new value and leave the old value untouched, if it is supported by references. I know, that BnF 13898981 doesn't exist anymore. But the reference is only valid for 13898981 and not 138989812, because in VIAF only 13898981 is linked to Dick Richards (Q1209782) --Floscher (talk) 20:37, 29 December 2014 (UTC)

Strunz 8 ed

Hello Ivan
Have a nice 2015, for u and ur family. Could you edit the constraint violations to accept IDs ending with a lower case letter please?
Wikidata:Database reports/Constraint violations/P711#Format
Property talk:P711 Thx --Chris.urs-o (talk) 14:26, 31 December 2014 (UTC)
Hello Chris, you too. Done. 'b' letter can be replaced to '[a-z]' if needed. — Ivan A. Krestinin (talk) 09:09, 1 January 2015 (UTC)
Thx --Chris.urs-o (talk) 11:30, 1 January 2015 (UTC)

Q404976

Why does this bot keep removing P41 (flag)?--Roger Camotes (talk) 19:36, 2 January 2015 (UTC)

There is no file named c:File:Flag of Garcia Hernandez, Bohol.png on Commons. Sjoerd de Bruin (talk) 19:40, 2 January 2015 (UTC)
Thank you for quick response. Is this local file for some Wikipedia maybe? — Ivan A. Krestinin (talk) 19:54, 2 January 2015 (UTC)
Seems to be: en:File:Flag_of_Garcia_Hernandez,_Bohol.png. --Stryn (talk) 20:01, 2 January 2015 (UTC)

Mineralogy

Hello Ivan
Constraint violations: is it possible to express:
"If Property:P579 is approved mineral and/or valid name (A) (Q13406835) then items with this property should also have Property:P690"
and
"If Property:P556 is triclinic crystal system (Q376927) then items with this property are allowed only to have Property:P690 with 'unknown', triclinic-pedial (Q13364996) or triclinic-pinacoidal (Q13365008)"
Regards --Chris.urs-o (talk) 08:29, 4 January 2015 (UTC)
Hello Chris, constraints templates does not support "if" constructions now. But AutoList tool supports its, for example see [28]. — Ivan A. Krestinin (talk) 08:43, 4 January 2015 (UTC)

Request

Could you refresh the report of input device (P479) for me? Will make it easier to process the remaining items. Sjoerd de Bruin (talk) 10:26, 11 January 2015 (UTC)


Pseudo-duplicate on Q84445

Hello Ivan,

At the above, I reverted the bot as the listing of places is sequential. --- Jura 22:59, 18 January 2015 (UTC)

Hello, thank you. I exclude this property from processing. More correct way is qualifying the values by start time (P580) / end time (P582) or another qualifier. — Ivan A. Krestinin (talk) 05:47, 19 January 2015 (UTC)

Hochschule

Moved to Talk:Q38723

Wikidata:Database reports/Constraint violations/P625

At Wikidata:Database reports/Constraint violations/P625 "Conflicts with" violations are listed (maybe even computed) twice. Any idea why ? Talkpage with config seems to be OK to me. --Jklamo (talk) 22:38, 4 February 2015 (UTC)

Hello, bot has a little bug: it see templates inside comments, pre and nowiki tags. Workaround is to add something more symbols to hide template name from bot. I do this for P625. — Ivan A. Krestinin (talk) 04:31, 5 February 2015 (UTC)
Thanks. --Jklamo (talk) 16:24, 6 February 2015 (UTC)

NLI removal

Hello Ivan! Why did you remove this? As far as I can tell, it is the correct identifier, see [29]. Or is there some consensus not to list NLI identifiers on Wikidata? Jonathan Groß (talk) 11:36, 6 February 2015 (UTC)

I think you're right. Looks like I'll have to be more careful with links to non-Latin alphabet catalogues. At least in this case, I found the right identifier: [30] Jonathan Groß (talk) 09:14, 9 February 2015 (UTC)

BnF error?

Hi, I think that KrBot is setting the wrong BnF in the item thatgamecompany. Could you please check it? The value cb16759917m corresponds to this page at bnf.fr, but for some reason KrBot tries to remove the first two letters and setting "16759917m", which is wrong if I understand correctly. (At least if you check the URL to bnf.fr, it will not work with the value cb16759917m replaced by 16759917m.) Thanks for your time. Fred Johansen (talk) 12:27, 6 February 2015 (UTC)

@Johansen.fred: Please see Property talk:P268.
Notice n° : FRBNF16759917
= http://catalogue.bnf.fr/ark:/12148/cb16759917m/PUBLIC
= 16759917m [on WD: P268]
If you click on the number you've added, you can see that the link is not working. --Kolja21 (talk) 15:11, 6 February 2015 (UTC)
Ah, ok, then I see! I have reverted the item back to KrBot's last revision. As for why I didn't understand it before, I think the link in the BnF field value was not (always) active for me earlier. Even now, it is active only when I load the page, but when I click the BnF link and then use the Back button in my web client (Chrome) to go back to the item, the BnF field value has no working link anymore. In that instance the value appears just like any other text. Fred Johansen (talk) 16:56, 6 February 2015 (UTC)
The last symbol is control sum. viaf.org uses IDs without this symbol. catalogue.bnf.fr uses IDs with control symbol and cb prefix. Wikidata uses something in the middle: with control sum, but without cb prefix, see Property talk:P268. — Ivan A. Krestinin (talk) 18:28, 6 February 2015 (UTC)


Undo

Hi Ivan,

Would you undo edits like this. They were done following an incorrect merge. In generally, I don't think the bot should update links to point to disambiguation items. --- Jura 12:00, 13 February 2015 (UTC)

Hello Jura, done. — Ivan A. Krestinin (talk) 20:31, 13 February 2015 (UTC)
I don't think the merge was wrong. Jura can't understand that page about personal name can be sometimes disambiguation page. --Infovarius (talk) 18:18, 15 March 2015 (UTC)

Statement on properties

Hi Ivan,

We are still thinking about how the constraints should look like when they are migtrated to statements on properties. Your proposal seems to work perfectly so this approach should be used. We are working on a script to migrate the constraints from the property talk pages to the properties. By doing this we asked ourselves, what do you mean with this "comment-property" you proposed. Is it only for clarity and has no relevance for checking the constraints? When we migrate, do we have to pay attention on those or are they added in the future? Jonas.keutel (talk) 17:06, 26 February 2015 (UTC)

One example: [31]. It will not be used by algorithms, but it is needed to clarify non-obvious constraints. — Ivan A. Krestinin (talk) 18:31, 26 February 2015 (UTC)

new LCCN property removal

Thank you for correcting my errors here and here. Am I correct in understanding that my mistake was adding subject headings with a sp prefix and only "n|nb|nr|no|ns|sh|sj|sn" prefixes are allowed in Property:P244, and if so is there a property I should use to connect this info to or not? Thanks, DGtal (talk) 10:31, 1 March 2015 (UTC)

P360

Care to explain this reversion: [32] ?

The strong sense from eg this Project chat discussion and eg @Jura1:'s comments in this PfD discussion is that P360 not only is appropriate for categories, it delivers exactly what is needed to make machine-readable sense of categories.

(See also this discussion at WD:PC on how far one should be able to go mining content fully automatically from a category based on the category's P360; and this new property, plus the previous two and the following one, currently being proposed at WD:PP/GEN to facilitate the process.)

Rather than making a blind revert, please join the discussion. Jheald (talk) 10:58, 1 March 2015 (UTC)

Please see Property talk:P360#P360 and categories for details. English, Russian or Polish chart is not correct place for such discussions. — Ivan A. Krestinin (talk) 16:54, 1 March 2015 (UTC)

Лудовико

Почему вы подчиняетесь латиноалфавитному участнику Jura и отсоединили статьи (например, Ludovico Ariosto (Q48900)) от русской статьи про имя (в Ludovico (Q6698984))? --Infovarius (talk) 15:18, 15 March 2015 (UTC)

  • Участник Jura пытается использовать для свойства given name (P735) не дизамбиги, а статьи про имя. По-видимому поэтому ему не очень нравится, когда айтем соответствующий дизамбигу объединяют с айтемом про имя. Не берусь судить насколько хороша его идея, вижу в ней и недостатки и достоинства. Дискутировать с ним на эту тему не хочется, другой работы хватает. — Ivan A. Krestinin (talk) 20:28, 15 March 2015 (UTC)

Обмен

Здравствуйте! Почему вы удаляете статьи "half-brother" и "half-sister" из статей редактированных мной, и обмените их с не существующими статьями Q19595226, Q19595227 и Q19595228? Если вы хотели бы использовать ваши статьи, пожалуйста маркировать их по-английски и входить существующие "interwiki" статьи, потому что ваши статьи еще пустые! Спасибо. Пока! Borgatya (talk) 15:33, 17 March 2015 (UTC)

  • Добрый день, не всегда для какого-то класса объектов есть статья в Википедии. Элементы я не удаляю, а заменяю более точными. Как перевести "единоутробный брат" и "единокровный брат" на английский я к сожалению не знаю. Мои познания в английском весьма ограничены. Надеюсь, что на английский это переведёт кто-нибудь лучше знающий этот язык. — Ivan A. Krestinin (talk) 19:51, 17 March 2015 (UTC)
Дорогой Иван! Спасибо за ваш ответ! Хорошо, но пожалуйста не править єти статьи на будущее с KrBot-ом, потому что єти статьи полностью хорошие с статьями "half-brother" и "half-sister". Надеюсь, что вы понимаете мое мнение.Borgatya (talk) 14:57, 18 March 2015 (UTC)
half-brother не очень точно отражает родство. Лучше использовать более точные классы. — Ivan A. Krestinin (talk) 19:44, 18 March 2015 (UTC)
например???Borgatya (talk) 12:13, 19 March 2015 (UTC)
half-brother (Q15312935) говорит лишь о том, что один из родителей одинаков. paternal half-brother (Q19595226) о том, что одинаковым является именно отец, а матери разные. — Ivan A. Krestinin (talk) 17:55, 19 March 2015 (UTC)

Russian swamp

Could you please check your bot's last editions on swamp (Q166735)? I don't think swamps are only in Russia and have postal codes ;) Thanks! --Quico (talk) 19:35, 22 March 2015 (UTC)

Please see ce:Болото (Бокситогорскан кӀошт) linked to swamp (Q166735). We have many interwiki conflicts like this. Please see this or this list for another conflicts. — Ivan A. Krestinin (talk) 19:48, 22 March 2015 (UTC)

KrBot и индексы

Ограничьте, пожалуйста, работу бота по добавлению индексов только для элементов со свойством населённый пункт или подобными (если такое возможно). Сейчас из-за наличия совершенно некорректных ссылок на ce.wp и других бот проставляет индексы дизамбигам, станциям метро и т. д.[33][34][35] Да, проблема конечно в этих ссылках, но и к новым некорректным правкам это не должно приводить. YLSS (talk) 20:33, 22 March 2015 (UTC)

Сейчас эти элементы к сожалению в равной степени и о населённых пунктах, а не только о станциях метро, потому говорить о новых некорректных правках не совсем правильно. Ограничить бота конечно можно, но тогда мы так и не узнаем о наличии конфликтов в интервиках. Эффективнее на мой взгляд конфликтные элементы всячески выделять и работать над их исправлением, чем прятать и игнорировать их наличие. Остальные элементы с конфликтными интервиками можно посмотреть в этом или этом списках, там кстати и до начала работы моего бота полторы тысячи элементов было. — Ivan A. Krestinin (talk) 20:48, 22 March 2015 (UTC)

Russian postal addresses in the outer solar system

Mother Russia is Great, but your bot is the first who has claimed a Russian postal address in the outer Solar system. [36] -- Innocent bystander (talk) 07:24, 23 March 2015 (UTC)

Hello, interwiki conflict causes so strange statements. See ce:Плутино linked to this item. — Ivan A. Krestinin (talk) 07:33, 23 March 2015 (UTC)
I link ce:Плутино with Plutino (Q12141106). -- Sergey kudryavtsev (talk) 07:46, 23 March 2015 (UTC)
Thank you for cleaning Wikidata from conflicts. — Ivan A. Krestinin (talk) 08:00, 23 March 2015 (UTC)

fictional character subclas of person

@Ivan A. Krestinin, Haplology, Gymel:

Removed statement

Do you have any reason to this? -- Innocent bystander (talk) 08:59, 26 March 2015 (UTC)

Ok, it was a duplicate, we really need the possibility to modify the edit comments ourself. -- Innocent bystander (talk) 09:04, 26 March 2015 (UTC)

How to exclude items from your "Suggested for merge" page

Hi. On your page User:Ivan_A._Krestinin/To_merge/ukwiki I see suggestion

Numbers

'Мі-#' — 'Ми-#' (rating 10, based on Q155635, Q155651, Q169427, Q242806, Q320900, Q336150, ...)

While this is wrong suggestion, what should I do to exclude this item from furher processing ? Probably I need to add some string in Wikidata:Do not merge, like this one ? --Movses (talk) 06:26, 28 March 2015 (UTC)


Personal names

Hi Ivan,

Would you update the reports when you get a chance? It has been some time since. --- Jura 07:11, 29 March 2015 (UTC)

Updated, but not all. Some reports are too large. Also unresolved redirects make update process very long due errors. — Ivan A. Krestinin (talk) 12:54, 29 March 2015 (UTC)
Thanks. I did some work on the labels. I also had a look at the redirects to disambiguations. Maybe one could update automatically, if there are less the five links. If there are more, it might be worth looking into them. In any case, eventually one need to fix manually the links on Special:WhatLinksHere/Q2426827 (or remove them). --- Jura 14:13, 29 March 2015 (UTC)

Hi Ivan, some taxa have now a postal code (P281) and belong to a country (P17) (list). :( Seems these values came from Tatar Wikipedia (Q60819), eg. red fox (Q8332) is linked to tt:Руч. Could you please fix these issues. Maybe checking the value of instance of (P31) could help next time. Thank you very much. --Succu (talk) 16:54, 1 April 2015 (UTC)

We have too many invalid items unfortunately. I can check for instance of (P31) of cause. But this just hide existing invalid items, not resolve conflict. I am working on invalid items fix, but Wikidata contains too many errors to resolve its by single person. For example see list: Wikidata:Database reports/Constraint violations/P281#Conflicts with. — Ivan A. Krestinin (talk) 19:49, 1 April 2015 (UTC)
Ivan, I ask for help resolving existing conficts like this: ce:Клюква linked to cranberry (Q13181). A move to another existing item is necessary or a creation of a new one. Removing these statements won't help. Regards --Succu (talk) 21:07, 1 April 2015 (UTC)
I separated ~200 items yesterday. Today we have ~1000 items like this. I am working on it, but the list is too long and contains too various cases for quick fix. — Ivan A. Krestinin (talk) 21:20, 1 April 2015 (UTC)

При расстановке индексов - русской фамилии "Никитин" был присвоен индекс

Что-то пошло не так?

https://www.wikidata.org/wiki/Q1389775 - здесь тоже самое "Мельников" - фамилия

Для рувики Q1389775 - это фамилия, для чевики - уже населённый пункт. Таких элементов у нас к сожалению много, будет здорово, если поможете их исправить: Wikidata:Database reports/Constraint violations/P281‎#Conflicts with. — Ivan A. Krestinin (talk) 13:26, 6 April 2015 (UTC)
угу. я руками сейчас по фимилиям бегаю и категории оптимизирую и убираю неправильные алиасы, если силы останутся пробегусь по именам тоже.

Reports: titles missing

Hi Ivan, your bot has changed the daily reports from:

{{Q|49133}} (old) -> [[Q49133]] (new)
= Museum of Fine Arts Boston (Q49133) -> Q49133

Can you change it back to the old version with titles? Cheers --Kolja21 (talk) 22:40, 8 April 2015 (UTC)

Hi, bot uses {{Q}} on small reports only (items count < 500). This template generates "Lua error: not enough memory." on large reports. — Ivan A. Krestinin (talk) 22:58, 8 April 2015 (UTC)
Wikidata:Database reports/Constraint violations/P227 has less than 500 items. Would it help if we delete "Qualifiers" violations? (It's less important.) --Kolja21 (talk) 23:23, 8 April 2015 (UTC)
13 unique value violations were resolved yesterday and bot uses {{Q}} again. {{Constraint:Qualifiers}} section can be reduced by adding valid properties to its list. — Ivan A. Krestinin (talk) 19:09, 9 April 2015 (UTC)
Perfect. Thanx! --Kolja21 (talk) 22:51, 9 April 2015 (UTC)


Redirects

Hi Ivan,

If the below are redirects, would you run your bot to update the items using them in given name (P735) them to the target?

I merged about 300 given names and there are about 50'000 items that need fixing. --- Jura 11:55, 10 April 2015 (UTC)


list of items

Q18038709 Q18038848 Q15208017 Q18002648 Q18012369 Q2424149 Q1158394 Q18191037 Q4927713 Q14629919 Q18410867 Q18199687 Q18176395 Q18145462 Q18157157 Q18551845 Q14946485 Q18075850 Q3195138 Q4178117 Q18420753 Q18181772 Q17864520 Q16281837 Q18609788 Q19688723 Q18814779 Q18067255 Q18111066 Q10541107 Q18608876 Q16618163 Q15849752 Q18342070 Q19688779 Q18188038 Q3801794 Q3766002 Q13584814 Q16422886 Q18524102 Q18190927 Q10544842 Q18220098 Q18608522 Q18526591 Q18177247 Q16280912 Q18180853 Q11887738 Q14438051 Q13405605 Q17854918 Q1158506 Q16280131 Q18220772 Q3229170 Q18978403 Q18195741 Q19688402 Q18707653 Q16275784 Q18193752 Q11220941 Q4817761 Q19689377 Q18220017 Q16430008 Q19688788 Q3991600 Q18116812 Q19689420 Q16276257 Q18194764 Q16154333 Q3094170 Q18587703 Q16549390 Q2874826 Q18082046 Q17641352 Q5209168 Q7900718 Q19688778 Q18180431 Q18978392 Q16323502 Q6751262 Q16275742 Q9301036 Q15732145 Q5831468 Q5528781 Q3885930 Q16275878 Q18195535 Q16280529 Q8758158 Q2755787 Q16278305 Q18950544 Q16276289 Q18147464 Q18844258 Q18030007 Q18130793 Q3719512 Q19411671 Q10479874 Q18227986 Q17770018 Q18039564 Q18145467 Q17518420 Q16274977 Q18218344 Q3759041 Q18562424 Q16282244 Q18571949 Q18520545 Q18385819 Q15823286 Q18218731 Q18324312 Q19729807 Q18978359 Q18584936 Q12796682 Q18182056 Q16419548 Q19688418 Q18385152 Q18823055 Q19688400 Q17518406 Q11725845 Q4354045 Q19689401 Q16861467 Q5932709 Q18218247 Q19688330 Q15829976 Q18180814 Q3482156 Q18177335 Q15816654 Q16277100 Q2963143 Q18611851 Q19689390 Q17142382 Q10482032 Q18222592 Q16275583 Q11977320 Q3491121 Q16422423 Q18503079 Q18171079 Q18551713 Q5088587 Q18449855 Q12059666 Q2689397 Q15809351 Q18023649 Q7308423 Q2827913 Q18978414 Q14901750 Q8346272 Q12672893 Q1209268 Q3623099 Q18218143 Q16282503 Q18038759 Q6124292 Q4256413 Q18978415 Q18986522 Q16161598 Q15111386 Q19347737 Q19689479 Q9258196 Q1822320 Q6836888 Q6762645 Q18156204 Q5829589 Q19688618 Q14770770 Q18978288 Q16430488 Q19345740 Q16276004 Q19688703 Q18452229 Q18198822 Q18198086 Q16282229 Q18448448 Q19687118 Q19688743 Q16275767 Q17764440 Q9267344 Q17327128 Q16425929 Q19688408 Q6434189 Q16423546 Q18978361 Q1820197 Q11684764 Q15130287 Q1233977 Q320865 Q16278300 Q19001706 Q16419976 Q19689528 Q12050883 Q9366507 Q11957650 Q16002716 Q3802190 Q10594119 Q16276743 Q18697129 Q19731088 Q18449597 Q1977784 Q16429346 Q12765136 Q11958147 Q16427546 Q19688881 Q1425137 Q18642351 Q18809207 Q16002097 Q16576135 Q18641773 Q16280473 Q16571606 Q16281029 Q2684485 Q16274716 Q14945139 Q2313170

Q12034998
Hi Jura, bot will start automatically after 24 hours elapsed. This delay is needed to take ability to revert merge. — Ivan A. Krestinin (talk) 22:36, 10 April 2015 (UTC)
Ok, sounds good. Shouldn't be an issue with the above, but it is safer. I thought the delay was shorter so I wondered if the bot was still running. --- Jura 08:11, 11 April 2015 (UTC)

Wrong bot action

Hi! Your bot chose not the correct action, losing data. [37] --JulesWinnfield-hu (talk) 10:18, 17 April 2015 (UTC)

KrBot deleted the VIAF number "ferrari". [38] Reason: "Format" violation. The bot don't know that an IP has changed the correct VIAF ID. But this form of vandalism is no big problem, since we have constraint violation reports (... P214#Format) that are checked daily. --Kolja21 (talk) 00:22, 18 April 2015 (UTC)
No violation, no problem. But you deleted the property, instead of reverting the vanadlism. If you delete all VIAF values we will have no violations. No problem, as you say. --JulesWinnfield-hu (talk) 10:34, 18 April 2015 (UTC)
It's not my bot. I'm just checking the constraint violation reports. I've tried to explain that this lists help to do the part of work you are missing. Step 1: KrBot puts the wrong IDs on the list. Step 2: KrBot deletes wrong numbers. Step 3: Humans (single users) correct vandalism. Result: No data get lost. BTW: Most of the cases are not vanalism but misspelled numbers. --Kolja21 (talk) 11:08, 18 April 2015 (UTC)
You're not right. In this case the data was lost. I discovered it only accidentally. --JulesWinnfield-hu (talk) 16:42, 18 April 2015 (UTC)
Hello Jules, I am sorry for delayed response. You are right, the data was lost from actual element version. There are two positive moments: this case occurs not often and the data is saved in element`s history (and in Constraint violations/P214 history). I will try to detect vandalism cases in bot`s code and do not delete values in this case. Thank you, Kolja, for the case clarification. — Ivan A. Krestinin (talk) 17:34, 18 April 2015 (UTC)
I add more checks to bot`s code. Bot will not delete such values. — Ivan A. Krestinin (talk) 07:26, 20 April 2015 (UTC)

VIAF changed from enWP to Wikidata

Hi Ivan, can you take a look at en:Wikipedia talk:Authority control/VIAF#OCLC Mechanisms? There is the question how often your bot deletes redundant VIAF IDs (redirects). Cheers --Kolja21 (talk) 00:05, 18 April 2015 (UTC)

Question regarding the KrBot's operation on VIAF numbers

Hi Ivan, at en:Wikipedia talk:Authority control/VIAF#OCLC Mechanisms a representative of OCLC is asking about details of your bot's operation with respect to constraint reports and detection / automatic tracking of changes in VIAF. I have tried to answer this based on my empirical knowledge but you might want to correct me or give additional comments. (As I understand the discussion takes place on en:WP instead here since it evolved from VIAF's announcement to switch their linking from en:WP articles to Wikidata items: Originally on a user page there and recently moved to the wikipedia namespace there). -- Gymel (talk) 00:08, 18 April 2015 (UTC)

Solving redirects

Hi, your bot automatically resolves redirects in items. However, per Help:Redirects#Links to redirects those links should stay as links to redirects. Then a wrong merge+redirect can be undone very easily, which becomes quite impossible after your bot "solved" all those redirecting links (cf. Wikidata:Project chat#Mess with japanese foster parents). Furthermore, I don't see any RFP for this task. Please also translate your bot's user page into English because I don't understand the language there and thus don't know what you write about your bot here. Best regards -- Bene* talk 13:07, 20 April 2015 (UTC)

Quite a conflict: In the case of qualifiers to properties wrong redirects are especially nasty to undo (I presume) and on the other hand it is very important for query purposes that redirects are "resolved": Querying all x with Py = z would become very hard if the query engine would have to expand z to the set of all z' which are sources of redirects to z... The example in question was a very isolated item, used only for ja:WP and therefore a) the wrong edit went unnoticed for the 24 hours the bot deliberately hesitates before executing the redirects and b) the effects were quite small in number. Unfortunately I do not see any other solution than to quickly breed more human users which are fluent in several languages. -- Gymel (talk) 13:36, 20 April 2015 (UTC)
@Bene*: Please do not stop this bot. The constraint violations reports as wells as WDQ rely on this. If you stop it, both systems won't work any longer properly which is, imho, worse than some wrong merges.--Pasleim (talk) 14:07, 20 April 2015 (UTC)
@Gymel: The problem concerning queries does not exist because the system can resolve redirects internally quite easily. So in the database tracking all statements the redirects would already be resolved. I think that being able to undo a merge, which is already quite hard, shouldn't be made even more difficult.
@Pasleim: Yes, stopping the bot wasn't my intention. I striked that sentence out as it is too harsh. Sorry. -- Bene* talk 14:17, 20 April 2015 (UTC)
@Bene*: I have no insight into the internals of "the system" but to recognize redirects in queries one usually needs a "reverse redirect" or "explosion" functionality and I honestly do not know how efficient this (i.e. substitution single items by unions at every single stage of a process) can be handled even in theory. And - are you sure that this functionality is implemented already? Some sections further up a user complains just because the bot is not resolving redirects in a timely manner ... -- Gymel (talk) 14:28, 20 April 2015 (UTC)
I don't know why User:Jura1 requested this "fixing". Concerning queries, the functionality to query for property-value-pairs does not exist at all but redirects will be handled of course. -- Bene* talk 14:32, 20 April 2015 (UTC)
Hello, strange section in help. Is this section was discussed anywhere before adding to help? My bot resolves redirects since introducing redirects in Wikidata. There was no problems with this. Bot delays resolving to 24 hours to allow users rollback changes. Also I can revert bot`s changes if needed. Redirects create many problems. Basic problem is keys compare algorithm: Q-key values can not be compared directly this simple action needs additional requests. — Ivan A. Krestinin (talk) 19:38, 20 April 2015 (UTC)
Afaik the api can automatically resolve redirects for you, otherwise this should be implemented. However, I hope you understand why I think not resolving redirects is better to be able to revert a merge. Maybe this reason isn't weighty enough but I think we should consider this. Some discussion is at least necessary imo. :-) -- Bene* talk 17:41, 21 April 2015 (UTC)
API cover only one of many cases. It does not cover simple data import case for example. Bot needs check existing values before adding new value to avoid data duplication. Usually bot does this using comparing new Q-key with exist ing Q-keys. Redirects makes this algorithm more complex. Many complex scenarios present in data consistency analysis and error fixing areas. — Ivan A. Krestinin (talk) 19:33, 21 April 2015 (UTC)
Hmm, ok, I see your point and it is valid. Maybe, as an alternative, you provide an easy way to undo the redirects resolving by your bot so that in case of a wrong merge, your bot can go through all redirects it resolved and revert them as well? This would be really cool. :-) If that is possible, we should change the guideline accordingly. -- Bene* talk 09:09, 22 April 2015 (UTC)
I concut with Bene*, undoing merge is a nightmare and if somewone wanted for some reason to hurt the project it would be really easy to use this tool. If it is catched late it becomes a mess, and a Wikipedia is likely to wrongly merged as for languages reason he can not see the difference in meaning, or because it in old style wiki habits want to merge articles and merge the items, likely wrongly to recreating Bonny and Clyde problems at the item level, which we do not want here. So it's important to get this right. Can you maintain for your bot a table of redirects to speed up things ? I don't think that solving redirects is a good idea and doing this removes almost completely the interest of the redirect feature. As far as Wikidata is concerned we could as well remove the item after the redirects are solved, so we come back the old situation for which we precisely introduced redirects, it is a nonsense.
Something that could make us solve this in a way that could satisfy everybody is a good unmerge tool that follows history back and propose solutions for unmerge ... TomT0m (talk) 09:54, 22 April 2015 (UTC)
  • I think the current delay in processing the redirects is helpful. It allows easy reverts. I think we should take the help page for what it is: help for new editors. --- Jura 10:17, 22 April 2015 (UTC)
    @Jura1: The guideline is not specialized for new editors. It's useful for anyone, especially for the one who will catch a wrong merge later, and this happens and is a nightmare. TomT0m (talk) 10:32, 22 April 2015 (UTC)
    Interesting point. We should probably add a list of tools and features that don't work with redirects. --- Jura 10:36, 22 April 2015 (UTC)

@Addshore, Magnus Manske: Can you please give us your view on how not resolving redirects affects WDQ and constraint validation? --JulesWinnfield-hu (talk) 10:48, 22 April 2015 (UTC)

  • I think most tools don't work with redirects. So if you want things to work, you have to fix the links. --- Jura 10:59, 22 April 2015 (UTC)
    @Jura1: If we're stuck with tools that don't work with new Wikidata versions, there is no point in adding new feature to the software. Those tools will have to evolve or be marked obsolete. Wikibase is not even feature complete when we compare to the initial development plan ... TomT0m (talk) 11:12, 22 April 2015 (UTC)
    Primary goal of redirects is permanent stable links for external data clients. Its are not so important for our internal usage. — Ivan A. Krestinin (talk) 18:59, 22 April 2015 (UTC)
    The api at least supports resolving redirects on the server. See Special:ApiSandbox#action=wbgetentities&format=json: redirects: Whether redirects shall be resolved. If set to "no", redirects will be treated like deleted entities.. I'm not sure how the dumps handle this but afair there also exists a mechanism to easily resolve them. -- Bene* talk 14:09, 23 April 2015 (UTC)
    As I say above, this mechanism covers only very limited number of cases. — Ivan A. Krestinin (talk) 19:23, 23 April 2015 (UTC)
    @Ivan A. Krestinin: I explicitely remember talking with Lydia on project chat or D:DEV my concerns about merge vandalism. This is also a motivation to have redirect, to be able to cancel it (removing the items was a catastrophy, as it at least would require an admin to undo the request ...) TomT0m (talk) 14:28, 23 April 2015 (UTC)
    I understand your concern, it is possible, but it is not popular vandalism type as I see. Also it is not so dangerous. Popular items usually are in watch lists of many users. Vandalism will be reverted inside 24 hour timeframe. Non-popular items have no backward links or have very limited number of links, its can be undo easily. The worst case with massive bot edits can be reverted using bot. — Ivan A. Krestinin (talk) 19:23, 23 April 2015 (UTC)
    Me and Bene* experience showed that we had to deal with several cases lately, so I'd say this is your opinion :) TomT0m (talk) 06:50, 24 April 2015 (UTC)
    My example : https://www.wikidata.org/w/index.php?title=Q11514315&action=history merged in october, catched in April. Only three sitelinks, but potentionally all the statements that linked to periodisation resolved. Must admit I moved the sitelinks but did not check if old statements that uses this item has been solved in the interval. Also note that someone had corrected the french label, not understanding he had to change the label because of a wrong merge. So merges are prone to errors and could be hard to notice. TomT0m (talk) 07:05, 24 April 2015 (UTC)
So what do you propose? Will you provide an easy way to revert the bot changes on redirects or will you stop that bot task and let Wikibase itself resolve the redirects? -- Bene* talk 12:34, 2 May 2015 (UTC)
The way exists already: just notify me in any form. I will automate this if request count will large. — Ivan A. Krestinin (talk) 13:17, 2 May 2015 (UTC)
Ok, so I may add this information to the help pages, ok? -- Bene* talk 13:47, 2 May 2015 (UTC)
Ok. Also we did not discuss one important thing: links created after redirect creation. Looks like its must be corrected manually in any case. — Ivan A. Krestinin (talk) 13:53, 2 May 2015 (UTC)
Of course, that's what makes it a nightmare in the first place :) Let's save us at least what can be done automatically, for a start. Another point : do you publish your code somewhere ? I don't like when important side of the project depends on a single user, if anything happens or if you want to live ... TomT0m (talk) 14:16, 2 May 2015 (UTC)
My code is not very simple for deploy. It is written on C++ and depends on a number of libraries. From other side the algorithm is trivial. It can be reimplemented easily using more popular frameworks like pywiki. — Ivan A. Krestinin (talk) 14:25, 2 May 2015 (UTC)
I was not speaking of just this algorithm, but you also runs pretty much all the constraint checks and so on :) That would be nice if you could push all that on github or somewhere, just deploying some code is different from recoding everything. (I'd like a code publishing policy for bots :))TomT0m (talk) 14:35, 2 May 2015 (UTC)
In the long term I think it would be even better to integrate thoese important tools like constraints and redirect resolving into Wikibase itself. For Constraints it is already worked on. The redirects should perhaps be resolved by Wikibase itself everywhere it is needed so we don't have to replace them any longer. I'm not sure where there still exist issues but as soon as we reached a proper support for resolved redirects in Wikibase the bot should obviously no longer replace them. -- Bene* talk 15:11, 2 May 2015 (UTC)
Preparing full code for publication is large job. It has five year history. Now it contain too many deprecated fragments and other issues. To Bene: moving redirects resolving task to wikibase is good way. Now wikibase does not resolve redirects in item-type property requests and in dumps. — Ivan A. Krestinin (talk) 15:42, 2 May 2015 (UTC)
It does not matter, something imperfect is something. I would be happy with just a push of the code in the current state in github. Maybe someone would do the cleaning for you (unlikely though :)) Perfection is an enemy sometimes. TomT0m (talk) 16:03, 2 May 2015 (UTC)
Thanks, I will try to fix those use cases where redirects don't get resolved properly. What exactly do you mean with "item-type property requests"? Can you provide an example? -- Bene* talk 09:15, 3 May 2015 (UTC)
For example https://www.wikidata.org/w/api.php?action=wbgetclaims&entity=Q8577237&property=P301 returns 6449034 instead of 12190402. — Ivan A. Krestinin (talk) 09:26, 3 May 2015 (UTC)

Metaclasses and constraints

Hi Ivan, did not see you comment yet on Wikidata:Requests_for_comment/Adopt_Help:Classification_as_an_official_help_page. This would be cool if you could comment something and give input on whether or not it is suitable for the constraint system, for example. TomT0m (talk) 09:56, 22 April 2015 (UTC)

Constraint reports for P1727 and P1747

Hi Ivan! The constraint reports for Flora of North America taxon ID (P1727) and Flora of China ID (P1747) are not up to date. Both have more than 10,000 usages for a long time. Any idea why? Regards --Succu (talk) 18:54, 30 April 2015 (UTC)

From the numbers on Wikidata:Database_reports/Constraint_violations/Summary, it seems that they might just have no errors. --- Jura 19:00, 30 April 2015 (UTC)
Hi, bot does not update report then the only change is item count. This approach saves a lot of time. You can use something like this to force report update. — Ivan A. Krestinin (talk) 19:14, 30 April 2015 (UTC)
Thanks. --Succu (talk) 06:30, 1 May 2015 (UTC)

Old edit

Do you remember why you have setted "no value" for Commons category (P373) in Q57771 --ValterVB (talk) 09:46, 1 May 2015 (UTC)

Hello, because Commons does not have category "Luna 24" and to avoid repeated invalid value import by bots. — Ivan A. Krestinin (talk) 09:54, 1 May 2015 (UTC)

P373 "no value"

Hello, I was wondering if you could help here. Thanks in advance, --Horcrux92 (talk) 09:50, 1 May 2015 (UTC)

Wikidata:Database reports/List of properties/datatype/string

Could your bot update Wikidata:Database reports/List of properties/datatype/string including a column for P1793? Also if it is possible to obtain data from talk pages, then add a column for "has unique value"? FreightXPress (talk) 18:34, 4 May 2015 (UTC)

This needs a time for programming, but I have another jobs now, sorry. Try to use existing reports instead. — Ivan A. Krestinin (talk) 18:44, 4 May 2015 (UTC)

Bot swing and a miss

Take a look at this bad edit. --Izno (talk) 14:37, 8 May 2015 (UTC)

There are a lot of wrong linked items like that. Revert the bot and move the sitelink to a different item. Sjoerd de Bruin (talk) 14:41, 8 May 2015 (UTC)
As if I didn't know that. I am usually happy to do so but I'm not editing at a location I should be... :) Additionally, the difference in language is as-always questionable. --Izno (talk) 20:32, 8 May 2015 (UTC)
Fixed. — Ivan A. Krestinin (talk) 18:12, 8 May 2015 (UTC)
Thanks! --Izno (talk) 20:32, 8 May 2015 (UTC)

I like this one! Jared Preston (talk) 15:36, 16 May 2015 (UTC)

Category:Manji is redirect to commons:Category:Swastikas, bot resolved it. — Ivan A. Krestinin (talk) 19:03, 16 May 2015 (UTC)


Hi Ivan,

Would you add a temporary category to ru:Шаблон:Фотомодель? Something like the following should do:

[[Категория:Фотомодель - Цвет волос - {{lc: {{{Волосы|}}} }}]]

I'd like to transfer the information to Wikidata. It should just take a couple of minutes, afterwards it can be removed again. I would do it myself, but for some reason, I can't add it in this language (some filter). Please ping me once it's done. --- Jura 12:41, 19 May 2015 (UTC)

@Jura1: I try this construction in preview mode, but it is incompatible with ref-tag, example article: ru:Евангелиста, Линда. I try to import this property using my framework, but I do not have enough ru-color-name <-> Q-item mapping. My bot error list: User:Jura1/Hair color import. — Ivan A. Krestinin (talk) 20:54, 19 May 2015 (UTC)

Thanks. I guess that explains why the category thing didn't render much in en.wikipedia Here is a mapping for a few remaining ones:

brown	Unknown hair color: шатенка	50	Q2367101
	Unknown hair color: русые	23	Q2367101
	Unknown hair color: коричневые	7	Q2367101
	Unknown hair color: шатен	2	Q2367101
blond	Unknown hair color: светлые	36	Q202466
	Unknown hair color: блонд	5	Q202466
black	Unknown hair color: чёрные	15	Q1922956
	Unknown hair color: черные	6	Q1922956

Maybe one of the "brown" ones should rather be chestnut (Q5094009). --- Jura 06:23, 20 May 2015 (UTC)

@Jura1: done using your mapping. Also two more templates were processed: ru:Шаблон:Порноактриса and ru:Шаблон:Порноактёр. Updated error list: User:Jura1/Hair color import. — Ivan A. Krestinin (talk) 21:22, 20 May 2015 (UTC)
Great. Thanks for your help. Seems I need to get started on natural hair colors (Q19955568), sample: Q479451. --- Jura 04:58, 21 May 2015 (UTC)

Null VIAFs

Was this deliberate? There are several others like that. Has any data been lost? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 11:31, 21 May 2015 (UTC)

Outdated VIAF data?

Hi Ivan, I was expecting the usual mass-substitutions of new VIAF redirects after 2015-05-15 (the date this month's dump 20150512 was ready for download). And now there en:Template talk:Authority control#VIAF different on Wikidata is the first case of a VIAF number in the english wikipedia which is more up to date than the one on Wikidata... Is there a central copy or database of the VIAF dumps which any bot can query or has KrBot a private one? Anyway, somewhere and somehow this month's fresh VIAF dump seems to have failed to become available to you. -- Gymel (talk) 16:51, 28 May 2015 (UTC)

Hello Gymel, thank you for the message. Bot stops working 14.05.2015 die specific issue. Bot uses 4 statically defined VIAF items for self test procedure. One of these items was changed and self test procedure did not allow bot to start main algorithm. Now all is fine and bot is processing 2136 Wikidata items. It will finish his work during ~1 hour. — Ivan A. Krestinin (talk) 20:49, 28 May 2015 (UTC)
Thanks a lot! -- Gymel (talk) 21:39, 28 May 2015 (UTC)
Not sure if you do VIAF imports as well. Now that VIAF redid their website, there are quite a lot of identifiers that that have Wikidata links, but where Wikidata has no VIAF identifier. Maybe these could be imported. --- Jura 09:14, 7 June 2015 (UTC)
Many information can be imported from VIAF, but I am not sure that VIAF and Wikidata licenses are compatible. — Ivan A. Krestinin (talk) 09:35, 7 June 2015 (UTC)
I was thinking of the VIAF identifier itself. I might be mistaken, but it seems that VIAFbot/Maximilianklein used to do that on some VIAF sponsored project. --- Jura 09:41, 7 June 2015 (UTC)
Then VIAF did a matching between english Wikipedia and VIAF, the results were subsequently loaded in en:WP. Since the data was provided for that purpose I think that was quite o.k. licence-wise. Now (cf. en:Wikipedia talk:Authority control/VIAF since April or so VIAF "follows" Wikidata on a regular schedule, i.e. Wikidata information is clustered and reclustered on a monthly base like other constituent files. Exact birth and death dates are important for that, to a certain extend also the VIAF number(s!) recorded in the Wikidata item. Last month's VIAF dump contained about 917.000 Wikidata matches, whereas Wikidata only contains 568.000 VIAF numbers (for 565.000 items). The quality of VIAFs matchings is usually very high, so I would have no objections against a one time import of excess VIAF numbers (maybe under certain restrictions like "persons only" or "persons with birth dates only" or "persons from VIAF clusters with at least n members". The license issues would of course have to be negotiated first). However there are weak spots, like ISNI entries of which IMHO many were created in a VIAF-like process by OCLC themselves but now are VIAF constituents of their own with a certain tendency to "pollute" VIAF clusters, the smaller the cluster the higher the impat of bad data. Now there are considerable delays on both sides (we react on VIAF changes based on their dumps, they possibly consider Wikidata dumps) and I fear that ingesting VIAF numbers on a regular base would create a loop of wrong associations being read back in from outdated data and it's uncertain whether human contributors will be fast enough to struggle against that.
The current situation is as follows: In Wikidata VIAF numbers are entered (semi-)automatically by contributors using the Authority Control gadget or other tools consulting VIAF. Or are imported from wikipedia instances, where they had been created in a comparable process. There are wrong numbers in articles (from Copy-and-Paste jobs), and if they show up in a constraint report they are going to be inspected by humans: I have spotted many non rivial merge canditates based on identical VIAF numbers in the respective entries. Thus VIAF is helpful for Wikidata operations, which I consider on a "per-item-basis under intellectual control" (importing all matches recorded by VIAF will increase the detection probability by constraint reports, but we must retain the ability to simply delete numbers from wikidata items for nonsensical matchings without fear that they'll be coming back ever and again). On the other end VIAF operates almost purely automatically but as I understand their statement cited above OCLC is willing to invest some intellectual effort in disambiguating wrongly clustered items on a case-by-case base: There is the point where Wikidata can help VIAF (my impression is that they have given up hope that the library authority files involved are able to clean up the mess by "sharpening" or disambiguation their records in a timely manner). IMHO this is a quite interesting interplay with great potential, but the roles should not be confused: VIAF is responsible for the "machine thing" and Wikidata for the intellectual side of matchings and we should make sure that we do not endanger this by using VIAF as a kind of "trusted upstream source". So both parties maintain their own matching and additionally have a close look at the other parties matchings but never accept them blindly. -- Gymel (talk) 11:24, 7 June 2015 (UTC)
Interesting, detailed note, thanks for that. Personally, I find it most helpful how my additions here eventually fix their data.
So there are two issues to tackle: (1) how to deal with current backlog (Only 2/3 of VIAF numbers for items are at Wikidata). (2) how to work going forward, given the various interactions between Wikipedia/Wikidata and VIAF. --- Jura 11:52, 7 June 2015 (UTC)
For either, we might want to take into account if a given match had already been removed from Wikidata (where do we find this?) and what is the age of a VIAF's match with WD. --- Jura 12:02, 7 June 2015 (UTC)
The monthly VIAF reclustering can be regarded as recalculating everything at the same time (so "what changed" might not be a question which can be easily answered for individual records), and we should not assume that VIAF follows Wikidatas log files (they might follow incremental dumps, at least for adding data). Also, there was the attempt to have a VIAF error report page at en:WP, but IMHO this did not have much impact. Establishing reporting workflows ("see, we have deleted that association, please consider the same") also would imply to set up a back channel ("ack, we proccessed everything up to X" or even individual "no, we still think this association is correct") and never was part of the current discussions. We mutually have access to "quite recent" complete dumps, wikidata has its reporting mechanisms and both sides have for individual, known records or numbers online access to live data. I imagine the two operations operating on a series of "snapshots" for maintenance (like KrBot substituting VIAF numbers after merges), but escalating to live access for "manual" operations. It seems that the mutual committment to use the data (on Wikidatas side: trying to keep constraint reports small by removing erroneos numbers and correcting genuine Wikidata errors by merging/splitting items, on VIAFs side: handling Wikidata data like that of any other "regular" constituent and additionally use the residue in our constraint reports as checklist for issues they might be able to fix manually) should suffice to keep this going - even without a mutual reporting workflow.
So, reappearing deleted values are a problem here even without VIAF mass imports: When coming from mass imports from some Wikipedia instance, they should be accompanied by a reference to the respective wikipedia, and things have to be fixed here and there. Problems arising from user errors here (things named to similar) or VIAF errors (notorious to have problems when persons with roughly the same name also were born and died the same year) can sometimes prevented from repetion by providing "novalue" - with that they at least show up almost immediately in a constraint report when some other value is added. Faulty VIAF clusters may nudge users here to perform wrong item merges, we have different from (P1889) and several "do not merge lists" for robots. Stuffing items with more information like exact to the day birth and death dates, places of birth and death, professions does help Wikidata and VIAF alike to keep different items apart.
A strategy I've been adopting during the last weeks when working through the constraint reports (P214, P227, P224, P245) was not only correcting the issues directly involving the authority control identifiers but also checking the birth and death properties whether one of the connected wikipedia articles would have some more specific information not yet in the item. Obviously this helps Wikidata but I think it's probably the single measure which is also most beneficary to VIAF. -- Gymel (talk) 14:01, 7 June 2015 (UTC)
Well, I don't think we should try to re-import automatically VIAF identifiers that had already been removed (manually) from Wikidata. Personally, I try to add VIAF to items I edit and sometimes find that VIAF already has a QID. Looking at the numbers you provided, this is fairly frequent, making it pointless to do it manually. --- Jura 14:52, 7 June 2015 (UTC)
Now bot is ready and I am working on improving data validation algorithm. Test edit: [39]. — Ivan A. Krestinin (talk) 04:49, 9 June 2015 (UTC)
Validation algorithm is ready too. Bot will not violate any of constraints from Property talk:P214 and have some additional checks. @Gymel, Jura1: need we discuss something else before start? — Ivan A. Krestinin (talk) 22:05, 9 June 2015 (UTC)
I had a look at the 1 test and it looked fine. Personally, I don't use VIAF for location though. If you also do some checks to avoid that we re-add things that had been removed from Wikidata, I'd go ahead. --- Jura 22:15, 9 June 2015 (UTC)
The most newly added codes are for geographical objects as I see. I do not made special check for removed codes, but I check for non-unique values. So moved IDs will not added to old element, but really removed codes will be re-added. History analysis needs additional ~350000 requests to WD. Is it really popular case? — Ivan A. Krestinin (talk) 22:27, 9 June 2015 (UTC)
I don't know. Usually, I just add them .. then VIAF merges them and you remove them ;)
Gymel might have a better view. Just in case someone check and removed them manually, we shouldn't re-add them. Maybe checking the history of the constraint report is sufficient? (avoid any that appeared in there) --- Jura 22:45, 9 June 2015 (UTC)
I'm busy enough operating on the items popping up in the "unique value" constraint reports. Flooding that section would knock me out. Reinserting numbers into that section I had already dealt with would be really bad. That means the bot should not add any VIAF number already mentioned in Wikidata. The "single value" violations mostly seem to be VIAF's business, the numbers in the report are too huge to inspect them here at the moment. So adding a VIAF number to an item which already has some (other) VIAF number might help spotting errors but we wouldn't have the capacity to process that. Therefore not adding VIAF numbers to any item which already has P214 will be o.k. (Both together should sum up to Ivan's statement "The bot will make sure that constraints are not going to be violated")
Leaving out not only VIAF numbers currently known to Wikidata but also those which had historically used here would be a very conservative approach. Appoximating that history to the history of the constraint reports page will make the bot blind for the interesting case where a VIAF number was removed (no violations involved, so perhaps it was simply wrong?) but I can't imagine that this has been the case very often. There were not many "unique value" violations where I had to remove the number from both items (I can't remember, there probably have been some cases) and VIAF clusters change over time: Maybe we had to throw out a VIAF number two years ago because it was not clear what it stood for but would decide differently this year. There remain the cases where a VIAF number was reported as "excess" for some item and subsequently removed and now is going to be added to another item - I would consider that to be "independent", the history gives us an indication that this VIAF element might be particularly problematic (since they import the complete content of library authority files VIAF of course has a certain amount of pure garbage they're trying to map anyway), but not more. So for me it would be fine when the bot operates without taking any historical data into account. -- Gymel (talk) 06:53, 10 June 2015 (UTC)
Ivan, Jura1
  • For recording the source I would recommend imported from Wikimedia project (P143) or even based on heuristic (P887) instead of stated in (P248).
  • The next VIAF reclustering round will probably happen within the next five days, data will be available shortly after - we probably should wait for that (just another idea: What about handling only those VIAF - Wikidata mappings which did demonstrate some stability by not having changed between the current dump and next week's?)
  • User:KasparBot is still importing VIAF numbers from the language versions to Wikidata (currently from ja:WP). I have no opinion wether that matters: If your Bot operates early, KasparBot will not import a differing number and not remove it from the Wikipedia where its recorded. If your Bot operates late it will not import a differing number. In the former case, a conflict resolution will be left to the Wikipedia in question, in the latter case it will go unnoticed.
  • I have not much experience with geographic items, typically a small village and some rural administrative unit surrounding it go by the same name and neither libraries nor Wikipedia articles perform the necessary distinction Wikidata or TGN would need. So I expect good matches from bad data to bad data. Why not... We'll have similar issues with Buildings vs. Corporate Bodies (neither libraries nor Wikidata has a good way to cope with the difference between Historical places and one or more museums located in these place), events (especially with series of them like biennals) and works (is this "the book" or "the movie"? Is this the Bach cantata or the aria lending its name to the cantata?). Let's just hope that VIAF does not match too aggressively...
  • We should probably solicit comments from a broader community and especially the VIAF representatives (@Ralphlevan, ThomasBHickey:. -- Gymel (talk) 06:53, 10 June 2015 (UTC)
  • Maybe we should concentrate on items with P31:Q5 for VIAF. It avoids adding a new set of issues we haven't dealt with before. --- Jura 10:15, 10 June 2015 (UTC)
  • I added history analysis to the code. It found two deleted codes per 6000 values. One of the codes was detected by "unique value" checker also: Q4092, another - Q34446. "Unique value" checker will block adding VIAF IDs present in another WD items. Reclustering procedure is not a big problem, it affects 1000-5000 elements only. My bot updates items after such procedure ~once per month. I can use imported from Wikimedia project (P143), but VIAF has strong expression "this VIAF ID is correspond to this WD item", bot`s code is accurate too, so stated in (P248) is better, as I see. So bot is ready, I am waiting for your response before start. — Ivan A. Krestinin (talk) 19:54, 10 June 2015 (UTC)
    • For the removed ones, maybe this can help. --- Jura 01:01, 11 June 2015 (UTC)
    • imported from Wikimedia project (P143) is the usual one for bulk imports but you are right, this is the first time we process third-party mappings which were specifically done with Wikidata as a target, so indeed they are stating exactly the statement we are going to import. But for me already the meaning of a VIAF cluster and of a Wikidata item are "based on heuristics" and certainly the process which assigned a Wikidata item to a certain VIAF cluster is purely heuristic. based on heuristic (P887) can both express that the mapping is based on heuristics and that VIAF performed or was responsible for it, not us.
    • Since it's a one time action the choice of the right point in time should matter. We have been massively cleaning up Template:P214 during the last weeks and the next VIAF reclustering will be their chance to catch up.
    • Like Wikidata items the entities in VIAF are clusters. Their meaning is determined by the members. Sometimes there are merges of complete clusters, sometimes only individual members are moved to different clusters and the original cluster may become empty and therefore meaningless (on VIAF it will be therefore deleted, Wikidata keeps them, since there are properties present which could give a meaning to the item). From the perspective of a single member (authority file entry, Wikidata item on VIAF's side, or a specific sitelink here) the corresponding cluster identifiers changes as the result of a merge or a move. We only can detect ids of deleted and merged clusters, so we will not be notified of changes in the case where a wikidata item will be individually taken out of an existing cluster and assigned to another one (not necessarily so: Unlike real authority files VIAF does not try to represent all of Wikidata, thus a given item might change its status from "mapped in VIAF" to "unrecognized by VIAF"). So the quality of the initial mapping does matter unless we implement additional reports to flag items which in VIAF are living in different clusters than recorded here. This could be dangerous since some users will be inclined to keep the report small by all means and therefore will "repair" the "wrong" VIAF id here. At that stage we simply would mechanically track VIAF mappings and therefore would (at least partially) loose our value of providing "intellectually" confirmed VIAF links. -- Gymel (talk) 06:41, 11 June 2015 (UTC)

At Wikidata:Database reports/Deaths at Wikipedia, I added links to list items that are likely to have VIAFs, but I seem to come across:

  1. items that don't have a VIAF
  2. items where VIAF.org already has QIDs, but WD doesn't have the property.

Maybe we could use a query like this and import corresponding VIAF. If it's preferred, I can make a request at Wikidata:Bot requests. --- Jura 06:45, 2 August 2015 (UTC)

Hello, I think we can simple import all missing VIAF codes. We waited for new dump during previous discussion. Now the dump is ready and I will start import today. @Gymel, Ralphlevan, ThomasBHickey: please notify me if something wrong. — Ivan A. Krestinin (talk) 07:01, 2 August 2015 (UTC)
Sure, please do. It's so annoying to find the QID is already at VIAF.org ;) --- Jura 07:10, 2 August 2015 (UTC)
Thanks for adding them! BTW, if you want to free up processing time for VIAF, maybe this can be discontinued. Otherwise, people will never start using the sitelink feature. --- Jura 12:39, 18 August 2015 (UTC)
Peoples do not use sitelinks because current model have big problems. — Ivan A. Krestinin (talk) 18:34, 18 August 2015 (UTC)
How about concentrating on VIAF till it has been sorted out? The report just perpetuates it. --- Jura 03:47, 19 August 2015 (UTC)

Bot errors

Your bot is making many errors. For example:

Please stop it; and fix them. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 22:20, 31 May 2015 (UTC)

Don't worry about it, I will take care of it. Someone redirected one item to the other by error. --- Jura 22:25, 31 May 2015 (UTC)

Unredlink your userpage ?

It's annoying that your page, for example with the {{Ping}} template, appears as a redlink, always wondering if I made a mistake ... could you please create it ? TomT0m (talk) 09:51, 7 June 2015 (UTC)

I dislike to spend time to personal page maintaining. Usually copy-paste is used to insert user name, so mistake is not probable. — Ivan A. Krestinin (talk) 10:02, 7 June 2015 (UTC)
You actually spend time making sure it stays deleted, what's the difference ? You just have to put "hey, this is Ivan page" and revert all changes not by you, that's all. Maybe you can ask an admin to semi protect it or something. I think it would make a lot of people more confortable, seeing a redlink in discussions always trigger a mistrust neuron. TomT0m (talk) 10:31, 7 June 2015 (UTC)
Why are you bothering him about this? There is no requirement for this. As there is no requirement for you to contribute before posting to talk/forum pages. --- Jura 10:45, 7 June 2015 (UTC)
@Jura1: I explained why, did'nt I ? TomT0m (talk) 11:01, 7 June 2015 (UTC)
It was deleted without my participation as I remember. Also blue link usually means something interesting, but I have no any interesting information for this page. Maybe user namespace must be deleted at all, there are another services for user pages hosting. — Ivan A. Krestinin (talk) 10:53, 7 June 2015 (UTC)
You would still need a way to link this services to your Mediawiki account :) Besides, I care for people I collaborate with. A little bit of human relationships does not hurt. TomT0m (talk) 11:04, 7 June 2015 (UTC)

Your bot added date of death (P570) : 3 February 1831 to this item, with the source imported from Wikimedia project (P143) Russian Wikipedia (Q206855). However, the article did not say that the person died on that date at the time. Where did the bot get this statement from? --Yair rand (talk) 21:29, 7 June 2015 (UTC)

Hello, see template "Персона", value "дата смерти = 3.2.1731". — Ivan A. Krestinin (talk) 21:54, 7 June 2015 (UTC)

Constraint:Inverse с двумя разрешёнными Property

Иван, привет! У меня такой вопрос. Можно ли настроить Constraint:Inverse так, чтобы он учитывал, что обратным может быть одно из двух свойств? Конкретно речь о Property:P1344, к которой обратным может быть как Property:P710 (участник-человек), так и Property:P1923 (участница-команда). Artem Korzhimanov (talk) 22:30, 8 June 2015 (UTC)

Добрый день, сейчас Constraint:Inverse поддерживает работу только с одним свойством. Если таких случаев будет много, то надо будет подумать о расширении функционала Constraint:Inverse. — Ivan A. Krestinin (talk) 04:36, 9 June 2015 (UTC)

P465

I see, you have removed P465 from some political parties, while you obviously let some other stay. I do not oppose it, I only added these statements as a part of an experiment here. Do you want me to remove the other statements too? -- Innocent bystander (talk) 12:04, 12 June 2015 (UTC)

Hello, these edits was based on the property domain from documentation. I think we must use color items for parties, not hex numbers. Also look like special property is needed, because relation between color and party is not obvious. What color is means? Color of flag, color of party emblem, color of paper with party memorandum, color of leader`s necktie or something else. Party is human group. What is color of human group? — Ivan A. Krestinin (talk) 13:01, 12 June 2015 (UTC)
I am not sure the colors you reverted exists anywhere but on Wikipedia. These are local political parties, and not all such political parties have choosen a color for themself. Q19978367 had some simlair ideas as the later "green" parties, therefor we have dressed their piecharts in green colors, but we do not know what color (if any) they used themself. If we always choose the same colors as the party choosed for themself, then we will have problems, when we try to describe the elections of the 1930 and 1920 Sweden. This, since it then used to be a bouquet of different communist-parties, all using the same version of red. -- Innocent bystander (talk) 13:23, 12 June 2015 (UTC)

GND non-values for non-persons

Hi Ivan, does this edit mean we don't need any more to fix manually this kind of erroneous GND ID (P227) values? That would be great (in the past I have contacted VIAF and DNB as "stakeholder" several times concerning this issue, unfortunately to no avail). -- Gymel (talk) 13:33, 18 June 2015 (UTC)

Hello, yes, bot fixes typical errors from "Format" violations section for this and some other properties. This error kind is covered by bot now.— Ivan A. Krestinin (talk) 20:24, 18 June 2015 (UTC)
This is really great news. Thanks a lot! -- Gymel (talk) 20:40, 18 June 2015 (UTC)

Unnecessary drama

Your bot recently tagged a number of factual BBC radio programmes as instances of radio dramas (e.g.). I've reverted those which were on my watchlist, but please check the rest. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 11:53, 19 June 2015 (UTC)

Hello, @Pigsonthewing, Andreasmperu: the edits was caused by this merge. Please revert the merge if it is incorrect. I will revert bot`s edits after it. — Ivan A. Krestinin (talk) 19:03, 19 June 2015 (UTC)
Thank you for spotting that. It is incorrect - we need to be able to distinguish between drama and non-drama series. I'm not sure how best to revert or resolve this. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 13:13, 20 June 2015 (UTC)
You can just revert Andreasmperu`s merge edits. — Ivan A. Krestinin (talk) 16:57, 20 June 2015 (UTC)

Шаблён:Населены пункт/Беларусь

А Вы можете ботом вытянуть из карточки be-x-old:Шаблён:Населены пункт/Беларусь координаты, численность населения (+год переписи) и сельсовет (для located in the administrative territorial entity (P131))? --Чаховіч Уладзіслаў (talk) 20:20, 22 June 2015 (UTC)

  • Вроде бы особых трудностей не вижу, могу попробовать, очень быстро не обещаю, но за неделю могу успеть сделать. — Ivan A. Krestinin (talk) 20:26, 22 June 2015 (UTC)
Было бы чудесно :) --Чаховіч Уладзіслаў (talk) 21:02, 22 June 2015 (UTC)
Кстати можете помочь, если составите табличку соответствий свойства "Статус" элементам Викиданным. Первая строчка такая: "горад — city (Q515)". — Ivan A. Krestinin (talk) 21:20, 22 June 2015 (UTC)
горад — city (Q515)
места — city (Q515)
гарадзкі пасёлак — list of urban-type settlements in Belarus (Q2599457)
мястэчка — list of urban-type settlements in Belarus (Q2599457)
пасёлак — posyolok (Q2514025)
аграгарадок — agrotown (Q1989945)
аграмястэчка — agrotown (Q1989945)
вёска — hamlet (Q5084)
хутар — khutor (Q2023000)--Чаховіч Уладзіслаў (talk) 09:03, 23 June 2015 (UTC)
Спасибо! — Ivan A. Krestinin (talk) 20:31, 23 June 2015 (UTC)
  • Бот закончил импорт из be_x_oldwiki, сейчас запустил ещё тоже самое для ruwiki и bewiki, может завтра-послезавтра бот закончит и с ними. — Ivan A. Krestinin (talk) 21:29, 27 June 2015 (UTC)
  • ruwiki и bewiki также импортированы. В bewiki кстати недавно была большая ботозаливка, для большинства статей из неё удалось извлечь только тип населённого пункта и русское название. Пример такого элемента: Sukremna (Q20537007). — Ivan A. Krestinin (talk) 18:59, 30 June 2015 (UTC)
Благодарю! А могли бы Вы еще в завершение вытянуть из белорусских сельсоветов в рувики (они там вроде все-все) вышестоящую адм. единицу (район) в Викиданные?--Чаховіч Уладзіслаў (talk) 20:25, 30 June 2015 (UTC)
Готово. — Ivan A. Krestinin (talk) 05:06, 1 July 2015 (UTC)
Чудесно, спасибо! --Чаховіч Уладзіслаў (talk) 17:03, 8 July 2015 (UTC)

Update "to merge"

Hi Ivan, no update of the "to merge" project today? How should I spend my Tuesday (Q127) now? :-) Could you perhaps teach me how KrBot creates these pages so I run and tweak my own updates? - FakirNL (talk) 09:50, 23 June 2015 (UTC)

Hello, bot updates the pages every time then new dump appears on http://dumps.wikimedia.org/other/wikidata/ The latest update were made today. Please note, bot does not touch page if no changes detected. For example see https://www.wikidata.org/w/index.php?title=User:Ivan_A._Krestinin/To_merge/dawiki&action=historyIvan A. Krestinin (talk) 18:52, 23 June 2015 (UTC)

No value

Should "no value"/"unknown value" necessarily be reported as a constraint failure? See e.g. Wikidata:Database reports/Constraint violations/P1157#"Format" violations.

Gut instinct says no, probably not. --Izno (talk) 13:52, 23 June 2015 (UTC)

For "no value" I'd say "probably yes": Besides Roosevelt there were some billion more people on earth who never were congressmen and therefore don't have a bio there. If maintenance for a property (like VIAF ID (P214)) makes active use of "no value" then it can be formulated as "empty pattern" into the constraint. For "unknown value" I have no experience: Probably there is some evidence that there should be a value, thus listing it somewhere (why not in a constraints report) indicates that more work is to be done. -- Gymel (talk) 14:35, 23 June 2015 (UTC)
Do we believe that we should build that into the constraint pattern or rather our assumption (or any consumer's assumption) of the constraint pattern? I'm probably a yes also after those "duh" comments. Unknown value--the most obvious use case is birthdays. Not sure of the best way to turn such a default off, but it seems to me that you would add the item to the list of exceptions for that particular constraint at that point. --Izno (talk) 14:41, 23 June 2015 (UTC)
@Jura1, UV:. --Izno (talk) 14:43, 23 June 2015 (UTC)
For the Roosevelt problem: Wouldn't it be easiest to remove the US Congress Bio ID (P1157) statement entirely from the Roosevelt item for now, and re-add it once he gets a bio there (if ever)? --UV (talk) 18:30, 23 June 2015 (UTC)
Technical part of the question: "no value" and "unknown value" are represented by bot as empty strings. Empty string can be accepted or rejected by regexp. For example Property talk:P214 regexp accepts novalue and somevalue. Non-technical part of the question: these values allows to model some difficult cases, but usually its are broke data uniformity. These values require special processing for data clients. This make data more hard to use. So it is good practice to avoid its usage where it is possible. — Ivan A. Krestinin (talk) 20:30, 23 June 2015 (UTC)

Bug in the last two updates. 85.212.22.136 22:45, 23 June 2015 (UTC)

Thank you for the notification, the bug is fixed, please wait the next update. — Ivan A. Krestinin (talk) 22:04, 24 June 2015 (UTC)

Bot edits which don't seem to be changing anything

Several of your bot's changes have appeared in my watchlist, e.g. [40] - what's going on there? The two values look the same to me. - Nikki (talk) 13:43, 25 June 2015 (UTC)

The bot removed the underscores--Pasleim (talk) 18:31, 25 June 2015 (UTC)
Thank you, you are right, bot normalizes file name. — Ivan A. Krestinin (talk) 19:39, 25 June 2015 (UTC)


Disambiguation item

Hi Ivan,

Would you avoid adding additional values in P31 by bot to disambiguation items (sample). We are trying to clean them up, so this just throws us back. --- Jura 05:12, 26 June 2015 (UTC)

Коды профилей на теннисных сайтах

Я уже говорил коллеге, что эти значения используется статьями в рувики; если ты удаляешь их в статьях здесь, то неплохо бы тут же сходить в рувики и заново прописать их там.--Дос Сантуш (talk) 11:33, 27 June 2015 (UTC)

Пока у нас обычно лишь копируют значения, а не перемещают, потому не глянул. Сейчас тогда пройдусь по рувики, восстановлю, заодно пропишу ссылку на правильный элемент на будущее. — Ivan A. Krestinin (talk) 12:05, 27 June 2015 (UTC)
Прошёлся по всем элементам, вы везде уже восстановили, спасибо. В сами статьи прописывать ссылки не стал, т. к. там оказывается не карточка, а несколько шаблонов, вместо этого добавил везде of (P642). В рувики должны через некоторое время включить доступ к произвольным элементам, тогда можно будет воспользоваться этим свойством для извлечения идентификаторов. — Ivan A. Krestinin (talk) 12:32, 27 June 2015 (UTC)

Suggestion

When doing {{Constraint:Single value}} and {{Constraint:Unique value}} check, statements with deprecated rank should be discarded. See Q153048#P569 and User_talk:Pasleim#Birth-_and_death_date_errors.--GZWDer (talk) 15:50, 27 June 2015 (UTC)

Wouldn't it be better to mark deprecation and analyze qualifiers: if all qualifiers have different values it still might be an issue but a minor one, c.f. Q2851070#P244? -- Gymel (talk) 17:04, 27 June 2015 (UTC)
It is interesting case. It has some other problems: how to specify Russian label for "Gudule" pseudonym? How to specify "Gudule" as author (P50) for Au Gringo's bar (Q2870261)? Separate item for pseudonym can solve these 3 problems, but this way has another issues... — Ivan A. Krestinin (talk) 21:21, 27 June 2015 (UTC)
Current bot implementation ignores ranks at all. This feature presents in my TODO list. But this case is looked more special. Need Wikidata collect all misprints in all sources? Such misprints are fixed often. So this can create many unreferenced claims (or referenced by archive.org only). I think we need to collect notable errors only. So my suggestion is simple remove invalid values from Q153048#P569 and add this item to bot`s ignore list. — Ivan A. Krestinin (talk) 21:21, 27 June 2015 (UTC)

Double given names

Is there any way you can avoid deleting this? I know: it's a very unique situation, but I'm just following the instructions on the property talk page. Sjoerd de Bruin (talk) 10:19, 30 June 2015 (UTC)

Hello, the world is greater than we think :-). I added this item to ignore list. Thank you. — Ivan A. Krestinin (talk) 18:54, 30 June 2015 (UTC)

John Flett Q20630482

Hi, probably some obvious explanation, but can you let me know the reason why the image was removed from this item? Thanks. Eagleash (talk) 09:28, 2 July 2015 (UTC)

Hello, it was removed as link to non-existent file. Often this happens then the file is located in Wikipedia, not in Commons. — Ivan A. Krestinin (talk) 20:05, 2 July 2015 (UTC)
Thanks; had to be something like that. Is there any way of picking this up before inputting. (I imagine all the underlines between bits of the link is a big clue?) Regards, Eagleash (talk) 22:25, 2 July 2015 (UTC)
Eagleash No, "non-free" is a BIG clue... Simply note that https://en.wikipedia.org/wiki/File:John_Flett_fashion_designer.jpg is a file local to English Wikipedia, as non-free images do not qualify to be uploaded to https://commons.wikimedia.org/ aka Commons. And only the names of images on Commons can be used for image (P18), cf. the usage notes on Property talk:P18. -- Gymel (talk) 22:36, 2 July 2015 (UTC)
Gymel So en.wikipedia etc. would indicate that an image is not on commons? You'll have to forgive me here, I've only ever used images which are own-work uploaded via commons, as the licensing criteria are somewhat baffling... :P Then again, the mark-up on Flett's wiki page just reads "File:John_Flett_fashion_designer.jpg" (on the face of it no different to normal) so I am again wondering how to spot similar problems in future. Eagleash (talk) 23:04, 2 July 2015 (UTC)
Compare with File:Nicholas Winton in Prague.jpg (from today's featured article on the English Wikipedia): The logo in the upper left corner tells you quite clearly who is serving the image. -- Gymel (talk) 05:53, 3 July 2015 (UTC)

──────────────────────────────────────────────────────────────────────────────────────────────────── Yes Wiki Commons file pages have the logo in the top left hand corner. However when on a wikipedia page there is no file name to click on, but by clicking on the image on a page it will open in a new screen with a button at the bottom right "more details" with (or without) a plain version of the commons logo; so one can tell that way where it is hosted. Eagleash (talk) 07:14, 3 July 2015 (UTC)

Hi,

I think your bot made a mistake here.

Regards, Thibaut120094 (talk) 15:19, 3 July 2015 (UTC)

Have a look at the item back in April: https://www.wikidata.org/w/index.php?title=Q4199132&oldid=213129198 --- Jura 15:51, 3 July 2015 (UTC)
Hello, I reverted my bot`s edits. But I am not sure that all is correct because two different pronunciation present in ruwiki: "Икеда" and "Икэда". — Ivan A. Krestinin (talk) 20:25, 3 July 2015 (UTC)

Constraint checking about proposed property

Hi, I've got a few properties in the pipe,

   Under discussion
DescriptionMISSING
Data typeMISSING
Example 1MISSING
Example 2MISSING
Example 3MISSING

that could require a check.

  1. For union of, there would be a potential constraint break if, for example, an item i is an instance of a class A, with
    ⟨ A ⟩ union of (P2737) View with SQID ⟨ B ⟩
    together with (P1706) View with SQID ⟨ C ⟩
    , but there is neither
    ⟨ i ⟩ instance of (P31) View with SQID ⟨ B ⟩
    nor
    ⟨ i ⟩ instance of (P31) View with SQID ⟨ B ⟩
    .
  2. For disjoint union of, there would be a potential constraint break if the condition in 1 are not met, plus i can't be both
    ⟨ i ⟩ instance of (P31) View with SQID ⟨ B ⟩
    and
    ⟨ i ⟩ instance of (P31) View with SQID ⟨ C ⟩
    .

Do you think it's manageable with your current codebase ? TomT0m (talk) 18:29, 6 July 2015 (UTC)

Hello, this check has another structure than existing. It is class-based, not property-based. Also existing constraints are described using templates, not using properties. So additional code is needed. I have a question about the suggested constraint: construction
⟨ A ⟩ union of (P2737) View with SQID ⟨ B ⟩
together with (P1706) View with SQID ⟨ C ⟩
is equivalent to
⟨ A ⟩ union of (P2737) View with SQID ⟨ C ⟩
together with (P1706) View with SQID ⟨ B ⟩
as I understand. Why B and C are described using different properties? — Ivan A. Krestinin (talk) 20:34, 6 July 2015 (UTC)
It's because the statement is on a list of classes, A can be the union of C, D, E, ... I use qualifiers because they can safely have a close semantics, ie. we can assume that the list is complete, which is not usually the case with several statements (per discussion with Markus on the mailing list, I can provide references if you want). This is also useful if we use different classification scheme, for example if A is both the disjoint union of A, B on one hand and the disjoint union of C, D on the other. See the example on the property proposal. But yes, the order is not important. But the property has a meaning and express something even if it is not checked. TomT0m (talk) 07:59, 7 July 2015 (UTC)
But yes, you're maybe right,
⟨ A ⟩ union of (P2737) View with SQID ⟨ B ⟩
union of (P2737) View with SQID ⟨ C ⟩
union of (P2737) View with SQID ⟨ D ⟩
would fit, I don't know. TomT0m (talk) 08:11, 7 July 2015 (UTC)


Family relation constraints

Hi Ivan,

Currently, the following remain unchecked:

  1. child (P40) on an item without father (P22) or mother (P25) on the value
  2. P7 (P7) or P9 (P9), without the reverse.

Inverse seems to require only one property. Could these be added? To add the result, sex or gender (P21) would be needed on the report.

The first one is obviously more important than the other. --- Jura 17:12, 7 July 2015 (UTC)

Hello Jura, the first constraint will appear in the report tomorrow. — Ivan A. Krestinin (talk) 04:43, 8 July 2015 (UTC)
Thanks. That was quick! --- Jura 15:57, 8 July 2015 (UTC)
I went through them. Looks fine. Even lists those that have several mothers and fathers. --- Jura 16:05, 9 July 2015 (UTC)
✓ Done for P7/P9 too. — Ivan A. Krestinin (talk) 18:58, 10 July 2015 (UTC)

Constraint report type (by property, by QID, both)

For some constraints, it might not be useful to output them by property and output by Qid is sufficient (sample: P569 on Property_talk:P27).

For others, the output by Qid isn't helpful (single constraint on P569).

Shall we add an option to the constraint template hat de-activates the list by property? --- Jura 15:57, 8 July 2015 (UTC)

I do not understand the idea, sorry. group property parameter can be useful maybe, see Template:Constraint/doc#Parameters. — Ivan A. Krestinin (talk) 19:31, 8 July 2015 (UTC)
What I meant to say is that the violations on Wikidata:Database_reports/Constraint_violations/P27#.22Item_date_of_birth_.28P569.29.22_violations aren't that useful to be listed, but at Q3996 (the first item on that list) the two violations at Special:ConstraintReport/Q3996 can help. --- Jura 16:09, 9 July 2015 (UTC)
Are you mean reordering Wikidata:Database_reports/Constraint_violations/P27 from per-constraint list to per-item list? Something like:
Q3996 violates:
Q19856472 violates:
...
Or want you merge all property reports and generate one long per-item report? Also this report can be useful maybe. — Ivan A. Krestinin (talk) 20:54, 9 July 2015 (UTC)

"properties for this type" constraint violation reporst ?

Hi, I was wondering if you would be able to create reports similar to what you do for properties using properties for this type (P1963). For instance, in geographic location (Q2221906), there should be a list of items with coordinate location (P625). If it takes into account all items that are instances of a subclass of geographic location (Q2221906) it may be very long. I suppose the ideal solution would be a separate report. For instance, núcleo de población (Q11939023), so there should be a constraint violation report for núcleo de población (Q11939023), and it should list items that do not have P625. --Zolo (talk) 06:14, 9 July 2015 (UTC)

Hello, it is interesting idea. I added this to my TODO list, but my code is optimized for per-property processing, so implementation will take a time. Also our classes tree is too unstable and contains many errors like loops. This will create additional troubles. — Ivan A. Krestinin (talk) 21:08, 9 July 2015 (UTC)

Hello Ivan,

thanks for updating the property violations on Database of Classical Scholars ID (P1935). I noticed your bot startet to make exceptions for double entries of DBCS identifiers. Could you please not do this? I need these double entries to show up on the violations page in order to keep track of them and notify the DBCS administrators.

Thanks! Jonathan Groß (talk) 13:33, 9 July 2015 (UTC)

Hello Jonathan, as I see you found the way to resolve this. — Ivan A. Krestinin (talk) 21:12, 9 July 2015 (UTC)
Yes, sorry for bothering you. I'm still learning. Jonathan Groß (talk) 11:35, 14 July 2015 (UTC)

Revert changing links from Q18094344 to Q323415

Wrong merge, but your bot started changing the links already. Can you revert? Sjoerd de Bruin (talk) 20:53, 25 July 2015 (UTC)

Hello, ✓ Done. — Ivan A. Krestinin (talk) 10:37, 26 July 2015 (UTC)

revert removing the DOID prefix from the claim Disease ontology ID (P699)

The KRBot is removing DOID prefixes from Disease Ontology Identifiers added by the ProteinBoxBot. By removing the prefix from these identifiers, the links to Disease Ontology no longer resolves. For example:

in Q2713959 [41] does not resolve, whereas [42] does. Could you please revert all previous prefix removals from Disease Ontology ID claims? Andrawaag (talk) 17:56, 27 July 2015 (UTC)

Hello, I fixed link, all must be fine now. KrBot uses property documentation for values normalization. This property documentation has pattern without DOID: prefix. Also I found discussion about this prefix: Property talk:P699#DOID prefix, but it does not contain arguments pro or contra. Please provide some arguments in this discussion if needed. — Ivan A. Krestinin (talk) 19:06, 27 July 2015 (UTC)
I have added some arguments on Property talk:P699#DOID prefix Andrawaag (talk) 20:49, 27 July 2015 (UTC)
Ok, good, lets wait 3 days for other users comments. — Ivan A. Krestinin (talk) 20:56, 27 July 2015 (UTC)

Find a Grave removal

Hello Ivan, why did your bot remove the Find a Grave number from Clarence Powers Bill (Q18745815)? It is the correct number. Jonathan Groß (talk) 10:58, 28 July 2015 (UTC)

The code is specified twice in this item without any qualifiers. — Ivan A. Krestinin (talk) 20:02, 28 July 2015 (UTC)
Oops, I didn't notice. Thank you! Jonathan Groß (talk) 20:17, 28 July 2015 (UTC)

Thank you

Hoi, I noticed that you add VIAF identifiers to items I added manually.. THANK YOU !! GerardM (talk) 07:42, 4 August 2015 (UTC)

Template:Constraint:Source

It looks like the constraint reports don't support Template:Constraint:Source, e.g. Wikidata:Database reports/Constraint violations/P854 only has a section for the format constraint and Wikidata:Database reports/Constraint violations/P143 claims to not have any constraints at all. Could that be fixed? - Nikki (talk) 09:30, 4 August 2015 (UTC)

Hello, I added this constraint support with one limitation: bot will not check qualifiers. I added this limitation removing to my TODO list, but it is not simple task. — Ivan A. Krestinin (talk) 19:55, 4 August 2015 (UTC) P. S. {{Constraint:Qualifiers}} can be used for detecting inappropriate usage of P854 as qualifier.
Thanks, it seems to be working. :) - Nikki (talk) 10:16, 13 August 2015 (UTC)
@Nikki: qualifiers support was added too. — Ivan A. Krestinin (talk) 19:19, 16 August 2015 (UTC)

Bot adding random VIAF

Moved to: Wikidata talk:WikiProject Authority control#Locations

Category:Commons removed

Hello,

Could you tell me why your bot removed this commons:category [43]? Is it a mistake? --El Caro (talk) 16:29, 13 August 2015 (UTC)

The category needs a capital L: Category:Dolmen_de_La_Borda --Floscher (talk) 18:14, 13 August 2015 (UTC)
OK, thank you! --El Caro (talk) 05:40, 14 August 2015 (UTC)

selfref / p301

Is a problem here: https://www.wikidata.org/w/index.php?title=Q2601433&type=revision&diff=242163101&oldid=241623284 --Steenth (talk) 08:34, 16 August 2015 (UTC)

Was caused by this: https://www.wikidata.org/w/index.php?title=Q7482744&action=history Almost looks like vandalism this way, really damaging our data. Sjoerd de Bruin (talk) 08:38, 16 August 2015 (UTC)
Yes, the bot`s edit is caused by invalid merge. Only two items were touched in this case. Its were reverted manually. I have code for many items revert. So invalid merge revert is not a problem. But current case highlights another problem: the merge is looked like unsuccessful attempt to avoid old Wikidata-Commons integration problem. — Ivan A. Krestinin (talk) 09:06, 16 August 2015 (UTC)

P1959

Could Wikidata:Database reports/Constraint violations/P1959 be updated again? It shows things I've already fixed yesterday. Sjoerd de Bruin (talk) 08:44, 23 August 2015 (UTC)

✓ Done, dumps have 12 hour time lag unfortunately. Small reports can be updated using direct data access, but this way is not usable for regular runs. — Ivan A. Krestinin (talk) 09:35, 23 August 2015 (UTC)
Ah, is there a task on Phabricator for that? Sjoerd de Bruin (talk) 09:41, 23 August 2015 (UTC)
I do not know such task. This behavior is described at http://dumps.wikimedia.org/other/incr/ and looked as expected. I started discussion Wikidata:Contact the development team#Wikidata incremental dumps time lag. — Ivan A. Krestinin (talk) 10:25, 23 August 2015 (UTC)

Daily report

Hi Ivan,

Good know you can still edit! BTW the 20,000 error from P2036 might have broken the daily report. Special:Diff/245790938 should fix it. --- Jura 12:17, 27 August 2015 (UTC)

Hello Jura, thank you for the fix. I updated the report manually. Tomorrow it will be updated in regular mode. — Ivan A. Krestinin (talk) 19:50, 27 August 2015 (UTC)

Regex validator

Hi Ivan, will your bot handle this regular expression if I add it as format validator to format as a regular expression (P1793)?

((?:(?:[^?+*{}()[\]\\|]+|\\.|\[(?:\^?\\.|\^[^\\]|[^\\^])(?:[^\]\\]+|\\.)*\]|\((?:\?[:=!]|\?<[=!]|\?>)?(?1)??\)|\(\?(?:R|[+-]?\d+)\))(?:(?:[?+*]|\{\d+(?:,\d*)?\})[?+]?)?|\|)*)

Source: [44] Matěj Suchánek (talk) 09:54, 30 August 2015 (UTC)

My bot uses PCRE regular-expression library. It is mentioned as supported this regexp on stackoverflow. So lets try it. — Ivan A. Krestinin (talk) 12:40, 30 August 2015 (UTC)

Value type for 'property' datatype

Sorry for bothering again. Could you please find some time to prevent this (ie. be able to check {{Constraint:Value type}} for properties as well, as it's possible with {{Constraint:Type}}). Matěj Suchánek (talk) 12:32, 30 August 2015 (UTC)

✓ DoneIvan A. Krestinin (talk) 13:22, 30 August 2015 (UTC)
Thank you very much! Are you able to update the affected reports? If not, just let it be for tomorrow. Matěj Suchánek (talk) 13:44, 30 August 2015 (UTC)
✓ DoneIvan A. Krestinin (talk) 14:22, 30 August 2015 (UTC)

Удаление ссылок

Здравствуйте. А почему у элементов Q16320304 (ТЭФИ-2014) и Q20127450 (ТЭФИ (премия, 2015)) были удалены следующие заявления: ссылки на официальные сайт и аккаунты в соц. сетях конкурсов (диффы — [45], [46])? KAV777 (talk) 21:27, 30 August 2015 (UTC)

Добрый день, удалил, т. к. эти аккаунты и сайты относятся не к TEFI 2014 (Q16320304), а к TEFI (Q1961504). Насколько я понимаю, у TEFI 2014 (Q16320304) нет собственного аккаунта или сайта. — Ivan A. Krestinin (talk) 21:33, 30 August 2015 (UTC)

IMDb and person groups

Hello, sometimes page contains multiple {{imdb name}} templates. Bot has troubles with this case: [47]. — Ivan A. Krestinin (talk) 19:46, 30 August 2015 (UTC)

Hello, thank you for the warning. I paused the bot. I will add more checkings if I restart it (control of sex or gender (P21), instance of (P31) and/or value on en.wp. --Sisyph 21:41, 30 August 2015 (UTC)

Constraint on properties for items with "novalue"

If an item has a property and the custom value "novalue", I wonder if we should apply the constraints that come with that property or if we should skip them.

For a constraint on position held (P39), I added applies to jurisdiction (P1001) to a series of items. For some of these, the property would just get "novalue" (sample: abbot (Q103163)). For abbot (Q103163), I don't think the constraints for applies to jurisdiction (P1001) should be checked. Currently there are not any on Property talk:P1001, but we might want to add some. However these are unlikely to apply to novalue items.

Shall we skip constraint checks for items with novalue? --- Jura 10:01, 10 September 2015 (UTC)

  • Hi Jura, I do not fully understand your case. I will try to show one idea using another example: {{Constraint:Type|class=Q5}} and two hypothetical statements:
  1. Mikhail Lermontov (Q46599) <P7 (P7)> novalue
  2. Lena River (Q46841) <P7 (P7)> novalue
Statement 1 is looked fine from point of view the constraint, but statement 2 is looked incorrect. So bot can not simple ignore items with novalue. — Ivan A. Krestinin (talk) 20:28, 11 September 2015 (UTC)
Good point. One needs to account for people using them on anything ;)
P26 might be a more frequent sample for "novalue".
I suppose I will have to try with indirect constraints. --- Jura 17:52, 12 September 2015 (UTC)


Duplicates (same birth/death dates)

Jheald is currently experimenting with WQS to find duplicated persons by date of birth or date of death (two items with claim[31:5] and A:P569=B:P569 and A:P570=B:P570). It seems that it keeps timing out unless some other criterion is added. Samples results are: Olavi Merinen (Q11885292)/Olavi Merinen (Q17382548) (same person, but can't be merged as there are two articles). False positives would be: Constantin de Faucher (Q18707766)/César de Faucher (Q18707768) (two persons) or Cyril Marcus (Q17149153)/Stewart Marcus (Q17149152) (two persons, but date precision = 9 for P570).

If the items can't be merged (or dates corrected), we should probably link them with said to be the same as (P460) or different from (P1889) to exclude them from future reports.

I wonder if this could be included in the daily constraint reports or a special one to be added to Wikidata:Database reports/identical birth and death dates. --- Jura 07:15, 13 September 2015 (UTC)

If the two pages are definitely about the same subject/person like with Olavi Merinen/Olavi Meriläinen above, don't be afraid to add the local version of Template:Merge (Q6919004) to the pages like I just did with this example. There is no reason the Finnish Wikipedia should have duplicate entries. - FakirNL (talk) 09:32, 13 September 2015 (UTC)
Sure, but please don't merge it, otherwise the test case is gone ;) --- Jura 09:33, 13 September 2015 (UTC)
I created the report in test mode, please see Wikidata:Database reports/identical birth and death dates. I will update and recheck it latter. — Ivan A. Krestinin (talk) 10:36, 13 September 2015 (UTC)
Excellent. I tried some formatting, but there are too many items (selection).
What shall I use for false positives? --- Jura 12:54, 13 September 2015 (UTC)
Please use Wikidata:Database reports/identical birth and death dates/False positives for false positives. — Ivan A. Krestinin (talk) 19:11, 14 September 2015 (UTC)
Seems to work. Thanks. Could the bot update the selection with 150 items from "Can be merged" or "Sitelink conflicts" as well? Maybe TheGame could be fed from the lists as well. --- Jura 11:20, 15 September 2015 (UTC)
Selection page now contains top 150 items from main list. I know a little about TheGame, but can prepare the data in some format if needed. Also please note, TheGamers did many inaccurate edits in the past. — Ivan A. Krestinin (talk) 20:19, 15 September 2015 (UTC)
Thanks. The problem with TheGame might be that the standard selection is on titles, and people might merge them merely on seeing that the dates are identical. As there are a few oddities in the list, maybe this isn't a good idea. --- Jura 20:33, 15 September 2015 (UTC)
It seems the update stopped working. --- Jura 04:47, 17 September 2015 (UTC)
You are right. Double redirects are not expected for bot. I resolved one and bot works now. I need improve bot`s code to make this situation expected. — Ivan A. Krestinin (talk) 05:42, 17 September 2015 (UTC)
That's what it was? Possible, as there were a couple of groups of 3 items. I suppose after the initial clean-up (if I ever get there ;) ), this might be less likely. Thanks for your help. --- Jura 10:11, 17 September 2015 (UTC)
It stopped again, I think there was another triplet. I'm getting closer to the end of the initial clean-up. --- Jura 09:43, 18 September 2015 (UTC)
Finally, I got to the end of the first section. Works quite well. Thanks for this. --- Jura 10:51, 19 September 2015 (UTC)
Great job! I added "Sitelink conflicts" section to selection. — Ivan A. Krestinin (talk) 12:27, 19 September 2015 (UTC)
Thanks, these are harder to do, as many need to be tagged for merging at Wikipedia. The quicker ones are those with a redirect that can be deleted (sample Q5499184) and items merged afterwards. --- Jura 12:48, 19 September 2015 (UTC)
The last column at Wikidata:Database reports/identical birth and death dates/1 now includes different from (P1889). Is it possible that these don't get detected if there is also a sitelink conflict? It seems that earlier adding different from (P1889) made them disappear. BTW, I made a proposal for an change to the merge script at MediaWiki_talk:Gadget-Merge.js#Feature_suggestion:_advanced_mode_with_merge_tagging_at_Wikipedia.3F. --- Jura 10:52, 21 September 2015 (UTC)
✓ DoneIvan A. Krestinin (talk) 20:04, 21 September 2015 (UTC)

──────────────────────────────────────────────────────────────────────────────────────────────────── Thanks. I'm done with the list as well (maybe 10 are left) .. we just need to wait for the articles to be merged and deal with new items (yesterday there were about 5-10).

There is another problem I came across:

  • Initially I had relied on the bot to detect Template:Merge and move items to "Sitelink conflicts, marked to merge".
  • Once the merge was done, the item would go back to
    • "sitelink conflict" (if a redirect remained)
    • or "can be merged" (if the wiki did a "history merge", deleting one item).

However, it appears that some wikis (eswiki) remove the merge template several days prior to the actual merge. Also, in a few wikis, the templates didn't get detected.

To avoid that we loose track of these, I added said to be the same as (P460) to all items that were "marked to merge". Now we can't detect anymore when they did the actually the merge ;). To avoid this, the section "different from (P1889), said to be the same as (P460)" would ideally specify if an item is "can be merged", "sitelink conflict" or "sitelink conflict, marked to merge" plus said to be the same as (P460) and/or different from (P1889). --- Jura 06:04, 24 September 2015 (UTC)

Somehow the following two went undetected Q6150075 and Q6150269. I suppose I shouldn't have fixed them to check. --- Jura 07:31, 28 September 2015 (UTC)
Hello, Jura, I am sorry for delayed response. The report was renewed, is it OK now? Items Q6150075 and Q6150269 had different birth month. — Ivan A. Krestinin (talk) 12:19, 28 September 2015 (UTC)
No hurry, thanks for improving it. I think we probably already went beyond perfect. Completed mergers like Q19975965 are now visible. As it seems to work, eventually we might want to apply the same approach to other selections. --- Jura 12:48, 28 September 2015 (UTC)

Массовые переносы

Здравствуйте, может Вы знаете: если в Википедии есть определенное количество статей с неправильными названиями, но есть таблица, где друг напротив друга приведены эти неправильные и правильные названия, можно ли техническими средствами (ботом) автоматизированно перенести статьи под правильные названия? --Чаховіч Уладзіслаў (talk) 16:25, 15 September 2015 (UTC)

Добрый день, задача массового переименования статей по списку выглядит вполне решаемой для бота, по-моему кто-то в рувики такое уже делал, возможно с помощью чего-нибудь стандартного вроде AWB или pywikipedia. Думаю стоит спросить на ru:ВП:РДБ, если там не откликнутся, то могу этим заняться, если возьмёте на себя необходимые обсуждения. — Ivan A. Krestinin (talk) 20:25, 15 September 2015 (UTC)
Пойду спрошу --Чаховіч Уладзіслаў (talk) 07:47, 16 September 2015 (UTC)

Why did KrBot remove the geneAtlas image (P692) claim on the RREB1 gene item ?

https://www.wikidata.org/w/index.php?title=Q18031325&oldid=252177993 --I9606 (talk) 19:19, 17 September 2015 (UTC)

Hello, it was duplicate value. — Ivan A. Krestinin (talk) 19:22, 17 September 2015 (UTC)
Okay. Weird edit history. anyway, looks all good now. thanks --I9606 (talk) 21:09, 17 September 2015 (UTC)

Database reports/Constraint violations/P625

Hi, thanks a lot for daily updating Wikidata:Database reports/Constraint violations/P625 and the others constraint violations. At Wikidata:Database reports/Constraint violations/P625 there is a "single value" constraint, but usually multiple coors with qualifiers is OK (like at Q550). It is technically possible to filter out these (with qualifiers) from single value constraint results? --Jklamo (talk) 23:42, 19 September 2015 (UTC)

Hello, this is possible technically, but I have some troubles with my code to implement this. Also qualifiers usage for P625 is not simple question. Primary usage of P625 is displaying coordinates in Wikipedia page header ({{coord|...|display=title}}) and similar tasks. coord template needs non-trivial algorithm for selecting correct coordinated in cases like Q550. One approach is free from both problems: creating additional properties like coordinates of the beginning and coordinates of the object end. P625 will be used as coordinates of center point only in this case. — Ivan A. Krestinin (talk) 09:37, 20 September 2015 (UTC)

image value

Hello, I've noticed several times that when I've added an image, shortly afterwards your bot has set a claim value. When I look at the edit I can't see what the difference is. Clearly I'm doing something not quite right but I don't know what it is. Can you clarify? Thanks. Regards, Eagleash (talk)

Hello, bot normalizes file names. This process includes several things: replacing '%20' and '_' to spaces, removing trailing non-printable characters, resolving redirects and etc. Please provide link to some edit and I will describe it in details. — Ivan A. Krestinin (talk) 18:35, 22 September 2015 (UTC)
Hi, thanks. The most recent one was here. I just wanted to know if there is anything I should be doing differently or additionally. Thanks. Eagleash (talk)
Bot replaced underscores ('_') to spaces in this case. This will not happens if you will copy file name instead of file URL. Or you can just do not worry about this. Bot makes all needed replacements for filename unification. — Ivan A. Krestinin (talk) 22:31, 22 September 2015 (UTC)
Thanks. I open the (image) file in the Wikipedia page I'm sourcing from and copy from the url at the top I.e. the bit after 'File:'. As you say best not to worry about it if there's nothing else I need to do. Thanks. Regards, Eagleash (talk) 22:48, 22 September 2015 (UTC)

KrBot

Hi, I think that your bot is not working properly. See my rollback of an edit of your bot. Best regards, --★ → Airon 90 19:57, 24 September 2015 (UTC)

Hello, do this item really need two identical Sognando Cracovia (Q20962354) values without any qualifiers? It is looked very strange. — Ivan A. Krestinin (talk) 20:34, 24 September 2015 (UTC)

Mergers

Hello, I was wondering what action should be taken when 2 pages are merged on (say) en.wiki? I can follow the merger instructions OK, but there are instances where, although a page may be merged into another on en.wiki, there may still be entries on other (geographical) wikis. Do we just delete the link to the merged and redirected en.wiki page and check/edit if necessary the wikidata item for the destination en.wiki page? Thanks. Regards, Eagleash (talk) 10:55, 25 September 2015 (UTC)

Hello, I do not fully understand your problem, sorry. Items about same object should be merged both in Wikipedia and Wikidata. You can use Merge template to suggest merging in difficult cases. Also you can add said to be the same as (P460) to merge pending items. Redirect is need to be deleted from item for success items merging. Also you can request assistance from Jura. He has large experience in merging. — Ivan A. Krestinin (talk) 21:53, 25 September 2015 (UTC)
Hi, well for example the item Peugeot Speedfight 2 Q7179487 has 2 links to Wikipedia pages one for en:wiki and one for sv:wiki. On en:wiki, speedfight 2 has been merged into Peugeot Speedfight, but not as far as I can tell on sv:wiki. What is the procedure please? Thanks. Eagleash (talk) 22:24, 25 September 2015 (UTC)
Update; thanks again. I've contacted Jura as you suggest. Eagleash (talk) 22:40, 25 September 2015 (UTC)

wrong merge

Can you please undo the redirect resolving of consultant (Q15978655) -> Andy Harris (Q15055923). The merge was wrong. Thanks.--Pasleim (talk) 09:48, 26 September 2015 (UTC)

Would you undo this as well? --- Jura 16:08, 30 September 2015 (UTC)

Hello, can you please complete undo? There was more page who linked here before merge. Thanks. Metamorforme42 (talk) 17:25, 10 October 2015 (UTC)

Question about constraints

Hi, I'm looking for a constraint that I don't know if it exists but I don't know very well how to explain the constraint type, so I'll explain it as an example: I want a constraint that makes a list of all the elements which have given name (P735) which have linked in instance of (P31) an element whose instance of (P31) is occupation (Q13516667) or a subclass of it. This is because at least in the cases that given name (P735) is a valid statement the elements subclass of occupation (Q13516667) should be put as occupation (P106), and not as instance of (P31). I hope that you could understand what I mean. Thanks. -- Agabi10 (talk) 02:41, 11 October 2015 (UTC)

Hello, I am not sure that fully understand your idea. The task can be executed using two constraints:
{{Constraint:Item|property=P106}} for given name (P735) and
{{Constraint:Value type|class=Q13516667|relation=instance}} for occupation (P106).
Ivan A. Krestinin (talk) 11:01, 11 October 2015 (UTC)
To some extent, this is already checked by this, unless they are used together with P31:Q5. As I checked it the other day, there shouldn't be any left. --- Jura 11:15, 11 October 2015 (UTC)
The idea is that it should be detected as a conflict. And as an answer for Jura1 the problem is that they are used together with P31:Q5, so they can't be detected correctly with the current constraints. -- Agabi10 (talk) 11:41, 11 October 2015 (UTC)

Chemical structures

Dear Ivan,

your bot has added many chemical structures from German Wikipedia. May you let KrBot run again, to add the structures of the articles created since then?--Kopiersperre (talk) 09:45, 13 October 2015 (UTC)

Hello, I started KrBot. It finds 34417 unprocessed and previously failed pages in different wikies. So the task is needed several days to complete. — Ivan A. Krestinin (talk) 19:41, 13 October 2015 (UTC)
Thanks. May you please explain, how KrBot is working?--Kopiersperre (talk) 08:56, 14 October 2015 (UTC)
The bot complete his work faster than I expected. Bot creates templates and wikies list using elements Template:Chembox Identifiers (Q6871512), Template:Chembox (Q52426), Template:Infobox drug (Q6033882) and Template:Chembox entry (Q10804919). After this bot finds all templates inclusions. Bot extracts predefined property list from pages source, for example numero_CAS or chemical_formula. Bot normalizes and verifies the properties values. After this it writes the values to corresponding Wikidata item. Bot does not override existing Wikidata values and fails item processing if some identifier (like PubChem) differs from existing value in Wikidata item. — Ivan A. Krestinin (talk) 19:11, 14 October 2015 (UTC)

Change in inverse of constraint

I would like to know if it is possible to change the constrait inverse of or create a constraint that checks the inverse of one of a list of properties. This is for cases like Adrian Mole series (Q373620) that has has part(s) (P527) with each of the books but in the target element they have the relation defined with part of the series (P179) instead of part of (P361) and the constraint as it is now detects that kind of cases as incorrect when, at least for my point of view, should be also considered correct. -- Agabi10 (talk) 17:41, 13 October 2015 (UTC)

I have put this in the archive thinking I was in the talk page, so I moved to the main talk page. -- Agabi10 (talk) 20:30, 13 October 2015 (UTC)
Hello, it is technically possible of cause, but is it really needed...? Maybe more consistent way is creating separate fully symmetric property for part of the series (P179)? — Ivan A. Krestinin (talk) 21:01, 13 October 2015 (UTC)
I agree that it would be more consistent a property named "part of the serie", but in the property creation usually one of the main reasons I read for not creating a property is that it is similar to an existing one. Wouldn't it be the same problem with this one? And also maybe we don't want to have all the elements that can be in the serie listed in the element page of the serie, but I imagine that is possible to add the inverse to "serie" but not to the "part of the serie"... -- Agabi10 (talk) 21:58, 13 October 2015 (UTC)

First names

Hi Ivan,

Not sure if you are interested in the topic, but I'd be glad to have your input for Wikidata_talk:WikiProject_Names#Cyrillic_-_values_for_given_name_.28P735.29.

Currently, the way we include Russian names and Cyrillic transliterations (including Russian) is "suboptimal". --- Jura 17:40, 14 October 2015 (UTC)

Объеденение дубликатов

Здравствуйте. Как я понял, вы умело можете с ботом массово объеденить разные итемы об одном и том же, содержаюших ссылок к разным википедиям (дубликаты). Могу я попросить вас о помощи? Вот например я здесь создал список страниц итемов которые должны быть объединены. Там же я и патерны нарисовал для каждого случая. Посмотрите если можете что нибудь сделать. Если что-то неясно, спросите, разясню. Заранее спасибо. --XXN, 01:01, 15 October 2015 (UTC)

Я обновил страницу с новыми даными: те две азиатские ботопедии (ceb и war) имеют сотни одиноких категорий дупликатов (про таксонов). Сравните их между ними если название одинаково (без префикса) - тогда надо объеденить. Можете даже сравнить со шведской вики и будете поймать больше дупликатов. Надеюсь не обижаетесь что я вам пишу прямо здесь, может надо было в Bot_requests. Я пока сам не научился сделать Merge с ботом. Ах да, и извините меня за мои ошибки, я очень редко использую русский для обшении, и если когда-то знал его почти на отлично, сейчас маленько подзабыл его. :)--XXN, 14:18, 15 October 2015 (UTC)
Добрый день, попробую сделать что-нибудь с этими элементами, если не получится то тогда уже на Bot_requests. — Ivan A. Krestinin (talk) 18:20, 15 October 2015 (UTC)
Поработал над списком, для случаев с цифровыми подстановками у меня был практически готовый код, прошёлся им, там где нашлись пары объединил. Более сложные подстановки сегодня наверное обработать не получится, возможно стоит дать линк на этот список на Bot requests, возможно кто-нибудь успеет быстрее меня. Таксоны кстати кажется уже кто-то объединил. — Ivan A. Krestinin (talk) 19:52, 15 October 2015 (UTC)
Спасибо! --XXN, 20:19, 15 October 2015 (UTC)

Constraint for country (P17) *or* continent (P30)

Hi :)

GeoNames ID (P1566) has a constraint saying that the item should have country (P17). There are over 30,000 constraint violations for it on Wikidata:Database_reports/Constraint_violations/P1566#.22Item_country_.28P17.29.22_violations. Some of them are countries themselves (instance of (P31) country (Q6256)) but there aren't many, so those can easily be added as exceptions. About 10,000 of them, however, are in Antarctica, which seems like too many to try and add as exceptions.

What would be useful is a constraint saying the item should have one of country (P17), continent (P30) Antarctica (Q51) or instance of (P31) country (Q6256). Is that something which could be added?

- Nikki (talk) 14:03, 15 October 2015 (UTC)

Maybe P17 should have a value for Antarctica and international waters. --- Jura 15:58, 20 October 2015 (UTC)
I am sorry for delayed response. I try to analyze P17 violations. Antarctica is not the only case, some historical regions has GeoNames ID (P1566) and other cases present. Maybe P17 violations maintaining will be too hard task without Antarctica too. Currently I am trying to create bot for fixing simple cases. We will have more detailed understanding of the situation after this work completion as I hope. — Ivan A. Krestinin (talk) 19:32, 22 October 2015 (UTC)

Strange addition and deletion

I do not understand the KrBot edits. This one and the next. Is the bot misbehaving? — Finn Årup Nielsen (fnielsen) (talk) 13:55, 20 October 2015 (UTC)

VIAF

Not sure what happened here. VIAF had merged two identifiers (of 100?), but the others are still valid. KrBot removed all except one. --- Jura 15:58, 20 October 2015 (UTC)

Hello Jura, this situation is strange. VIAF database does not treat the items as redirects now. Something was changed after bot`s edits, but I do not see any related changes in items history sections. I have only one explanation: the items was merges in VIAF database, but the database was restored from backup copy after it. — Ivan A. Krestinin (talk) 19:48, 22 October 2015 (UTC)

Contraint violation

Hello

I'am looking for page Wikidata:Database_reports/Constraint_violations/P53#"One_of"_violations and I do'nt understand why House of Lomagne (Q21122351) and House of Armagnac (Q16034353) are not accepted as values for family (P53).

Sincerly -- Odejea (talk) 10:54, 21 October 2015 (UTC)

The "one of list" is hidden on the property talk page, so you can ignore them for now. Mbch331 (talk) 16:12, 21 October 2015 (UTC)


Merge into Q1294670

Hi Ivan,

Would you be so kind as to undo the edits of the bot? --- Jura 02:25, 25 October 2015 (UTC)

Would you do the same for the merge into Q17524135 ? --- Jura 13:07, 5 November 2015 (UTC)
And Q390? Thanks again. --- Jura 12:10, 17 November 2015 (UTC)
✓ Done, please check remaining links. — Ivan A. Krestinin (talk) 21:42, 17 November 2015 (UTC)
thanks. checked. --- Jura 23:48, 17 November 2015 (UTC)
Would you look into Q1294670 as well? --- Jura 15:42, 22 November 2015 (UTC)
✓ DoneIvan A. Krestinin (talk) 16:53, 22 November 2015 (UTC)

Hello, this page have not been updated for two months. Could you please induce a refresh? Regards --Färber (talk) 16:02, 28 October 2015 (UTC)

Thank you, it works. --Färber (talk) 06:16, 29 October 2015 (UTC)

Thanks for updating this report. Would it be possible to generate also a list of values of instance of (P31) and heritage designation (P1435) for all the labeled items? E.g., there is, say, 321 churches, 150 castles, 70 human settlements etc. among the labeled items, and 100 monuments of national significance, 200 monuments of local significance etc. This would really help to create an overview of the coverage of Slovene heritage. A monthly report would suffice if there's a problem of computing resources. — Yerpo Eh? 12:05, 30 October 2015 (UTC)

  • Hello, I added some statistics to the report, please review. Current bot implementation does not load labels to reduce memory usage, so I can not implement label analysis quickly. https://query.wikidata.org/ can be used to extract this information maybe, but I have no experience in SPARQL currently to help with request writing. — Ivan A. Krestinin (talk) 16:37, 31 October 2015 (UTC)
This is great, thank you. Labels aren't so important. — Yerpo Eh? 09:13, 3 November 2015 (UTC)

No entiendo

esto. Perdón por escribir en español, category Commons. --Jcfidy (talk) 09:05, 6 November 2015 (UTC)

"Can de Palleiro" is something different than "Can de palleiro". Sjoerd de Bruin (talk) 09:07, 6 November 2015 (UTC)
Sí, es cierto. No me había fijado que en Commons la p está en mayúsculas, lo corrijo ahora. Gracias --Jcfidy (talk) 18:42, 6 November 2015 (UTC)

This edit of your bot does not make sense IMO. Could you please check? --Leyo 09:22, 6 November 2015 (UTC)


Property proposals for constraints

Hi Ivan,

Here there are still a series of proposals for constraints. I asked Jonas about them, but he doesn't work on them any more. I think one or the other property could be useful merely for querying the information with SPARQL, but unless you decide to use them in your tool, it might not be worth creating all properties. The Special page will probably not change soon either. What do you think? --- Jura 09:41, 7 November 2015 (UTC)

Hi Jura, I can migrate my bot to the properties as soon as its will be created. This migration will make constraints querying more simple. But some negative points are present also:
  1. Templates visualize constraints. Its add human-readable description, links to report section, icons and etc. This allows to review and check constraint parameters using "Show preview" button before saving changes. We can try to visualize constraint properties on property talk page, but this visualization will not be available in constraint adding time. Also I do not have enough experience in LUA for creating such visualizer in short time.
  2. Some values are commented in page source. I suggested special property for constraints commenting, but it can not be used for individual values.
  3. Wikidata editor works very bad with large items lists. We need to test it in cases like Property talk:P225.
  4. Properties will limit constraint mechanism extension. For example the next construction is not representable using properties:
 {{Condition|property=P21|value=Q6581097|constraints=
   {{Constraint:Inverse|property=P7}}
   {{Constraint:Item|property=P31|item=Q5}}
 }}
So migration is possible, but we need to be ready for constraint mechanism degradation. — Ivan A. Krestinin (talk) 11:16, 7 November 2015 (UTC)
For readability, the current solution is preferable. Given that Special:Constraint report keeps using them, we would have to maintain them anyways. We could add checks to the templates to verify if at least part of the constraint is in the properties. There was an attempt to do this for Template:Constraint:Format/P1793.
In any case, I think the exceptions for Property talk:P225 should be handled on the items, not on that page.
I'm looking forward to see the conditional constraints in action. --- Jura 12:41, 8 November 2015 (UTC)

4. Anything is representable, for complex constraint we might need to use an additional item(s). For example for a set of constraint, create an item Qxyz with two statements

Inverse property of “P7 (P7):
if [item A] has this property (Ivan A. Krestinin/Archive) linked to [item B],
then [item B] should also have property “P7 (P7)” linked to [item A].
Exceptions are possible as rare values may exist. Exceptions can be specified using exception to constraint (P2303).
List of violations of this constraint: Database reports/Constraint violations/Ivan A. Krestinin/Archive#Inverse, SPARQL

 ;

Item: items with this property should also have instance of (P31) = human (Q5)
Exceptions are possible as rare values may exist. Exceptions can be specified using exception to constraint (P2303).
List of violations of this constraint: Database reports/Constraint violations/Ivan A. Krestinin/Archive#Item P31 = Q5, SPARQL

, and in the condition use for example : Template:Condition. Please also join the discussion on WikiProject Reasoning on the way to represent inference rule which are structurally very similar to conditional constraints. author  TomT0m / talk page 18:02, 8 November 2015 (UTC)

I think Ivan meant the representation with LUA, not with the constraint templates he had made. Besides, I think he reasons well even outside a specific project. --- Jura 12:31, 10 November 2015 (UTC)
No, actually for those both LUA and the properties. --- Jura 12:36, 10 November 2015 (UTC)

KrBot added incorrect instance of human

https://www.wikidata.org/w/index.php?title=Q3069985&type=revision&diff=269558095&oldid=269165036

This report was updated last on October 28th and is in the category unused properties. However it's used on 5 items. Could you refresh the report? Mbch331 (talk) 10:33, 9 November 2015 (UTC)

On Wikidata:Database reports/Constraint violations/P2241 the item count is wrong. I forced an update of the report and it ran today, but it keeps saying it's unused. Could it be that deprecated statements are ignored when making this report? If so, this report will never be used because this is a qualifier for deprecated statements. Or is there a way for you to also check deprecated statements for this report? Mbch331 (talk) 08:37, 12 November 2015 (UTC)

VIAF update

Hi Ivan, VIAF dumps 20151102 have been available for some days now but I'm missing the usual activity on P214. Is the dump yet to be imported or did again some pivot record change its association? -- Gymel (talk) 09:09, 12 November 2015 (UTC)

  • Hello, VIAF server is working unusual for some items, for example viaf:122613301. Bot`s self check procedure blocked processing in this situation to avoid possible data damage. I disabled part of self checks and now bot is operational. Thank you for your attention. — Ivan A. Krestinin (talk) 20:36, 13 November 2015 (UTC)

Constraint report tables broken

The statistics tables used for the constraint reports are broken and that makes it almost impossible to use the reports. For example see Wikidata:Database reports/Constraint violations/P1040 or Wikidata:Database reports/Constraint violations/P1199. Mbch331 (talk) 06:47, 15 November 2015 (UTC)

I noticed the main problems: you're not adding a new row marker before each new row (|-) and the closure mark of the table (|}) should be on a separate line. Mbch331 (talk) 07:08, 15 November 2015 (UTC)
Fixed, thank you. — Ivan A. Krestinin (talk) 17:37, 15 November 2015 (UTC)
Thanks for fixing and updating the reports again. Mbch331 (talk) 18:06, 15 November 2015 (UTC)

Contraint reports for P1367

Hi! Would it be possible to make the "single value" constraint report for Art UK artist ID (P1367) to ignore any values marked as deprecated ?

For example, Dora Meeson (Q5297435) has a value which has now been retired and which is marked as deprecated, as well as an ongoing value which is current.

Would it be possible to exclude cases like this from the report ? It would be really useful if we leave only the cases where we think the site really may have two different pages for the same individual.

Thanks, Jheald (talk) 21:58, 15 November 2015 (UTC)

cf. #Suggestion above. At Wikidata:Project chat#How to make "unique value" constraint violations report ignore deprecated values ? I pledge to make the behavior configurable in order to implement "workflows". -- Gymel (talk) 22:29, 15 November 2015 (UTC)
Current bot implementation ignores ranks. The P1367 case is very similar to VIAF case. Redirect VIAF IDs are resolved and deleted in Wikidata currently. This make our data clean, easy for understanding and simple for usage. Are we really need to collect P1367 deleted IDs in Wikidata? How to reference "messon-dora-18691955" value from Dora Meeson (Q5297435)? Archive.org is not very good source. Is current deleted identifiers set full or is it random subset of all deleted IDs? — Ivan A. Krestinin (talk) 07:13, 16 November 2015 (UTC)
There are different opinions on the P1367 example but I don't think it's worth dicusssing it here. The VIAF case is not optimal, too: I'm mainly relying on processing the unique value report every day (thus checking everything in person) with fallback on the feature of my browser to mark all links visited within the last 12 months. Neither of that scales to report processing done by more than one person. Any idea is welcome how a simple mechanism (like tagging values by means of reason for deprecated rank (P2241)?) in combination with constraint reports can support the different tasks of dealing with "fresh" reports and those issues that already have been inspected here. -- Gymel (talk) 07:26, 16 November 2015 (UTC)
IMHO we don't have to think along the lines of ranks like actual "deprecated rank" in Wikidata: There seems to have been the impression among a group of mailing list participants that deprecated values are automatically disregarded from constraint reports. In case of several VIAF numbers for one item marking n-1 of them as "deprecated" would often be quite arbitrary. But tagging them by means of setting a certain qualifier to the value "Wikidata verified" (some flavors of that will be needed depending on the workflow one wants to set up) could be worth the additinal effort. -- Gymel (talk) 07:35, 16 November 2015 (UTC)
As for VIAF, maybe this report can be helpful for "unique value" section processing. It can be updated every 2-4 hours. — Ivan A. Krestinin (talk) 13:09, 16 November 2015 (UTC)
Interesting. But there should be more items marked as Wikimedia duplicated page (Q17362920) (but maybe the VIAF id's already have been removed there). But in what respect are Taiwan (Q865) and Taiwan (Q22502) "marked to merge"? In theory there is also a won't fix case Wikimedia permanent duplicate item (Q21286738) -- Gymel (talk) 19:34, 16 November 2015 (UTC)
"marked to merge" = "marked by Template:Merge (Q6919004), Template:Merge to (Q5625904) or Template:Merge from (Q6041546)". Items Wikimedia permanent duplicate item (Q21286738) and Wikimedia duplicated page (Q17362920) are not very useful because its do not indicate duplicate of what. Properties said to be the same as (P460), different from (P1889) take more information. — Ivan A. Krestinin (talk) 20:32, 16 November 2015 (UTC)
(I still don't see where the "marked to merge" for the two aspects of Taiwan might come from) There is the qualifier of (P642) sometimes. And together with the actual duplication of P214 instances of wikimedia duplicated pages might indicate that someone already has looked at that situation.
But anyway: The report Wikidata:Database reports/Identical VIAF ID was not very helpful (except for the fact that it does show labels): The many known reasons for duplicate value violations are quite evenly distributed over the sections the report constructs. But thank you for letting me try it out. -- Gymel (talk) 22:12, 16 November 2015 (UTC)
Gymel: you might want to have a look at Wikidata:Database_reports/identical_birth_and_death_dates. Items to be checked are at Wikidata:Database_reports/identical_birth_and_death_dates/1. As we generally remove dates from items that should be merged, these disappear from the list. --- Jura 16:54, 23 November 2015 (UTC)

Property for constraint

The property for constraint is ready. See property constraint (P2302). Mbch331 (talk) 07:36, 16 November 2015 (UTC)

Good, now we need to extend {{Property documentation}} for this property displaying. — Ivan A. Krestinin (talk) 07:55, 16 November 2015 (UTC)
@Mbch331: As I see you create only one property. property constraint (P2302) cannot be used without another properties. Will you create its? — Ivan A. Krestinin (talk) 19:21, 16 November 2015 (UTC)
That property was set to Ready by Jura1, that's why I created it. Didn't have time to check the others out. Will look into them. Mbch331 (talk) 19:27, 16 November 2015 (UTC)
They have all been created. Mbch331 (talk) 20:39, 16 November 2015 (UTC)

constraint violations, link layout mystery

Hello Ivan, do you know why all the links for items on the P809 report (and other such reports) can randomly change their layout? i mean the following:

  • in the source the items have only "simple links" with square brackets like * [[Q112375]]
  • sometimes i see them like Q112375 (as normally expected),
  • but sometimes i see them like Galápagos National Park (Q112375) (which is great if it happens, very helpful to have the labels visible there)
  • the mystery: i don't understand why this can change, because in the meantime your bot did not change the link format there, and i have changed nothing too (neither in my preferences, nor in my browser). i noticed this behaviour only on the constraint violations pages and nowhere else. it happens for about 2 weeks now. i had no other "problems" on Wikidata in this time, and I am not aware of using any gadgets/scripts which could be responsible for the changing link display. i use german interface. changes from normal to extended links with labels (or in the other direction) can occur several times a day (= every time i visit a report page, i do not know before how links will appear).

i am really confused! Hope you can help (if possible, i would prefer to always see the item links in extended format, of course) Holger1959 (talk) 12:08, 18 November 2015 (UTC)

Somehow pages look differently if you are logged in or not, if viewed directly or in a diff, if it was just updated by a bot or not. Mysterious indeed. --- Jura 13:00, 18 November 2015 (UTC)
forgot to say: the above is always and only logged in. Holger1959 (talk) 00:13, 19 November 2015 (UTC)
I think you get the text either in the diff or when the bot just updated the page. --- Jura 00:15, 19 November 2015 (UTC)
Hello, the mystery is started some time ago. I think development team introduced new feature, but the feature has bug and works sometimes only. — Ivan A. Krestinin (talk) 03:30, 19 November 2015 (UTC)
thank you both for the replies! @Jura: right now, 4 hours after last bot update, i see the extended links again on Wikidata:Database reports/Constraint violations/P809 (not in diff view). @Ivan: from your answer i conclude that it was not you who added some hidden code especially for the report pages. so i will try to find out more by ask someone of the development team. Holger1959 (talk) 10:35, 19 November 2015 (UTC)
Sometimes I have the same "feature" in pages created by my bot, for example here, also in preview mode. It's appear randomly. --ValterVB (talk) 17:16, 19 November 2015 (UTC)
thank you for the additional notice, so its not only on the report pages. and exactly in this moment and on this talk page i also see this "feature" in my original question above: two links where i wanted to show the difference look the same. really confusing. Holger1959 (talk) 07:25, 20 November 2015 (UTC)

"Викимедиa" with Latin "a"

We seem to have 1.6 million descriptions which say "категория в проекте Викимедиa" with a Latin "a" at the end instead of a Cyrillic "а". Some of the most recent were added by your bot, e.g. Special:Diff/264508801 and Special:Diff/264510882. Is that correct? If not, would you be able to fix them? (I have a list of items, if it would help) - Nikki (talk) 22:20, 22 November 2015 (UTC)

The error is minor, but it must be fixed. I will do this. The list is not needed, thank you. — Ivan A. Krestinin (talk) 23:21, 22 November 2015 (UTC)
Thanks! :) - Nikki (talk) 08:06, 23 November 2015 (UTC)

Using statements for regex violation reports

Hi, now that format as a regular expression (P1793) exists, would it be possible to use it regex constraint violation reports so that we no longer need to store the constraint at two different places (in the template of the talk page and in the statements). --Zolo (talk) 08:19, 25 November 2015 (UTC)

Undid erroneous bot edit

Hi, I undid this erroneous bot edit of KrBot. The village of Schellingwoude is part of the municipality of Amsterdam. Please make sure your bot does not do this again. Gpvos (talk) 23:38, 27 November 2015 (UTC)

Constraints

That's great we are migrating to property statements! I had already started a module but then let it be because I didn't know how it should look like. Is there anything you would use in your system? I'm asking just because it is obsolete now and I will delete it.

Do you also plan migrating using bot? Matěj Suchánek (talk) 14:28, 30 November 2015 (UTC)

I have no enough LUA experience to understand all features of your module. Migration tool/bot is ready. But one issue blocks this process now. — Ivan A. Krestinin (talk) 15:00, 30 November 2015 (UTC)

НП-Беларусь

Добрый день! Можете ботом добавить на Викиданные все статьи из этой категории? --Чаховіч Уладзіслаў (talk) 12:29, 1 December 2015 (UTC)

  • Приветствую, по категории мне не очень удобно. Запустил бота по всем статьям из be_x_oldwiki, bewiki и ruwiki, которые содержат Q5625611. Надеюсь это покроет категорию. — Ivan A. Krestinin (talk) 20:26, 1 December 2015 (UTC)
Да, можно и так. Раз уж Вы вытягивает всё из шаблонов в статьях, то может и названия из шаблона lang-be-trans вытяните в качестве дополнительных для крупных языковых разделов, которые используют латинский алфавит (en, de, fr, it, es, sv, pl, pt, можно еще lt, lv, et)? Это бы упростило поиск и объединение статьей с категориями на Викискладе, например. --Чаховіч Уладзіслаў (talk) 21:19, 1 December 2015 (UTC)
Не уверен, что это правильно, всё-таки названия на разных языках нередко пишутся по-разному. — Ivan A. Krestinin (talk) 16:30, 4 December 2015 (UTC)
Поэтому я и не предлагаю добавлять эти названия в качестве основных. К сожалению, свойства "транслитерация" в Викиданных нет (либо я не знаю про него).--Чаховіч Уладзіслаў (talk) 21:53, 4 December 2015 (UTC)
Неосновные названия тоже должны быть написаны на определённом языке. Не хочется мне что-то такое делать, по-моему это не слишком правильно. — Ivan A. Krestinin (talk) 22:10, 4 December 2015 (UTC)
Создание элементов кстати бот закончил, плохо что в статьях нет какого-нибудь уникального идентификатора, какого-нибудь аналога кода ОКАТО, ОКТМО или хотя бы почтового индекса, а то можно было бы автоматически объединить элементы. — Ivan A. Krestinin (talk) 19:44, 5 December 2015 (UTC)
Благодарю! Есть список всех-всех населенных пунктов Беларуси (в том числе ликвидированных) с указанием кода Code System for Administrative-territory Division Objects and Settlements (Q6538410), но, к сожалению, он содержит только русские названия и немного устаревшее административное деление, поэтому пока что затруднительно использовать информацию из него для заливки в статьи в bewiki, хотя один участник взялся его адаптировать. Также на сайте Госкартгеоцентра есть странички для каждой административной единицы (и не только) Беларуси (пример - Вауки, искать можно здесь). Пока могу выставить заявку на создание соответствующих свойств. --Чаховіч Уладзіслаў (talk) 10:38, 12 December 2015 (UTC)
Свойства — это хорошо, только если в статьях не указаны эти идентификаторы, то проставить их ботом будет довольно сложно. — Ivan A. Krestinin (talk) 17:53, 12 December 2015 (UTC)

Expansion depth limit exceeded Constraint:Units

Your latest edit to {{Constraint:Units}} made it impossible to use {{Q}} for the values for list. It now gives a Expansion depth limit exceeded. See for example http://www.wikidata.org/w/index.php?title=Property_talk:P2370&oldid=278251799. Could you check how to fix this? This makes pretty display of units impossible. Mbch331 (talk) 15:56, 1 December 2015 (UTC)

It is strange behavior, I have no idea how to fix this:
  • {{#invoke:Constraints|getCaption|units|123456}} has expansion depth 2.
  • {{Q|123456}} has expansion depth 6.
but
  • {{#invoke:Constraints|getCaption|units|{{Q|123456}}}} has expansion depth 41.
I do not understand why 2 + 6 = 41. If you do not have any ideas too then please rollback my edit. The edit is not important. — Ivan A. Krestinin (talk) 20:48, 1 December 2015 (UTC)
I have no idea either how it's possible. So I reverted your last edit. But it might be worth to let someone with more knowledge of lua take a look. Mbch331 (talk) 21:16, 1 December 2015 (UTC)

Database reports

Hi Ivan, just found your Wikidata:Database reports/Identical VIAF ID. This list is a great help. Is there a simular list for GND ID (P227)? If not, can you create one? --Kolja21 (talk) 01:29, 3 December 2015 (UTC)

Awesome. Thanks! --Kolja21 (talk) 17:05, 7 December 2015 (UTC)

When Mohammed do not go to the mountain!

I would like to invite you to sv:WP:WF#Wikidata to responde to questions about KrBots edits to the VIAF-property! I invited Annika64 to come here to talk, but she says she do not want be in such discussions! Fell free to write in English! -- Innocent bystander (talk) 19:31, 3 December 2015 (UTC)

Revert

FYI Q319537. Please double check. --- Jura 12:58, 8 December 2015 (UTC)

Constraint Type

Здравствуйте. Скажите пожалуйста, почему "Constraint:Type" здесь не работает? Danneks (talk) 12:24, 12 December 2015 (UTC)

Constraint is defined to check the value for subclass of (P279) on the item that uses statement describes (P2384). However in the violation cases there is no subclass of (P279). So either the constraint needs to be changes or the items that violate the constraint. Mbch331 (talk) 12:42, 12 December 2015 (UTC)
Thank you. I'm not quite sure which solution is right, but it is easier to change the constraint. Danneks (talk) 12:55, 12 December 2015 (UTC)

Vandalism

Your bot seems to add data without any human verification. Please stop this vandalism... Sumurai8 (talk) 16:23, 12 December 2015 (UTC)

@Sumurai8: The source is stated for the edits. Wikidata works a bit different than Wikipedia. Importing data from Wikipedia's is currently still accepted. As long as the community doesn't say it's not allowed, it's still allowed to import data from Wikipedia. Mbch331 (talk) 16:33, 12 December 2015 (UTC)
Thank you, Mbch331. I can skip dates from future. But most probably this case is single. Vandalism detection problem does not have robust general solution unfortunately. — Ivan A. Krestinin (talk) 17:51, 12 December 2015 (UTC)
I think it's a single issue too, but to be safe it is indeed best to skip future dates for birth and death. Mbch331 (talk) 18:05, 12 December 2015 (UTC)
It would help if you used the latest tested/checked version of wikipedia's that uses that. You imported an unchecked vandalised page edited by an anonymous user. The tested/checked version was correct. In the worst case the tested/checked version is incomplete, but at least a human has read through it. Alternatively, you can possibly log pages where the current version is not the tested/checked version and manually check the content before parsing it and importing the data. Sumurai8 (talk) 18:36, 12 December 2015 (UTC)
Wikipedia:Reviewing (Q11107170)-based approach is looked reasonable and implementable. But unfortunately the most Wikipedias have too bad patrolling rate. The only exception is dewiki as I know. Additional think: popular well-patrolled articles are imported already. My bot is working with low popular articles now. Patrolling rate has additional decrease in this area. — Ivan A. Krestinin (talk) 19:34, 12 December 2015 (UTC)
These kind of errors show up in some reports, not the baddest vandalism we have here. Sjoerd de Bruin (talk) 22:12, 12 December 2015 (UTC)

Self-reference

Have you any idea why this edit was done by KrBot? -- Innocent bystander (talk) 17:48, 13 December 2015 (UTC)

Ok, thanks. That merge was probably wrong, but I am not in the mood to argue about it today. -- Innocent bystander (talk) 08:29, 14 December 2015 (UTC)

Error on marking church as human Q21706793

Hello. Please fix your bot, because of follow issue: https://www.wikidata.org/w/index.php?title=Q21706793&type=revision&diff=282684636&oldid=282684625 --Movses (talk) 09:39, 25 December 2015 (UTC)

Wikidata:Database reports/Constraint violations/P373

It is possible to fix that error? --Jklamo (talk) 22:19, 25 December 2015 (UTC)

Возера и Рака

Здравствуйте, можете "выпотрошить" шаблоны Возера и Рака? --Чаховіч Уладзіслаў (talk) 08:34, 26 December 2015 (UTC)

Qualifier or reference constraint

Currently, there are constraints for value-only, qualifier-only, reference-only constraints. For P1683, a "qualifier or reference"-constraint might be useful. --- Jura 14:06, 28 December 2015 (UTC)

Adding information

Hi again, I wanted to share with you an idea how to make database reports even more useful. You know, in Types Statistics, the bot lists how many instances are there of each type. You could program the output with a link to AutoList2 which would then show the reader a list of all those instances. It's quite simple, AutoList2 has a permalink option, so it's only a matter of formatting the link correctly.

Like this: http://tools.wmflabs.org/autolist/index.php?language=en&project=wikipedia&category=&depth=0&wdq=claim%5B{{{1}}}%5D%20AND%20claim%5B31%3A{{{2}}}%5D&pagepile=&statementlist=&run=Run&mode_manual=not&mode_cat=not&mode_wdq=and&mode_find=not&chunk_size=10000, where {{{1}}} is the property reported and {{{2}}} is type.

For example, the Types Statistic in report for P1587 would look like this:

What do you think? — Yerpo Eh? 09:08, 29 December 2015 (UTC)

Without "language=en" of course. And may be "chunk-size=1000" as it loads quicker. --Infovarius (talk) 21:30, 30 December 2015 (UTC)
Many reports are very large. Adding many long URLs will increase reports size. Is the links really useful? What tasks require the links? — Ivan A. Krestinin (talk) 19:33, 31 December 2015 (UTC)
I think it's useful to see which items are of some type, but it's your decision. For the example I gave above, it's useful to cross-check with official list of monuments. — Yerpo Eh? 19:05, 2 January 2016 (UTC)

ISNI

The use of spaces in the ISNI number is disupted.. Please stop your bot.  – The preceding unsigned comment was added by GerardM (talk • contribs).

Constraint report update/KrBot

Hi Ivan,

I guess the bot has a day off .. dumps would be available.
--- Jura 11:37, 15 January 2016 (UTC)

List/category of failed constraint violation reports

Hi! I'm new to Wikidata and I just happened to discover several property constraint violation reports that had been failing for some time now, e.g. [48] and [49] (I fixed all the ones I found).

That's all fine, people make mistakes. But I couldn't find any existing pages about failing reports. To aid fixing these problems, could you create a list page with links to all failed property reports? That way interested people can watch it and fix problems quickly when they show up.

If that's too complicated, maybe just include a template or add a category to failed report pages. The downside is that categories/transclusions can't be watched using the watchlist (AFAIK?), but it's still better than having to search for failed reports. Intgr (talk) 11:50, 28 January 2016 (UTC)

Constraint violations/P1447

Hi! It looks like I screwed up it a little bit :) Maybe your bot could recognise {{Q|Qx}} as a valid expression because {{Q}} accepts it? --Edgars2007 (talk) 11:13, 29 January 2016 (UTC)

Дубликаты

Здравствуйте. Нашёл в русской википедии ~700 статей про украинских сельсоветов, которые имеют свой личный элемент Викиданных и не связаны с аналогичным статей из украинской википедии (а они существуют и тоже имеют свой элемент. т.е. итемы-дубликаты). Список здесь. Можете найти какой-нибудь удобный алгоритм для вашего бота что-бы найти и объеденить пары этих элементов? --XXN, 15:59, 9 February 2016 (UTC)

В похожем ситуации (но чуть сложнее) около 700 статей про украинских сёл в румынской википедии,[50][51] статьи которые оказались изолированы здесь в викиданных и нуждаются в связании с украинсками аналогами.
И если уже речь о дубликатов, заметив эту вашу инициативу, пользуясь случаем принесы к вашему вниманию факт что, на этот момент здесь в викиданные есть ~109.000 изолированных (одиноких) категорий немецкой википедии[52] и ~68.000 подобнух категорий французкой википедии[53]. Среди них конечно куча дубликатов для объединения, но проблема как бы найти им пару. Может сумеете зделать вы что нибудь с частью из них, а вообше-то при таком масштабе уже и немецкие девелоперы должны помочь, вовлечся как нибудь. Простите если побеспокоил. --XXN, 19:57, 9 February 2016 (UTC)
По сельсоветам попробую объединить используя КОАТУУ, практически готовый код для этого имеется. К концу предстоящих выходных наверняка сделаю, может раньше. С категориями надо ещё посмотреть. — Ivan A. Krestinin (talk) 04:53, 10 February 2016 (UTC)
@XXN: На данный момент сделал следующее: из шаблонов uk:Шаблон:Селищна рада, uk:Шаблон:Сільрада України, ru:Шаблон:Административная единица Украины импортировал KOATUU ID (P1077), located in the administrative territorial entity (P131), украинское название. После этого объединил те элементы, у которых совпадали KOATUU ID (P1077) и названия на украинском. Дополнительно проверял непротиворечивость located in the administrative territorial entity (P131). Завтра ещё раз пройдусь алгоритмом объединения по всем элементам, возможно что-то в сегодняшний дамп не попало. Румынской википедией пока занимаюсь, у них нет специализированных карточек для Украины, потому извлекать данные намного сложнее. Категориями не занимался. — Ivan A. Krestinin (talk) 12:06, 14 February 2016 (UTC)
По rowiki я смотрю ваш бот начал работать, скажите тогда, как он закончит, я тогда своим ботом ещё пройдусь, может ещё что-нибудь извлеку. Кстати у нас в метки похоже попали ударения, причём похоже не в том формате, в Bezimianka (Q4081102) метка "Безім'янка", в тоже время в rowiki указано название "Безім`янка". Не знаете случаем, это точно просто ударение, которое можно удалить из метки? Ещё у многих сёл в качестве located in the administrative territorial entity (P131) указан район, хотя в rowiki имеется и сельсовет, например Balky (Q3617635), нужно такие случаи исправлять? — Ivan A. Krestinin (talk) 14:16, 14 February 2016 (UTC)
Про located in the administrative territorial entity (P131) - я не знаю точно какой здесь консенсус; наверное надо указывать низшее административное единица. Но посколько большентсво википедий не имеют статьи про сельсоветов, у меня лично на данный момент нету единой мнений что надо "исправить" ту значению из района в сельсовета, но при этом и не могу и не буду возражать если будем произведена однородность.
Да, как заметили, в последние дни я начал при помоши бота (а местами и вручную) объеденить изолированные итемы статей rowiki - а зделал это только там где мог автоматически/массово найти 100%-ые пары для румынских статей. И там было немножко хаоса: кроме тех более 1500 статей (если не ошибаюсь) про украинских Н.П. и сельсоветов имея собственный элемент WD, несвязанные с украинскими/русскими статьями-аналогами; были также, румынские статьи про сёл одиночки-в ихнем сельсовете - связаны неправильно с украинскими статьями про сельсоветов а не про село (это исправил до конца). Мой метод работы с ботом для слияний элементов не очень продвинут - нужно иметь точные паралельные списки элементов/названий статей; так что где я мог - формировал эти списки при помоши интрументов как PagePile, Tabernacle и Linked items, а потом приступал к слияний. Короче говоря мучался немало и на этот момент я исчерпал свой списки над которыми было поработать а остались ещё 1100 несвязанных статей - итемов для объединения [54]. Но думаю что вам может быть трудно незная румыснкий язык. В крацех, в rowiki статьи созданы по данным 2001 переписи и есть статьи и про исчезнувших сёл, для которых украинци не имеют аналог :), например про ru:Заречье (Белогорский район, Крым)ro:Zaricicea, Bilohirsk. Некоторые сёла, сельсоветы и даже один район созданы с старыми названиями (2001) а в украинском разделе (и остальные) у них уже новые, изменёные. А также, не знаю почему, не во всех статей rowiki указан код КОАТУУ, то что может вам создавать трудности. Мда, и похоже что мой коллега румынский ботовод как-то перепутал и поменял украинский апостроф с каким-то другим символом. Если вас ещё не испугали все эти трудности можете взяться и за этот таск, попробовать и посмотреть что получется. --XXN, 15:30, 14 February 2016 (UTC)
Накосячил этот ботовод не только с апострофом, но и с самими названиями. Большая загадка почему в качестве украинского названия в статье ro:Comuna Novomariivka, Solone значится Новомар`ївка вместо Новомар'ївська сільська рада. Похоже из rowiki украинские названия лучше и не пытаться импортировать. — Ivan A. Krestinin (talk) 20:25, 14 February 2016 (UTC)
Да, да. Как нибудь подскажу ему об этом. Дело в том что в румынском разделе нередко недооцениваются префиксы/дополнительные слова из словосочетаний названий админ. единиц. Вот например как выглядит верхушка статьи об Ясском уезде — "подзабылось" слово "Județul" и в вступлении и в карточке. Точно также и в случае коммун. --XXN, 22:21, 14 February 2016 (UTC)

Importing content from Wikipedia GeneWiki info boxes to Wikidata

Hi! Your bot seems to occasionally do imports from the English Wikipedia Gene Wiki infoboxes. On on hand, your bot seems to be malfunctioning, as it put the gene ontology term protein serine/threonine kinase activity (Q14326094) as molecular function (P680) and as biological process (P682) on item [55] and many other items (and also many other gene ontology terms), which is wrong. As we, the ProteinBoxBot team, have been importing large amounts of biological data from their primary sources, e.g. Uniprot and many of the data in the GeneWiki infoboxes are outdated, it might be the right time to stop importing data from the Wikipedia GeneWiki infoboxes. Furthermore, we are currently rolling out a new Wikipedia GeneWiki infobox template script, which pulls the data directly from Wikidata to populate these infoboxes [56]. Re-importing these data to Wikidata is therefore not required anymore and only creates a lot of overhead. More info on what we currently do and plan can be found here: [57]. Thank you! Sebotic (talk) 09:32, 10 February 2016 (UTC)

Oh, I just realized that your bot only fixed problematic redirects and is not responsible for the wrong Gene Ontology terms. Sorry for that! But in general, imports from GeneWiki infobox to Wikidata should not be required anymore. Best Sebotic (talk) 10:45, 10 February 2016 (UTC)

Undo

Hi Ivan,

Would you undo these changes. I fixed the corresponding merge.
--- Jura 15:52, 14 February 2016 (UTC)

If you have an easy way to undo those merge-related edits, would you also undo those that followed the erroneous merger between Istrup (Q1674886) and defensive wall (Q57346) 8 Feb ? --Zolo (talk) 11:35, 27 February 2016 (UTC)

Confusing reference

Could you please explain what this edit is about? --Yair rand (talk) 12:05, 22 February 2016 (UTC)

Hello, bot detected duplicate values. After this bot copied all references to the first value and delete all another duplicate values. Please see the next bot`s edit. — Ivan A. Krestinin (talk) 12:32, 22 February 2016 (UTC)
Ah, I see. Thank you. --Yair rand (talk) 15:53, 22 February 2016 (UTC)


Year 0

Hi Ivan,

I undid a couple of these. They showed up with the "calendar model" complex constraint on P570.
--- Jura 13:10, 24 February 2016 (UTC)

Thank you

My thanks to you and your bot for [58] and [59] changes.

I put these with Magnus' Quick Statements, with a statement like:
LAST P569 +00000001946-08-03T00:00:00Z/11 P248 Q21087031
which is not putting the reference in the correct place.

Do you have any idea what would be the correct syntax to put it in the correct place? --FocalPoint (talk) 14:38, 20 March 2016 (UTC)

Do NOT change redirects

Your bot just made a wrong merge worse - instead of reverting just two edits I now have to find all those edits of your bot where it changed the link to the redirect to the new page, e.g. Bangkok Noi (Q806486). There is absolutely no technical reason to change any uses of redirected item. Ahoerstemeier (talk) 22:12, 1 April 2016 (UTC)

@Ahoerstemeier: quick answer, Ivan will give you more detailed answer. a) AFAIK, bot is able to revert these changes, you simply have to say which item has to be "reverted" b) this is not Wikipedia, there are reasons, why redirects should get fixed. There are 24h delay before doing that, so wrong merges can be undone. --Edgars2007 (talk) 22:23, 1 April 2016 (UTC)
Thank you, all right. Links are needed to be checked anyway after merge revert. Some links can be created during merged item existing time. — Ivan A. Krestinin (talk) 22:29, 1 April 2016 (UTC)
@Edgars2007: Help:Redirects#Links to redirects clearly states these links to redirect shall not be changed, which was also what I think to have seen as a discussion result in the project chat. Besides, 24h is way to short to spot a wrong merge if its not a prominent item. Ahoerstemeier (talk) 22:37, 1 April 2016 (UTC)
Help page was written without any discussion and situation analysis. — Ivan A. Krestinin (talk) 22:59, 1 April 2016 (UTC)


Merged items

Hi Ivan,

Just to let you know: I merged a series of items you created for articles a Kazakh wiki. They appeared on Wikidata:Database reports/identical birth and death dates/1.

I think it might be worth looking into Wikidata:Bot_requests#Cyrillic_merges. Obviously, it's not my area of expertise. Even if writing of names differs, I think some items could be identified by accounting for different name orders and punctuation. I keep coming across samples once they get dates.
--- Jura 16:56, 4 April 2016 (UTC)

Template:Constraint:Units

Hi again :)

area (P2046) is a generic property, so any unit of area should be allowed (we have 58 different units at the moment although only 10 are listed on the property talk page).

Would it be possible to have something like Template:Constraint:Value type? e.g. {{Constraint:Units|class=Q1371562|relation=instance}} to mean "the unit must be an instance of unit of area (Q1371562) (or one of its subclasses).

- Nikki (talk) 09:45, 10 April 2016 (UTC)

Hello, it is possible of cause. But are we really need to spend efforts to implement and support this ability? Only 10-11 units are used now: Wikidata:Database_reports/Constraint_violations/P2046#Units_statistics. The list can be extended at any moment. Also suggested approach is less stable. — Ivan A. Krestinin (talk) 10:42, 10 April 2016 (UTC)
It's not specific to this property, there are quite a lot of properties like this. In fact, I think most properties using units are more accurately described as having a type of unit (e.g. unit of length) as the allowed values for the unit. More examples:
It would be particularly useful for currencies - someone else has asked on Property_talk:P2284 if it would be possibly to just say any currency instead of trying to list all currencies.
- Nikki (talk) 10:44, 13 April 2016 (UTC)
Hello Ivan, I came here, because I had basically the same request. I just created budget (P2769) and it would be useful to say "unit can be any instance of budget (P2769)", instead of listing currencies by hand. --Srittau (talk) 12:44, 23 April 2016 (UTC)
In the meantime, we have 270 units of type "area". I don't think these are necessarily all suitable.
--- Jura 13:56, 23 April 2016 (UTC)

TLDR

I am slowly working with the P1566-report. One problem, that causes me to edit less in a day than I could, is that the report is lllllloooooonnnnngggg (857316 bytes in the last version). It takes me half a minute every time I successfully load the page, and far from every time is successful! The report is already shortened, not everything is written in the report. Can you make it even shorter? -- Innocent bystander (talk) 07:14, 20 April 2016 (UTC)

If you just want a small selection of the many violations to fix to work on, you would probably be better off using a SPARQL query (e.g. 50 items with P1566 and no country). You can even select labels or other properties that way. The SPARQL endpoint is also more up to date, whereas the constraint reports seem to be a couple of days behind (at least I often have to wait a couple of days before things I've fixed get removed from a constraint report and I can continue working on it). To be honest though, there's so much to fix for P1566 that fixing things by hand is really inefficient. The problem was caused by bots (mostly lsjbot importing GeoNames into svwiki and cebwiki as far as I can tell), so it should be possible to add lots of the missing statements with bots too. - Nikki (talk) 11:22, 20 April 2016 (UTC)
The errors added by Lsjbot are the most easy to solve, since every Lsjbot-article matches one and only one GeoNames-ID, while the man made articles have several topics and are difficult to match in our items. Many of the items in the unique-part of the list also have a wrong GeoNames-ID introduced by a local bot here. They have been imported directly from a GeoNames-dump. But the links to Wikipedia in GeoNames are often wrong, only matching the name of the article, not the topic.
I cannot see any other way than fixing this by hand. Adding P625/P17 can probably be added by bot, but such bot-runs have already been done, leaving a lot of items behind.
I am not familiar with SPARQL at all. Can you provide me a link for the unique-constraint? -- Innocent bystander (talk) 14:13, 20 April 2016 (UTC)
This should be what you need. Of course, we can add there (almost) anything else, that helps find violations better. P.S. Here is some help about SPARQL. --Edgars2007 (talk) 17:13, 20 April 2016 (UTC)
@Edgars2007, Nikki: The links to the WD-items in the SPARQL-report goes to json-versions of the pages, not to the same page links like Q1 takes you! -- Innocent bystander (talk) 06:28, 21 April 2016 (UTC)
That's the same problem you mentioned on Wikidata:Contact the development_team#Identify_the_unit. I've added a comment there. - Nikki (talk) 06:54, 21 April 2016 (UTC)
{{Q}} is overused on the report. The template will be excluded from long tables during next run. — Ivan A. Krestinin (talk) 18:49, 20 April 2016 (UTC)

ru:ПРО:ВО: бардак с реками

Приветствую!

Не поможешь навести порядок с реками?

  1. Для 53-х тысяч речных объектов в ВД не указано свойство страна: [60]
    1. Проблема, по-видимому, создана криворукими участниками, заливавшими от шведов (46 тысяч рек без стран): [61]
    2. Залить 800 штук можно из немецкого раздела: [62]
    3. Залить 1200 штук можно из английской вики: [63]
  2. В шведском разделе тысячи статей о реках с «ручными» интервиками. Элементы на ВД я им создал, а вот как туда привязать статьи из интервик, не знаю (там привязки к ВД ещё нет). Вот собственно и нужно привязать все статьи по «ручным» интервикам к объектам на ВД. Мастер теней (master of shadows), 18:27, 2 May 2016 (UTC)
  • У меня был какой-то алгоритм импортирующий страны, запустил его на шведском материале. Потом можно будет и по остальным пройтись. Чего-то готового для удаления интервик у меня нет, но у других ботоводов должно остаться со времён миграции на Викиданные. — Ivan A. Krestinin (talk) 15:06, 3 May 2016 (UTC)
Там можно и не удалять, в принципе. Главное связать с уже имеющимися объектами в ВД. Мастер теней (master of shadows), 16:04, 3 May 2016 (UTC)
У меня даже такого кода нет, статьи с локальными ивиками я обычно обходил стороной т. к. там велика вероятность неразрешённых ивики-конфликтов. Попробуйте спросить у участника Emaus, он в своё время занимался переносом интервик, если мне не изменяет память. — Ivan A. Krestinin (talk) 16:44, 3 May 2016 (UTC)
  • Бот прошёлся по странам, но получилось у него не везде. Бот не смог разобраться там, где в одном параметре была указана и страна и регион. Также он не справился там, где указано сразу несколько стран. — Ivan A. Krestinin (talk) 04:54, 6 May 2016 (UTC)

Ваш бот объединяет элементы, которые не должны быть объединены

Q17682271 и Q21672401 не должны быть объединены, это разные населенные пункты. --Alex NB IT (talk) 04:05, 4 May 2016 (UTC)

Вероятно потому, что кто-то указал одинаковые почтовые коды. Мастер теней (master of shadows), 08:54, 4 May 2016 (UTC)
Действительно, т. к. названия и почтовые индексы были одинаковыми, и противоречий в каких-либо других свойствах не было, то бот "решил", что это один и тот же населённый пункт. — Ivan A. Krestinin (talk) 09:23, 4 May 2016 (UTC)

Проставление кода ОКТМО

Ваш бот проставил код ОКТМО на страницу Q18769731. Похоже, эта страница - ошибка ботозаливки чеченской вики, никаких следов существования Васюково в Андреевском сельсовете или сельском поселении я не нашла. Единственное Васюково Вашкинского района это Q4105177 (и в нем правильно указан тот же код ОКТМО). Я бы стерла это значение ОКТМО из Q18769731, но не проставит ли его бот заново? --Daryona (talk) 10:18, 4 May 2016 (UTC)

  • Добрый день, похоже этот НП перестал существовать. Хорошо бы заполнить свойство "дата прекращения существования", тогда бот точно его трогать больше не будет. Но вот чем его заполнить - непонятно, я сходу тоже не нашёл упоминаний. Вообще с сёлами-призраками беда, их слишком легко перепутать с одноимёнными ныне существующими сёлами в этом же районе. — Ivan A. Krestinin (talk) 11:07, 4 May 2016 (UTC)
    • Да, я тоже сейчас нашла следы, что был такой НП раньше (в справочниках почтовых индексов, в частности, встречается). Но я выверяла по Вологодской области все НП с 1999 года, так что его еще раньше ликвидировали, и вряд ли удастся найти что-то сейчас. Может, указать "1999", чтобы просто была какая-то отметка об упразднении? Все равно вряд ли кто-то будет реально редактировать эту статью и страницу викиданных. --Daryona (talk) 12:49, 4 May 2016 (UTC)
  • День добрый. Бот начал проставлять коды ОКТМО и почтовые индексы в реки. Проблема с чеченской викой. Мастер теней (master of shadows), 14:11, 22 May 2016 (UTC)
    • Да, видел немало таких ситуаций, но тут уж ничего не поделать, нужно разгребать неправильные интервики. — Ivan A. Krestinin (talk) 14:29, 22 May 2016 (UTC)
Почему? Можно спрашивать P31. Если не подкласс НП — зачищать левые интервики. Мастер теней (master of shadows), 14:56, 22 May 2016 (UTC)
Код, который достаёт ОКТМО и без того излишне сложный, ещё и этим его нагружать не стоит. Проще как сейчас - импортировали данные, увидели, что наблюдаются конфликты, исправили проблемные элементы. — Ivan A. Krestinin (talk) 14:59, 22 May 2016 (UTC)
Хорошо. Мастер теней (master of shadows), 14:56, 22 May 2016 (UTC)
Вот все НП, привязанные к рекам: [66]. Мастер теней (master of shadows), 20:16, 22 May 2016 (UTC)
Сократил этот список до трёх элементов. — Ivan A. Krestinin (talk) 20:34, 22 May 2016 (UTC)

Wrong P17 on bot edit

This is wrong https://www.wikidata.org/w/index.php?title=Q464980&diff=prev&oldid=328852199Finn Årup Nielsen (fnielsen) (talk) 20:34, 11 May 2016 (UTC)

Литва

Добрый день! Можете потянуть данные из литовских lt:Šablonas:Ltgyv и lt:Šablonas:Seniūnija? --Чаховіч Уладзіслаў (talk) 17:22, 14 May 2016 (UTC)

Да, всегда Литва. Сенюнии (староства) - нижний уровень адм. деления Литвы, поэтому импортировать нужно sen, хотя, возможно, некоторые населенные пункты имеют районный статус, т.е. нет sen, только sav. Этих свойств вполне достаточно. --Чаховіч Уладзіслаў (talk) 07:47, 15 May 2016 (UTC)
Староства импортировал вместе с country (P17), located in the administrative territorial entity (P131), instance of (P31). А вот с населёнными пунктами не совсем хорошо получается, для многих НП в качестве located in the administrative territorial entity (P131) уже указан район и непонятно: не то это ошибка и нужно исправлять на староство, не то этот НП действительно районного подчинения. Буду импортировать без located in the administrative territorial entity (P131). — Ivan A. Krestinin (talk) 14:20, 15 May 2016 (UTC)

Отчёты

Приветствую.

Предлагаю заменить в списках точки на нумерацию.

К примеру, Wikidata:Database reports/Constraint violations/P884 — если посещал страницы, фиг поймёшь, какие только что исправлял, а какие нет. С номерами будет попроще. Мастер теней (master of shadows), 11:01, 24 May 2016 (UTC)

  • Доброго вечера, в принципе можно, но всё равно будет не вполне удобно. Возможно при разгребании удобнее использовать SPARQL запросы, которые есть в шаблонах ограничений, их результаты обновляются довольно быстро. Правда иногда они возвращают ерунду. — Ivan A. Krestinin (talk) 18:45, 24 May 2016 (UTC)

imported from:Tajik Wikipedia

Do you remember where this comes from? I have seen some of these statements now, and they have had in common that they were about something in Brazil and they had Tajik Wikipedia as source, but the claims stated something (very) different. -- Innocent bystander (talk) 06:50, 30 May 2016 (UTC)

GNDs imported from VIAF

Hi Ivan, KrBot seems to have stopped to correct format violations of GNDs that have been imported from VIAF, see: Wikidata:Database reports/Constraint violations/P227. There are to many erros to correct them by hand. Can you restart the bot? --Kolja21 (talk) 14:04, 4 June 2016 (UTC)

Hello, GND site request format was changed. Fixed. — Ivan A. Krestinin (talk) 08:29, 5 June 2016 (UTC)
Great. Thank you! --Kolja21 (talk) 10:13, 5 June 2016 (UTC)
Hi Ivan, thanks a lot, indeed. In the GND ID (P227) case there are many manual re-imports of the unsuitable GND identifiers from VIAF, so they also show up as duplicate values in the report and then vanish again after your bot has taken care of their form and then removed them as duplicate value. Even when everything goes smooth this takes about four days until all reports are neutral again. Would it therefore be possible to couple your bot's repair actions of "number is formally wrong but I know how to fix" with the much more frequent updates on Wikidata:Database reports/Constraint violations/Mandatory constraints/Violations ? -- Gymel (talk) 05:59, 7 June 2016 (UTC)

The bot sets Per Boye Hansen (Q2069926) to a wrong VIAF (85219782) [67] Per Boye Hansen is not the same as Per Christian Hansen. — Finn Årup Nielsen (fnielsen) (talk) 08:30, 6 June 2016 (UTC)

The identifier was imported from VIAF. This was single pass import. So you can just delete wrong identifier. Bot will not re-import it. — Ivan A. Krestinin (talk) 17:14, 6 June 2016 (UTC)

"Single value" violations

Hi Ivan, Wikidata:Database reports/Constraint violations/P227 has now more than 1.000 "single value" violations. If GND ID (P227) is used with a qualifier like pseudonym (P742) the violation is o.k. (= no error). Is there a way your bot can list violations with qualifiers separately? --Kolja21 (talk) 11:44, 11 June 2016 (UTC)

Constraint:Contemporary

Hi, Ivan.

Would you like to share your comments about a proposal for a new constraint? If you don't have the time, you can simply ignore this message.

Thanks in advance, and best regards, --abián 20:15, 9 July 2016 (UTC)

Please, feel free to tell me if I can help with this issue in some way. --abián 16:40, 15 July 2016 (UTC)
We need to close this issue and distinguish every {{Constraint:Diff within range}} from the rest here. Anyway, I'm still thinking that the discussion page and the constraint violations page get too long and complex by this system (the good point is that I've discovered a new bug thanks to that xD), and a {{Constraint:Contemporary}} would significantly improve this point if it was included in link properties (A-B-C), where it would appear only once and no parameters would be needed. --abián 11:31, 23 July 2016 (UTC)
Hi again. I understand you don't have the time to help with this, and I would like to help you, but I cannot find your bot's code. What can we do? --abián 10:14, 29 July 2016 (UTC)
Hello, you are right, I have no enough time to implement this now. Bot`s code is not published yet. But you can create separate report for now. — Ivan A. Krestinin (talk) 16:28, 29 July 2016 (UTC)
I fear that managing a separate report system that works in a completely different way wouldn't be the best idea. I think it's important for Wikidata that you publish KrBot's code as soon as possible so that the community can help you and that we can prevent any possible data loss. KrBot is too important for all of us! :-) --abián 20:35, 25 September 2016 (UTC)

Your bot currently generates on the constraint pages HTML of the form <span />. Can you please change this behavior either to use {{Anchor}} or to use <span></span>? --Izno (talk) 13:53, 15 July 2016 (UTC)

One of constraint on P17

Hey. Q55 is in the list of the "one of" constraint on Property talk:P17. But on WD:Database reports/Constraint violations/P17#One of all items with P17=Q55 are listed as violations. --Pasleim (talk) 09:13, 21 July 2016 (UTC)

Incorrect constraint violations list

this list generated by your bot does not seem to be correct. According to queries to check this constrain on Property talk:P1472, for last few days, there are ~5 pages that violate that constrain. --Jarekt (talk) 12:41, 27 July 2016 (UTC)

PetScan

Ivan, could you translate for PetScan for me 'P579 and P2695'? thx --Chris.urs-o (talk) 11:08, 2 August 2016 (UTC)

Thx. AutoList 2 is off and I can not write PetScan or WDQ. But thank you very much for the programm. Regards --Chris.urs-o (talk) 16:21, 4 August 2016 (UTC)
@Chris.urs-o: you can make WDQ requests here, if you need some. --Edgars2007 (talk) 16:49, 4 August 2016 (UTC)
Thx, sry --Chris.urs-o (talk) 19:02, 4 August 2016 (UTC)

Whitelists for constraint violations

Regarding Wikidata:Database reports/Constraint violations/P2880, would it be possible for some single value violations to be whitelisted as valid? Sometimes an article in the database has multiple entries for some reason, so having multiple values is valid. Marking off the valid ones would help make sure I don't repeatedly look at them while screening the ones that genuinely are constraint violations. Thanks, James Hare (NIOSH) (talk) 16:34, 4 August 2016 (UTC)

@James Hare (NIOSH), Harej: that is possible with |exceptions=. See for example Property talk:P2593. Note, that then you have to remove |mandatory=true, if it's included in constraint. --Edgars2007 (talk) 16:47, 4 August 2016 (UTC)
Edgars2007, that should work. Thank you! James Hare (NIOSH) (talk) 20:19, 4 August 2016 (UTC)

Названия НП-Беларусь

Добрый день! Можете обновить белорусские названия населенных пунктов Беларуси в Викиданных, использовав для этого названия из карточки НП-Беларусь, а те названия, которые поселения имеют сейчас в Википедии и Викиданных (а они разные), перекинуть в дополнительные варианты? А то Вёска NNN, NNN сельсавет выглядит в карточках жутко, а Аграгарадок NNN в карточке персон, рожденных, например, в 19 в. - нелепо. --Чаховіч Уладзіслаў (talk) 07:04, 7 August 2016 (UTC)

  • Здравствуйте, это уже не вполне тривиально, т. к. названия каких-то НП участники могли задать руками и весьма вероятно, что будет некорректным менять его ботом на какой-то ещё вариант. Но если вы можете как-то сформулировать правило по которому бот мог бы выделить "плохие" названия, то их я могу поменять. — Ivan A. Krestinin (talk) 08:32, 7 August 2016 (UTC)
По-идее участники должны были задавать руками то же название, что и в карточке статьи, если же они задавали названия до заливки населенных пунктов в Белорусской Википедии, то вполне вероятно, что такие названия могут быть ошибочными, несогласованными с нормативным справочником названий, по которому велась заливка. В любом случае однозначно необходимо исправить названия ключами Вёска, Хутар, Горад, Гарадскі пасёлак, Аграгарадок, а также все названия с запятыми. --Чаховіч Уладзіслаў (talk) 08:56, 7 August 2016 (UTC)

Identical birth dates (no P570)

Hi Ivan,

Wikidata:Database reports/identical birth and death dates seems to work quite well. I wonder if you would want to expand it to living people: Wikidata:Database reports/identical birth dates (living people)/1

On Wikidata talk:Database reports/identical birth and death dates/1, there is a SPARQL query that found some.

Maybe as exclusion we could attempt to skip those that have different nationalities (P27) and occupations (P106). So adding these properties would get them removed from the list.
--- Jura 08:47, 7 August 2016 (UTC)

The following query uses these:

  • Properties: date of birth (P569)  View with Reasonator View with SQID, date of death (P570)  View with Reasonator View with SQID, instance of (P31)  View with Reasonator View with SQID, sex or gender (P21)  View with Reasonator View with SQID, country of citizenship (P27)  View with Reasonator View with SQID, occupation (P106)  View with Reasonator View with SQID, place of birth (P19)  View with Reasonator View with SQID, father (P22)  View with Reasonator View with SQID, mother (P25)  View with Reasonator View with SQID, Freebase ID (P646)  View with Reasonator View with SQID, position played on team / speciality (P413)  View with Reasonator View with SQID, P7, P9, different from (P1889)  View with Reasonator View with SQID, said to be the same as (P460)  View with Reasonator View with SQID
    SELECT DISTINCT ?item1 ?item2  ?d1
    {
      	?item1 wdt:P569 ?d1 .
      	FILTER (?d1 > "1980-01-01T00:00:00Z"^^xsd:dateTime && ?d1 < "1985-01-01T00:00:00Z"^^xsd:dateTime) .
      	OPTIONAL { ?item1 wdt:P570 ?dummy1 } FILTER ( !bound(?dummy1) ) .
      	?item2 wdt:P569 ?d1 .
      	OPTIONAL { ?item2 wdt:P570 ?dummy2 } FILTER ( !bound(?dummy2) ) . 
      	?item1 wdt:P31 wd:Q5 .
      	?item2 wdt:P31 wd:Q5 .
      	?item1 wdt:P21 ?g .
      	?item2 wdt:P21 ?g .
    	OPTIONAL { ?item1 wdt:P27 ?l1 }		OPTIONAL { ?item2 wdt:P27 ?l2 }		FILTER ( !( BOUND(?l1) && BOUND(?l2) && ?l1 != ?l2) )
    	OPTIONAL { ?item1 wdt:P106 ?n1}		OPTIONAL { ?item2 wdt:P106 ?n2}		FILTER ( !( BOUND(?n1) && BOUND(?n2) && ?n1 != ?n2) )
      	OPTIONAL { ?item1 wdt:P19 ?b1 }		OPTIONAL { ?item2 wdt:P19 ?b2 }		FILTER ( !( BOUND(?b1) && BOUND(?b2) && ?b1 != ?b2) )
      	OPTIONAL { ?item1 wdt:P22 ?f1 }		OPTIONAL { ?item2 wdt:P22 ?f2 }		FILTER ( !( BOUND(?f1) && BOUND(?f2) && ?f1 != ?f2) )
      	OPTIONAL { ?item1 wdt:P25 ?m1 }		OPTIONAL { ?item2 wdt:P25 ?m2 }		FILTER ( !( BOUND(?m1) && BOUND(?m2) && ?m1 != ?m2) )
      	OPTIONAL { ?item1 wdt:P646 ?z1 }	OPTIONAL { ?item2 wdt:P646 ?z2 }	FILTER ( !( BOUND(?z1) && BOUND(?z2) && ?z1 != ?z2) )
      	OPTIONAL { ?item1 wdt:P413 ?w1 }	OPTIONAL { ?item2 wdt:P413 ?w2 }	FILTER ( !( BOUND(?w1) && BOUND(?w2) && ?w1 != ?w2) )  
      	FILTER ( str(?item1) < str(?item2)  )
      	MINUS { ?item1 wdt:P7 ?item2 }		MINUS { ?item1 wdt:P9 ?item2 }		# twins
      	MINUS { ?item1 wdt:P1889 ?item2 }	MINUS { ?item1 wdt:P460 ?item2 }	# identified issues
    }
    LIMIT 50
    


I did some tests with SPARQL (see above). It seems that this could generate a lot of items. Items that have nothing more than P569/P21 turn out a lot of results. Maybe this is suitable for a game instead: add a statement to each item with a property that differentiates them.
--- Jura 08:23, 17 August 2016 (UTC)

Constraint violation summary

Hi! Maybe it would be possible to create something like one constraint violation summary page for all props? Section for P1146 it would return (in table format or list format):

  • P1146 - "Format" violations - 0
  • P1146 - "Unique value" violations - 0
  • P1146 - "Single value" violations - 18 (+ 2 exceptions)
  • etc.

It would be probably useful to know, which props and constraint violations has the biggest problems. I'm not saying, that I would personally be very active, but if it's pretty trivial to create such report (and I don't see anything very hard there, but I may be wrong), then why not? --Edgars2007 (talk) 09:32, 13 August 2016 (UTC)

Update frequency of Wikidata:Database reports/Constraint violations/Mandatory constraints/Violations

It seems to take very long to get this page updated every day, can you explain? I would love to clean some constraints violations, but it is full of old values. Sjoerd de Bruin (talk) 19:55, 16 August 2016 (UTC)

Wikidata:Database reports/Constraint violations/P625

There is a timeout at Wikidata:Database reports/Constraint violations/P625. I already proposed, that single value constraint may be omitted or limited, that can ease the calculation. --Jklamo (talk) 08:27, 17 August 2016 (UTC)

Constraint violations report

Добрый день! Иван, не могли бы вы в отчётах "Single value" violations около каждого из значений показывать его ранг? Как правило в таких случаях, что показывать редактор указывает именно рангом. См., например, Q25939342 попавший в Wikidata:Database reports/Constraint violations/P1973. -- Sergey kudryavtsev (talk) 07:00, 27 August 2016 (UTC)

PS: Можно использовать иконки вроде s:ru:Файл:Statement prefered rank icon.gif, s:ru:Файл:Statement normal rank icon.gif и s:ru:Файл:Statement not recommended rank icon.gif. -- Sergey kudryavtsev (talk) 07:08, 27 August 2016 (UTC)

  • Добрый день, сейчас бот не зачитывает ранги совсем, сделать конечно можно, но определённые трудности есть. Кстати можно немного по-другому поступить: для устаревшего ID указать какой-нибудь квалификатор, вроде "дата деактуализации" и это свойство указать в параметре "separators" шаблона {{Constraint:Single value}}. Ещё, как вариант, можно не пытаться коллекционировать устаревшие и невалидные идентификаторы, а просто удалять их, так делается для VIAF, GND. — Ivan A. Krestinin (talk) 11:48, 28 August 2016 (UTC)
    А что за параметр "separators"? Как он действует? -- Sergey kudryavtsev (talk) 03:54, 30 August 2016 (UTC)
    Всё не хватает времени задокументировать его. В этом параметре перечисляются свойства-квалификаторы. Если в каком-то элементе два значения, но они отличаются одним из указанных квалификаторов, то ограничение {{Constraint:Single value}} пропустит такой элемент. — Ivan A. Krestinin (talk) 20:14, 1 September 2016 (UTC)
    Да, полезная возможность. А отсутствие квалификатора, перечисленного в separators будет считаться особым значением, отличным других? Судя по Wikidata:Database reports/Constraint violations/P947#Single value это не так, поэтому я добавил там в separators ещё и excluding (P1011), посмотрим как ваш бот отработает это ограничение. -- Sergey kudryavtsev (talk) 08:58, 15 September 2016 (UTC)

Вы не могли бы заставить бота обновить Wikidata:Database reports/Constraint violations/P947, а то он последний раз несколько месяцев назад обновлялся (2016-07-17)... -- Sergey kudryavtsev (talk) 11:45, 23 September 2016 (UTC)

Спасибо, что обновили этот отчёт, но похоже, что ваш бот не воспринял параметр |separators=P407, P1011, который вы рекомендавали выше. Например, в отчёт попал Galileo Galilei (Q307), у которого значения отличаются значениями квалификаторов. -- Sergey kudryavtsev (talk) 17:40, 24 September 2016 (UTC)
Оба значения должны иметь квалификатор одного типа и значения этого квалификатора должны отличаться. Возможно в данном случае правильнее будет вместо задания квалификатора P1011, создать специальный элемент с заголовком "все языки, кроме русского" и выставлять его в качестве значения свойства P407. — Ivan A. Krestinin (talk) 12:07, 25 September 2016 (UTC)
По-моему отсутствие квалификатора должно считаться отдельным значением, отличным ото всех других (как значение NULL в базах данных). Или, если это сделать технически сложно, значением somevalue (неизвестно). А сейчас бот работает так, будто отсутствие квалификатора — это значение, равное любому другому. -- Sergey kudryavtsev (talk) 18:49, 27 September 2016 (UTC)

Units redir

Hi Ivan. Q15639371 was redirected to Q13147228, but is still used in units (538 times). Would you consider adding this to your bot or shall I ask at WD:Bot requests?
--- Jura 16:40, 29 August 2016 (UTC)

KrBot - handling unknown values

Hi again, I have a question about KrBot's reports. For Slovene Cultural Heritage Register ID (P1587) (report here), I implemented a change, entering heritage designation (P1435) of some objects with the system setting unknown value. This is on purpose, the registry actually has some objects that have officialy "unknown status" as monuments. The question is, what to put on the property's description page so that the bot will label this as an accepted value? As you can see in the report now, it labels them as violations and "somevalue". — Yerpo Eh? 07:32, 2 September 2016 (UTC)

Thanks, this works. The label "somevalue" is a little confusing, but I guess it's not so easy to modify that. — Yerpo Eh? 12:31, 9 September 2016 (UTC)

Value types

Hi, somehow the value type violations on P3096 seem odd. There is a similar report on project chat. You might want to look into it.
--- Jura 04:07, 9 September 2016 (UTC)

Something went wrong with dumps?

Hi! Yesterday the bot made such update, lowering number of items, that has this prop. Excluded was also for example Luisa Mercado (Q24701033), that hasn't been changed for a while. And today the number of items hasn't been restored to approx. previos number, so I assume this problem is still actual. This prop of course isn't the only one, there was some weirdeness on other covi pages, too. --Edgars2007 (talk) 05:11, 9 September 2016 (UTC)

Удаление прежних кодов ОКТМО

Добрый вечер, Иван! Не могли бы пояснить, почему вы удаляете устаревшие коды ОКТМО? В пояснении написано, что Marking erroneous statements as deprecated instead of simply deleting such statements has three benefits. Игорь Темиров (talk) 19:50, 9 September 2016 (UTC)

Wikidata:Database reports/Constraint violations/P625 [2]

Hi, there was no update of that constraint since 11 September. It may be related to addition of constraint (here). --Jklamo (talk) 07:13, 14 September 2016 (UTC)

If I'm the cause, then I'm sorry and we can revert that edit of mine. Didn't intentionally broke update :) But isn't there some way to generate value statistics table for P625? At least, a one-time update somewhere else, not on covi page. I think it would really be interesting to see, what the situation is. --Edgars2007 (talk) 08:03, 14 September 2016 (UTC)
Maybe not, it just mine hypothesis. I think your table may be crated using query, maybe this is what you looking for (autolist query, as I am still not familiar enough with sparql). --Jklamo (talk) 11:38, 14 September 2016 (UTC)
No, using WDQ is not possible to get the table of value statistics. changed the header, so one can get to right section --Edgars2007 (talk) 15:26, 14 September 2016 (UTC)
The report is too large now. I fixed error processing code to highlight the issue. But I need to think about working with too large report. — Ivan A. Krestinin (talk) 12:25, 25 September 2016 (UTC)

SetReference

При изменении ОКТМО добавление ботом нового ОКТМО происходит без проблем, а вот при добавлении к нему ссылки пишет, что такого claim нет. Должно пройти какое-то время (не засекал, но не меньше 10 минут) прежде чем он даёт добавить добавить ссылку. У вас это происходит одномоментно. Не подскажете, в чём здесь загвоздка? Игорь Темиров (talk) 05:18, 15 September 2016 (UTC)

  • Не знаю, впрочем я уже какое-то время не редактировал ничего, возможно с каким-нибудь очередным обновлением у меня тоже будут проблемы. — Ivan A. Krestinin (talk) 12:09, 25 September 2016 (UTC)

Wikidata:Database reports/Constraint violations/P1792

The "target required claim" reports on Wikidata:Database reports/Constraint violations/P1792 seem to be wrong. Basically all items with a category of associated people (P1792) claim are listed. Strangely, on Wikidata:Database reports/Constraint violations/Mandatory constraints/Violations the report length for P1792 is shorter but there still many false positives. --Pasleim (talk) 23:56, 20 September 2016 (UTC)

Thanks. --Pasleim (talk) 14:42, 25 September 2016 (UTC)

KrBot changing halfbrother to martial half-brother

See the discussion in the Project Chat: https://www.wikidata.org/w/index.php?title=Wikidata:Project_chat&action=view&section=37 ChristianKl (talk) 16:26, 27 September 2016 (UTC)

Update stuck?

Identical dates hasn't been updated.
--- Jura 09:23, 29 September 2016 (UTC)

Bot had some troubles with Q16086943. Now all is ok. It is difficult to identify the issue now. — Ivan A. Krestinin (talk) 10:23, 2 October 2016 (UTC)
Three items that became one? I don't think I have had that before. Thanks for having looked into it.
--- Jura 10:27, 2 October 2016 (UTC)
Is it possible that the main constraints reports have gotten so long to run that it just doesn't get to update this? Is there something we can do to limit the main constraint reports? Maybe for some of the constraints, it's not necessary to compute the violation and leave it to Special:ConstraintReport?
--- Jura 08:02, 6 October 2016 (UTC)
Would you check it once more?
--- Jura 15:33, 11 October 2016 (UTC)
Fixed, constraint report task took too much time. — Ivan A. Krestinin (talk) 13:50, 15 October 2016 (UTC)

Database Constraint report for 3217

Should I do something to get the constraint violation report starting to run? - Salgo60 (talk) 13:37, 4 October 2016 (UTC)

No, you did everything you had to. That page will be updated in next 24 h. --Edgars2007 (talk) 13:42, 4 October 2016 (UTC)

Single value constraint

For some identifiers we had deprecated values. Some contributors like to add these old data but this is a problem for the constraint Single value often used with identifiers. Do you know if it is possible to consider only value with a normal rank when perform the Single value check ? Thanks Snipre (talk) 07:28, 24 October 2016 (UTC)

Or rather, not to consider deprecated rank (since preferred rank is also possible @Snipre:). --Izno (talk) 12:01, 24 October 2016 (UTC)
@Izno: If we assume that an identifier is unique then we can't choose between several values so the preferred rank should not be used. Again the single value constraint implies that you have only one valid value at a certain date. Using preferred rank and deprecated rank only is not correct in this case. Preferred rank should be used only when you have several normal rank values. This is not the case here. Snipre (talk) 12:15, 24 October 2016 (UTC)
@Snipre: My point is that only "deprecated" should be excluded if there are multiple claims with a single value constraint. In other words, "preferred" should also produce a constraint error, if there are normal ranks also present. --Izno (talk) 12:18, 24 October 2016 (UTC)
@Izno: Ok with you if your "possible" is a possibility but not the acceptance as a normal case. Snipre (talk) 14:33, 24 October 2016 (UTC)
Current bot implementation ignores ranks at all. You can use parameter "separators=start time (P580), end time (P582)". Collecting deprecated identifiers is not looked as very good idea. The identifiers can be unverifiable. Are you know data client who need this information? Deprecated values are need to be processed in all data clients. Are you want to import all deprecated values to Wikidata? — Ivan A. Krestinin (talk) 20:46, 24 October 2016 (UTC)
The objective is first to know if it is possible and what are the technical consequences. We started a discussion in Project Chemistry about one identifier, the CAS number. We don't have a lot of deprecated values per item so it doesn't concern hundred thousands values.
The problem of the source is not a real problem: most of databases about chemicals using CAS numbers already implement some kind of redirection for deprecated values because CAS number is a worldwide identifier used in research as well than in the chemicals trading. So most of use still keep a trace of old values. Then the idea is not to import all possible deprecated values for CAS number because the original database is a proprietary database. So we have access only to the CAS number of the chemicals which the most use in the economy.
This idea is not supported by 100% of the contributors who took part to the discussion so the question is only technical. Snipre (talk) 19:41, 26 October 2016 (UTC)

Some structures of identifiers seem to escape the single value constraint; for example, HMDB changed it's ID structure from HMDBxxxx to HMDB00xxxxx. If one wikidata ID has the old structure, and another the new structure, this seems to go unnoticed by the single value constraint. See for example https://www.wikidata.org/wiki/Q24745328 and https://www.wikidata.org/wiki/Q27265681 (for the latter I removed the HMDB ID). Do you have any solution for this? Perhaps add a warning if the old ID structure is filled into Wikidata (like CHEBI:yyyyy is no longer an option, the "CHEBI:"-part should be removed).... DeSl 15:50, 22 August 2018 (UTC)

You can detect old format using format constraint (Q21502404) and fix it using {{Autofix}}. — Ivan A. Krestinin (talk) 19:02, 22 August 2018 (UTC)

Bn description

Hi, Currently your bot (KrBot) automatically adds description ("Wikimedia category") in 8 languages. Is it possible to add bengali description ("উইকিমিডিয়া বিষয়শ্রেণী") also? --Aftabuzzaman (talk) 15:57, 26 October 2016 (UTC)

Is it also possible to add hebrew? The description will be ("דף קטגוריה")--Mikey641 (talk) 17:43, 26 October 2016 (UTC)
This algorithm is not very efficient now. My bot adds every description using separate edit. I saw bots that add many descriptions using single edit. Please ask another botmasters. — Ivan A. Krestinin (talk) 19:09, 15 February 2017 (UTC)

Here your Bot produced an error. Steak (talk) 20:41, 2 November 2016 (UTC)

External use bot

Hi, Ivan. I noticed that your bot is adding to "external use" template usage on ruwiki. I appreciate that as it is important to have that template as complete as possible. Do you have a script for that and it is possible to extend that to other wikis? --Jklamo (talk) 08:54, 3 November 2016 (UTC)

+1 to Jklamo request. It would be great to have a similar bot (or as script) to run also in hewiki. Eran (talk) 19:17, 14 November 2016 (UTC)
The code is dump-based and resource-consuming. It is not simple to scale it to all wikies. Do you know some service or API that allows to search {{#property... like constructions in all wiki templates? — Ivan A. Krestinin (talk) 19:45, 18 November 2016 (UTC)
Why not mw:API:Search? There you can search also by regex. --Edgars2007 (talk) 02:41, 25 December 2016 (UTC)
@Edgars2007: Sorry, I miss your post. mw:API:Search does not allow to search in page wiki-text. It searches in generated text. Try to find {{#property: construction. — Ivan A. Krestinin (talk) 12:31, 15 January 2017 (UTC)
There is insource: command for that, I thought you know about it, that's why I didn't mention it. See this example in API sandbox. --Edgars2007 (talk) 12:45, 15 January 2017 (UTC)
I think that https://quarry.wmflabs.org is able to search specific text in wiki templates.--Jklamo (talk) 12:43, 15 January 2017 (UTC)
No, page wikitext isn't stored in SQL databases (that means, it's not available in quarry.wmflabs.org). --Edgars2007 (talk) 12:45, 15 January 2017 (UTC)
Thank you for "insource" keyword. Bot processed the most projects. Please notify me if bot miss something. — Ivan A. Krestinin (talk) 12:41, 23 January 2017 (UTC)
The thing bot "missed" is using other constructions than #property: to get data. You come from ruwiki, which is an excellent example. And then there are those Lua modules :) Much fun. Just loud thinking, no complains. --Edgars2007 (talk) 08:48, 26 January 2017 (UTC)
Also not complain, but profound thanks. What about deleted properties, did you process them as well? If not, I am very interested at least in P7 (P7) and P9 (P9).--Jklamo (talk) 09:10, 6 February 2017 (UTC)
Please provide examples of missed usages. Bot processed deleted properties too. But it has troubles with {{ExternalUse}} adding. Bot does not recreate property talk pages. As for P7/P9 bot failed to find appropriate place to the template. I added it manually and will restart bot. — Ivan A. Krestinin (talk) 09:37, 6 February 2017 (UTC)

Why?

Why did you do this edit? /ℇsquilo 10:21, 7 November 2016 (UTC)

Because the claim is exact dublicate of another claim for this property at that item. Read the edit summary. --Edgars2007 (talk) 10:40, 7 November 2016 (UTC)

Constraint violation reports

Your usually helpful bot hasn't updated any of the Wikidata:Database reports/Constraint violations/P* pages since 5 November. Is there an issue? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 15:21, 9 November 2016 (UTC)

Just adding myself to the list of those who miss constraint violations updates.--Jklamo (talk) 00:22, 14 November 2016 (UTC)
+1. Lymantria (talk) 14:29, 14 November 2016 (UTC)
It's a pity that wikidata depends in a critical functionality on one user. Why cannot the Foundation run a reliable bot? Steak (talk) 20:07, 14 November 2016 (UTC)
Really, Steak? This outage is unusual and I hope Ivan is well. --Succu (talk) 21:34, 14 November 2016 (UTC)
Steak! Most (but not all) constraint reports can be seen if you follow the SPARQL-link in the constraints templates. Having a set of at least two bots for the most critical reports is not a bad idea at all. -- Innocent bystander (talk) 11:58, 15 November 2016 (UTC)
Not all reports are "SPARQLed" and there is anther added value of daily reports - if you are checking reports daily there is possibility to compare daily reports and detect bad imports or newbie users edits. --Jklamo (talk) 13:15, 15 November 2016 (UTC)
I had bad troubles with hardware. Now all is fine and bot will update the reports during next 5-10 hours. Pasleim reimplement the bot as I see. So the system is reserved now. Thank you, Pasleim. — Ivan A. Krestinin (talk) 19:50, 18 November 2016 (UTC)
That are good news. Thanks a lot. --Succu (talk) 20:55, 18 November 2016 (UTC)
Thanks much, Ivan and Pasleim!! Lymantria (talk) 10:07, 20 November 2016 (UTC)
Good to hear. Thank you. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 15:10, 20 November 2016 (UTC)

Thank you

Hello, this is amazing. I made a mistake and am unable to fix it using tools that I can control. There are identical problems in most Czech municipalities. Could you run through all of them and do the same? Thank you, --Vojtěch Dostál (talk) 14:16, 16 November 2016 (UTC)

Could you provide some unfixed item? — Ivan A. Krestinin (talk) 19:52, 18 November 2016 (UTC)

Publish the source of KRBot please

Hi Ivan, you're running a bot without it's source published. You vanished and the bot broke down so nobody could recover it. Please publish your source somewhere (https://github.com/?) as soon as possible. No matter that it might be dirty, clean up needed, etc etc. Just dump online somewhere and add the link to User:KrBot. Pasleim now spend(/wasted) a lot of time on trying to recover the service, I would hate to see that happen again. Multichill (talk) 10:50, 21 November 2016 (UTC)

Hi Ivan, it's been a month since I posted this message. It would be nice if you could respond. Multichill (talk) 09:14, 21 December 2016 (UTC)
It looks like the response is no. I think similar code is already in Mediawiki (Special:Constraint report), it's just that WMDE doesn't want to work on it.
--- Jura 11:24, 21 December 2016 (UTC)
Let's not do assumptions and have Ivan respond. Multichill (talk) 11:29, 21 December 2016 (UTC)
The code is written on C++ and uses custom libraries. It does not have any documentation. It has unused parts, for example for property-described constants. The reports update algorithm is resource consuming. It eats 32 Gb of operating memory. Do you really need this code? And I will publish the code in case of my exit from the project of course. — Ivan A. Krestinin (talk) 18:13, 21 December 2016 (UTC)
Sounds like you forked wdq. ;-)
The current bus factor (Q1812883) is one so yes, you should definitely publish it somewhere.
I don't want you to leave, but if you do decide to take a long holiday, at least the code is available to fork. It's also nice for other people who are re-implementing parts to be able to have a look at existing code how it's solved there. Multichill (talk) 22:39, 21 December 2016 (UTC)
 Support would be great. please! --Marsupium (talk) 13:31, 15 March 2019 (UTC)

Hi Ivan, I concur with Multichill's comments. It would really help if you could publish your code. There are a lot of benefits to that - for instance, people could propose new implementations for the constraints they want to add support for in KrBot. It would also be possible for people to know which constraint is supported, and how. − Pintoch (talk) 14:25, 13 June 2018 (UTC)

This required too many efforts from my side. — Ivan A. Krestinin (talk) 15:31, 13 June 2018 (UTC)

Deception

If you look at this edit, everything seems fine. But what you don't see in the difference is that the bot moved one claim from a slot that had no references into a slot that had two references. The edit therefore implies that the newly moved value is the one with references. This is deceptive, as the new claim value was not the one to which those references apply. --EncycloPetey (talk) 18:50, 11 December 2016 (UTC)

Identifier 179917357 is equal to 101760867. Both values identify single record in VIAF database. So reference applied to one value can be applied to the second value. — Ivan A. Krestinin (talk) 12:47, 12 December 2016 (UTC)

Unmerge

Откатите, пожалуйста, перенаправления типа этого. --Infovarius (talk) 21:28, 20 December 2016 (UTC)

И эти. --Infovarius (talk) 21:40, 20 December 2016 (UTC)
✓ DoneIvan A. Krestinin (talk) 04:35, 21 December 2016 (UTC)
Может, где-то есть страница запросов? :) Вот ещё: [70]. --Infovarius (talk) 13:00, 26 December 2016 (UTC)
✓ DoneIvan A. Krestinin (talk) 05:13, 27 December 2016 (UTC)

Batch rollback request

Your bot changed a large number of instance of (P31)MediaWiki configuration setting (Q21481290) to instance of (P31)Manuel Candamo (Q170059), where it doesn't know the merge action of Q21481290 is invalid. --Liuxinyu970226 (talk) 01:03, 24 December 2016 (UTC)

Would you do the same for Gottfried (Q16746939) > Q3566065. Thanks.
--- Jura 14:56, 26 December 2016 (UTC)

✓ DoneIvan A. Krestinin (talk) 05:13, 27 December 2016 (UTC)

Change the external link for CAS number

Hi. Just to inform you that the tool that lists all URL with a CAS number in order to access external databases is dead. Can you please please the links in Wikidata:Database_reports/Constraint_violations/P231 by https://chem.nlm.nih.gov/chemidplus/rn/* where * is the CAS number ? So https://tools.wmflabs.org/magnustools/cas.php?cas=63800-37-3 will become https://chem.nlm.nih.gov/chemidplus/rn/63800-37-3. Thank you. Snipre (talk) 20:36, 9 January 2017 (UTC)

@Snipre: the only thing you have to do is change formatter URL (P1630) at CAS Registry Number (P231) ($ there is kind of the same thing as * in your example here) --Edgars2007 (talk) 06:06, 10 January 2017 (UTC)
Edgars2007 Thank you. Snipre (talk) 09:07, 10 January 2017 (UTC)

Q27891162

Regarding Q27891162 I do not think that it is true, that it has a P:366 as Q949149

best regards to your bot, Scoid d (talk) 17:04, 20 January 2017 (UTC)

Bot in hollyday?

https://www.wikidata.org/w/index.php?title=Wikidata:Database_reports/Constraint_violations/P345&action=history -- MovieFex (talk) 16:45, 25 January 2017 (UTC)

KrBot {{ExternalUse}}

Hi. Thanks for maintaining KrBot. It seems that the bot hasn't been indexing property uses on Cantonese Wikipedia. Does it need to be set up separately for each wiki? Deryck Chan (talk) 15:40, 6 February 2017 (UTC)

Hello, bot processes all wikies. But it can miss some unknown constructions. Could you provide links to missed templates? — Ivan A. Krestinin (talk) 15:44, 6 February 2017 (UTC)
It seems that the template w:zh-yue:Template:英格蘭郡明細 (Infobox English county) is only indexed by Property talk:P625 (which ironically doesn't work, because zh-yue has a different implementation of {{coord}} that hasn't been ironed out yet). Your bot's latest batch of edits didn't index that infobox on any of the other properties it uses. Deryck Chan (talk) 11:31, 7 February 2017 (UTC)
I see that the latest run has indexed my template. Is there anything that needs to be done to ensure future uses of Wikidata by zh-yue.wp continues to be done properly? Deryck Chan (talk) 11:06, 8 February 2017 (UTC)
Different projects use different constructions to call Wikidata. Ruwiki uses {{Wikidata|P18}}, zh-yuewiki uses {{#invoke:wikidata|getvalue|P18}}. These constructions are need to be processed separately. So bot can miss some another call types. Just notify me if bot miss something. And I will add new calling convention to bot`s algorithm. — Ivan A. Krestinin (talk) 21:23, 8 February 2017 (UTC)
I wonder if syncing calls of something like w:Template:Uses Wikidata - and its equivalent in other projects - might be a better way of indexing uses of Wikidata. Eventually we'll need a central register of how each wiki calls Wikidata - it will be very useful if KrBot's input dataset (the constructions in each wiki) are listed on an editable wikipage. Deryck Chan (talk) 11:14, 9 February 2017 (UTC)

Rollback request

The merge from Murakami (Q21494521) to Murakami (Q1953526) was wrong. Could you please undo edits like this? --Pasleim (talk) 10:16, 10 February 2017 (UTC)

The merge from Juri (Q18932166) to Yuri (Q4995000) was wrong. Could you please undo edits like [71]? --Pasleim (talk) 17:04, 18 February 2017 (UTC)

KrBot problem with ExternalUse

Hi KrBot has duplicated several times ExternalUse for itwiki, examples Property talk:P3109, Property talk:P3513, Property talk:P3442. Could you check your work and fix it? Thanks in advance. --Rotpunkt (talk) 13:50, 11 February 2017 (UTC)

Mhm, the problem is in a lot of pages, you should fix your errors, reverting your bot. --Rotpunkt (talk) 20:54, 12 February 2017 (UTC)
Fixed. — Ivan A. Krestinin (talk) 19:05, 15 February 2017 (UTC)

Your feedback matters: Final reminder to take the global Wikimedia survey

(Sorry to write in Engilsh)

External use

There is some reason to add no existing template in Externale use like in P106? --ValterVB (talk) 20:06, 24 February 2017 (UTC)

✓ DoneIvan A. Krestinin (talk) 21:54, 24 February 2017 (UTC)

Do not delete the identifier

Don't delete the identifier IMDb without correcting the entry at the Wiki-Article. There is Wikidata:Database reports/Constraint violations/P345 and it will be fixed, all inclusive. --MovieFex (talk) 02:51, 25 February 2017 (UTC)

Talk pages of constraint reports

When your bot creates a report page like Wikidata:Database reports/Constraint violations/PNNNN, please can it then create a redirect from the corresponding talk page, Wikidata talk:Database reports/Constraint violations/PNNNN, to the talk page for the property, Property talk:PNNNN, where those constraints are set? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 17:31, 27 February 2017 (UTC)

Wow, I just saw the redirect creations, that a very good idea! Thank you Pigsonthewing and Ivan! Cdlt, VIGNERON (talk) 20:51, 1 March 2017 (UTC)

qualifier → reference

Здравствуйте! Спасибо за помощь,

Thank you for making corrections where I should be using the 'reference' field and not 'qualifier', I agree.

Kind regards, --2015.ww (talk) 05:25, 7 March 2017 (UTC)

Double entry on published in (P1433)

You bot introduces a double entry on published in (P1433) [72]. — Finn Årup Nielsen (fnielsen) (talk) 13:38, 10 March 2017 (UTC)

Mass revert

Hi, could you mass revert somehow your bot's "genre: drama" edits? Two items were merged that shouldn't have been. – Máté (talk) 13:24, 17 March 2017 (UTC)

✓ DoneIvan A. Krestinin (talk) 07:24, 18 March 2017 (UTC)

External use

Hey, based on what did you do that can u do it again for all properties in hewiki? btw: in hewiki we use a template that indicates what properties are used for example he:תבנית:קו רכבת תחתית. thanks--Mikey641 (talk) 11:36, 19 March 2017 (UTC)

Reaction on WD:DNM

Hi Ivan, I left a message on Wikidata talk:Do not merge where I pinged you. Could you please have a look at it? Q.Zanden questions? 13:44, 21 March 2017 (UTC)

KrBot not updating constraint violation reports

Hi Ivan, I noticed that KrBot did not update the constraint violation reports yesterday. Is everything working? Cheers. − Pintoch (talk) 09:14, 22 March 2017 (UTC)

Hello, bot started update process some hours ago. It had some troubles, but all is ok now. — Ivan A. Krestinin (talk) 21:05, 22 March 2017 (UTC)

Item labels on constraint report

Hi there, firstly thanks for running the bot for the constraint report, I've fixed many errors in both wikidata and articles because of it. I've been using Wikidata:Database reports/Constraint violations/P2697 and Wikidata:Database reports/Constraint violations/P2698 and I've noticed that P2698 just has the item numbers (ie Q123456) in the lists, but P2697 shows the item label and number (ie P.G. Wodehouse (Q207515)). Is there a setting somewhere to force the item labels to be shown on all reports? Thanks, The-Pope (talk) 00:45, 23 March 2017 (UTC)

It's because of length of report (if you use {{Q}} too much, you'll get Lua error). Because of relatevily new updates in Module:Wikidata Ivan may reevaluate the threshold, if it's set in bot's code. --Edgars2007 (talk) 13:56, 23 March 2017 (UTC)
Hello, you are right. Current threshold is 450 items. I increase it to 2000. Lets wait the next update. — Ivan A. Krestinin (talk) 18:28, 23 March 2017 (UTC)

Can't your bot do that in one edit? --Leyo 08:04, 6 April 2017 (UTC)

Three copies of the link

After this edit, Euripides (Q48305) now has three identical VIAF links to the same target. --EncycloPetey (talk) 23:36, 6 April 2017 (UTC)

Help with a regex for IPv4 routing prefix (P3761)

Hi again Ivan, Thanks for your fix to the bot last month!

I have tried to fix the regex for the format of IPv4 routing prefix (P3761). It seems to be correct now (as a PCRE), and the web interface accepts well-formed edits, but it looks like KrBot considers that all the statements have an incorrect format. I think the slash was not properly escaped before, which is now fixed. Maybe the regular expressions are cached somewhere and we need to update that cache too? Have a good week-end! − Pintoch (talk) 22:45, 7 April 2017 (UTC)

Updated VIAF

I have done this in connection with an update of VIAF. The "correct" number was already there. Maybe the bot should check if the number is already there and then maintain the old number and deprecate it? Or do you have any other opinion? — Finn Årup Nielsen (fnielsen) (talk) 17:15, 9 April 2017 (UTC)

Eventually it removes that duplicate.
--- Jura 17:19, 9 April 2017 (UTC)
✓ Done Bot will fix duplicates immediately after appearing. — Ivan A. Krestinin (talk) 05:21, 10 April 2017 (UTC)

Удаление информации

Чего это бот такое делает: [73] ?ShinePhantom (talk) 03:31, 19 April 2017 (UTC)

В коде бота был баг, большинство правок вызванных этим багом я откатил, но вот эту похоже упустил. — Ivan A. Krestinin (talk) 04:26, 19 April 2017 (UTC)

Удаление Административно-территориальных единиц

Добрый день! Обратил внимание на несколько правок, сделанных Вашим ботом в части удаления из населённых пунктов и административных образованиях информации об административно-территориальных единицах, как упразднённых, так и ещё вполне действующих. Данная информация имеет значение для викиданных, просьба не удалять её. В частности в русской Википедии от этого зависит корректное заполнение карточек персон, в противном случае получается, что-то вроде Иванов родился в 1781 году в Нижегородской области Российской Федерации. И если можно, прокомментируйте, пожалуйста, правки Вашего бота. Хотелось бы понять масштаб проблемы: это было разовое действие или правка, осуществляемая на регулярной основе? Речь идёт только о России или о других странах тоже? Можно ли её отменить (вернуть как было)? --Ksc~ruwiki (talk) 19:28, 19 April 2017 (UTC) P.S. Прошу прощения, не заметил, что коллега уже затрагивал эту тему, но если можно ответьте на вопросы, чтобы у меня было понимание происходящего. --Ksc~ruwiki (talk) 19:30, 19 April 2017 (UTC)

По поводу элементов, где приведена история принадлежности объекта различным образованиям, где указаны квалификаторы start time (P580)/end time (P582) был баг в коде бота, такие элементы он трогать не должен был. Подобные правки, насколько я вижу, все откатили. Второй случай другой: если упрощённо, то раз Arzamas (Q103789) входит в Arzamas Urban Okrug (Q27546988), а Arzamas Urban Okrug (Q27546988) входит в Nizhny Novgorod Oblast (Q2246), то нет смысла дополнительно указывать, что Arzamas (Q103789) входит в Nizhny Novgorod Oblast (Q2246). Для России на Викиданных сейчас такая модель используется, про другие государства не знаю. Хотя существуют и другие модели. Например, модель чисто-административного деления. Или модель чисто-территориального деления. — Ivan A. Krestinin (talk) 19:55, 19 April 2017 (UTC)
Смысл есть на самом деле. Дело в том, что Arzamas Urban Okrug (Q27546988) сравнительно новое образование и Nizhny Novgorod Oblast (Q2246) для Arzamas (Q103789) всё равно придётся добавлять в отношении периода более раннего, чем 2005 год, конечно, проставляя квалификаторы. Страшного в таком удалении возможно и нет, но лишнюю работу делать никому не хочется. P.S. Спасибо, что быстро и понятно ответили. --Ksc~ruwiki (talk) 20:52, 19 April 2017 (UTC)

Constraint report P274 (chemical formula): error code 'badmd5'

Hi. It seems the constraint report for chemical formula has been failing since September with the message "Edit failed, error code 'badmd5' (The supplied MD5 hash was incorrect.)." Prior to this, it was untouched for a long time -- the last successful KrBot report appears to be April 2015. Do you know what causes this? Perhaps some differences or weirdness in Unicode handling. Seems like something that needs to be fixed in User:KrBot. Intgr (talk) 15:17, 24 April 2017 (UTC)

Hello, the issue is caused by symbol with UTF-8 code 0xCE 0x87. MD5 checksum is b80d5a5d9193d69ce0b1009c31587da5, but it is rejected by https://www.wikidata.org/w/api.php?action=edit&bot=1&assert=bot&format=json&utf8=true&md5=b80d5a5d9193d69ce0b1009c31587da5&notminor&nocreate&basetimestamp=2017-04-24T17:37:01Z&starttimestamp=2017-04-24T17:37:01Z&title=Wikidata:Database%20reports%2FConstraint%20violations%2FP274 request. I will continue the investigation. — Ivan A. Krestinin (talk) 17:41, 24 April 2017 (UTC)
Posed to Phabricator. — Ivan A. Krestinin (talk) 18:51, 24 April 2017 (UTC)

Source code for redirect fixing script

Hello Ivan, do you have the source code for the script that fixes links to redirects on Wikidata statements? I merge items a lot and this creates a lot of redirects, and I would like to be able to fix them quickly. Thank you, Harej (talk) 03:40, 26 April 2017 (UTC)

Summary

[74]: Removed claim: Commons category (P373): Burston School Strike, несуществующий файл / missing file. Стоит пофиксить :) --XXN, 12:20, 29 April 2017 (UTC)


  • P.S. Вас могло бы заинтересовать этот таск. Даже если Infovarius работает/работал над этим, тут пахнет перманентным заданием. --XXN, 12:27, 29 April 2017 (UTC)
  • Описание поправил, а тот таск мне не по душе :) Не нравится мне самое главное искать глазами где-то в конце метки, зря что-ли во всех энциклопедиях фамилию первой пишут. — Ivan A. Krestinin (talk) 13:01, 29 April 2017 (UTC)

same object

Please, can you verify Q25710438 and Q25188681. It's the same

2A01:CB14:1E0:9900:9930:B901:4A5:2467 09:04, 5 May 2017 (UTC)

Merged. — Ivan A. Krestinin (talk) 20:25, 5 May 2017 (UTC)

Координаты из бевики

Добрый день! Не могли бы Вы вытянуть в ВД координаты из шаблона Славутасць (Храм, Помнік)? --Чаховіч Уладзіслаў (talk) 20:10, 5 May 2017 (UTC)

Please see Project Chat for spacecraft ontology discussion

Please see my comment here on your massive reverting. Thanks [75] - Fuzheado (talk) 01:44, 12 May 2017 (UTC)

Rodovid

Hello !
I am a member of the genealogical wiki Rodovid (FR) and I have seen that on 5 January 2015 KrBot added numerous links to records of our site.
If you are interested, there are other links that could be established towards pages of Rodovid, like the one for Ivan the Terrible.
Rodovid has a complete list of the records (some 30 000) that include a link to a Wikipedia page about the same person.
In Russian, you can find it starting here.
Some of these links are internal redirects to pages of Rodovid, but most are leading to Wikipedia and are beginning with "wikipedia:", "wikipedia:en:", "wikipedia:fr:", "wikipedia:ru:", "wikipedia:uk:" etc.
Normally, one person is identified by a unique number. Different records can be written about one person in several linguistic localizations of Rodovid but they will have the same number (a bit like Wikidata).
(In some temporary cases, duplicate records may have been written about the same person under different numbers.) Penkroff (talk) 22:02, 12 May 2017 (UTC)

Hello, Rodovid uses CC 2.5, WD uses CC0. Are these license compatible and allows data import? — Ivan A. Krestinin (talk) 06:58, 13 May 2017 (UTC)
I asked the question to Y.Boychuk (Baya), developer of Rodovid. Penkroff (talk) 18:00, 13 May 2017 (UTC)
My personal interpretation : CC 2.5 deals about the content, but just adding addresses with numbers probably does not involve any copyright; and the license says it does not prevent "fair use". Penkroff (talk) 23:59, 14 May 2017 (UTC)
Question from Baya on the Rodovid talk page : if these links were possible in 2015, why should the situation be different now ? 23:59, 14 May 2017 (UTC)
Previously bot imported data from Wikipedia. Now you suggest to import data from Rodovid site. — Ivan A. Krestinin (talk) 18:28, 15 May 2017 (UTC)
This list can not be used directly. It has many unreliable records. My bot need account for getting page source. Could you create it? — Ivan A. Krestinin (talk) 18:54, 15 May 2017 (UTC)
Of course, if Baya agrees. (Another procedure for membership is indicated on the Main Page of Rodovid EN.) Penkroff (talk) 01:00, 16 May 2017 (UTC)

Baya likes this idea of cooperation. Just send a message to : en.rodovid@gmail.com (indicating the name you have chosen for your account). Penkroff (talk) 18:00, 17 May 2017 (UTC)

Ok, I sent the email. — Ivan A. Krestinin (talk) 18:37, 17 May 2017 (UTC)
Done (see mail of answer). Penkroff (talk) 22:01, 17 May 2017 (UTC)

Rodovid uses very old Mediawiki version. I can not find a way to retrieve list like [76] via api.php. Could you help with this task? — Ivan A. Krestinin (talk) 07:53, 22 May 2017 (UTC)

Добавление описаний в элементах категорий

Здравствуйте. Не могли бы вы слегка изменить код бота что-бы он при добавлении множественных описаний в элементах категорий сделал бы это всё в одну правку (что-бы не перегрузить излишне историю элемента и списки наблюдения участников)? См. также обсуждение. --XXN, 16:14, 16 May 2017 (UTC)

Приветствую, давно бы это сделал, но несколько проблематично выполнить это пожелание из-за некоторых сложностей архитектурного характера. Вообще в плане категорий мой бот лишь подчищает то, что не сделали другие боты. Я не прочь, если другие боты будут более аккуратными и тогда моему просто не останется работы. — Ivan A. Krestinin (talk) 18:28, 16 May 2017 (UTC)

Hey Ivan, can you please look up how your bot decided to remove the URL in this diff? I don’t see anything wrong with it… Regards, MisterSynergy (talk) 15:52, 21 May 2017 (UTC)

Read the edit summary till end and then take a look at Property talk:P973. --Edgars2007 (talk) 17:20, 21 May 2017 (UTC)
Oh I see. That wasn’t self-explaining … Thanks! —MisterSynergy (talk) 17:36, 21 May 2017 (UTC)

Сколько можно? ... ОКАТО и ОКТМО это про Урал (Буздякский район) Каранский сельсовет (Буздякский район)). А этой деревни уже нет...--User№101 (talk) 19:31, 23 May 2017 (UTC)

Поправил, бот не будет пытаться подобрать коды, если указано свойство dissolved, abolished or demolished date (P576) или в графе "население" значится "0". — Ivan A. Krestinin (talk) 21:12, 23 May 2017 (UTC)

Constraint reports

In the constraint reports for P345 (IMDb ID (P345)) I realized that there are values that shouldn't be reported for the conflicts with report. For example, in the second section of the conflicts with constraint there is Spierig brothers (Q7577128), whose value for the IMDb ID (P345) is no value. This is usually done to prevent some bots from importing the value incorrectly and to indicate that the element doesn't had value for that property. I don't know if you agree or not with this usage of the no value option, but regardless of that the "no value" claim shouldn't be listed as a conflicting value. It's already ignored by the constraint report special page for that entity and it would be nice if it could be ignored by the constraint reports generated by your bot. Could you fix it for future reports? -- Agabi10 (talk) 19:34, 25 May 2017 (UTC)

I do not think that it is good idea to add novalue as workaround for bugs in bot`s code. This will generate unexpected number of =novalue claims. The claims will be similar to garbage. Somebody will add P345=novalue to every group of humans (Q16334295) item, somebody will add it to every taxon (Q16521)... I think bots must be more intellectual. It is very simple to add check constraint like "Conflicts with" to bot`s code. I have one more idea. @Lucas Werkmeister (WMDE): Could we integrate wbcheckconstraints to wbcreateclaim? Now the call supports checks "assert=bot", "assertuser=KrBot". Additional parameter like "assertconstraints=Q21502838,Q21510859" or "assertconstraints=all" will improve quality of the project data. — Ivan A. Krestinin (talk) 20:09, 25 May 2017 (UTC)
Regardless if it's a good workaround or not the entities with novalue shouldn't appear in the constraint report as conflicting values. If we want to prevent a property with novalue set I rather add an additional constraint to generate a report for the claims with novalue set for a given property. Anyway this is only for conflicting constraints, obviously novalue shouldn't be ignored for example in the single value constraints (I've seen today an entity with novalue and other two values for the same property). About the assertion of the constraints I don't really understand what type of checks it would be doing... The first two assertions are quite straightforward, but the last one should it be checking a given constraint type, should only check for the conflict constraint types or should work for any constraint? If the creation of the claim would for example trigger a constraint for a check if the instance is for a given type and the instance is not specified should that claim creation be prevented or not? -- Agabi10 (talk) 20:46, 25 May 2017 (UTC)
I think novalue and inapplicable are different cases. For example I can create claim Oon Chong Hau (Q1426006) IMDb ID (P345) novalue because IMDb collects information about peoples. And this man can have this identifier, but currently he does not have it. Maybe he will have the ID in the future. But claim Oon Brothers (Q7095351) IMDb ID (P345) novalue is incorrect because IMDb does not collect information about sibling group (Q16979650). This item will have the ID never. The first case is novalue case. The second is inapplicable case. {{Constraint:Conflicts with}} is needed for inapplicable case detection. — Ivan A. Krestinin (talk) 08:01, 28 May 2017 (UTC)
@Ivan A. Krestinin: As far as I know, our current plan is to add support for checking constraints on statements that you are about to add (so the user script / gadget would warn you before you click “save”), but I don’t think we would do that as part of wbcreateclaim. But I’ll try to keep this suggestion in mind when we actually get there. --Lucas Werkmeister (WMDE) (talk) 10:40, 29 May 2017 (UTC)

LC Auth ID

According to the property descriptions, this edit is incorrect.

The description of Library of Congress authority ID (P244) says: "Library of Congress ID for authority control for persons, organizations and subject headings (for books use P1144)".

Before your edit, Cape Cod (Q27979486) was using Library of Congress Control Number (LCCN) (bibliographic) (P1144) because it is an instance of a book. You changed it to Library of Congress authority ID (P244) which is for persons, organizations and subject headings. --EncycloPetey (talk) 23:38, 25 May 2017 (UTC)

Bot did such edit, because (according to description at property talk page) it was using inccorect format (Bibliographic records do not have prefixes n, nb, nr, ns, sh, sj, sn). So it automatically moved to Library of Congress authority ID (P244). It was wrong anyway :) --Edgars2007 (talk) 06:42, 26 May 2017 (UTC)
But the change is also wrong. This property should not be used for books. --EncycloPetey (talk) 14:50, 26 May 2017 (UTC)
Ok, lets fix the issue. My bot changed ~60 items. But Library of Congress authority ID (P244) was widely used for books before my bot edits, see [77]. Also format constraint on Property talk:P1144 does not allow codes like n85201132. Lets fix and improve constraints as the first step. Could you fix format constraint, add type constraint and conflict with constraint? I do not understand difference between these properties enough. — Ivan A. Krestinin (talk) 16:27, 26 May 2017 (UTC)
No one understands the difference in these properties! If you look at the talk pages, people keep asking question because they are confused, but no one gives answers. We may need to have a community-wide discussion to get answers. --EncycloPetey (talk) 17:35, 26 May 2017 (UTC)
I added {{Constraint:Type}} for Library of Congress authority ID (P244), please review results: Wikidata:Database reports/Constraint violations/P244#Types statistics. Do we need move some values to Library of Congress Control Number (LCCN) (bibliographic) (P1144) based on item type? — Ivan A. Krestinin (talk) 08:15, 29 May 2017 (UTC)

Selective reflection of constraint violation changes

KrBot's update of Wikidata:Database reports/Constraint violations/P1422 from 2017-05-28T04:41:08 reflects this change from 2017-05-26T18:03:56 but not this change from 2017-05-27T15:27:10. Perhaps that is normal behaviour of KrBot. I just wanted to let you know for the case it is not … Cheers, --Marsupium (talk) 16:01, 28 May 2017 (UTC)

Constraint reports are always 12 hours outdated afaik. Sjoerd de Bruin (talk) 16:16, 28 May 2017 (UTC)
You are right. Details can be found on https://dumps.wikimedia.org/other/incr/ This lag was created for Wikipedia, it allows to hide some changes. Wikidata is a bit different project. I think you can create Phabricator task to remove this lag for Wikidata. — Ivan A. Krestinin (talk) 16:32, 28 May 2017 (UTC)
OK, thank you, good to know! --Marsupium (talk) 18:27, 28 May 2017 (UTC)
Took some time to create the task. Sjoerd de Bruin (talk) 11:11, 31 January 2018 (UTC)

Dublicates on covi reports

Can't show you an example, but when some item has dublicate value for some property, it's get listed in both single value violation section and unique value violation section. I could agree, that it's a single value violation, but don't you think, that it shouldn't get reported as unique value violation? --Edgars2007 (talk) 17:08, 28 May 2017 (UTC)

  • My bot removes such values automatically as one of the tasks. So it appears only once in the reports :) I can filter such values, but this will create dependencies between separate constraints. It will increase bot`s code complexity. Also this dependency will be not obvious for another users. So I prefer do not make such changes. — Ivan A. Krestinin (talk) 21:39, 28 May 2017 (UTC)

Edits with no change

Why did your bot do this? There was no change after it was done. --EncycloPetey (talk) 21:04, 28 May 2017 (UTC)

  • It was manual edit. I experimented with query
    SELECT ?value (STRLEN(STR(?value)) AS ?stringLength) WHERE
    {
      wd:Q3785162 wdt:P950 ?value
    }
    
    Try it!
    I do not understand why stringLength is 10. — Ivan A. Krestinin (talk) 21:48, 28 May 2017 (UTC)

Constraints

Hi, I tried using {{CURRENTYEAR}} in a constraint for date of official closure (P3999), and your bot didn't recognize it. Would it be possible for your bot to accept this value? Thanks, Jc86035 (talk) 14:25, 30 May 2017 (UTC)

Incorrect postal code change

Moved to Property talk:P281

See here and the subsequent revert. —Justin (koavf)TCM 19:51, 5 June 2017 (UTC)

Wrong edit about Reddit

This change is wrong. "AndreaGibsonOfficial" is a username, not a subreddit ID. Robin van der Vliet (talk) (contribs) 12:25, 9 June 2017 (UTC)

Still plenty of VIAF clean up

I see plenty of commentary of incorrect VIAFs from KrBot, and there is still plenty of mistakes to tidy up. Some of the errors are just weird, this is more recent than the last commentary, so just pointing to https://www.wikidata.org/w/index.php?title=Q991872&action=history. It would be useful in additions to know the basis of how the bot applied its addition.  — billinghurst sDrewth 06:04, 12 June 2017 (UTC)

This quite recent edit has picked up wrong data from somewhere, can you work out the origin? I hacked back a range of data to cleanup and reimport (easier than one-to-one matching), and at first blush it seems that only the VIAF data is incorrect, which is what the bot has imported. Can we identify what is going on? We need to stop the error rate of the bot which is problematic in the incorrect application, and a PITA to have to clean up. Thanks.  — billinghurst sDrewth 07:37, 13 June 2017 (UTC)

Not helpful https://www.wikidata.org/w/index.php?title=Q56206&diff=516166225&oldid=489376419 already utilised, maybe you can get the bot to check whether something is used prior to addition. Even if it means creating an error list elsewhere for checking. Forcing a wrong addition is not the right way to go.  — billinghurst sDrewth 12:42, 10 July 2017 (UTC)

https://www.wikidata.org/w/index.php?title=Q788749&diff=516188102&oldid=468413955 with no indicators of why added  — billinghurst sDrewth 12:50, 10 July 2017 (UTC)
https://www.wikidata.org/w/index.php?title=Q162757&diff=516175452&oldid=471303097 and noting that I am not including every error, I skip some.  — billinghurst sDrewth 12:56, 10 July 2017 (UTC)
Edit without comment is redirect resolving. Just see both IDs (old and new) on VIAF site. So please see the items history or reference section for original ID source. The edits are needed to highlight errors and bring its to "Unique value" report. — Ivan A. Krestinin (talk) 13:03, 10 July 2017 (UTC)
For these, I think redirect resolving is useful. Interesting to watch when you add 3 or 4 identifiers for the same person and then see it reduce as they merge one after the other.
--- Jura 13:08, 10 July 2017 (UTC)

Small optimization for User:Ivan A. Krestinin/To merge/*

Hi Ivan, these days there are a lot of items listed in your automatically generated merge lists. It would somewhat reduce the number of false positives if you could slightly improve the algorithm as follows: do not list a matching pair of items if one refers to the other using different from (P1889). Amongst many other examples: 32 dekabrya (Q17351696) and December 32 (Q2816177). It would be some kind of item-defined white-list. Hope you can add that feature. Anyhow, thanks for these very useful lists. -- LaddΩ chat ;) 21:12, 5 July 2017 (UTC)

✓ Done, please wait the next update. — Ivan A. Krestinin (talk) 20:44, 7 July 2017 (UTC)

Patriarch ≠ Patriarchat

There is a difference beetween patriarch as person and patriarchat as diocese. --Enzian44 (talk) 14:47, 9 July 2017 (UTC)

For example this: https://www.wikidata.org/w/index.php?title=Q359466&oldid=516747467 . --Enzian44 (talk) 15:25, 10 July 2017 (UTC)
Item Patriarchate of Venice (Q23776433) and Patriarchate of Venice (Q740783) were merged by ArTrix. — Ivan A. Krestinin (talk) 22:13, 10 July 2017 (UTC)

Exceptions on Property:P21

Hi Ivan,

The migration seems to go fine, mostly. Good work!

Just a minor thing: on Property:P21 an exception that was defined for one thing go added to other statements as well.
--- Jura 08:10, 13 July 2017 (UTC)

Thank you. Source template does not contain information about exact property. — Ivan A. Krestinin (talk) 16:03, 16 July 2017 (UTC)
Indeed. I hadn't noticed that.
--- Jura 16:26, 17 July 2017 (UTC)

You have imported a incorrect qualifier to difference-within-range constraint (Q21510854). The correct is minimum value (P2313) and maximum value (P2312) with unit year (Q577).--GZWDer (talk) 15:50, 13 July 2017 (UTC)

I have corrected it, but please update your bot if possible (see Help:Property constraints portal/Diff within range).--GZWDer (talk) 15:55, 13 July 2017 (UTC)
Migration script failed for several properties. Now I am working on it. — Ivan A. Krestinin (talk) 20:12, 13 July 2017 (UTC)

Double comments

I reverted two edits of your bot commenting out once-commented templates. Please check all edits and revert if needed. --Tacsipacsi (talk) 10:54, 14 July 2017 (UTC)

несуществующая категория

В этом элементе с 2014 г. до сегодняшнего дня три раза была добавлена Commons category (P373) и три раза ваш бот удалил её как несуществующюю. Всё правильно ) Но могли бы вы добавить немножко кода что-бы бот попытался искать русскую интервику в элементе и по ней идти в русской статьи, искать и удалить и там параметр commonscat или как его там? Как альтернативу можно просто постить сообжение на СО русской статьи о наличии такой ошибки, или же генерировать журналы и предоставить их куда-то в рувики для дальнейших действий других участников. А то думаю что такие круговороты могут продолжаться долго в частности из-за гаджета инфобокса экспортера который упрощает процес; может даже стоит там предлаогать для него новую фичу - распознавание несуществующих категорий с викисклада что-бы народ не тратил время зря добавить их сюда. Или же одноразовый или периодический таск для ботов в рувики - проверить действительны ли категории викисклада указанные в статьях. --XXN, 20:26, 15 July 2017 (UTC)

  • Проблема не специфична для рувики, подобный мусор периодически танется из многих разделов. Если в рувики я ещё что-то могу сделать роботом, то в других разделах флага бота у меня нет. Вообще, на мой взгляд, проблема в кривых гаджетах и ботах, которые ленятся осуществлять хотя бы минимальную проверку импортируемых данных. Что стоило гаджету проверить существование категории и если она не существует, то выдать пользователю сообщение? — Ivan A. Krestinin (talk) 16:11, 16 July 2017 (UTC)

Reports broken

Both Wikidata:Database reports/Constraint violations/P350 and Wikidata:Database reports/Constraint violations/P650 are broken since 13 July 2017‎. Can you please restore them? Multichill (talk) 10:57, 16 July 2017 (UTC)

Couldn't your bot do this in one edit? --Leyo 14:33, 17 July 2017 (UTC)

Hi Ivan,

Some properties have this as a statement as well. I wonder if the syntax clarification (P2916) qualifiers could be moved to the constraint statement and these deleted. I'm aware that this isn't part of the constraint migration. There also many P1793 that were not in sync with the constraints.
--- Jura 16:30, 17 July 2017 (UTC)

https links

I noticed KrBot has been adding links to xkcd comic strips using https instead of https. It would be a good idea to only add https links whenever possible. ~nmaia d 13:02, 22 July 2017 (UTC)

Hello, could you provide link to the edit? — Ivan A. Krestinin (talk) 13:54, 22 July 2017 (UTC)

Thanks for your edit

Thank you. for this and other similar edits. I set them wrong and was just trying to figure out how to fix them, but you beat me to it. --Jarekt (talk) 17:03, 26 July 2017 (UTC)

Novalue as units

Hi Ivan,

Somehow the new "novalue" doesn't seem to match "no units", so there are many exceptions on Wikidata:Database reports/Constraint violations/Mandatory constraints/Violations. Sample: Property:P1457#P2302 at Q107556#P1457.
--- Jura 13:40, 27 July 2017 (UTC)

Новая заглушка

Смотрю что бот не распознал и не убрал её сразу,[78] наверное "не знает" о ней (отсутствует в определенном списке?), по этому решил уведомить вас (может нужно обновить/дополнить список). --XXN, 17:42, 27 July 2017 (UTC) P.S. Написал и тому кто добавил их.

Добрый день, бот считает заглушками то, что находится в категории commons:Category:Image placeholders. — Ivan A. Krestinin (talk) 21:44, 27 July 2017 (UTC)
Понял. Добавил тот файл в нужную категорию и бот убрал заглушки. XXN, 17:54, 29 July 2017 (UTC)

Этнохоронимы

Нашел этот список этнохоронимов. Может захотите залить их в элементах стран где отсутствует demonym (P1549) для русского языка. --XXN, 20:06, 27 July 2017 (UTC)

Прямо сейчас времени нет, лучше на Wikidata:Bot requests. — Ivan A. Krestinin (talk) 21:46, 27 July 2017 (UTC)
Не в этом дело; я больше написал вам по тому что для вас русский - родной язык, и там нужно было бы перепроверить всё тщательно перед заливкой (источник так себе). Тем временем постил список в ВП и в словаре, и нашлись пару-тройку значений которых лучше пропустить. XXN, 17:54, 29 July 2017 (UTC)
На перепроверку у меня времени точно не хватит. — Ivan A. Krestinin (talk) 12:11, 31 July 2017 (UTC)

Range constraint

[[79]] has currently more violations than items processed, but the values seem to be within range of focal height (P2923).
--- Jura 08:39, 31 July 2017 (UTC)

Fixed, some other properties are need fix like this. — Ivan A. Krestinin (talk) 18:02, 31 July 2017 (UTC)
Thanks. There were 3 left: ✓ Done
--- Jura 18:27, 31 July 2017 (UTC)

P2916

At Wikidata:Database_reports/Constraint_violations/P3450, would you change it so that it gets processed?
--- Jura 06:46, 1 August 2017 (UTC)

DoneIvan A. Krestinin (talk) 17:14, 1 August 2017 (UTC)
Pasleim changed it to something else, but at P2572 it's still used and likely to stay. Neither work.
--- Jura 05:50, 2 August 2017 (UTC)

If the constraint check on some of the constraint statements fail, maybe the bot could just report that at the beginning of the report and then process what works. With constraint statements, malformed constraints should be less frequent than with templates. Currently constraint with the "contemporary constraint" and the qualifiers for annotations break the reporting
--- Jura 05:40, 22 August 2017 (UTC)

I have some troubles with error processing architecture. Also I think we need to cleanup malformed, unimplemented and randomly defined constraints. Contemporary constraint is not ready for migration from template as I think. This kind of constraint must be better defined firstly, see https://phabricator.wikimedia.org/T141859. Situation with annotation qualifiers must be resolved too. Two different properties with similar descriptions and purposes are unsuitable. syntax clarification (P2916) has incompatible labels and descriptions for constraint commenting. — Ivan A. Krestinin (talk) 20:10, 22 August 2017 (UTC)
At some point, I asked for the properties for constraints to be created to enable us to query some of these even though neither your reports nor Special:ConstraintReport supported it. It helped us sort out some things and discover some limitations. Even if "contemporary constraint" currently doesn't work, I think we should let its proponent experiment with it. The idea seems good, maybe eventually someone will find a way to fully specify and implement it. In the meantime, we could query properties to which it might apply. As for annotations, I think they can just be ignored for automated processing (they are for users). P2916 will probably be displayed in the error message. All these currently pass constraint checks for property constraint (P2302)
--- Jura 06:39, 23 August 2017 (UTC)

P3450

Related question: Wikidata:Database reports/Constraint violations/P3450 has not been updated since 1 August 2017, and it says ERROR: Constraint Target required claim does not accept qualifier P2916. However, syntax clarification (P2916) has not been used in the constraints of sports season of league or competition (P3450) anywhere since 20 August 2017. Do you have any idea why that constraint report does not update? Could it be related to the use of Wikidata usage instructions (P2559) qualifiers in the constraint statements? Thanks, —MisterSynergy (talk) 12:48, 16 March 2018 (UTC)

Hey Ivan, that mentioned constraint report now receives updates again, but it still has issues as you can see from the page history. Can you please have a look? Thanks, MisterSynergy (talk) 06:20, 5 May 2018 (UTC)
The issue is on Mediawiki core side. Maybe the core fails on overriding one long report (~ 1 megabyte) to another long report. Reducing the report size can help. — Ivan A. Krestinin (talk) 07:02, 5 May 2018 (UTC)
Maybe it could try to blank the page first and then add the new one. I will try to reduce the size a bit, maybe not today though.
--- Jura 07:20, 5 May 2018 (UTC)
An alternative would be to review the plenty item/value requires statement constraints, and maybe remove some of them. They are not crucial for the functionality of this property, but result in a large amount of violations. On the other hand it might be possible to fix a substantial amount of them with imports … —MisterSynergy (talk) 07:28, 5 May 2018 (UTC)
Well, it is a very bulky field. Anything needs to be done on thousands of items. I think the property helped streamlining it.
--- Jura 07:56, 5 May 2018 (UTC)
Started a discussion at Property talk:P3450#Constraints review. —MisterSynergy (talk) 08:19, 5 May 2018 (UTC)

Constraint - unique values combination

Hello, Is possible to add some not simple constraints to category combines topics (P971)?

  • Must be at least two values
  • Must be unique combination (Q123 + Q456 should be only in one category)
  • maybe add too: P971 must not be the same as P301?

JAn Dudík (talk) 07:51, 2 August 2017 (UTC)

@JAn Dudík: I removed the multi-value constraint from that property, not really because I don't think it applies, but merely because I tend to fill this property for one value at a time. (e.g. all categories about churches get P971=church). To make use of the category, the full combination is generally not needed either.
Obviously, none of this prevents us from having such a constraint. It may be useful in other contexts. BTW with Template:Complex constraint, you could design it now.
--- Jura 08:25, 2 August 2017 (UTC)

VIAF values get deleted by the bot

I just saw that the bot deleted a VIAF authority reference from Eleanor Criswell Hanna (Q26973029). What's the reasoning for deleting here with the bot instead of deprecating the value? ChristianKl (talk) 09:35, 3 August 2017 (UTC)

Some questions: Is this value usable for some applications? Can this deprecated value be validated? Can we create full set of such values? — Ivan A. Krestinin (talk) 19:17, 3 August 2017 (UTC)
As they redirect they can be validated I think. --Marsupium (talk) 06:30, 25 September 2017 (UTC)

Size of talk page

You talk page is currently 188,904 bytes long; please archive most of it. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:26, 7 August 2017 (UTC)

Now 400,348 bytes and 287 sections. Please archive most of it, in several sub-pages. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 22:00, 1 February 2020 (UTC)

Instance or class

Ivan, can you please implement relation = instance or subclass of (Q30208840) ? It is used e.g. for "физическая величина". --Infovarius (talk) 09:37, 16 August 2017 (UTC)

The case is looked like bad defined property domain. I am worry that implementing such abilities will reduce data quality. Single property will have many meanings depending on exact item. — Ivan A. Krestinin (talk) 10:19, 16 August 2017 (UTC)

Заливка Property:P625

Приветствую! Хотим ускорить выверку многострадальных рек РФ в Рувики, вытащив координаты из себуанской вики. Если есть возможность, был бы благодарен помощи:

  1. чтобы отделить статьи на выверке от прочих, необходимо залить координаты устий из статей Рувики таким образом (все реки без координат (19578 штук))
  2. чтобы в дальнейшем получить координаты из себуанской вики необходимо проделать такую операцию (там координаты не обязательно в устье) (все реки без координат (26692 штук))
  3. если есть возможность, хорошо бы проставить код GeoNames, он есть во всех статьях себуанской вики о реках РФ.

Предыстория вопроса: ru:Википедия:Форум/Викиданные#Связать себуанскую ВП, ru:Википедия:Форум/Викиданные#Связать себуанскую вику: реки Республики Алтай (само обсуждение в skype). Мастер теней (master of shadows), 00:53, 22 August 2017 (UTC)

Вечер добрый, изучаю что могу сделать. Ленивые ботозаливщики в cebwiki вместо того, чтобы сделать специализированный инфобокс для заливки зачем-то использовали универсальный geobox, это вносит определённые трудности. — Ivan A. Krestinin (talk) 20:16, 22 August 2017 (UTC)
Запустил по категории ceb:Kategoriya:Rusya paghimo ni bot. — Ivan A. Krestinin (talk) 21:18, 22 August 2017 (UTC)
Если есть возможность, то заливка координат из Рувики на данный момент приоритетнее. Там требуется заполнить порядка 13 тысяч элементов. Так мы сможем сразу заняться слиянием пласта на выверке. Мастер теней (master of shadows), 12:29, 23 August 2017 (UTC)
По-быстрому не получилось, в шаблоне Река использован нестандартный способ указания координат, готовый код не подошёл. Может на выходных сделаю. — Ivan A. Krestinin (talk) 21:21, 23 August 2017 (UTC)
Есть код на C#, JS и php. Подойдёт какой? Мастер теней (master of shadows), 22:38, 23 August 2017 (UTC)
Запустил. — Ivan A. Krestinin (talk) 08:24, 26 August 2017 (UTC)
Если не сложно, поправьте реки у которых оказалось по четыре набора координат: https://query.wikidata.org/#SELECT%20%3Fitem%20(COUNT(%3Fval)%20AS%20%3Fcount)%20%7B%0A%20%20%3Fitem%20wdt%3AP625%20%3Fval%20.%0A%20%20%3Fitem%20wdt%3AP31%20wd%3AQ4022%20.%0A%7D%20GROUP%20BY%20%3Fitem%20HAVING%20(%20%3Fcount%20%3E%203%20)Ivan A. Krestinin (talk) 08:46, 26 August 2017 (UTC)
Подчистил. — Vort (talk) 10:27, 26 August 2017 (UTC)
Vort: там уже ботопроход требуется... Как минимум удалить все себуанские при наличии наших. Мастер теней (master of shadows), 17:29, 29 August 2017 (UTC)
Наши бывают грубее себуанских. Тут скорее дедупликация по критерию расстояния нужна. И не только по этому списку, а по всем водным объектам. — Vort (talk) 04:49, 30 August 2017 (UTC)

"imported from" references where nothing was imported

In Special:Diff/545635381, your bot added a reference saying it was imported from the Cebuano Wikipedia when it wasn't. I added it manually in 2015, the Cebuano page didn't even exist then.

This is misleading because an "imported from" reference implies that the statement was added by a bot. If you want to add "imported from" references to existing data which wasn't imported from Wikipedia, I think it should be discussed first.

- Nikki (talk) 10:17, 25 August 2017 (UTC)

I noticed similar at Special:Diff/546483579. As "imported from xxwiki" is not considered as reliable source, it is useless to add it all cases except during the real import.--Jklamo (talk) 09:30, 26 August 2017 (UTC)
It's a known phenomenon. It happens when the same information gets imported from several sources and merged into the same statement.
--- Jura 10:09, 27 August 2017 (UTC)
It both cases there was no merger.--Jklamo (talk) 18:26, 28 August 2017 (UTC)
Bot import the data from specified wiki. All data exists except the reference. Bot adds missed reference. You can specify any other reference. Bot does not replace any references. It does nothing if any other reference exists. "imported from" is useful for error source tracking, especially during splitting items. — Ivan A. Krestinin (talk) 21:04, 28 August 2017 (UTC)

Autofix on DOI (P356)

I tried one for this, but it doesn't seem to work. Possibly due to the use as reference. Even if you decide not to do these in general, would you run this one?
--- Jura 10:09, 27 August 2017 (UTC)

{{Autofix}} works for mainsnacks only currently. — Ivan A. Krestinin (talk) 21:06, 28 August 2017 (UTC)

The template have clear security issues. Everyone may use it to make large-scale disruption in Wikidata. Please describe how to prevent it from being misused, otherwise I will let community review this task.--GZWDer (talk) 16:35, 28 August 2017 (UTC)

All edits can be easily reverted. Also bot has some validation algorithm. So I do not see large-scale security issue. Do you have some ideas how to improve the autofix mechanism? — Ivan A. Krestinin (talk) 21:15, 28 August 2017 (UTC)

Micah

Please undo edits like this, wrong merge. Sjoerd de Bruin (talk) 20:10, 30 August 2017 (UTC)

Missing commons categories - a suggestion

Hi. Your bot made this edit removing an incorrect commons category that was due to a copy-paste error I made. Rather than removing cases like this, it might be better to check for the existence of a site link to a Commons category, and to replace the typo with that category instead. Thanks. Mike Peel (talk) 00:38, 31 August 2017 (UTC)

It is interesting idea, but I have too limited amount of free time now. — Ivan A. Krestinin (talk) 19:23, 6 September 2017 (UTC)

Constraint violations

Seems that KrBot is on constraint violations holiday since 31 August. Everyone needs a break, but I hope he will be back soon.--Jklamo (talk) 23:50, 2 September 2017 (UTC)

I think it will be back today. — Ivan A. Krestinin (talk) 19:20, 6 September 2017 (UTC)

ExternalUse

From what I can see, your bot adds templates to {{ExternalUse}}, but not from svwiki. That is understandable since our templates maybe are designed a little differently from many others. But what can I do to make your bot add svwiki-templates? Only sv:Template:Ortsfakta WD uses: P17,P18,P31,P94,P131,P180,P237,P281,P282,P407,P473,P507,P525,P571,P580,P582,P585,P610,P625,P775,P776,P794,P1082,P1376,P1449,P1566,P1589,P2044,P2046,P2096 and P2370 and probably even some more. And then I have not even tried do list all the properties that can be used in sources. I can design a template that can be added to the doc-pages of templates and modules, and you can harvest from them, if you like?! -- Innocent bystander (talk) 06:52, 4 September 2017 (UTC)

Bot searches direct usages. For example P17 is not used in sv:Template:Ortsfakta WD, it is used in sv:Modul:Ortsfakta WD. But my bot has no code for module parsing. — Ivan A. Krestinin (talk) 19:19, 6 September 2017 (UTC)

Кто прав, кто неправ...

Бот второй раз подряд прописывает в d:Q15704031 не существовавшее в те годы государство. Или так и надо, и имеется в виду не тогдашнее место действия, а нынешнее его название? но событие-то политическое, всё вертелось именно вокруг тогдашнего трона тогдашней монархии. Retired electrician (talk) 11:38, 11 September 2017 (UTC)

Так дело в том, что Англии тогда тоже не существовало... Исправил. --Infovarius (talk) 12:07, 12 September 2017 (UTC)
Бот действует в соответствии с этим заданием, пожалуйста скорректируйте его, если в нём есть ошибки. — Ivan A. Krestinin (talk) 18:23, 12 September 2017 (UTC)

P734 constraints violations page

Hi! I see KrBot hadn't be able to update since mid-August, but there is no constraint about synchronicity anymore. Any idea why it doesn't work now? --Harmonia Amanda (talk) 09:39, 18 September 2017 (UTC)

Bot throws error if constraint definition contains any unexpected value. Bot know nothing about contemporary constraint (Q25796498). I have some troubles with this constraint type implementation, see https://phabricator.wikimedia.org/T141859. — Ivan A. Krestinin (talk) 22:05, 18 September 2017 (UTC)
The thing is, there is no constraint like this anymore. What can be the problem now? --Harmonia Amanda (talk) 13:08, 21 September 2017 (UTC)
Bot does not use templates on property talk pages anymore. It uses property constraint (P2302) from property page now. — Ivan A. Krestinin (talk) 21:07, 21 September 2017 (UTC)

… is empty. Is there anything wrong with Property:P1480#P2302? Thanks in advance, --Marsupium (talk) 12:23, 23 September 2017 (UTC)

Current implementation of one-of constraint (Q21510859) constraint does not check qualifiers and references. Bot checks mainsnaks only. — Ivan A. Krestinin (talk) 17:54, 23 September 2017 (UTC)
Ah, sure, thank you! --Marsupium (talk) 18:23, 23 September 2017 (UTC)
@Marsupium: I implement qualifiers check for one-of constraint (Q21510859), please see the report. — Ivan A. Krestinin (talk) 20:39, 24 September 2017 (UTC)
Thanks a lot, this is much more convenient than using WDQS … --Marsupium (talk) 06:28, 25 September 2017 (UTC)
Hey, yet another request … It would be cool to have this for value-type constraint (Q21510865), too. So kinship to subject (P1039) could be checked that has also used as qualifier constraint (Q21510863). I don't know how much work it is, but the implementation would be appreciated. Thank you, --Marsupium (talk) 18:41, 2 October 2017 (UTC)

Constraint bot runs seeming to take >> 24 hours

Hi. Thanks for the constraint violations bot runs of KrBot. It would seem that bot runs are now taking 24 to 48 hours to run, which means that when one fixes something that it doesn't appear in the next published report, but instead in the subsequent report. This is somewhat confusing where one is trying to manage multiple reports in trying to remember what was fixed, and whether you missed something, or something else is broken.

I am wondering whether it might be time to move constraint reports to be undertaken every second day. Having every second day removes the report overlap, and presumably it will reduce load on the system.  — billinghurst sDrewth 22:09, 3 October 2017 (UTC)

Wikidata:Database reports/Constraint violations/P625

There is contenttoobig error.--Jklamo (talk) 20:11, 7 October 2017 (UTC)

Still no change. If needed, I think we can leave behind single value constraint, as i already proposed in past.--Jklamo (talk) 13:18, 13 October 2017 (UTC)
Hello, I am sorry for delayed response. I reduce listed items count per constraint for long reports. Single value constraint is specified with qualifier separator (P4155) = applies to part (P518). So it is looked good. — Ivan A. Krestinin (talk) 18:03, 13 October 2017 (UTC)

Odd "instance of" edit

This was way back in January, but I thought I'd report it: [80]. The bot apparently decided my team at WMF was an instance of the Employees' Provident Fund Organisation of India :) — MusikAnimal talk 02:11, 20 October 2017 (UTC)

Why

https://www.wikidata.org/w/index.php?title=Q285392&type=revision&diff=591746235&oldid=591611994--Oursana (talk) 20:12, 10 November 2017 (UTC)

Groeningemuseum described at URLs

The KrBot deleted a lot of 'described at URL's' from the records of the artworks of Groeningemuseum. At some point the URL's were offline, but they are back online now. So:
1. Can this bot please stop deleting not-working URLs without contacting the source first? The URL's are active again now, but the important data about the source was lost from Wikidata. Now we will have to add it again. This is not a good practice.
2. The bot seems to delete de links again, even after I manually add them by using 'undo'. Example: https://www.wikidata.org/w/index.php?title=Q22240332&action=history Can you explain why that happens? This also has to stop. Thank you! Alina data (talk) 12:50, 1 December 2017 (UTC)

1. The bot doesn't care if the link is dead or not.
2. The bot removes described at URL (P973) because the same ID for this website is used at Groeningemuseum work PID (P2282). --Edgars2007 (talk) 14:55, 1 December 2017 (UTC)
Ah thank you. I didn't know that. The links used in described at URL (P973) and Groeningemuseum work PID (P2282) are different though, only the last part is the same. Shouldn't both links be able to stay?Alina data (talk) 15:36, 1 December 2017 (UTC)
Bot uses {{Autofix}} rule from Property talk:P973. Please delete or change the rule on this page if needed. — Ivan A. Krestinin (talk) 19:43, 1 December 2017 (UTC)
Thank you! Hopefully I have done this correctly. - Alina data (talk) 10:26, 4 December 2017 (UTC)

Problematic merge

Hi Ivan, would you undo the edits following the merge for the ones this was part of? Thanks for your help.
--- Jura 22:52, 1 December 2017 (UTC)

✓ DoneIvan A. Krestinin (talk) 18:38, 2 December 2017 (UTC)

contemporary constraint

Hi, is there any way to solve problem with contemporary constraint causing none generation of reports. I know that removing property constraint (P2302) contemporary constraint (Q25796498) is working, but that is the loss of information a bit. What about setting deprecated rank (and set up bot to ignore these values)?--Jklamo (talk) 09:53, 11 December 2017 (UTC)

Hello, see https://phabricator.wikimedia.org/T141859 for details. Parametrized algorithm is need to be defined for this constraint to start implementation. Also we need create properties for this constraint parameters. — Ivan A. Krestinin (talk) 19:58, 13 December 2017 (UTC)
On Wikidata:Database reports/Constraint violations/Errors, there is also the problem of the qualifier of Commons gallery (P935). Would you consider changing that (the absence of the qualifier should be read as namespace=0) or should I ask Lucas to change it on his side (novalue should be read as namespace=0)?
Would you allow syntax clarification (P2916) as qualifier?
Remains the instance of/subclass of thing to be sorted out.
--- Jura 18:15, 16 December 2017 (UTC)
The parameters of Commons link constraint (Q21510852) are documented on Help:Property constraints portal/Commons link#Parameters, and I see no need to change their interpretation. I discussed this at the time with Ivan A. Krestinin (see talk pageMissing Wikimedia Commons namespace (P2307) means namespace "0"), so I’m not sure why we now seem to disagree on the interpretation… --Lucas Werkmeister (WMDE) (talk) 13:04, 18 December 2017 (UTC)
It hasn't been working since ..
--- Jura 13:25, 18 December 2017 (UTC)
@Jura1: sorry, since when? --Lucas Werkmeister (WMDE) (talk) 17:08, 19 December 2017 (UTC)
August 4: https://www.wikidata.org/w/index.php?title=Wikidata:Database_reports/Constraint_violations/P935&action=history
--- Jura 18:43, 19 December 2017 (UTC)
The first and current implementation of my bot requires Wikimedia Commons namespace (P2307) = novalue for Commons gallery (P935). Original version of Module:Constraints required such construction too. But the module was changed. Now it supports both variants as I see. I think Wikimedia Commons namespace (P2307) = novalue requirement is better than missed qualifier. Commons link constraint (Q21510852) requires Wikimedia Commons namespace (P2307) always. Same as format constraint (Q21502404) requires format as a regular expression (P1793). Missed by mistake qualifier is easily detected and highlighted. So I prefer to fix documentation. — Ivan A. Krestinin (talk) 11:10, 5 January 2018 (UTC)

Creating broken redirects

Hey, your bot is creating broken redirects like this and my bot keeps deleting them. Amir (talk) 14:44, 21 December 2017 (UTC)

Fixed for this property. Property creators create property talk page too usually. — Ivan A. Krestinin (talk) 11:16, 5 January 2018 (UTC)

KrBot adding and deleting dups

Not sure why the bot is doing this, but I've noticed that one at least a few different items I'm watching, KrBot is adding a duplicate VIAF value and then deleting it because it's a duplicate. Seems like a waste of processing power.

Example: https://www.wikidata.org/w/index.php?title=Q43139096&curid=44376084&action=history

Hazmat2 (talk) 02:55, 22 December 2017 (UTC)

This happens sometimes. I have two parts of algorithm. One part resolves redirects and another part removes duplicates. I can combine these parts of cause. But this requires some efforts for coding, debugging, bugfix and etc. I think the issue is too minor to spent time to this. — Ivan A. Krestinin (talk) 11:21, 5 January 2018 (UTC)

The page Wikidata:Database reports/Constraint violations/P1472 was blanked with an error code: ERROR: Command 'wget -c https://dumps.wikimedia.org/commonswiki/20171220/commonswiki-20171220-page.sql.gz -O dumps/incomplete/25a3e1ad4e0c13c87a15e3f465ae66d4_commonswiki-20171220-page.sql.gz' failed with code 2048. --Jarekt (talk) 14:39, 22 December 2017 (UTC)

✓ DoneIvan A. Krestinin (talk) 11:22, 5 January 2018 (UTC)

2018-01 KrBot and P4342

Bonjour et bonne année.

The page Wikidata:Database reports/Constraint violations/P4342 is not updated by your bot KrBot. Is there something to do? Visite fortuitement prolongée (talk) 21:06, 2 January 2018 (UTC)

Hello. Bot does not change the report if the only think that is changed is elements count. This saves time and limits the reposts history growing. — Ivan A. Krestinin (talk) 11:26, 5 January 2018 (UTC)
Thank you. Visite fortuitement prolongée (talk) 21:12, 6 January 2018 (UTC)

Your edit

I see that you changed country (P17) [from Wales to United Kingdom at 4.30am this morning]. Country = Wales was correct; your edit was incorrect as you have created a duplicate value (UK twice). Why? Secondly, if this is automated, then how many similar edits have you made? P17 can have duplicate values, as villages in the country of Wales are also in the UK. As Wales is the detailed value (the local value), then Wales is better. 2.28.38.22 07:18, 12 January 2018 (UTC)

Sometimes rules need to be broken, as in the case of this edit. It refers to the country Wales, and not the UK state. Please stop this bot from editing this page. Llywelyn2000 (talk) 22:02, 12 March 2018 (UTC)
Remove corresponding {{Autofix}} template from Property talk:P17 and bot will stop. — Ivan A. Krestinin (talk) 05:52, 13 March 2018 (UTC)

Removal of VIAF

This edit removed a VIAF, claiming it was deleted, but the target VIAF still exists. --EncycloPetey (talk) 07:07, 13 January 2018 (UTC)

[81] Looks like your bot is still unsupervised and removing data. --EncycloPetey (talk) 14:14, 13 January 2018 (UTC)

novalue as a format constraint violation

On, for example, Wikidata:Database reports/Constraint violations/P236#Format, 'novalue' should not trigger a format constraint warning. Can you exclude such cases, please? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 17:20, 19 January 2018 (UTC)

Undo merged values

Пожалуйста, верните русские имена на место! --Infovarius (talk) 13:44, 26 January 2018 (UTC)

  • ✓ Done, только оставшиеся страницы всё равно нужно проверять. Кто-то мог задать свойство "имя" в период, когда элемент был редиректом. — Ivan A. Krestinin (talk) 21:16, 28 January 2018 (UTC)

update?

Would there be an update of User:Ivan A. Krestinin/To merge subpages, please? Thanks, Lymantria (talk) 10:33, 28 January 2018 (UTC)

Fixing a series of edits

Hello, is it possible to change back all occupation (P106)Category:Model railroader (Q32075377) to occupation (P106)model railroader (Q22812942). Your bot had changed them because the two items had been merged, but the merge was wrong. Thanks ! -Zolo (talk) 20:45, 7 February 2018 (UTC)

Placeholder may be used several times in formatter URL

Hello, I noticed on P1670 that the "$1" placeholder may be used several times. But only the first occurrence is replaced KrBot's reports, thus the links don't work. Best regards, Od1n (talk) 23:13, 9 February 2018 (UTC)

Wikidata:Database reports/Constraint violations/Mandatory constraints

Seems like the amount of mandatory constraints is too much for displaying all the information. Maybe split into identifiers and general properties? Sjoerd de Bruin (talk) 17:00, 10 February 2018 (UTC)

RKD constraint reports

Both Wikidata:Database reports/Constraint violations/P650 and Wikidata:Database reports/Constraint violations/P350 haven't been updated for a while. Can you have a look? Thanks, Multichill (talk) 23:43, 1 March 2018 (UTC)

P1330 constraint violations

Hi there! I see your bot is the one who has done updates on https://www.wikidata.org/wiki/Wikidata:Database_reports/Constraint_violations/P1330 I've recently done some work on these and was wondering if you could update that report as last was on 27 feb. No rush, thank you --CatQuest, The Endavouring Cat (talk) 11:44, 12 March 2018 (UTC)

KrBot2

Hello Ivan A. Krestinin, I have granted its bot flag and noticed you have not used it yet. Perhaps you didn't notice. Kind regards, Lymantria (talk) 06:51, 13 March 2018 (UTC)

@Lymantria: no, there are other issues. --Edgars2007 (talk) 07:29, 13 March 2018 (UTC)
Bot downloaded new dumps for a long time. But now it is working. Performance is very bad die insufficient volume of RAM. — Ivan A. Krestinin (talk) 19:26, 13 March 2018 (UTC)

Can you look into Wikidata:Database reports/Constraint violations/P1472 and Wikidata:Database reports/Constraint violations/P1612 which had only 2 updates in last 30 days. Can we switch them back to daily updates? --Jarekt (talk) 13:04, 14 March 2018 (UTC)

Did you read last two threads, Jarekt? :) --Edgars2007 (talk) 13:17, 14 March 2018 (UTC)
No I did not. Sorry. I will wait. --Jarekt (talk) 13:45, 14 March 2018 (UTC)

Important!

We (I and Pasleim) decided, that you can shot down your bot for now :) Till the moment, when all things settle down. --Edgars2007 (talk) 06:36, 20 March 2018 (UTC)

Just to clarify, I implemented a function that my bot is not overwritting your bot edits. So you can let your bot run but as soon as a report wasn't updated for multiple days DeltaBot steps in. --Pasleim (talk) 09:56, 20 March 2018 (UTC)
I think nothing bad if both bots will update the reports in current situation. Also it is good to improve DeltaBot`s report to deduce diffs with KrBot2. — Ivan A. Krestinin (talk) 20:26, 20 March 2018 (UTC)
Pasleim: looks like there are formating discrepancies. --Succu (talk) 21:07, 20 March 2018 (UTC)
Sorry Ivan, but I don't think it is right. The main idea was to give your bot some rest. Deltabot is putting more or less up-to-date data, while your bot is putting currently 11 days old report. And then Deltabot won't update the report for 5 days... --Edgars2007 (talk) 10:24, 22 March 2018 (UTC)

@Ivan A. Krestinin, KrBot2, pasleim, DeltaBot: Please sort this out ASAP. Bots edit warring do not help those of us who monitor these constraint reports. --99of9 (talk) 06:29, 10 April 2018 (UTC)

@pasleim: I restore back KrBot2 performance to previous level. So we can stop DeltaBot for now. For the future: DeltaBot uses more actual data, but part of functionality is missed. For example "Conflict with", "Commons link" constraints, summaries... Could you implement missed functionality to stop my bot at all? — Ivan A. Krestinin (talk) 09:07, 10 April 2018 (UTC)
DeltaBot is only updating a report if KrBot2 did not update it for 5 days, thus no actual edit wars can happen (I would not call this an edit war). I can implement the missing functions but all DeltaBot reports are based on SPARQL queries. It is likely that queries for heavily used properties time out and the reports don't get fully updated. This can only be resolved if phabricator:T179879 or a related task is done. --Pasleim (talk) 09:23, 10 April 2018 (UTC)
Ok, thank you. Please note that KrBot2 does not update report if the only change is the first line (date or items count). This makes watchlist usage more comfortable and reduce edits count. — Ivan A. Krestinin (talk) 10:22, 10 April 2018 (UTC)
This is still not a good situation. Even if you don't call it an edit war, the fact that the diffs are so messy means it is hard to monitor and quickly see any change. I monitor many of these reports, and was using them a lot, but I cannot use them well anymore. I suggest that the two bots either publish to different pages, or one is taken off this task, or Deltabot also adopts the convention of not editing if the only change is date or item count, or KrBot2 makes an update within every 5 days (even if there is only a change in the number of items). --99of9 (talk) 04:37, 16 April 2018 (UTC)
Here is an example showing that Kr2bot should not edit if it receives an error. --99of9 (talk) 04:42, 16 April 2018 (UTC)
P705 issue fixed: [82], [83]. — Ivan A. Krestinin (talk) 06:37, 16 April 2018 (UTC)
My revert of the instance issue didn't have a good edit summary. But I do think that would change an awful lot of items (and may be hard to reverse), so should be discussed by protein people. I assume you'd agree that methane (Q37129) is rightly an instance of chemical compound (Q11173). So why shouldn't a (specific sequence) biomolecule also be an instance of chemical compound (Q11173) (via its subclass protein (Q8054)). It looks like a discussion was started, but didn't go far. In practice most protein items are not necessarily even about specific sequences (or are at least tolerant to variation), and seem to get labelled as both instance and subclass of protein. --99of9 (talk) 12:05, 17 April 2018 (UTC)
Lets assume I turn on my electron microscope and see large but single biomolecule. What is it? I think it is instance of (P31) of protein (Q8054). Another case: for example Hypothetical protein CCNA_01158 (Q38143346). It was marked as instance of (P31) of protein (Q8054). But there is L-form of Hypothetical protein CCNA_01158 (Q38143346) and R-form of Hypothetical protein CCNA_01158 (Q38143346) (I mean chirality (Q214514)). So Hypothetical protein CCNA_01158 (Q38143346) is not something concrete with exact structure. It describes class of a bit different proteins. — Ivan A. Krestinin (talk) 07:47, 18 April 2018 (UTC)

Edits to Q50531489

Hi, for two time KrBot edited item Q50531489 and deleted the property reference URL (P854) and replaced with full work available at URL (P953), but that URL don't take to the complete text but just to a description in the Italian Library Service Catalogue, so P953 is not correct --moz (talk) 10:21, 21 March 2018 (UTC)

Spaces changed to underscores in OSM tags

Hi, I added OSM tags that contain spaces to state forest (Q7603683) and wilderness area (Q2445527). The bot changed the spaces to underscores.

I think spaces are correct, you can see them in taginfo: https://taginfo.openstreetmap.org/keys/protection_title#values

If I change back to spaces will the bot revert my edit? Edward (talk) 16:49, 21 March 2018 (UTC)

Please slow down

Hi, can you please slow down editing with KrBot? You are causing noticeable dispatch lag. Your bot should stay well below 50 edits per minute. Cheers, Hoo man (talk) 11:44, 11 April 2018 (UTC)

Thanks! :) -- Hoo man (talk) 19:09, 12 April 2018 (UTC)
Dispatch lag is high now. Could you identify the user who use DB aggressively? — Ivan A. Krestinin (talk) 19:35, 12 April 2018 (UTC)
I already did… we should recover in a couple of minutes. Cheers, Hoo man (talk) 19:41, 12 April 2018 (UTC)
@Hoo man: Hello, could you clarify, why large dispatch lag really critical for the project? 5 minutes or 30 minutes, does it matter? I ask because now my bot is blocked too often die large value of the lag. — Ivan A. Krestinin (talk) 15:03, 23 April 2018 (UTC)
Please see the description of phabricator:T184948 for all the issues this is causing. --Lydia Pintscher (WMDE) (talk) 15:40, 23 April 2018 (UTC)
Thank you. I changed my bot algorithm a bit to avoid blocking. Bot will make 5 second delay after every item if average dispatch lag is more than 5 minutes. — Ivan A. Krestinin (talk) 16:01, 23 April 2018 (UTC)

Category combines

Hi Ivan, I noticed your bot occasionally adds them. I thought you might be interested in the similar Wikidata:Property_proposal/list_combines_topics. I find it would be an improvement over the existing properties and could help putting them to better use.
--- Jura 05:35, 23 April 2018 (UTC)

Removing mandatory constraints with a bot

Hi Ivan, you probably shouldn't remove constraints like these with a bot unless you also add the mandatory constraint when something is cleaned out. The whole constraint work flow can probably use a bit of attention. It's quite hard right now to keep track of the things that have been cleaned out. Multichill (talk) 10:42, 23 April 2018 (UTC)

Watch the references

When you made this change, you left the attached reference in place. The reference was for the former value, so if you change the value, you need to remove the reference, since it does not support the new value. --EncycloPetey (talk) 16:42, 23 April 2018 (UTC)

@Ivan A. Krestinin, EncycloPetey: Presumably this is because a 'personaje de ficción' may not be human, unlike a 'personaje humano de ficción', and the eswiki article does not explicitly say that Chris Redfield is human (perhaps this is meant to be implied? I am not too familiar with the Resident Evil universe). Mahir256 (talk) 04:28, 25 April 2018 (UTC)
Also, when the reference says "imported from...", if you change the value, then the new value will not be the one that was imported. --EncycloPetey (talk) 14:56, 25 April 2018 (UTC)

Update of constraint violation report failed

Hi, the update of constraint violation report for P235 failed with the comment "ERROR: Too many P2309 values." Can you explain the reason of that problem ? Thanks. Snipre (talk) 13:07, 25 April 2018 (UTC)

  • Hello, Jira1 improved the situation: [84]. But my bot does not support instance or subclass of (Q30208840) too. I think we need to make classes system more strong. Why you need specify both relation types? — Ivan A. Krestinin (talk) 19:41, 25 April 2018 (UTC)
    • This constraint was set up to be sure that the property is correctly used, but as this property is a kind of identifier for chemical, I think we can delete it. Few people take care about constraint violation and I prefer to use the other constraints to cure data. So I delete this instance/subclass constraint. Thanks to add the property in your list for update and thanks a lot for your job in WD. Snipre (talk) 11:51, 26 April 2018 (UTC)

Bot edit war

https://www.wikidata.org/w/index.php?title=Q6678817&offset=20180423100050&action=history Mateusz Konieczny (talk) 13:46, 26 April 2018 (UTC)

Hey Ivan, there is a new constraint type no-bounds constraint (Q51723761), introduced to Wikidata recently based on phab:T170610. It is in use in only a very few items right now, but your bot apparently does not like it: Wikidata:Database reports/Constraint violations/P1086.

Can you please implement it, or at least ignore this constraint type with a short notice on the constraint violation report page as long as you cannot implement it properly? Thanks, MisterSynergy (talk) 14:14, 26 April 2018 (UTC)

It apparently works, that was quick! May I suggest a small tweak of the output? Right now it lists just the value in the covi report, but “value±diff” might be more appropriate as the value itself is not the issue here. You can calculate the diff as diff = (upper - lower) / 2. Thanks for your efforts, MisterSynergy (talk) 05:12, 27 April 2018 (UTC)
✓ DoneIvan A. Krestinin (talk) 10:19, 30 April 2018 (UTC)

Format regex

\d in general also matches digits that are not ASCII, but in the formatters it seems what is meant is ASCII. Would be good to replace all \d with [0-9].

Grouping and count currently incorrect, since some use \d others [0-9]:

SELECT ?regex (COUNT(?regex) AS ?count) (GROUP_CONCAT(?propertyLabel; separator=', ') AS ?propertyLabels) WHERE {
  ?property wikibase:propertyType wikibase:ExternalId; wdt:P1793 ?regex .
  OPTIONAL { ?property rdfs:label ?propertyLabel . FILTER(LANG(?propertyLabel) = 'en') }
} GROUP BY ?regex ORDER BY DESC(?count)
Try it!

77.179.61.171 21:37, 4 May 2018 (UTC)

  • You are right, some regexp implementations can include non-ASCII characters to \d. But I think this is not very critical because non-ASCII characters are very rare case. — Ivan A. Krestinin (talk) 06:15, 5 May 2018 (UTC)
    Also, sometimes it is [a-f0-9] and sometimes [0-9a-f], could all be ordered as in ASCII: [0-9a-f]. If the regexes are standardized more, then it is easier to find special cases and maybe errors in regexes. 78.55.25.245 10:46, 5 May 2018 (UTC)
First complete(?) clean up. In the above SPARQL page one can now use the filter to the top right of the list of results and insert 0-9A-Za-z to find all regexes that allow these values. No more A-Za-z0-9 or any combination with \d. I get 64. 78.55.25.245 11:56, 5 May 2018 (UTC)

Hi Ivan, would you consider adding support in the reports for the above? Some discussion of it at [85]. Sample use at [86]. I will try to do a complex constraint that does the same.
--- Jura 09:33, 6 May 2018 (UTC)

Autofix edit description

Hi Ivan. I just stumbled upon this edit by your bot, which removes a statement without a replacement, and the additional edit description only was "autofix". If someone wants to understand this, they have to identify the affected property, go to the talk page (as the template is only put there and not on the property page itself), identify the rule which triggered the edit (in this case, there are four different, relatively complex rules on Property talk:P243) and then understand that by moving the statement to a different property together with a duplicate check this may in the end mean that the statement is simply removed. This process is fine, it's just hard too understand what's going on. Could the edit description include the specific rule? And maybe even the consequences (like "trying to move OCLC-ID:lccn-n83228442/ to LCAuth:n83228442, but skipped as this was already there")? --YMS (talk) 10:50, 7 May 2018 (UTC)

  • Hi! The specific rule does not have some short identifier. We can add something like "name" or "description" parameter to {{Autofix}}, but I think it is too hard to maintain all these names only for good descriptions. I made edit descriptions a bit more detailed. See [87], [88]. — Ivan A. Krestinin (talk) 18:56, 7 May 2018 (UTC)

New constraint type: integer constraint (Q52848401)

Hey Ivan, there is again a new constraint type for which you need to add support in your bot code: integer constraint (Q52848401). I just saw for example Wikidata:Database reports/Constraint violations/P1086 on my watchlist. Can you please have a look? Thanks, MisterSynergy (talk) 05:14, 12 May 2018 (UTC)

Bot error: wrong GND

For an unknown reason KrBot keeps adding a wrong GND to Charles Christian Erdmann, Duke of Württemberg-Oels (Q479755): [89] (6 May), [90] (12 May). KrBot2 lists this error on the page Wikidata:Database reports/Constraint violations/P227. --Kolja21 (talk) 00:05, 15 May 2018 (UTC)

Bot error: YouTube channel

This edit on Q48816788 is a mistake... the TV series in question is a YouTube Red Original, so its official website is a YouTube channel. Your bot shouldn't be "correcting" that. Could you get it to stop? —Joeyconnick (talk) 04:37, 15 May 2018 (UTC)

Tgwiki categories

Hi! Could you bot import "cat main" values (e.g. tg:Гурӯҳ:Вилояти Пактиё) from tg.wikipedia to category's main topic (P301)? Thank you very much, --Epìdosis 11:56, 17 May 2018 (UTC)

Big merge tg.wiki - ru.wiki

Hi! I've just discovered a lot of items to be merged:

Could you do it or could you have someone do it? Thank you very much! --Epìdosis 12:57, 17 May 2018 (UTC)

      • I did not find any unique identifiers like postal codes, OKATO, OKTMO or something like this to make the merge. The only parameter is settlement name, but there are too many settlements with exact the same names in Russia. — Ivan A. Krestinin (talk) 20:38, 18 May 2018 (UTC)

Hey Ivan again, your bot does not support constraints involving instance or subclass of (Q30208840), and as far as I understand two sections further above, you are not convinced that this is a good constraint type. Could you please consider to fix it anyway, at least in a way that individual constraints using this condition are ignored, rather than breaking the entire report? This list indicates that there are ~44 reports not updated at all due to the appeareance of instance or subclass of (Q30208840) in one of the constraints. Thanks, MisterSynergy (talk) 09:19, 18 May 2018 (UTC)

  • Hi, some time ago I reviewed the reports and replaced some instance or subclass of (Q30208840) to instance of (Q21503252) or subclass of (Q21514624). Others looks like errors too. Maybe better way is fix badly defined properties and make our classes system more strong? Also there are some troubles in my bot architecture that makes ignoring hard task a bit. — Ivan A. Krestinin (talk) 20:49, 18 May 2018 (UTC)
    • Thanks for the answer. I think whether we like it or not is not relevant as long as it is there and a legit constraint condition. In a perfect Wikidata maybe we wouldn’t need it indeed, but in real Wikidata I have (unfortunately) seen a enough cases where both relations are used, and I can’t decide which one to prefer over the other.
      Can’t you hard-code a condition into the code that catches these cases just before they are about to be evaluated, which always yields an empty result or an error message that is displayed in the covi report section instead? This would avoid that we have so many broken covi reports. —MisterSynergy (talk) 20:59, 18 May 2018 (UTC)

One more: Wikidata:Database reports/Constraint violations/P4045 does not seem to be listed at Wikidata:Database reports/Constraint violations/Errors, although it never was evaluated successfully. Can you please add it? There are no other covi reports with this problem, except for those of Property:P1450, Property:P3464, and Property:P578 which do not have constraints any longer and are thus dispensable on the update errors list. —MisterSynergy (talk) 09:52, 18 May 2018 (UTC)

Mass addition of "Used on values constraint"

Hi Ivan,

Lucas Werkmeister (WMDE)
Jarekt - mostly interested in properties related to Commons
MisterSynergy
John Samuel
Sannita
Yair rand
Jon Harald Søby
Pasleim
Jura
PKM
ChristianKl
Sjoerddebruin
Fralambert
Manu1400
Was a bee
Malore
Ivanhercaz
Peter F. Patel-Schneider
Pizza1016
Ogoorcs
ZI Jony
Eihel
cdo256
Epìdosis
Dhx1
99of9
Mathieu Kappler
Lectrician1
SM5POR
Infrastruktur

Notified participants of WikiProject property constraints

I just saw that you added a lot of used for values only constraint (Q21528958) (for instance there): what is the motivation behind this, and was this discussed anywhere? Many people use identifiers in references for instance, so I would prefer not to add these constraints on identifier properties by default.

Cheers − Pintoch (talk) 06:48, 21 May 2018 (UTC)

  • Hello, bot adds the constraint to properties that are not used as qualifiers or references. Motivation is improving data quality. Many users make errors like this or this. — Ivan A. Krestinin (talk) 06:57, 21 May 2018 (UTC)
    • Yeah, but particularly identifiers are explicitly allowed in references per Help:Sources#Databases. I don’t think any of them should have a used for values only constraint (Q21528958) constraint. —MisterSynergy (talk) 07:09, 21 May 2018 (UTC)
      • This practice is not looked very good. Because identifiers are changed from time to time. Bots and users update identifies specified as main value. But nobody looks into references. Also identifiers in references are not controlled by distinct-values constraint (Q21502410), single-value constraint (Q19474404), conflicts-with constraint (Q21502838). — Ivan A. Krestinin (talk) 07:13, 21 May 2018 (UTC)
        • In references, identifiers are not used to identify the entity; they are used to provide a link to a source (together with the formatter URL), thus they don’t necessarily have to be used in the item which they identify. —MisterSynergy (talk) 07:17, 21 May 2018 (UTC)
          • But its are used to identify information source. And its are need to be validated and updated same as main value. The most usages that I saw just duplicate identifier in main value section. — Ivan A. Krestinin (talk) 07:48, 21 May 2018 (UTC)
            • Well, if there were reference URLs in the references, which technically directly include the formatter URL and the identifier, one would have to check each of those references as well, but it would be much more difficult as accessibility is much worse in that case. It is also by far not as easy to repair sources when only the formatter URL changed (i.e. identifiers remained stable), since one nevertheless had to touch each and every link individually, rather than updating the formatter URL only. Identifier properties in sources are widely used for that reason; not to identify the entity, but to manage sources efficiently. —MisterSynergy (talk) 08:21, 21 May 2018 (UTC)
    • I think that constraint is frequently forgotten. Maybe it's better to add them and later remove them when it's found as being not applicable.
      --- Jura 07:16, 21 May 2018 (UTC)
      • This could make sense for properties with some datatypes, but for external-ids I find it really counter-productive, especially when marked as mandatory constraints. − Pintoch (talk) 08:15, 21 May 2018 (UTC)
        • Maybe we can add a scope limitation for external identifiers. constraint scope (P4680)="checked on qualifiers only". At least for the ones where this is useful.
          --- Jura 08:41, 21 May 2018 (UTC)
          • Yes, at least for the ones where this is useful, not blindly for all external identifiers, marked as mandatory and without any consultation with the community, maybe? − Pintoch (talk) 08:45, 21 May 2018 (UTC)
            • Can you think of any identifiers useful as qualifiers that haven't been used yet?
              --- Jura 08:51, 21 May 2018 (UTC)
              • On the top of my mind, no. But would it actually be useful? If you want to do that, then it would make sense to first experiment with that on a *few* properties, and then propose it somewhere more visible to gather some consensus around it. Other people might have use cases for identifiers in qualifiers. Constraint violations are important and affect a lot of people now that the gadget has been fully deployed, so we cannot just let people decide on their own that they are going to change constraints on thousands of properties, especially when this breaks the existing recommendations about referencing. − Pintoch (talk) 11:00, 21 May 2018 (UTC)
                • There are very few identifier properties which are substantially in use as qualifiers, see this query. Maybe the first three to eight cases are valid, but most of the rest is probably to be fixed. I think a general assumption that identifier properties should not be used in qualifiers (with exceptions as mentioned) is fine. For references, as outlined earlier above, it is not. —MisterSynergy (talk) 11:18, 21 May 2018 (UTC)
                • Can every external ID property be used as reference? I think no. Social networks can not be used as source. VIAF, IMDb and others can not be used as sources too. So user needs to decide is the source acceptable for referencing or not. If user decide that exact identifier is acceptable then he can delete or improve the constraint. — Ivan A. Krestinin (talk) 11:28, 21 May 2018 (UTC)
                  • Most identifiers can be used as references. They refer to a database, where some information can be consulted, and therefore used as a reference. We don't want newcomers to be faced with a mandatory constraint violation if they try to follow the guidelines and add identifiers in references, just because it happens that the property has not been used yet as such. So, by default, these constraints are unwelcome. Ivan, I really value your work on constraints otherwise (thanks a lot for it) and I understand this edit batch was made in good faith. But it turns out that it contradicts guidelines on referencing, so can I please ask you one more time to fix this? Thank you! − Pintoch (talk) 11:55, 21 May 2018 (UTC)
  • +1. A mistake to add this to external IDs. I have just removed it from BHL bibliography ID (P4327). If you don't like the practice, start an RfC. Don't just use a bot to add constraints without discussion, breaking current references. Jheald (talk) 14:03, 21 May 2018 (UTC)

Incorrect change of country from Scotland to the UK on Inventory of Historic Battlefields in Scotland (Q6059933)

Hi

Today your bot made this edit to Inventory of Historic Battlefields in Scotland (Q6059933), this is incorrect as the database (as the name suggests) only covers Scotland. I'm likely to be making a lot more of these kind of edits. Is it possible for you say why your bot did this? I'm not complaining, I just want to understand and try to reduce to chances of it happening again.

Thanks

--John Cummings (talk) 21:51, 23 May 2018 (UTC)

дублирующее значение

Разве при поиске дублирующих значений свойства квалификаторы не учитываются? [91]? --Infovarius (talk) 11:20, 24 May 2018 (UTC)

Hi! You say that this is type of administrative entities in Russia Federation and possible in another countries. This item however is not defined as being specific to Russia. It is a generic item used for all kinds of settlements all over the world. So its statements shouldn't be specific to Russia or some other particular country. If an entity is both a settlement and an administrative territorial entity then there are two possible solutions for classification purpose, either 1) indicate that an item like Q28514115 is an instance of both Q486972 and Q56061, or 2) use country specific class-item like Q183342 or Q21672098. --90.191.81.65 14:53, 27 May 2018 (UTC)

For most of the time it hasn't been a subclass of Q56061. It was it for some time in January and for past a few days. I think it'd be more constructive not to use it as a some sort of workaround. Isn't the purpose of constraints to help getting the statements right rather than simply setting the statements so that they would fullfil the constraints. With constraint violations being suppressed it's harder to know that something is wrong and needs fixing. Why not let people fix it gradually region per region while constraint violations are shown? I don't know whether different classes are needed for all countries, or whether using multiple class-items would be sufficient for some instances. Also, checking the contraint violations should reveal that some non-administrative entities (settlements) have been used with located in the administrative territorial entity (P131). I've been fixing related stuff for Estonia, but processing the whole world would be too much for me alone. --90.191.81.65 16:45, 27 May 2018 (UTC)

Constraints migration

This edit by KrBot on the page Property talk:P227 is a while ago, but still needs to be fixed. The page shows parts of text that should be hidden:

"... Albert Kitzler (Q16054255), keine Bezeichnung (Q16941268) }} -->"

--Kolja21 (talk) 11:37, 29 May 2018 (UTC)

✓ Done found the solution and did the clean up. --Kolja21 (talk) 15:08, 29 May 2018 (UTC)


Autofix

Hi Ivan,

At Property_talk:P973, I added one for P1258. The query currently matches three items, including Q27590909, but KrBot wont replace it. What's not working?
--- Jura 13:16, 30 May 2018 (UTC)

Error message: GND ID (P227)

Hi Ivan, I've added to "distinct values constraint" the separator object named as (P1932) like before to "single value constraint". Why does the bot not accept it?

--Kolja21 (talk) 00:53, 3 June 2018 (UTC)

There are mainly two types of items:

GND 4039237-5 = microscope (Q196538) + subclass optical microscope (Q912313)
GND 4027249-7 = intellectual (Q58968) + synonym intelligentsia (Q381142)

If we use the qualifier object named as (P1932) these items should not be shown in the error list "distinct values constraint". --Kolja21 (talk) 23:52, 5 June 2018 (UTC)

✓ Done, please notify me if something works wrong. — Ivan A. Krestinin (talk) 06:40, 11 June 2018 (UTC)

The error message is gone but the list contains items that are marked as ok:

Greetings --Kolja21 (talk) 22:21, 11 June 2018 (UTC)

Any idea how to solve the problem? --Kolja21 (talk) 21:53, 14 June 2018 (UTC)
Apologize, I missed your question. The issue is not related to the latest changes. separator (P4155) works only both values have the same qualifier property with different qualifier values. — Ivan A. Krestinin (talk) 20:42, 15 June 2018 (UTC)
Is it possible to improve the bot? pseudonym (P742) is very popular and it would be difficult to explain why we have mulitple allowed qualifiers but only one for one item can be used. --Kolja21 (talk) 23:36, 16 June 2018 (UTC)
This algorithm is not good for other cases. For example total produced (P1092), let assume one value has country (P17) qualifier, another one - point in time (P585). This case is not acceptable. As for Friedrich Liechtenstein (Q1251322): the problem happens because two entities are mixed in the same item: human and pseudonym. I see better way to model the case: see Kate Cary (Q465326) and Erin Hunter (Q641740). Qualifiers are not needed for this model. — Ivan A. Krestinin (talk) 09:21, 17 June 2018 (UTC)
It's about one item with one entitiy: A human with a name and a pseudonym. A joint pseudonym (= a group of persons) is a different case. I don't think that we want to create a second or third item for every person who is known by more than one name. Just think of all the popes, kings, people that got married etc. Ten thousands of duplicate items only to solve a technical issue? ---Kolja21 (talk) 21:25, 17 June 2018 (UTC)

Distinct on P2534

Somehow I think the distinct value constraint could be useful on Property:P2534. Currently KrBot outputs an error message.
--- Jura 11:56, 6 June 2018 (UTC)

+1. Same problem as above. --Kolja21 (talk) 19:03, 6 June 2018 (UTC)
✓ Done, but current update is in progress, please wait the next. — Ivan A. Krestinin (talk) 16:19, 11 June 2018 (UTC)

Wikidata:Database reports/Constraint violations/P2580

Wikidata:Database reports/Constraint violations/P2580 - There has been a major clean-up of the BBLd IDs. Would be nice if you could start an update of the report. The IDs should now match /[-.()0-9A-Za-z]+/ 92.227.230.108 12:36, 7 June 2018 (UTC)

Simply wait the next update. Should be in next 24h. --Edgars2007 (talk) 18:41, 7 June 2018 (UTC)
Edgars2007, no change. 85.180.59.131 20:18, 8 June 2018 (UTC)
Now all is fine as I see. — Ivan A. Krestinin (talk) 16:21, 11 June 2018 (UTC)

Last 500 visible changes is not enough

Dear Ivan, I think some time ago there was that the last 1000 changes could have been seen, but have a look at latest changes P161: the day is not over but the limit has been reached. Similar to other maintenance pages. -- MovieFex (talk) 18:33, 7 June 2018 (UTC)

And what Ivan has to do with that? :) Simply change 500 to any number in URL. --Edgars2007 (talk) 18:42, 7 June 2018 (UTC)
Good hint, but it would be not bad also if that works when to choose. -- MovieFex (talk) 19:24, 7 June 2018 (UTC)
I do not understand the issue. Cold you explain it in details? I it related to me or my bot? — Ivan A. Krestinin (talk) 16:24, 11 June 2018 (UTC)

Undo merge Abkhazia

Пожалуйста, можно отменить эти ужасы: [92]. --Infovarius (talk) 20:15, 7 June 2018 (UTC)

✓ DoneIvan A. Krestinin (talk) 16:25, 11 June 2018 (UTC)

ERROR: Unknown constraint type: Q21528958.

Не уверен репортили ли это выше, но после вот такого и подобных изменений легли репорты (например Wikidata:Database reports/Constraint violations/P3876) --Ghuron (talk) 04:57, 13 June 2018 (UTC)

  • Да, репортили, один из участников по не вполне понятным мне причинам вернул назад устаревшие ограничения, которые были заменены их более совершенным аналогом, см. Wikidata talk:WikiProject property constraints#Change of format for scope constraints. — Ivan A. Krestinin (talk) 15:27, 13 June 2018 (UTC)
    • Насколько я понял, там целая плеяда причин, наврядли я чем-то смогу помочь :( М.б. навернуть бота так, чтобы отражал факт непонятного ему констрента в соответствующей секции, но все же заполнял остальные? Мой workflow заточен на регулярный анализ свойств, которые я сильно меняю, без Вашего отчета как без рук :( --Ghuron (talk) 18:26, 13 June 2018 (UTC)
      • Не хочется кардинально переделывать архитектуру бота дабы заткнуть временную ситуацию. Пока ситуация не нормализовалась можете просто удалить и старые и новые ограничения касающиеся способа использования свойства. — Ivan A. Krestinin (talk) 16:45, 14 June 2018 (UTC)

Please revert reverted merge redirect resolving for stained-glass artist (Q2205972)

They are too many, I don't want to fix them manually. Less than 30h is really short to review such a merge unfortunately … :-/ Thanks in advance! --Marsupium (talk) 12:55, 18 June 2018 (UTC)

✓ DoneIvan A. Krestinin (talk) 21:25, 18 June 2018 (UTC)
Thanks a lot! --Marsupium (talk) 22:54, 18 June 2018 (UTC)

VIAF misplaced

FYI https://www.wikidata.org/w/index.php?title=Q18672073&action=history - this was incorrrect 2.247.48.113 18:59, 26 June 2018 (UTC)

Deprecated values should not trigger 'single value' constraint violations

On, for example, Wikidata:Database reports/Constraint violations/P496, your bot lists a constraint violation for Reinhardt Kristensen (Q63514), based of the two IDs at Q63514#P496. However, one of those IDs is marked as deprecated, because it's a withdrawn identifier value (Q21441764).

Please can you exclude deprecated values from such reports? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 15:08, 28 June 2018 (UTC)

  • Hi, my bot ignores ranks at all. Generally collecting invalid and deprecated identifiers is not goal of Wikidata. We can not collect all such IDs because this dataset is unlimited. Collecting only some IDs is not looked as goal. For another IDs (for example VIAF, GND) we have special cleanup procedure. It automatically resolves redirect IDs and removes invalid IDs. ORCID DB has API. So the same procedure can be created for this identifier. I can help with this task. — Ivan A. Krestinin (talk) 13:50, 29 June 2018 (UTC)

ERROR: Unknown constraint type: Q21528959.

Hello, your bot seems to malfunction at this page Wikidata:Database_reports/Constraint_violations/P813

"type constraint" to "item requires statement constraint"

Hey Ivan, I see that you are moving lots of "type constraints" to the less intuitive "item requires statement constraint". Which advantage does this change give us? On first impression I think we should not touch "type constraints" in this way. Thanks, —MisterSynergy (talk) 22:23, 8 July 2018 (UTC)

P4224:Q5 на метакатегории

Я тут обнаружил что бодаюсь с Вашим ботом на элементах типа Q9684548. Это же классическая метакатегория, не предполагается что она содержит что-либо кроме подкатегорий. Если про человека известно, что он умер где-то в Афганистане, то его включат в Q6343289 --Ghuron (talk) 09:51, 24 July 2018 (UTC)

А, это аутофиксы на Property:P971. Идея отличная, но мне кажется Q1322263, Q18658526 и Q19660746 там лишние --Ghuron (talk) 09:55, 24 July 2018 (UTC)
Убрал соответствующие автофиксы. Также отметил эту категорию как метакатегорию и добавил ограничение на применение свойства category contains (P4224). — Ivan A. Krestinin (talk) 18:22, 24 July 2018 (UTC)
c указанием P31:Q30432511 к сожалению всё сложно - см. Wikidata talk:WikiProject Categories#Metacategory vs. Category --Ghuron (talk) 08:05, 25 July 2018 (UTC)

Hey Ivan, regarding this change by KrBot: can you please use based on heuristic (P887) as a reference qualifier instead of imported from Wikimedia project (P143) in future? The latter is to be used for Wikimedia projects as values only, and recently we started to tidy up this situation. based on heuristic (P887) was exactly made for what you want to say, but unfortunately it wasn’t used very much until recently and it is therefore not yet well-known. Thanks, MisterSynergy (talk) 07:47, 26 July 2018 (UTC)

Hi, fixed. Thanks. — Ivan A. Krestinin (talk) 11:00, 26 July 2018 (UTC)

Bot is incorrectly merging statements!

Ivan, your User:KrBot is incorrectly merging non-identical statements. These statements have the same value but different qualifiers and should not be merged. Here is an example change: https://www.wikidata.org/w/index.php?title=Q353172&type=revision&diff=717998682&oldid=716750878 . This has been done incorrectly across thousands of disease items. Please revert all of these edits. Thank you Gstupp (talk) 18:30, 1 August 2018 (UTC)

Could you clarify meaning of the following statement? — Ivan A. Krestinin (talk) 20:32, 1 August 2018 (UTC)
UMLS CUI
Normal rank C0033740
0 references
add reference
Normal rank C0033740
mapping relation type close match
0 references
add reference


add value

Ivan, this statement is without references, and so doesn't have the same meaning. I've added one below with more closely follows the statements in question.

UMLS CUI
Normal rank C0033740
1 reference
Normal rank C0033740
mapping relation type close match
1 reference
add reference


add value

This means that the Disease Ontology (DO) has a cross reference to the item in question (from the class DOID:8888 in DO), but does not specify the mapping relation type. See here for a discussion on mapping relation type. The second statement says that MONDO has a xref between to the item (from the class MONDO:1234) and specifies that it is a close match. If the statements are merged, then both references get added to the statement and you cannot tell which reference belongs to which (value, qualifier) pair. Furthermore, if two statements have different qualifiers and references (but the same value), and after the KRBot's merging, the statement will have both qualifiers and both references and you cannot tell which reference goes with which qualifier. Please LMK if this is clear.. Gstupp (talk) 23:00, 1 August 2018 (UTC)

One database has only information about identifier value (C0033740). Another database has additional information about this value (close match). Do we really need to represent these details in Wikidata? Is this minor details needed to the data users? Usually single source is more than enough for external ID value. Do we need to show that another source exists with incomplete information about the value? Also duplicate values violate single value and unique value constraints. — Ivan A. Krestinin (talk) 21:49, 2 August 2018 (UTC)
I'd say that is a question that can be open to discussion, but that is a separate issue from the statement merging. Either way, the two statements mean different things and shouldn't be merged. Gstupp (talk) 22:44, 2 August 2018 (UTC)
I think we are discussing this already. Is it ok if I delete the first value instead of merge? — Ivan A. Krestinin (talk) 21:25, 5 August 2018 (UTC)
I don't see why it should be deleted. Multiple sources have complementary or partially overlapping data about an item. The value with the qualifier may be wrong and the first value correct? Regarding the single value constraints, the constraint should be modified to check for uniqueness instead of len()==1. Gstupp (talk) 17:27, 7 August 2018 (UTC)
Primary Wikidata goal is providing data to external users. We are interesting in involving additional external users too. Because more users will validate and improve our data. To involve the users we need make our data interface user-friendly. Because nobody like to spent a time to investigating overcomplicated datasets. Duplicate external identifiers are looked very strange. User need to spent a lot of time to understand the reason why its are needed. — Ivan A. Krestinin (talk) 19:26, 7 August 2018 (UTC)

suggestion: using edit groups for solving redirects.

Hi Ivan. Thanks for always fixing links pointing to redirected items. As from time to time people are merging wrong items which only gets dedected after a few days, some redirect fixing have to be undone. So far this is only possible by contacting you on your talk page, but with the new tool Edit groups by @Pintoch the process could be simplified. What do you think? --Pasleim (talk) 07:59, 6 August 2018 (UTC)

Hi Pasleim, good idea! Documentation recommends links like:
[[:toollabs:editgroups/b/CB/89ead4fe|details]]
Do you know, can I use something like:
[[:toollabs:editgroups/b/KrBot/Merge/Q12345_Q67890|details]]
Sometimes a job is executed using several bot runs. Random numbers requires some cross-run storing. So some informative link is preferred. — Ivan A. Krestinin (talk) 08:40, 6 August 2018 (UTC)
Random numbers are not required but each bot run (or group of but runs) needs an unique ID, so something like Q12345_Q67890 should be okay. You will need to provide to Pintoch a regular expression how to extract the ID from the edit summary. --Pasleim (talk) 13:54, 6 August 2018 (UTC)
I added required comment. Sample edit: [93]. Now we waiting for Pintoch: User talk:Pintoch#Edit groups for KrBot. — Ivan A. Krestinin (talk) 17:24, 6 August 2018 (UTC)

Hi Pasleim and Ivan A. Krestinin, that's a great idea! We cannot have / characters in tool ids or batch ids, so the pattern will not work in the current state. What about something like [[:toollabs:editgroups/b/KrBotMerge/Q12345_Q67890|details]]? With this pattern there is a risk that two batches are merged if two items are merged several times, so ideally it would be better to add a timestamp or a random hash to disambiguate them. But if that is too hard to code, we can also try like that. − Pintoch (talk) 08:43, 12 August 2018 (UTC)

Changed, example edit. — Ivan A. Krestinin (talk) 21:05, 12 August 2018 (UTC)

Note: we have an issue with lexemes. — Ivan A. Krestinin (talk) 13:45, 13 August 2018 (UTC)

Redirects in lexical categories

Hi Ivan, lexical categories occasionally get merged. It seems that redirects there aren't updated automatically yet. Would you kindly consider adding it to your bot's tasks?
--- Jura 09:33, 11 August 2018 (UTC)

Hi Jura, I try to implement this, but I did not find any items to test this. — Ivan A. Krestinin (talk) 13:51, 13 August 2018 (UTC)
Thanks. I noticed. Sorry, I fixed a couple before commenting here. Today I found à droite (L796)
--- Jura 13:55, 13 August 2018 (UTC)
✓ Done, thank you for the example item. — Ivan A. Krestinin (talk) 20:56, 13 August 2018 (UTC)

Hey Ivan, the single-best-value constraint (Q52060874) is available for a while, but I just figured out that your bot apparently does not support it yet. Can you please consider integrating it into your code? Thanks, MisterSynergy (talk) 23:25, 13 August 2018 (UTC)

Now bot ignores rank information at all. It does not load it, does not process... Adding this constraint is hard task.
Your question got me thinking about ranks. I see some redundancy in coexisting ranks and qualifiers... For example what is the preferred value for population (P1082)? The latest value maybe. But the latest value can be extracted using point in time (P585) qualifier values.
My bot executes many different tasks in Wikidata. But it never need ranks information. This fact says about redundancy too.
Redundancy in data model is very bad thing. It makes data client algorithms more complex. It allows to break data integrity. So we need to remove redundancy. I think deleting ranks at all is good idea. What do you think? — Ivan A. Krestinin (talk) 21:23, 16 August 2018 (UTC)
You ask some unexpected questions here . At first glance, I do not think that ranks and qualifiers can completely replace the other one so that we could get rid of one of them (here in particular: ranks). I guess we need to live with the existence of both, and we need to accept that constraints may consider aspects of ranks as well.
However, this does not mean that you necessarily need to support any type of constraints with your bot. You don’t look at qualifiers and references in many situations as well, which is well accepted among editors (and constraint violation reports are long enough right now anyway). If you don’t load ranks, it seems indeed very complicated to add this type to the constraints reports. As an alternative, I suggest to implement a “pass” routine (do nothing) for all official constraint types (or constraint qualifier relations) that your bot does not support (yet). You could just add a section to the covi page, stating that this particular constraint type has been seen by your bot, but it is not (yet) supported for technical reasons or complexity; at the same time, your bot could evaluate all the other supported constraints. Right now, unfortunately, reports like Wikidata:Database reports/Constraint violations/P297 are (not) produced just because there is one unsupported (but formally correct) constraint type defined on the property. This leads to the situation that editors tend to avoid these unsupported (by your bot) constraint types, although they have an official status. I think we should shorten this list this way. —MisterSynergy (talk) 21:42, 16 August 2018 (UTC)

Hey Ivan, did you think about this one, particularly the idea to pass formally valid constraint evaluations if for some reasons you cannot do it yet? It would be really great if there was some progress here… Thanks and regards, MisterSynergy (talk) 12:47, 25 August 2018 (UTC)

Yes, pleaase. 0:-) --abián 13:53, 21 September 2018 (UTC)

Hi Ivan. For value requires constraints, this might work better if the grouping was by properties on the value rather than the item. [94]
--- Jura 12:30, 20 August 2018 (UTC)

Hi. Your bot is blindly (of course) replacing a claim by another on numerous pages. In the case of Paul Henri Mallet’s page this should not be done, as the new link links to a page containing nothing (whereas the old suppressed link contains valuable information). Can you please do something about this? Thank you. Sapphorain (talk) 21:19, 21 August 2018 (UTC)

Thank you for your answer, but I know nothing about bots and the terminology you mention. All I can do is revert, over and over again, what your bot is doing on this precise page, because it is removing valuable information. Sapphorain (talk) 22:33, 21 August 2018 (UTC)
I removed the autofix templates here. They need to be discussed on the property talk page before being added again. —MisterSynergy (talk) 11:19, 22 August 2018 (UTC)
Thank you. — Ivan A. Krestinin (talk) 19:03, 22 August 2018 (UTC)

Isadore/ ORCID errors

Your bot has added a number of invalid ORCID iDS to items about long-dead people; see Wikidata:Database reports/Complex constraint violations/P496. Please can you revet the edits, or simply delete the bad values? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 22:28, 21 August 2018 (UTC)


Constraint reports update

Hi Ivan, it looks like the updates have stopped. I thought that they might be back today.
--- Jura 02:11, 30 August 2018 (UTC)

Hello Jura, I have some troubles with server where bot is running. It will be fixed today or tomorrow most probably. — Ivan A. Krestinin (talk) 05:50, 30 August 2018 (UTC)
Ok. Thanks. No urgency. I just thought it might have been accidentally switched off.
--- Jura 06:02, 30 August 2018 (UTC)

Реки Индонезии

Привет! Если вдруг будет нечего залить (да, смешно) за углом притаились 10 тысяч рек Индонезии в себуанской вики без элементов. Вопрос дубликатов не существенен, у индонезийцев всего 446 статей о водных объектах. В наличии тип объекта, код GeoNames, координаты, страна и провинция. А всего там ещё около 300-т тысяч водных объектов без элементов… Мастер теней (master of shadows), 17:34, 4 September 2018 (UTC)

Приветствую, пока бот идёт по категории ceb:Kategoriya:Coordinates not on Wikidata импортируя координаты, страну и GeoNames. Подумаю как импортировать ещё провинцию и тип объекта. — Ivan A. Krestinin (talk) 18:26, 4 September 2018 (UTC)
@Мастер теней: Импорт данных в существующие элементы продолжается, а вот создание новых элементов пока приостановил по просьбе участника Nikki. Пока нет времени разбираться что именно вызывает возражения. Обычно я импортирую коды и данные, а потом автоматически мержую элементы, где совпадают все данные. Раньше это возражений не вызывало. — Ivan A. Krestinin (talk) 14:18, 14 September 2018 (UTC)

Fake "imported from" references

Please stop your bot from adding "imported from" references to existing statements, e.g. here (and please also stop it from repeatedly redoing edits that have been undone). Those statements were added manually, not imported from Wikipedia. - Nikki (talk) 08:41, 13 September 2018 (UTC)

Just add any other reference and bot will not add "imported from". Something like this. Unfortunately Wikidata engine has some bugs. Write requests fail sometimes. This is the reason why bot adds all required claims again and again until its will be created successfully. — Ivan A. Krestinin (talk) 14:10, 14 September 2018 (UTC)
References are not required for all properties. You could make your bot check whether the request actually failed before trying again and/or you could make it add the reference in the same edit that adds the statement. I don't even know what reference I should add in some cases (what's the reference for "based on manually comparing the source and target data"?). Also, how am I supposed to find those statements before your bot does? - Nikki (talk) 09:58, 16 September 2018 (UTC)
Reference is not required, but it is good to have. "imported from" helps to track error source. I can change algorithms, but this can affect bot stability and/or performance. Is it really big issue? I think Wikidata is wiki-project. Somebody adds data. Another bot or user adds some reference. The third user replaces reference to better reference. And so on... I think it is normal wiki-life. Now I am thinking about bot that compares Wikidata items and data from Geographic Names Information System (Q136736). If the data is equal then replace imported from Wikimedia project (P143) to stated in (P248): Geographic Names Information System (Q136736). It is not data import, so licensing issues are not affected. — Ivan A. Krestinin (talk) 15:08, 16 September 2018 (UTC)
I think it's wrong to add "imported from Wikipedia" references to existing statements, especially when they can't be replaced, but I don't have the energy to argue any more.
There's no need to make a bot for GNIS references, my bot is already doing that.
I see you're merging items from cebwiki now. Could you explain how you decide that they're the same? Are you aware that we have a lot of wrong GeoNames IDs?
- Nikki (talk) 22:33, 21 September 2018 (UTC)
Jesberg (Q573235) went wrong at least, I think cebwiki (and thus GeoNames) has articles for the municipality and the municipality seat. Sjoerd de Bruin (talk) 08:42, 22 September 2018 (UTC)
Could you suggest some approach for importing instance of (P31) = municipality and instance of (P31) = municipality seat from cebwiki? This will prevent users and bots from incorrect merges. — Ivan A. Krestinin (talk) 08:58, 22 September 2018 (UTC)
Bot merges items if there are no conflicts in GeoNames ID (P1566), GNIS Feature ID (P590), instance of (P31), country (P17), located in the administrative territorial entity (P131), coordinate location (P625). Also it checks item names and merges only if items has the same name. — Ivan A. Krestinin (talk) 08:50, 22 September 2018 (UTC)

KrBot

Hello Ivan, sorry to bother you, but for some reason your bot removed instance of human from a Phycologist from South Korea]. Dan Koehl (talk) 21:21, 13 September 2018 (UTC)

Your mistake. The Bot did the right thing. --Succu (talk) 21:24, 13 September 2018 (UTC)
Yes @Succu:, I realized that the bot worked just perfect, the instance had been inserted twice, and removing a duplicate is of course correct. So please forget about this Ivan, your Bot seems to work perfectly. Dan Koehl (talk) 21:29, 13 September 2018 (UTC)

объединение вселенной и серии игр StarCraft

Привет, можете отменить перенаправления https://tools.wmflabs.org/editgroups/b/KrBotResolvingRedirect/Q9091680_Q290106/? У меня при нажатии "Undo entire group" выдаёт "Server Error (500)"... --Infovarius (talk) 15:30, 18 September 2018 (UTC)

@Pintoch: could you check https://tools.wmflabs.org/editgroups/b/KrBotResolvingRedirect/Q9091680_Q290106/? "Undo entire group" shows "Server Error (500)" on click. — Ivan A. Krestinin (talk) 22:15, 18 September 2018 (UTC)
@Infovarius, Ivan A. Krestinin: I have just tried to undo this batch, and did not get any server error. But thanks for the report, I will investigate. − Pintoch (talk) 08:06, 19 September 2018 (UTC)

Redirects in Lexemes

Hi Ivan,

Would it be possible to update these? https://www.wikidata.org/w/index.php?title=Special%3AWhatLinksHere&limit=500&target=Q13923816&namespace=146 Maybe they didn't show before as the Lexeme structure is still evolving. --- Jura 16:57, 8 October 2018 (UTC)

The two last entries are wrong there, because I changed the items already more than a week ago. Why is this not updated? Steak (talk) 08:12, 10 October 2018 (UTC)

There might be some data missing in the dumps Ivan is using for evaluation. I saw something similar on other covi pages this week, notably here. —MisterSynergy (talk) 08:15, 10 October 2018 (UTC)

resolving redirect

Do you think is possible revert all this type of edit? (example). The merge was reverted --ValterVB (talk) 08:52, 13 October 2018 (UTC)

@ValterVB: You can click on "details" at the end of the edit comment. Then the option shows up to undo the entire edit group by one click. --Pasleim (talk) 09:09, 13 October 2018 (UTC)
@Pasleim: Great, I did not know. But I receive "Server Error (500)" when I try to do the login or simply I click on "Undo on entire group". Some temporary problem? --ValterVB (talk) 09:16, 13 October 2018 (UTC)
@Jura1, ValterVB: thanks for letting me know. It is curious, the problem disappeared once I restarted the service. I can't figure out where the error comes from. Can you try again? − Pintoch (talk) 15:25, 13 October 2018 (UTC)
@Pintoch: ✓ Done it works fine. Thanks. --- Jura 15:30, 13 October 2018 (UTC)
Yes, work fine. --ValterVB (talk) 19:02, 13 October 2018 (UTC)

please rollback change from Greenland to Kingdom of Denmark

Moved to Property talk:P17#please rollback change from Greenland to Kingdom of Denmark

Deprecated rank

Hello Ivan, two years ago you wrote: "Current bot implementation ignores ranks at all." Is there a chance that this problem will be solved? Wikidata:Database reports/Identical GND ID lists items like Juan Cruz Ochoa (Q362424) and Juan López Díez (Q19863329) as "can be merged" even though one of the GNDs have been marked as deprecated rank. @MisterSynergy: This might be of interest for you. --Kolja21 (talk) 22:16, 15 October 2018 (UTC)

This limitation exists now. But are you sure that we need to collect such errors? Looks like VIAF ignores the rank too. — Ivan A. Krestinin (talk) 20:15, 20 October 2018 (UTC)

Scope saga

Hi Ivan, I probably missed part of the story, so I'm not sure how to get it right. At P5683, the "as reference" doesn't seem to work while at P2358 the "as main value" works fine. Any suggestions how to fix it? --- Jura 02:02, 16 October 2018 (UTC)

Constraint definitions with deprecated rank

Hi Ivan, the js feature ignores definitions with deprecate rank. It would be good if the reports of KrBot would do the same. The advantage is that the new suggestion feature could be used even without the constraint being active. Sample on P106 (the report fails for another reason). --- Jura 12:23, 19 October 2018 (UTC)


Hi Ivan,

In addition to:

I just noticed that this was changed in September to use also the items:

The meaning seems to be that the first one now allows use directly on the lexeme, the other two in the forms/senses parts. See also https://phabricator.wikimedia.org/T201078

Would you adjust the reports for these? --- Jura 10:50, 25 October 2018 (UTC)

Добрый день! Ваш бот проставил кучу чуши вроде "Пабло Пикассо хранится в коллекции Нью-Йоркский музей современного искусства", я её снёс - подскажите, как сделать, чтобы он её не вернул? Wikisaurus (talk) 11:48, 27 October 2018 (UTC)

Приветствую, вариантов несколько, какой из них правильный - не знаю:
  1. Удалить из этих элементов свойство Museum of Modern Art work ID (P2014).
  2. Удалить у свойства Museum of Modern Art work ID (P2014) ограничение: item-requires-statement constraint (Q21503247): collection (P195) - Museum of Modern Art (Q188740).
  3. Удалить у этого ограничения атрибут constraint status (P2316): mandatory constraint (Q21502408).
Ivan A. Krestinin (talk) 12:10, 27 October 2018 (UTC)
  • Очевидно, что правильное решение - это удалить Museum of Modern Art work ID (P2014), это свойство может быть только у произведений, а не у классов объектов и не у художников; от того, что в музее хранится какой-то humidor (Q921455), не следует на humidor (Q921455) вешать код этого элемента, как и наличие там картины Пикассо не отправляет трупик Пикассо валяться там.
  • Но, поскольку такие ошибки массовые, думаю, требуется как-то остановить вашего бота, чтобы он не портил элементы ещё сильнее, когда разные деструктивные элементы добавляют в них чушь. Вы можете сделать 2, 3 или ещё что-то, чтобы ваш бот перестал их расставлять? Wikisaurus (talk) 22:25, 21 September 2019 (UTC)

Visockis

Привет, интересно, что это было? --Infovarius (talk) 23:51, 4 November 2018 (UTC)

Приветствую, следствие объединения элементов. — Ivan A. Krestinin (talk) 21:56, 6 November 2018 (UTC)
Я подозревал это, просто мне казалось, что это как-то в описании правки отражалось... --Infovarius (talk) 17:03, 8 November 2018 (UTC)
Возможность задавать описания правкам появилась далеко не сразу. Даже сейчас это можно сделать не для всех типов правок. — Ivan A. Krestinin (talk) 22:12, 27 November 2018 (UTC)

Autofix

Hi Ivan,

It seems to be offline for some time. Would you restart it? --- Jura 06:21, 25 November 2018 (UTC)

Fixed. — Ivan A. Krestinin (talk) 22:11, 27 November 2018 (UTC)

P1482 (Stack Exchange tag)

Иван, добрый день! Страница обсуждения говорит, что значение должно иметь вид, например, https://stackoverflow.com/tags/wikidata, а не https://stackoverflow.com/questions/tagged/wikidata. Поправил везде, добавил ограничение на значение свойства. -- Luitzen (talk) 11:44, 26 November 2018 (UTC)

Добрый день, хорошо. Вообще я не слишком хорошо разбираюсь в этом свойстве. — Ivan A. Krestinin (talk) 22:14, 27 November 2018 (UTC)

Could you stop adding persons' data to this data? It is an insult to the people added to it. --Ohtani tanya (talk) 12:27, 1 December 2018 (UTC)

Md5 hash problem at P1476

Hi Ivan, would you have a look at Wikidata:Database reports/Constraint violations/P1476? Following the discussion on Property talk:P1476, I tried to get the type constraint for title (P1476) to work again. --- Jura 10:20, 2 December 2018 (UTC)

  • BTW, would you kindly run another update of this report? I modified a few constraints, so it's not necessary to use current data. --- Jura 05:58, 21 December 2018 (UTC)

Hello Ivan A. Krestinin, Last update has been 15 days ago. Can you look into it? Kind regards, Lymantria (talk) 10:02, 13 December 2018 (UTC)

Updates are there. Thank you. Lymantria (talk) 08:40, 15 December 2018 (UTC)

User:KrBot recently replaced described at URL (P973) (pointing to wikia domains) with Fandom article ID (P6262). Will this job be performed regularly? --Loominade (talk) 10:25, 17 December 2018 (UTC)

Hi, several pages in Category:Pages with script errors are created by your bot, which occasionally creates pages with too many items for page to render so we get "The time allocated for running scripts has expired." error. I would like to bring the number of pages in that category to zero so it is easier to spot new issues. Any ideas how to modify your bot not to create pages with Lua errors? Possibly we could limit each query to some number of items returned, or not use {{Q}} if number of items returned is over some limit. --Jarekt (talk) 21:09, 2 January 2019 (UTC)

@Jarekt: There is some sort of limit for the usage of {{Q}}Wikidata:Database reports/Constraint violations/P281, for example, doesn’t use it (see the source code). Maybe the limit should be lowered. —Tacsipacsi (talk) 09:56, 3 January 2019 (UTC)
{{Q}} is a lua based template and no matter how effivient it is, if you have several thousands of them than you will run out of allocated lua time per page. I do not think we can or should alter those thresholds, but I think we should either put a limit on number of items returned by a query or not use {{Q}} on pages like Wikidata:Database reports/Constraint violations/P3426 or split them into several pages. --Jarekt (talk) 12:41, 3 January 2019 (UTC)
@Jarekt: It seems you misunderstood my answer. The P281 limit report shows that the bot avoids using {{Q}} on some pages. It seems pretty probable that it’s based on some threshold built in the bot, and that threshold should be lowered (so that more pages fall in the category where plain links are used). I don’t think the list should be truncated (i.e. less violating pages listed), as having harder-to-read entries is still better than not having those entries. Splitting the report into multiple pages makes changes harder to follow through page histories and watchlists, and also raises the question what we want to do with the blank subpages if the violation count drops. —Tacsipacsi (talk) 20:58, 3 January 2019 (UTC)
Ah, now I get it. Yes if you are right than your solution makes the most sense. --Jarekt (talk) 03:35, 4 January 2019 (UTC)
You are right, bot has threshold. Currently it is 2000. Previously it was 5000. Now you suggest make it lower again. It looks like Wikidata engine or the template performance degradation. I can change the threshold again, but it looks like workaround, not as real fix. I think last changes in {{Q}} should be reviewed first. — Ivan A. Krestinin (talk) 23:18, 4 January 2019 (UTC)

Constraint bot

Hello - your bot is missing a new constraint: [95]. Notice that other constraint bots seems to be able to cope with it. Deryck Chan (talk) 16:45, 5 January 2019 (UTC)

It also has problems with single-best-value constraint (Q52060874), see [96] Ahoerstemeier (talk) 15:01, 18 January 2019 (UTC)
✓ DoneIvan A. Krestinin (talk) 09:22, 16 February 2019 (UTC)

External uses from zh-yue.wp

Would you please add the following format strings to your KrBot so it will sync external uses from zh-yue.wp more accurately?

  • {{#property:Pxxxx}}
  • {{#statement:Pxxxx}}
  • {{#invoke:Wikidata|getValue|Pxxxx|
  • {{#invoke:Wikidata|getRawValue|Pxxxx|
  • {{#invoke:WikidataIB|getValue|Pxxxx|
  • {{#invoke:Wikidata|getImages|Pxxxx|

--Deryck Chan (talk) 18:58, 18 January 2019 (UTC)

✓ DoneIvan A. Krestinin (talk) 15:16, 17 February 2019 (UTC)

Your bot update of this page seems to be failing and blanking the page, while @Pasleim:'s DeltaBot seems to be updating the page OK. Can you have a look, and at least stop it from edit-warring? Thanks. Mike Peel (talk) 10:41, 20 January 2019 (UTC)

Same issue is present on Wikidata:Database reports/Constraint violations/P937, and if I check the bot's edit history several others - Andre Engels (talk) 11:45, 15 February 2019 (UTC)
I also noticed this on Wikidata:Database reports/Constraint violations/P144. Please respond. Multichill (talk) 21:37, 17 February 2019 (UTC)
"Too many files for existance check" error is fixed. "Unknown relation: Q30208840" is more interesting case. Every time then I see instance or subclass of (Q30208840) usage it is looked like bad modeling issue. I fixed such issues for several properties. So the question is: what is the better spend time to bad models support or spend time to fixing bad models. — Ivan A. Krestinin (talk) 10:00, 18 February 2019 (UTC)
Ivan, with all respect for your merits regarding the constraints system: it is not yours. instance or subclass of (Q30208840) is now more than 1.5 years old, and its invention bases on a lot of demand from the community. The constraints gadget supports it, and it is documented at Help:Property constraints portal/Type and Help:Property constraints portal/Value type. You may not like it because "it lookes like a bad modeling issue" (I do not like it as well), but I think you should really accept its existence, and I also think your bot should reflect this. I would be fine if you do not compute any data for this constraint type and write something like "constraint type Q30208840 in not supported by KrBot" in the reports. That way, the reports would not be completely broken. This is what I have already suggested half a year ago in #single best value constraint (Q52060874) further up this page (“pass routine”). —MisterSynergy (talk) 10:34, 18 February 2019 (UTC)
Please either fix your bot, stop it or have it skip reports like Wikidata:Database reports/Constraint violations/P144. Just having it out empty out valid reports by DeltaBot is not acceptable behaviour. Multichill (talk) 17:22, 20 February 2019 (UTC)
FixedIvan A. Krestinin (talk) 11:52, 17 March 2019 (UTC)

Blatantly false statements for P6379

Please revert as soon as possible all statements your bot made with has works in the collection (P6379) Musée d'Orsay (Q23402). They are all without basis, backed by fake references and the vast majority of them is blatantly false.

Can't you see in Répertoire des artistes (Q21518713) that this is a directory of artists about which the museum has documentation which is a totally different thing than having an artwork in the collections?? There are thousands of entries in the directory that are architects, what would that mean: storing buildings? Yes some entries are for artists that do have pieces in the collections, but by indiscriminately adding thousands that don't you are just rendering the data unusable and making the whole property laughable, so better revert it.

I'm not usually the kind of person to make rants, but this is so out of thought that it just vindicates all the people in the wikipedias that don't want to have anything to do with Wikidata as they see it as a giant sh*tload dumped by brainless bots. @Hannolans: as you are the root cause of this mess! --Nono314 (talk) 15:52, 23 February 2019 (UTC)

I discovered yesterday that this porperty is not about artists in the Musee d'Orsay, but about artists in the library of the museum. This was the reason I altered the item Répertoire des artistes (Q21518713) and the property description yesterday because you wrote "many artists are in the authority control database without having artworks in the collections!!".
I was already suspicious, that's why I several days earlier asked on Property_talk:P2268 about the aim of this property, but no response. So based on the earlier descriptions I made the wrong assumption that musee d'Orlee had that artists in the collection. Next step is to remove the statements. I tried to remove the statements yesterday with https://petscan.wmflabs.org/ but this didn't work. --Hannolans (talk) 18:29, 23 February 2019 (UTC)
Those statements are being removed right now. --Hannolans (talk) 19:53, 23 February 2019 (UTC)
Thank you @Hannolans: for having followed up on this, much appreciated. It's a pity when bots rush to add tens of thousands of false statements based on wrong assumptions, without any prior check --Nono314 (talk) 14:27, 27 February 2019 (UTC)
I hoped this bot would write those statements as we want to fill a base set of artists with works in collection. I added the property as a requirement forcing the bot to add those, but next time I will do a double check if the property is what is looks like. --Hannolans (talk) 11:22, 1 March 2019 (UTC)

Bad bot edit

Please stop your bot from doing this again. Thanks, Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 22:17, 28 February 2019 (UTC)

Just remove corresponding {{Autofix}} from Property talk:P856. — Ivan A. Krestinin (talk) 11:55, 17 March 2019 (UTC)

Don't change P973 to other identifier statements

See https://www.wikidata.org/w/index.php?title=Q61926006&action=history --Anvilaquarius (talk) 09:09, 1 March 2019 (UTC)

Just remove corresponding {{Autofix}} from Property talk:P973. — Ivan A. Krestinin (talk) 11:58, 17 March 2019 (UTC)

P2862 autofix

Could you apply the autofix of Catalogus Professorum Academiae Rheno-Traiectinae ID (P2862) as well to references? This for some upcoming bot project. Sjoerd de Bruin (talk) 13:33, 25 March 2019 (UTC)

Review GNIS/Geoname constraints for lakes

The KrBot seems to have applied instance of Lake in the USA where they should be instance of spring. I've reverted a few, but checking the instance to match the type would be ideal cleanup of the bot's applied instance of. This could be pulled from either an associated GNIS ID or Geoname ID. Example revert. Thanks. Wolfgang8741 (talk) 23:50, 28 March 2019 (UTC)

P2907 UTC timezone offset range constraints

Many of not all time zones are failing their UTC timezone offset (P2907) range constraints; see e.g. UTC−05:00 (Q5390). I'm not sure what's going on, but it strikes me that your change [97] might either be the cause of this problem, or an attempt to fix this problem, but I'm not sure which. Scs (talk) 14:49, 7 April 2019 (UTC)

New constraint status

The other day, I read that in May, there will be a new constraint status ( Wikidata:Project_chat#New_status_of_constraint:_"suggestion" ). Supposedly the bot reports would or should just handle them like the current non-mandatory ones. --- Jura 10:44, 14 April 2019 (UTC)

Hello Jura, thank you, added. — Ivan A. Krestinin (talk) 23:46, 14 April 2019 (UTC)

Automatic replacements

Hi, I've added some autofixes to Property talk:P973 and tried testing them at Wikidata Sandbox (Q4115189). Unfortunately KrBot didn't edit the item today. Does KrBot skip that item? How could I verify that the autofixes work? Jc86035 (talk) 16:41, 18 April 2019 (UTC)

Never mind; the bot has already made the autofixes. Jc86035 (talk) 18:30, 20 April 2019 (UTC)
Bot had network connection. Now all is fine. — Ivan A. Krestinin (talk) 18:56, 20 April 2019 (UTC)

Wikidata:Database reports/Constraint violations/P496

Hoi, this report used to be updated once every three days. It has not been updated since 11-4. Is something wrong? Thanks, GerardM (talk) 21:30, 19 April 2019 (UTC)

Bot had network connection. Now all is fine. — Ivan A. Krestinin (talk) 18:57, 20 April 2019 (UTC)

Can KrBot stop deleting item from married couple (Q3046146) and duo (Q15618652)

Hi,

For exemple here and here ?

eru [Talk] [french wiki] 06:35, 22 April 2019 (UTC)

  • Hello, please see description of occupation (P106). The property is person-related, not person group-related. Extending scope of this property is not very good idea. Often some bots import occupation (P106) from Wikipedia. And many bots make the same mistake. Its import value that is applicable for one person only to person group item. Now my bot fixes such errors. The mechanism will be broken if scope of occupation (P106) will be extended. — Ivan A. Krestinin (talk) 00:07, 25 April 2019 (UTC)

Tűzhányó blokk

FIGYELEM! KengurúZsebe (talk) 18:15, 9 May 2019 (UTC)

Hello, I do not understand you. — Ivan A. Krestinin (talk) 18:53, 9 May 2019 (UTC)

Extra run requested

Hi. I would like to have an update on Wikidata:Database reports/Constraint violations/P2951. Can you trigger the bot? best regards, Thomas Ledl --09:19, 18 May 2019 (UTC)

thx! --Thomas Ledl (talk) 12:01, 18 May 2019 (UTC)

Hi Ivan,

I had a look at the above and tried to fix a couple. Maybe some need some adjustments from your side:

"synonym (P5973) Can not present value 'L3036-S4' (type lexeme sense) as entity key. "
probably from https://www.wikidata.org/wiki/Lexeme:L9#S3 . Not sure why though.
"genre (P136) Constraint Type does not accept qualifier P4680 "
"narrative role (P5800) Constraint Type does not accept qualifier P4680 "
could that be supported?
"spam black list"
maybe the reports for the two identifiers could be generated without links. (unless there is a way around the global list).
"preferred pronoun (P6553) Can not present Q7536 as lexeme."
not sure whats meant here.
"Commons quality assessment (P6731) Bad value of P2305 Q59712033"
if present as only value Wikibase MediaInfo (Q59712033) should lead the property to be usable only on Commons (eventually possibly there).

Hopefully, the 503s work next time. --- Jura 17:39, 18 May 2019 (UTC)


Hi Jura,

"synonym (P5973) Can not present value 'L3036-S4' (type lexeme sense) as entity key."
TODO for me: support symmetric constraint (Q21510862) for forms.
"genre (P136), narrative role (P5800) Constraint Type does not accept qualifier P4680 "
looks like redundant qualifier. subject type constraint (Q21503250) is checked for main values only always.
"spam black list"
looks like formatter URL (P1630) should be deleted to be consistent with spamlist.
"personal pronoun (P6553) Can not present Q7536 as lexeme."
incorrect usage of lexeme requires lexical category constraint (Q55819078). lexeme requires language constraint (Q55819106) and lexeme requires lexical category constraint (Q55819078) check lexeme there the property is specified, not target lexeme.
"Commons quality assessment (P6731) Bad value of P2305 Q59712033"
I did not find usage of this property to understand how it should work.
"503 error"
looks like Wikidata servers have some bug. Sometimes report writing request fails. Probability is higher for large reports.

Ivan A. Krestinin (talk) 21:45, 18 May 2019 (UTC)

Hi Ivan, Thanks for your feedback.

"Constraint Type does not accept qualifier P4680"
narrative role (P5800) makes an interesting distinction with the scope, depending on whether it's used as a qualifier or as main value. Maybe you could support at least one of the two checks.
"spam black list"
there was some lengthy discussion before it was added, but I'm not particularly interested in either report
"Can not present Q7536 as lexeme"
good point. Seems we lack a suitable constraints for that.
Wikibase MediaInfo (Q59712033)
It should mean that eventually only uses like this are possible. Unless also another entity type is specified.
--- Jura 09:20, 19 May 2019 (UTC)

Hi Jura, Wikibase MediaInfo (Q59712033) and lexeme-value-requires-lexical-category constraint (Q64006792) support were added. Spamlist issue was fixed also. — Ivan A. Krestinin (talk) 21:08, 25 May 2019 (UTC)

Krbot errors

Hi, the bot seems to have made a lot of errors in adding collection (P195) and V&A item ID (P3929) properties relating to the Victoria and Albert Museum (Q213322). For example, the properties have been added to Edna Best (Q445510) but the item in the collection is not the actress but a photograph of the actress. Albert Memorial (Q281465) has the properties, but the item in the collection is a photograph of the memorial, not the memorial. See also The Theatre Royal (Q1756626). capital (Q193893) has this issue, but the thing in the collection is a specific capital, not the concept of capital. The bot has given Shakespeare plays such as King Lear (Q181598) the properties, but the collection holds posters or designs advertising the plays, not the plays themselves. There are dozens of affected items: I haven't the time to identify all of them. Can you undo the relevant edits? Thanks in advance for any help, MartinPoulter (talk) 11:30, 12 June 2019 (UTC) Looking again, it seems other users on Mix'n'Match have made the false matches, and Krbot has added identifiers consistent with those matches. So the bot has acted sensibly within its own rules, but it's added false information. I'll raise this on project chat. MartinPoulter (talk) 11:34, 12 June 2019 (UTC)

Bot make such changes because the items have V&A item ID (P3929) property. The property was added by MistressData. Looks like the most of V&A item ID (P3929) usages should be deleted due reasons listed above. @MistressData: what do you think? — Ivan A. Krestinin (talk) 22:55, 13 June 2019 (UTC)

Edit comment for VIAF redirect resolving

Hi! Thanks for resolving VIAF redirects! However, such edit comments are not useful, I would have no idea what’s going on if I didn’t know it’s redirect resolving. Could you please add some edit summary after the automatic one that explains the intent of your edit? Thanks in advance, —Tacsipacsi (talk) 16:04, 16 June 2019 (UTC)

Constraint reports for P214

Hi! Since the beginning of May, your bot’s only edits to Wikidata:Database reports/Constraint violations/P214 are replacing DeltaBot’s constraint report with

ERROR: Edit failed, error code 'badmd5' (The supplied MD5 hash was incorrect.).

Could you please check what’s happening here? —Tacsipacsi (talk) 20:09, 22 June 2019 (UTC)

Nope, it still fails. (Except if it’s the nearest update, completed in more than two days…) —Tacsipacsi (talk) 14:47, 26 June 2019 (UTC)

P973>P214 for long identifiers

Hi Ivan,

At Q48338088#P973 (and a few others), the conversion doesn't seem to work. It seems autofix-regex and format constraint seem ok. --- Jura 02:31, 24 June 2019 (UTC)

Hi Jura, it is strange situation. Most probably it is bug in PCRE library. I split the autofix and it works as expected. — Ivan A. Krestinin (talk) 22:15, 25 June 2019 (UTC)


Property uses in templates

Hi Ivan,

From what I have seen for P6801, I think uses in the two above don't get detected. --- Jura 12:37, 1 July 2019 (UTC)

Too many uses ?

At Wikidata:Bot_requests#Fix_encoding_errors_in_author_name_string, I suggested autofix, but it occurred to me later that might not work given the number of uses of the property. --- Jura 19:26, 4 July 2019 (UTC)

Semi colon in author names

It seems to me you have told your bot to separte author name strings in case there is a semi colon. However, this yielded some errors like this, where the semi colon is part of a wrong encoding. I noticed also a few other errors. Can you fix this? Steak (talk) 17:36, 6 July 2019 (UTC)

Property uses count

Hi Ivan,

As we lack recent updates at https://www.wikidata.org/w/index.php?title=Template:Property_uses&action=history I was wondering if KrBot could occasionally update it. It's used on property documentation pages for current uses and "usage history".

As it uses the link table, the counts slightly different from those of KrBot, but I suppose that shouldn't matter that much. --- Jura 07:43, 12 July 2019 (UTC)

I need some data source for updating. I have dumps of course. But its are not very good way due several reasons. Maybe you can suggest something else? — Ivan A. Krestinin (talk) 21:01, 17 July 2019 (UTC)
Personally, I wouldn't mind the data from dump. I think deltabot used something like this, but it now times out. Maybe splitting the query into several and then combing its results works (sample for P3%). --- Jura 12:07, 18 July 2019 (UTC)

Same source, new statement

Hi Ivan, KrBot is resolving redirects what is great. Unfortunately it doesn't change the source. The result:

So if your bot is changing a stament, please remove the source or add a note. --Kolja21 (talk) 01:58, 14 July 2019 (UTC)

Hi Kolja, it is interesting question. My bot makes very minor change. It just resolves redirect. This is technical action. It does not affect original meaning of the link. So source of the identifier is not changed during resolving redirects. — Ivan A. Krestinin (talk) 05:38, 15 July 2019 (UTC)

Autofix of author names

Why did KrBot stop autofixing author name strings? At Property_talk:P2093 there are change prescriptions, and there are still a lot of author name string statements that need to be changed according to these prescriptions. Steak (talk) 08:55, 14 July 2019 (UTC)

Autofix works not very good for this task. I started another task. Now it is in progress. — Ivan A. Krestinin (talk) 06:49, 15 July 2019 (UTC)
It seems that it stopped again? Steak (talk) 10:11, 15 July 2019 (UTC)
Bot completed the task. Could you provide links to some items that were not fixed? — Ivan A. Krestinin (talk) 21:26, 15 July 2019 (UTC)
For example Q28083129, Q57725646. Steak (talk) 08:29, 16 July 2019 (UTC)
Bot does not touch item if result value violates some constraints. For Q28083129 result value "María E. López" violates format constraint ([^⊥∥†\*‡§©\|¶€±®¡;í]{1,250}). Looks like list of replacements provided on Property_talk:P2093 is incomplete. Could you provide full list? — Ivan A. Krestinin (talk) 20:58, 17 July 2019 (UTC)
No thats not possible, the "í" has to be replaced differently depending on the name, therefore one cannot give an autofix prescription. Can you modify the bot such that it also edits values with a constraint violation? Otherwise I could temporarily change the constraints. Steak (talk) 21:21, 17 July 2019 (UTC)

I have meanwhile changed the constraints, but still no autofix from the bot?! Steak (talk) 10:05, 27 July 2019 (UTC)

Error list: User:Ivan A. Krestinin/FormatFixErrors. Could you provide some rules how to fix items from the list. The rules can be more complex than allowed by {{Autofix}}. — Ivan A. Krestinin (talk) 11:31, 27 July 2019 (UTC)
I have added a few rules to User talk:Ivan A. Krestinin/FormatFixErrors. Can you fix those, and then we will see further. Steak (talk) 12:24, 27 July 2019 (UTC)
Those, one could either simply remove or set to <no value>. I have no preference in this case. Steak (talk) 09:15, 28 July 2019 (UTC)
Except the NAs, I have fixed the remaining problems by hand. Steak (talk) 14:04, 30 July 2019 (UTC)

Bot edits on fanqie (P5523)

Greetings. Can you revert all bot edits related to this property? For Chinese characters, fanqie readings always come in pairs, so the number of items under this property has to be in multiples of two, e.g. two items for a single reading, or four items for two readings, and so on. In some characters such as (Q3595028), there are only three items under this property so it does not make sense, especially if no references are provided. Things can get even more confusing if there are up to eight fanqie readings but only eleven items under this property such as the example below, so kindly revert the bot edits when you have the time to do so. KevinUp (talk) 14:28, 17 July 2019 (UTC)

Item Fanqie (initial) Fanqie (final) References
Guangyun (Q2189818)
Guangyun (Q2189818)
Guangyun (Q2189818)
Guangyun (Q2189818)
Jiyun (Q35792)
Jiyun (Q35792)
Jiyun (Q35792)
Jiyun (Q35792)
I stopped deleting duplicate values for fanqie (P5523). But I have some troubles with reverting edits now. I will try to revert edits latter. — Ivan A. Krestinin (talk) 21:03, 17 July 2019 (UTC)
If the edits cannot be reverted, try listing out items that have odd numbers of fanqie (P5523) so that I can clean them up manually. Thank you. KevinUp (talk) 06:05, 18 July 2019 (UTC)
@KevinUp: Even in the current model of data, it's impossible to recover the pairs from only set of upper characters and set of lower characters (assuming we use a format not persisting the order of the statements, like RDF). We should either:
  1. Create a new property "value group number" (qualifier) and add different value group numbers to each of pairs. the "value group number" will otherwise means nothing.
  2. Create two new properties "upper character of fanqie" and "lower character of fanqie", replace all values of fanqie with list of values as qualifiers (Q23766486) and add the data using these two qualifiers.

--GZWDer (talk) 16:13, 18 July 2019 (UTC)

I have created Wikidata:Property proposal/value group number. Please comment there.--GZWDer (talk) 16:25, 18 July 2019 (UTC)
Thank you for creating this proposal. Since there are only less than 50 items affected by the bot, I think I can manually fix the errors using Special:WhatLinksHere/Property:P5523. KevinUp (talk) 21:54, 18 July 2019 (UTC)

Question about a replacement of an GND-Id

Hello! It is about this change: Special:Diff/976919768 How did you determine that new number 118630202 ? Note: Just a question, no need for some change of the behaviour.

Background: I do a lot with these numbers in german WP. I read the dump-files of the Deutsche Nationalbibliothek for these GND-IDs and in these files I found a different replacement (see next change by user Kolaj21). However, the website redirects the user in the same way as your change reflects. Seems to be a bug somewhere in Deutsche Nationalbibliothek, I have already contacted some person there. --Wurgl (talk) 06:22, 19 July 2019 (UTC)

wbEntity config value to be dropped on July 24th

Hello,

We are about to drop the mw.config.get( 'wbEntity') config value, that is deprecated for two years. Starting on Wednesday, July 24th, scripts that use this value may encounter issues.

I noticed that your scripts located on User:Ivan A. Krestinin/Gadget-AuthorityControl-Local.js and User:Ivan A. Krestinin/Gadget-MoveClaim.js are still using this value. I suggest that you update it, for example by using the hook wikibase.entityPage.entityLoaded (see an example here).

If you have any questions or need help, feel free to leave a comment under the related task.

Thanks for your understanding, Lea Lacroix (WMDE) (talk) 09:02, 22 July 2019 (UTC)

Q33467482

Please revert redirect resolvations for Nijmegen (Q33467482). Sjoerd de Bruin (talk) 10:36, 22 July 2019 (UTC)

Constraints (Krbot2) not running?

Hi - the constraint reports are super useful... but it looks like KrBot2 hasn't run any since July 11? Any idea when it could go again? ArthurPSmith (talk) 17:28, 24 July 2019 (UTC)

I had connection stability issue. The reports are updating normally now. — Ivan A. Krestinin (talk) 09:17, 27 July 2019 (UTC)
Thanks! Is there a schedule (once a week or something like that) or just when you are able to do it? ArthurPSmith (talk) 20:24, 31 July 2019 (UTC)
There is no special schedule. Bot is working continuously. But sometimes bot has connection or some other troubles. — Ivan A. Krestinin (talk) 12:53, 7 August 2019 (UTC)

Can this report be updated?--GZWDer (talk) 21:47, 25 July 2019 (UTC)

Database reports/Constraint violations/P227

Hi Ivan, can you manage to exlude items where GND ID (P227) is marked as "deprecated rank" from the list Wikidata:Database reports/Constraint violations/P227? See for example: Jeffrey Williams (Q382202). This would be of great help. --Kolja21 (talk) 21:36, 27 July 2019 (UTC)

Currently bot bot ignores ranks at all. It is not simple to add ranks support. Currently you can add the item to ignore list. — Ivan A. Krestinin (talk) 22:55, 27 July 2019 (UTC)
I think you should use single-best-value constraint (Q52060874) here. Steak (talk) 08:27, 28 July 2019 (UTC)
Ranking recognition has been requested several times, but as far as I understand it is not at all simple to include it in Ivan's bot framework. My take is that there will probably never be support for anything which involves ranks in Ivan's constraint violation lists.
An alternative approach would be to query the lists either on-the-fly (such as this one, but can be altered of course), or to write the results from such a query to a wiki page with a small custom bot script with daily execution. —MisterSynergy (talk) 08:39, 28 July 2019 (UTC)
If ranks are not supported, may be it make sense to ignore single-best-value constraint (Q52060874), because right now we are getting quite useless sections like Wikidata:Database reports/Constraint violations/P6879#"Single value" violations --Ghuron (talk) 11:55, 14 August 2019 (UTC)
Коллега, я сейчас заливаю много астрономической инфы, и мне бы очень помогли ваши отчеты, но в таком виде они, увы, бесполезны. Давайте исключим single-best-value constraint (Q52060874)? --Ghuron (talk) 15:31, 2 September 2019 (UTC)

Could you please update this page? The property is now used in many more items than recently. --Leyo 21:28, 1 August 2019 (UTC)

Nicolas Théodore de Saussure

Hello. Can you stop KrBot from including Nicolas Théodore de Saussure (Q116058) in the Dutch Academy? On the link to the source provided the information is not available. Either the link should be mended, or the entry deleted. Thank you. Sapphorain (talk) 07:47, 18 August 2019 (UTC)

Hello, looks like KNAW site is not fully operational now. But it presents in archive: https://web.archive.org/web/20190421124258/https://www.dwc.knaw.nl/biografie/pmknaw/?pagetype=authorDetail&aId=PE00002767 Is this link ok? — Ivan A. Krestinin (talk) 08:52, 18 August 2019 (UTC)
Yes, thanks, this link is fine. In any case, I think in such cases an interactive search site is not appropriate: the direct link should always be provided. Sapphorain (talk) 09:28, 18 August 2019 (UTC)
I am not fully understand you. What do you mean under "interactive search site"? — Ivan A. Krestinin (talk) 06:52, 15 September 2019 (UTC)

Autofixes

As I started using formatter URL (P1630) to make autofixes, I figured Wikidata:Property proposal/mobile formatter could help build more of them. --- Jura 09:51, 25 August 2019 (UTC)

The belonging to some academy must be independently sourced; it is not sufficient to just report that one or several wikipedia pages say so. Sapphorain (talk) 21:29, 8 September 2019 (UTC)

Resolving redirects

Hello! Please, don't make your bot "resolve redirects" if the target identifier is already present in the item, otherwise we obtain duplicates. Furthermore, in the provided example the statement is marked as "deprecated" exactly because now the identifier is a redirect. --Horcrux (talk) 10:30, 13 September 2019 (UTC)

Hi! Bot removes duplicates also. Bot had some troubles with this exact item and did not remove duplicate. As I understand your idea, you want to add all old VIAF identifiers to Wikidata. But VIAF was too much redirects for every item. Are you sure that we need to collect all this information? Maybe better approach is focusing on actual identifiers? — Ivan A. Krestinin (talk) 07:00, 15 September 2019 (UTC)
There are also other items where the bot created the duplicate. Anyway, you should at least ping the users you are replying to. --Horcrux (talk) 09:37, 15 September 2019 (UTC)
FixedIvan A. Krestinin (talk) 11:15, 15 September 2019 (UTC)


Failed to connect to/No route to host

On Wikidata:Database_reports/Constraint_violations/Errors, there seems to be a new type of error: "Failed to connect to ..". In case you haven't noticed yet. --- Jura 07:36, 20 September 2019 (UTC)

2019-09 KrBot

Hello. FYI Ireverted today KrBot edits on Q67800615. Visite fortuitement prolongée (talk) 20:18, 28 September 2019 (UTC)

Hello, am I understand the situation correctly? https://onlinelibrary.wiley.com/doi/abs/10.1111/sena.12230 describes review of The Emergence of Iranian Nationalism: Race and the Politics of Dislocation (Q67800615), not the book. — Ivan A. Krestinin (talk) 11:43, 29 September 2019 (UTC)
Yes. Visite fortuitement prolongée (talk) 20:04, 29 September 2019 (UTC)
Review is separate work, it should be described in separate item and connected with original work using review of (P6977). For example I created Zia‐Ebrahimi, Reza The Emergence of Iranian Nationalism: Race and the Politics of Dislocation (Q69136730). — Ivan A. Krestinin (talk) 20:26, 29 September 2019 (UTC)

New page for catalogues

Hi, I created a new page for collecting sites that could be added to Mix'n'match and I plan to expand it with the ones that already have scrapers by category. Feel free to use, expand. Best, Adam Harangozó (talk) 19:59, 19 October 2019 (UTC)

suggest meta user page for user:KrBot

Hi. With this bot editing at wikidata, yet the user links showing on each wiki as a red link, can I ask for you to consider creating a user page for your bot at metawiki. It will enable users to better understand the edits being made. Thanks.  — billinghurst sDrewth 21:32, 31 October 2019 (UTC)

If you need assistance getting it done at meta, then please welcome to ping me.  — billinghurst sDrewth 21:33, 31 October 2019 (UTC)
Hi, I like red links :) But my bot has user page in every wiki where it makes some edits. What page shows the bot name using red color? — Ivan A. Krestinin (talk) 22:43, 31 October 2019 (UTC)

Bad authors on Q64861757

The authors imported from Semantic Scholar on Vector boson scattering: Recent experimental and theory developments (Q64861757) don't seem to be related to the authors listed at [98], and include a lot of organizations. Batavia (Q64861753) is also a nonsense item. Ghouston (talk) 07:44, 21 November 2019 (UTC)

The records were created by User:QuickStatementsBot, not by my bot. — Ivan A. Krestinin (talk) 22:00, 24 November 2019 (UTC)
Ah, your bot just fixed redirects. Sorry. Ghouston (talk) 05:31, 25 November 2019 (UTC)

Duplicates from The Peerage

Hello! Due to a massive import of data by GZWDer (flood) bot (hundreds of thousands of names from The Peerage (Q21401824)), many entries are receiving duplicate values of children or parents (the dumb, blind bot adds values that are already present). I don't know how widespread the problem is, but see for instance children of Benjamin Seymour Guinness (Q27942475), and father of Dorothy Gould Burns (Q5298437). This has instant results in infoboxes that display children and parents. It is also likely that many newly imported items will be merged, as identical to previously existing items. Could you perhaps prioritize KrBot to examine changes by GZWDer (flood), and delete duplicates? In cases where the existing value has more detailed information, such as multiple references, or start and end date qualifiers for spouses, the more detailed entry should obviously be retained. Thanks. -Animalparty (talk) 03:16, 27 November 2019 (UTC)

Hi, the task is done, but not so fast as you want. Sorry. — Ivan A. Krestinin (talk) 12:57, 13 January 2020 (UTC)

subclass vs instance of list of poker hands

Hi:


KrBot just went through and made poker hands be subclasses of list of poker hands (Q40742). I think this is wrong - list of poker hands (Q40742) is a collection of the kinds of poker hands, not a collection of actual poker hands. The poker hand Queen of hearts, Ace of spaces, Jack of diamonds, 2 of clubs, and 2 of clubs is not an instance of list of poker hands (Q40742). So one pair (Q20679894) is not a subclass of list of poker hands (Q40742), but is instead one of its instances.

Hi, are you about this edit? This edit was done because the elements were merged: [99]. Please discuss the merge reason with User:Aytrus. — Ivan A. Krestinin (talk) 12:54, 13 January 2020 (UTC)
The merge had to be made, the page was exactly the same. I don't know, if you read one pair — taking your example — as a subclass considering the different suits it could make sense, but feel free to change the statement to instance, I understand your point. — Aytrus (talk) 12:14, 22 January 2020 (UTC)

KrBot: Moving DOI URL to DOI is not always correct

Hi - this automated edit on an item I've been working on is not an appropriate edit, I believe. I added a URL to the "described at URL" field. However, I used a DOI link, which means your bot rewrote it as "DOI". That changes the semantics: I already have a DOI for the item, and it's a DOI for the item itself. I wanted the "Described at URL" field to be used to link to a paper which describes the item (as opposed to being the item). Could you comment please? --Mcld (talk) 18:33, 17 December 2019 (UTC)

I'm still concerned about this. Your automated edit is equivalent to assuming that "URL (P2699)" is the same concept as "described at URL (P973)", which I believe is incorrect. To put it a different way: you should not map described at URL (P973) to DOI (P356). You should map it, if at all, to a property "Described at DOI" (which does not exist!). What is your opinion? --Mcld (talk) 13:27, 1 January 2020 (UTC)
Bot uses {{Autofix}} rules from page Property talk:P973. You are free to remove some rules from Property talk:P973 if its are incorrect. Or discuss the rules on Property talk:P973. — Ivan A. Krestinin (talk) 12:40, 13 January 2020 (UTC)

Taxon mismatch

This edit [100] cannot be correct. One taxon is an order, the other is a family. --EncycloPetey (talk) 23:39, 14 January 2020 (UTC)

Commons:Category:Anthocerotales is redirect to Commons:Category:Anthocerotaceae. — Ivan A. Krestinin (talk) 19:03, 15 January 2020 (UTC)
That doesn't make it the correct target for the link. --EncycloPetey (talk) 15:50, 17 January 2020 (UTC)
Just fix redirect on Commons and after this fix link in Q5696912. — Ivan A. Krestinin (talk) 00:05, 18 January 2020 (UTC)
The redirect location is the correct target, see Category:Anthocerotales (Q6184206). Your edit means that the two will no longer match, and there will be two data items with the family link, and this data item will have two different associated categories. Please stop making the incorrect change. --EncycloPetey (talk) 16:29, 18 January 2020 (UTC)
Commons category (P373) should never link to a redirect, it should link to a Wikimedia Commons category containing files related to this item. Anthocerotales (Q5696912) is about an order that only has one family (Anthocerotaceae (Q1050205)) so actually all the related files about Anthocerotales (Q5696912) are in Commons:Category:Anthocerotaceae so the bot edits are correct. Multichill (talk) 17:07, 18 January 2020 (UTC)
If you cannot correct then problem, then I will remove the problematic value. As I say, you are creating a situation where the same taxon has two different categories listed, and the same Commons category appears on two distinct taxa data items. This is a problem. --EncycloPetey (talk) 16:37, 19 January 2020 (UTC)

Incorrect addition of "different from"

This edit is incorrect because resistance to alternating current (Q1048490) is a special case of, or superclass of (Q66088480), electrical resistance (Q25358), not different from resistance to alternating current (Q1048490). Jc3s5h (talk) 15:42, 19 January 2020 (UTC)


Constraints not updating?

Hi - I've been watching Wikidata:Database reports/Constraint violations/P356 and it updated 3 times in early January, but hasn't done so for over a week now. Something blocking the bot? ArthurPSmith (talk) 20:33, 22 January 2020 (UTC)

Hi Arthur, I fixed two issues. Update frequency will increased as I hope. — Ivan A. Krestinin (talk) 20:32, 1 February 2020 (UTC)
Out of curiosity, what would it take to run these updates daily? Is the bottleneck computing power, or something else? Thanks. Mike Peel (talk) 20:39, 1 February 2020 (UTC)
Bottleneck is computing power. More efficient algorithms may help also. — Ivan A. Krestinin (talk) 18:15, 4 April 2020 (UTC)

Code?

Out of curiosity, is the code for KrBot available somewhere? I'm interested in looking at how you find the constraint violations, as that could improve the efficiency of some of Pi bot's tasks. Thanks. Mike Peel (talk) 20:29, 25 January 2020 (UTC)

Hi Mike, the code is not available in public. Generally bot downloads Wikidata dumps, load its and check all values using different constraints. The code is written on C++ and uses several custom libraries. So it will not very useful for you as I think. But I can provide details about parts that are interesting for you. — Ivan A. Krestinin (talk) 20:26, 1 February 2020 (UTC)
Thanks for your reply. I used to code in C++, but nowadays I use Python, so perhaps I can't directly reuse your code, although I'd be interested in looking at it if you can make it publicly available. To pick a specific example, your bot updates the Commons link violations at Wikidata:Database reports/Constraint violations/P373. I wrote a Python script at [101] to try to remove the bad links automatically on a daily basis, but I re-run it through the constraint violation report every time your bot updates it, and that seems to find extra cases. I'd be interested in learning how you find those extra cases, so that pi bot can handle them quicker. Thanks. Mike Peel (talk) 20:34, 1 February 2020 (UTC)
Bottleneck of my bot is dumps management. Currently bot loads all values of all properties from dumps into memory. This is long process due to dumps size. I am improving this mechanism now. I hope the improvements will increase update frequency. You can try to load needed values directly from Wikidata DB. It should be possible on https://tools.wmflabs.org. This will allow you to get more actual data. About P373: I also have bot that fixes wrong Commons category (P373) links. — Ivan A. Krestinin (talk) 21:00, 1 February 2020 (UTC)
I hope you don't mind, but I've raised this at Wikidata:Contact_the_development_team#Increasing_the_frequency_of_constraint_violation_report_updates - you provide a vital service, and it would be good if this was better supported. I'll follow up about P373 soon. Thanks. Mike Peel (talk) 21:33, 1 February 2020 (UTC)

Unique/Single value constraints – one item/value listed twice as a violation

Hey, I just want to let you know that user:KrBot2 listed some violations in a wierd manner: [102]. One item is listed twice as a violation of a unique value violation and one value is listed twice as a single value violation. Wostr (talk) 11:38, 1 February 2020 (UTC)

…and in many other reports, some really extreme ones being Wikidata:Database reports/Constraint violations/P2639 (ca. 3400 false positives) and Wikidata:Database reports/Constraint violations/P2250 (1000+ false positives). (@Wostr: The fact that the same item shows up as a unique value violation and the same value shows up as a single value violation does not necessarily mean that the bot is wrong: sometimes people add the same value twice—usually using some sort of (semi)automated edit—, and then it’s right to list these as violations. But that’s not the case now, at least not for the most of the reported violations.) —Tacsipacsi (talk) 12:28, 1 February 2020 (UTC)
Hi, thank you for the report. The issue is fixed. Please wait for the next update. — Ivan A. Krestinin (talk) 20:17, 1 February 2020 (UTC)
Just a comment that I noticed this too, particularly on the report for DOI (P356) - suddenly there are over 400,000 violations of the single-value constraint! ArthurPSmith (talk)

Hi Ivan, if it's bug fixing time, could you please look at the discussion here? Not sure if you noticed my ping there. The community thinks that deprecated values should not trigger the unique value violation... Is that fixable or not? Cheers, Vojtěch Dostál (talk) 22:22, 1 February 2020 (UTC)

@Ivan A. Krestinin: You made an update on Feb 1. 2020 but the bug is still not fixed. -- MovieFex (talk) 12:59, 5 February 2020 (UTC)
The same problem in Wikidata:Database reports/Constraint violations/P6359. -- MovieFex (talk) 13:16, 5 February 2020 (UTC)
The February 1 update was the one that went wrong (see the above timestamps), of course it’s not fixed. The next one should be good. —Tacsipacsi (talk) 13:52, 5 February 2020 (UTC)

KrBot and maxlag

Hi! I see from Special:Contributions/KrBot that this bot is currently editing at about 100 edits per minute, but maxlag is currently at 17. How do you account for this? See also Wikidata:Administrators'_noticeboard#WDQS_lag_is_terrible_(over_9_hours_now) and mw:Manual:Maxlag_parameter. Cheers, Bovlb (talk) 16:57, 10 February 2020 (UTC)

Wrong merge and consequences

Hi,

Last November, an IP wrongly merged brigand (Q20650523) and Dacoity (Q17176963). The merge has been undone been meanwhile KrBot resolved the redirection (which was a good idea in theory but not in this specific case). Could KrBot now undo this batch https://tools.wmflabs.org/editgroups/b/KrBotResolvingRedirect/Q20650523_Q17176963/ ?

Cheers, VIGNERON (talk) 18:34, 11 February 2020 (UTC)

2020-02 KrBot2

Hello,

Could you run KrBot2 every 5 day? Visite fortuitement prolongée (talk) 14:19, 15 February 2020 (UTC)

Or at least, allow the bot to be triggered manually like ListeriaBot. That would be very useful, if you have time. Just noticed that the P7882 report is old. --Ysangkok (talk) 17:53, 27 February 2020 (UTC)
Bot cycle is ~5 days now. It works automatically. My current attempt to reduce the time fails. I hope the next attempt will be more successful. — Ivan A. Krestinin (talk) 18:08, 4 April 2020 (UTC)
Thank you very much. Visite fortuitement prolongée (talk) 22:39, 5 April 2020 (UTC)

Wrong ISBN-13 to ISBN-10 transfer

I have rolled back your update. You removed the correct ISBN-13 value and wrongly used it as a (wrong) ISBN-10 value. Geertivp (talk) 09:59, 18 February 2020 (UTC)

KrBot block

Hey Ivan, your bot ignores the maxlag parameter since 2 PM today, thus I blocked it now. You can see its edit pattern in these Grafana charts, particularly in the "Max Single User Edit Rate" panel. It is your bot which accounts for the ~100 edits/min in that chart, and it obviously does not stop during phases of high database load.

Since server resources are unfortunately very limited, I had to block the bot. All other bot operators do respect the maxlag parameter as indicated by the Wikidata:Bots policy, and it is only fair if you do so as well. Please let me know when you have implemented it properly, as the bot can then be unblocked again of course. In case of questions, feel free to ask. —MisterSynergy (talk) 17:48, 6 March 2020 (UTC)

Related to this, can you please collapse these edits into one edit? I assume you're using https://www.wikidata.org/w/api.php?action=help&modules=wbsetdescription , with https://www.wikidata.org/w/api.php?action=help&modules=wbeditentity you can just do it in one edit like in the example. Multichill (talk) 15:55, 7 March 2020 (UTC)

The last successful update to this page by your bot was over half a year ago. It should nowiki all links if it runs into the spam blacklist. (By the way, I think Facebook onion address (Q24590047)’s constraint violation causes triggers the spam blacklist, but that’s not important, as the bot doesn’t need to detect it, just nowiki everything.) —Tacsipacsi (talk) 00:59, 23 March 2020 (UTC)

Please remove **non**-violations from reports (green boxes)

The bot posts statistics about cases which are NOT violations. This complicates the task for checking properties as these properties then have links pointing to them as "Report/Pnnn/violations" when in fact there are none: we have to search each property by their Qid in some very long pages (very long because of violations reported there).

Please remove the "green" boxes for these report pages, only list the violations (red boxes). Notably for the generated tables for "Allowed types" listing all references (valid or not).

You may want to report instead on other pages (not "violation" pages) the cases where there are no violation detected, but I think it is just not needed at all.

Thanks. Verdy p (talk) 05:50, 23 March 2020 (UTC)

I think the green boxes are a quite useful feature, they help getting a picture about how the property is used. Maybe they could use external links instead of internal ones, although I don’t know how external links work with displaying labels (whether getting label generates a backlink, how much more complicated the module becomes with this extra feature etc.). —Tacsipacsi (talk) 00:01, 24 March 2020 (UTC)

Canadiana Authorities ID

Hi. I have seen your bot removing the suffixes E and F from the values. Please do not do that. The english entries with E are deprecated, as those have been merged into the NACO authority file. The french entries with F are still valid and are used as a base for the new "Canadian name authorities in french" file, which is part of VIAF. Example: [103] --Sotho Tal Ker (talk) 20:34, 15 April 2020 (UTC)

This job is stopped. — Ivan A. Krestinin (talk) 20:57, 15 April 2020 (UTC)

Hello Ivan, your last update 2020-04-20 was with old data from 2020-04-13. If you have such a long time between the update intervalls it would be nice to have an actual data record. At the moment it takes 2 updates to get data nearly 2 weeks old. This cannot be the sense of an update. -- MovieFex (talk) 11:37, 21 April 2020 (UTC)

Hello, current update interval is 5 days unfortunately. — Ivan A. Krestinin (talk) 18:36, 25 April 2020 (UTC)
You do not understand what I'm trying to say. Last update was today (2020-05-05) with data from 2020-04-29. Every correction from that status was not considered. Why didn't you took actual data from today or not more older than yesterday? Don't you see that the lists of constraint violations grow and grow? How should anyone work with an update which is absolutely outdated? -- MovieFex (talk) 22:22, 5 May 2020 (UTC)
Bot needs 5 days to process the data. So I understand the issue, but it is not simple to fix it. — Ivan A. Krestinin (talk) 16:24, 8 May 2020 (UTC)
Hello, is it possible to include that the labels are shown like User:Queryzo did here? -- MovieFex (talk) 18:17, 26 May 2020 (UTC)

category

Ivan, do you can to get the category in hebrew: קטגוריה:טקסונים שתוארו בידי תומאס הנרי האקסלי to the file: Category:Taxa named by Thomas Henry Huxley in another launguges? 2A01:6500:A051:379A:2008:67A5:30CE:7E4F 13:50, 23 April 2020 (UTC)
I have no idea that is wrong. Looks like bug. It is better to discuss it here: MediaWiki_talk:Gadget-Merge.js#Unexpected_error_while_merging. — Ivan A. Krestinin (talk) 18:32, 25 April 2020 (UTC)

VIAF updates

Hi! A little question: I remember that KrBot, among other extremely useful functions, also updates VIAF ID (P214) removing deleted clusters and correcting redirects; in general, 1) how frequent are the updates 2) when was the last update and 3) when will the next update take place? Thank you very much also from @Bargioni:, --Epìdosis 14:52, 8 May 2020 (UTC)

Hello! Update frequency is limited by appearing new dumps on http://viaf.org/viaf/data/. New dump appears today, so updating is in progress now. Usually this happens once per month. — Ivan A. Krestinin (talk) 16:11, 8 May 2020 (UTC)
Really great! Thank you very much! --Epìdosis 19:08, 9 May 2020 (UTC)
Thx, Ivan. -- Bargioni 🗣 09:48, 11 May 2020 (UTC)

Value Taipei, China (Q30940804) will be automatically replaced to value Chinese Taipei (Q216923)

Hi Ivan, I do not agree with the automatic deletion of Taiwan (Q865) from country for sport (P1532). Krdbot replaces Taiwan with Chinese Taipei. Wanted was to replace Taipei, China with Chinese Taipei. I do not agree with this one-sided following of Chinese politics and ignoring Taiwanese politics. --Florentyna (talk) 04:38, 14 May 2020 (UTC)

Hello, could you provide link to the edit? — Ivan A. Krestinin (talk) 08:54, 14 May 2020 (UTC)
See for instance [104]. At the end for all people with occupation [P106] badminton player [Q13141064] and country Taiwan [Q865], where everytime were in country for sport [P1532] both values: Taiwan [Q865] and Chinese Taipei [Q216923]. Taiwan was everywhere deleted from this property. --Florentyna (talk) 05:08, 15 May 2020 (UTC)
I disabled the autofix rules. — Ivan A. Krestinin (talk) 08:32, 15 May 2020 (UTC)
Thanks a lot! --Florentyna (talk) 09:15, 15 May 2020 (UTC)

GND ID replacement of redirected ids

Please stop [105]. The IDs are still valid and resolve. They are used in third party websites. Removing them, breaks links to WD by P227 and the ability to see which IDs are merged in GND database. Additionally in the case above the inserted value was left with "deprecated" label, existed already as preferred and the bot has been reverted before. Pinging @Kolja21: who works on GND. MrProperLawAndOrder (talk) 18:41, 18 May 2020 (UTC)

@Raymond: FYI: You made the same suggestion on deWP. --Kolja21 (talk) 19:22, 18 May 2020 (UTC)

Also, the edit has no edit group, no batch number so how can one see all edits done in the same run? And, can you in general provide more documentation of the bot runs? A page for each task? MrProperLawAndOrder (talk) 18:47, 18 May 2020 (UTC)

Thank you MrProperLawAndOrder and Kolja21 for bringing this up. I am working on GND together with the "Deutsche Nationalbibliothek" since > 10 years. The IDs are deprecated but still valid and resolve. This fact is an important information for all users incl. 3rd party users who use Wikidata as authority control data hub. Raymond (talk) 19:44, 18 May 2020 (UTC)
@Kolja21, Raymond: I think, they should even be imported by some bot. I don't know if this is in VIAF and KrBot can do it (monthly), or it has to be done by GND dump (not regular). Raymond, what is the exact terminology used by DNB for the merged/redirected values? Do they call it "deprecated". Because in WD there are three levels, in the case above one value is preferred, one deprecated and no value normal. How to do this exactly should probably discussed on P227 talk. Or better for all VIAF components since other libraries probably also merge and redirect. MrProperLawAndOrder (talk) 20:49, 18 May 2020 (UTC)
Imho we don't need to import redirects. Many of them had a low cataloguing level or had been Tns. These IDs should only be kept in Wikidata when they where added with a source and used by other databases. --Kolja21 (talk) 20:56, 18 May 2020 (UTC)
Kolja21, the source would be GND. DtBio also stores them, I have seen that old IDs are redirected on their website. Very professional. But I think that that excludes Tns. When I proposed importing GND IDs, I only meant the redirects, not Tns. MrProperLawAndOrder (talk) 23:17, 24 May 2020 (UTC)

Similarly, I reverted you here: Q64633427. Your bot changed the value of a deprecated VIAF statement!!! Vojtěch Dostál (talk) 13:54, 24 May 2020 (UTC)

@Vojtěch Dostál: ongoing also with GND [106], breaking any resolver. MrProperLawAndOrder (talk) 23:17, 24 May 2020 (UTC)

BTW @Vojtěch Dostál: I would differentiate between:
--Kolja21 (talk) 00:34, 25 May 2020 (UTC)
@Kolja21 Good point, will do in future. Vojtěch Dostál (talk) 05:47, 25 May 2020 (UTC)

Still doing it [107]. MrProperLawAndOrder (talk) 02:40, 30 May 2020 (UTC)

I deleted the IDs. VIAF contains huge number of redirects. Several IDs for each item actually. Adding only some deprecated IDs make our data inconsistent. Adding all deprecated IDs will increase amount of data significantly. Adding such IDs to Wikidata makes the data usage more hard and add nothing to our data quality. I think we should focus on data collection instead of garbage collection. — Ivan A. Krestinin (talk) 20:00, 30 May 2020 (UTC)
@Vojtěch Dostál, Kolja21, Raymond: "garbage collection", "add nothing to our data quality" - no insight he is actually listening to the concerns. I agree, it is inconsistent to have only some - for a given source, but then add the others. Here two kinds of IDs were mentioned, VIAF ID and GND ID. Replacing GND ID is not doing work on VIAF, since the GND IDs are contained in the GND DB and as explained above are still used by third parties. Yes, maintenance is harder, but it isn't "garbage", it is verifiable information, useful for some.
Last but not least: No response regarding "Additionally in the case above the inserted value was left with "deprecated" label". MrProperLawAndOrder (talk) 04:15, 31 May 2020 (UTC)
And, as if there is no need to care for consensus, now deleting IDs manually [108] MrProperLawAndOrder (talk) 04:26, 31 May 2020 (UTC)
Please stop doing this, Ivan, until we find consensus. We need to establish some community spirit in Wikidata and this behaviour is going exactly the other way... Vojtěch Dostál (talk) 06:47, 31 May 2020 (UTC)
Please stop deleting deprecated IDs. These are a valueable information for our users who looking them up and want to know what are the current IDs, not only for VIAF but for all kind of IDs. Raymond (talk) 08:36, 31 May 2020 (UTC)

Constraint violations P650

For Property:P650 (RKDArtists ID), single value constraint has recently been changed to single best value constraint to make sure that multiple values are not seen as a violation if one of them is set to preferred. However, the single value violation report still lists items with two values where one value is set to normal rank and the other is set to deprecated (for example Q6539264). Is there a way to exclude those from the report? Best, --RKDdata (talk) 07:50, 28 May 2020 (UTC)

Bot uses the same algorithm for single best value constraint as for single value constraint for now. It is not simple to fix this unfortunately. — Ivan A. Krestinin (talk) 20:03, 30 May 2020 (UTC)
Thanks for letting me know. I think I found a way around this. A regular query on items with P650 excludes all items with multiple RKDartists IDs if one of them is set to preferred and/or deprecated. That means that only the items on that list that still have multiple P650 statements need to be checked. --RKDdata (talk) 11:42, 8 June 2020 (UTC)

Using item template for constraint reports

Is there a reason why the constraint reports like https://www.wikidata.org/wiki/Wikidata:Database_reports/Constraint_violations/P21#Allowed_qualifiers use Wikilinks [[...]] instead of the templates {{Q|...}} and {{P|...}}. It seems to me the pages would be more readable when the templates would be used. ChristianKl17:04, 11 June 2020 (UTC)

Probably because so many instances of the template would (potentially) cause script errors. Tip: if you load the page through a diff (e.g. https://www.wikidata.org/w/index.php?title=Wikidata:Database_reports/Constraint_violations/P21&diff=0), MediaWiki automatically displays labels for bare links, although exactly because of this it’s more likely to time out and show only a WMFTimeoutException (e.g. P21 loads for me, but https://www.wikidata.org/w/index.php?title=Wikidata:Database_reports/Constraint_violations/P18&diff=0 times out). —Tacsipacsi (talk) 23:34, 11 June 2020 (UTC)
You are right. Bot uses {{Q|...}} and {{P|...}} when number of links is less than 2000. Large reports generate errors if the templates are used. — Ivan A. Krestinin (talk) 18:35, 28 June 2020 (UTC)

KrBot blocked

Hi Ivan A. Krestinin, Hope you noticed that KrBot has been blocked by Ymblanter. Jura1 was requested here to unblock the bot, but Ymblanter suggested to discuss on the project chat, If there is consensus then bot may unblocked. Kindly start a discussion on project chat. Regards, ZI Jony (Talk) 15:20, 14 June 2020 (UTC)


KrBot reactivated

Hi Ivan, For GND ID (P227), once the discussion here concluded with point d., would you skip the some 1200 entries in the query? Also, VIAF ID (P214), would you wait a few days before this is closed? If so, I don't see why KrBot shouldn't be deblocked. --- Jura 17:36, 27 June 2020 (UTC)

Hi Jura, sorry, I have no time to review all discussions. Am I right? Both GND ID (P227) and VIAF ID (P214) related tasks should be disabled. — Ivan A. Krestinin (talk) 18:13, 28 June 2020 (UTC)
Temporarily yes. Later VIAF could probably continue and GND should skip deprecated items. --- Jura 19:03, 28 June 2020 (UTC)
✓ DoneIvan A. Krestinin (talk) 19:12, 28 June 2020 (UTC)
Я разблокировал бота, спасибо--Ymblanter (talk) 20:07, 28 June 2020 (UTC)
For GND, the following change should be done: deprecated statements should be skipped (not updated or deleted), i.e. option (d) in the discussion. Can you also confirm to @Epìdosis: that for normal/preferred values, it will update redirecting values to the new value and delete deleted values?
Thanks for your patience with this. --- Jura 00:40, 17 July 2020 (UTC)
I join Jura in thanking you for all the great work you have been doing in these years, it is of great help for all of us! Looking forward to seeing again your bot editing also VIAFs and GNDs, --Epìdosis 07:37, 17 July 2020 (UTC)
I need to make significant changes in bot code to ignore deprecated values. But I have too few free time now. Maybe some other bot masters will create bot bot this job. — Ivan A. Krestinin (talk) 22:08, 18 July 2020 (UTC)
For VIAF ignoring deprecated values is not necessary, according to the result of the discussion: could you reactivate only VIAF for now, leaving GND in pause? --Epìdosis 22:31, 18 July 2020 (UTC)
✓ Done for VIAF. — Ivan A. Krestinin (talk) 06:57, 19 July 2020 (UTC)
Can you probably publish the source code? Just implementing a feature is probably easier than creating a bot from scratch. —Tacsipacsi (talk) 23:06, 19 July 2020 (UTC)

About the recent KrBot edits on Q83

Hi Ivan, not sure how your bot works but I think it needs some further refinement. On page Q83 it added statements and in a second step removed the same statements made by actual editors as duplicate. I guess this is not necessary. Moreover the bot does not consider rank information, i.e. it removed the preferred rank for e.g. 1.34.2 which caused the wrong date being shown (1.31.8) as the most recent preferred release. This is not correct in the end. Anyways thank for trying to make a difference, which is appreciated. Cheers --[[kgh]] (talk) 15:12, 3 July 2020 (UTC)

Bot is no so clever unfortunately.... Our data structure is looked as overcomplicated also (not normalized). What is latest version? It is version with the latest date from usual man point of view. But template uses another logic: it is version with preferred rank. Humans also may make similar errors during version data update. — Ivan A. Krestinin (talk) 22:14, 9 July 2020 (UTC)
> What is latest version? - All versions are released the same day, meaning that all three are the latest version. The version with the preferred rank is the latest version of the latest branch, the others are not. Thus I believe the same logic as for the templates should be applied, moreover so since the templates depend on the preferred label. If there is no preferred label all templates are broken. Probably the bot should back down from such subjects until enough cleverness was added? Cheers --[[kgh]] (talk) 20:39, 16 July 2020 (UTC) PS Humans make errors, too. Admittedly, my own doings over the day are full of them. :)

Convert external identifier to lower case

Hi! Filmportal ID (P2639) values are UUIDs, with lowercase version being canonical. Is there a way your bot can convert uppercase values (like this one) to lowercase? Preferably using a template on the property talk page instead of hardcoding the rule, so that community has greater control over it by being able to remove the template. {{Autofix}} seems unpractical in this case, as it would require adding possibilities in the regex. The last member of this sum alone is approximately 8×1024 times a 32-character string, which is more than 262 yottabytes, and is quite likely to exceed all limits in MediaWiki (page size limit, template size limit etc.), not to mention how inefficient this would be compared to a search for [A-Z] and applying std::tolower on its characters… —Tacsipacsi (talk) 11:21, 4 July 2020 (UTC)

Now I have only two abilities. Controlled {{Autofix}} and uncontrolled fixes for several properties. I added the property to the second code. — Ivan A. Krestinin (talk) 21:49, 9 July 2020 (UTC)
And would it be possible to create a new template (or an {{Autofix}} parameter) to control upper-/lower-casing? I think it’s a fairly common task, especially for hexadecimal identifiers, as consistent casing helps finding duplicates. Letting the community control the process makes everyone’s life easier and enabling/disabling the fix a lot faster. —Tacsipacsi (talk) 23:32, 9 July 2020 (UTC)
Bot contains custom fix code for 34 properties. Only 2 of its are simple lowercasing. So I think it is too specific case to create separate template. — Ivan A. Krestinin (talk) 06:32, 10 July 2020 (UTC)
This low number actually proves why a template is needed: people don’t reach out to you despite of the potential need for such fixes. Programmers don’t like to talk to other humans, placing a template on the property talk page is much more comfortable. :) With a fairly naïve regex I found over a hundred properties that expect something hexadecimal:
SELECT ?prop ?propLabel ?regex
WHERE {
	?prop p:P2302 ?stmt .
	?stmt ps:P2302 wd:Q21502404 .
	?stmt pq:P1793 ?regex .
	FILTER(REGEX(?regex, ".*[Aa]-[Ff].*")) .
	FILTER(!REGEX(?regex, ".*(A-Fa-f|a-fA-F).*")) .
	SERVICE wikibase:label { bd:serviceParam wikibase:language "[AUTO_LANGUAGE],en". }
}
Try it!
Tacsipacsi (talk) 23:03, 10 July 2020 (UTC)
✓ Done, please use:
{{autofix|operation=lower_case}}
{{autofix|operation=upper_case}}
Ivan A. Krestinin (talk) 06:14, 14 July 2020 (UTC)

Autofix P973->P535

I have defined an autofix on Property_talk:P973 which have ~80 to fix.--GZWDer (talk) 00:52, 7 July 2020 (UTC)

The pattern had small mistake. Now everything is fine. — Ivan A. Krestinin (talk) 21:12, 10 July 2020 (UTC)

Two updations on same date

Please check the history of Wikidata:Database reports/Constraint violations/P2002. In that, an updation is done by DeltaBot and then, the update done by your bot removes it and shows there is an error. Can this be resolved in any way? Adithyak1997 (talk) 15:02, 13 July 2020 (UTC)

Issue related to one-of qualifier value property constraint (Q52712340) is fixed. The next update should be fine. — Ivan A. Krestinin (talk) 22:50, 15 July 2020 (UTC)

KrBot2 check P3318

Hi:

I'm working on a big upload related with Guía Digital del Patrimonio Cultural de Andalucía ID (P3318) and would be nice to have checking feedback before uploading the big part. I've uploaded about 1000 updates and there are about 21K to be uploaded in the future. Thanks in advance. -- Olea (talk) 16:32, 13 July 2020 (UTC)

Hi, great job! I updated Wikidata:Database reports/Constraint violations/P3318. — Ivan A. Krestinin (talk) 06:48, 14 July 2020 (UTC)
Your bot is really useful. Thanks! Olea (talk) 09:38, 15 July 2020 (UTC)

Thanks

Hi, thanks for fixing my mistake at Charles Edward Balch (Q97319350). On another item created by me, DeltaBot did the same correction but using just one step instead of two. Your bot could gain some time if it did that, too. Regards, —capmo (talk) 00:56, 14 July 2020 (UTC)

Good idea! Maybe DeltaBot will process all items and my bot will have nothing to do) — Ivan A. Krestinin (talk) 22:52, 15 July 2020 (UTC)

Incorrect inference

You added these links based on this edit.

But your inference is incorrect. Virgil (Q1398) has no works in the Philadelphia Museum of Art. The ID indicates that there are works depicting Virgil in the Museum, not works created by Virgil. You will need to re-check all the automatic additions made because you made an incorrect inference about what the ID means. --EncycloPetey (talk) 14:43, 15 July 2020 (UTC)

It looks as though someone modified the identifier constraints incorrectly. All the automatic inferences about works being held in the Museum should be removed. The identifier does not imply that the Museum has works in their collection by that person or organization. --EncycloPetey (talk) 19:36, 15 July 2020 (UTC)

I started rollback process. Thank you for early issue detection. — Ivan A. Krestinin (talk) 23:18, 15 July 2020 (UTC)

KrBot and Google Books

Why does your bot keep making this edit? The item (Q63796565) is an item of a person, not a book, and the url is of a book that discusses that person. Gamaliel (talk) 23:49, 16 July 2020 (UTC)

Two times the same value in the same item

Hi! I've noticed that some weeks ago through QuickStatements some values of VIAF ID (P214) and ISNI (P213) have been inserted two times in items. Can you confirm me that your bot will delete them soon? Thank you very much! --Epìdosis 17:09, 24 July 2020 (UTC)

Hello, bot deletes duplicate values, but this may take up to two weeks. — Ivan A. Krestinin (talk) 11:25, 8 August 2020 (UTC)

On Wikidata:Database reports/Constraint violations/P535, Can you set KrBot2 to format the items as Q templates, rather than wikilinks, such that the labels, and not arbitrary Q-numbers, are visible to humans? This will make identifying, comparing and evaluating items much easier. For instance, comparing Dominguito (Q5812983) and Dominguito Fidel Sarmiento (Q5812984) is easier than Q5812983 and Q5812984. Please consider changing any other database reports that use only wikilinks without labels, as humans, not bots, are the ones who will ultimately perform merges or resolve the constraint violations. Thanks, -Animalparty (talk) 00:46, 1 August 2020 (UTC)

@Animalparty: See #Using item template for constraint reports. —Tacsipacsi (talk) 08:46, 1 August 2020 (UTC)

Bot problem with P973 & P953

According to this edit, your KrBot seems to be making wrong edits. In the case of paintings, I would imagine to find a copy of the painting in question, along with any archives of it, at the url with a label "full work available at". In this case, there isn't even a black and white copy of the painting available. You will need to re-evaluate the logic you are using for these edits. Jane023 (talk) 11:00, 5 August 2020 (UTC)

Problem with Q5508224

Hi Ivan! KrBot assigned Q5508224, fullback (position in rugby league football) to numerous Q937857, association football players.

SELECT DISTINCT ?item ?itemLabel ?position ?positionLabel WHERE {
  ?item wdt:P106 wd:Q937857; # association football player
        wdt:P413 ?position.
  FILTER(?position = wd:Q5508224)
  SERVICE wikibase:label { bd:serviceParam wikibase:language "en". }
}

This is probably due to some inconsitencies in Q5508224 at the time of the changes KrBot made. In the meantime @Mattythewhite: has created the accociation football specific item Q90173132 (full back) which is now clearly separated from Q5508224. I've tried to fix the wrong claims by hand but unfortunately did not succeed because of the large number of items. Therefore my kind question if you could fix the claims using your KrBot. If you decide to do so, please take care with association football players that are also rugby players. Trilemma2 (talk) 14:50, 9 August 2020 (UTC)

Could you place the request to Wikidata:Bot requests. I have too few free time now unfortunately... — Ivan A. Krestinin (talk) 21:25, 10 September 2020 (UTC)

P973 autofixes not working?

Hi Ivan. It seems they stopped working. --- Jura 06:53, 14 August 2020 (UTC)

Bot is working too hard on DOI (P356)... But this should be finished soon. Sorry for long response... — Ivan A. Krestinin (talk) 20:45, 10 September 2020 (UTC)

KrBot2

Hello Ivan! KrBot2 shut down on September 3rd and hasn't been updating anything since. Would you check it out? Thank you Palotabarát (talk) 10:34, 14 September 2020 (UTC)

Hi, I had troubles with hardware. Now all should be fine. Please wait day or two. — Ivan A. Krestinin (talk) 20:08, 17 September 2020 (UTC)
OK., thanks! Palotabarát (talk) 22:15, 17 September 2020 (UTC)
It worked for a day :( Stopped on September 18th. Palotabarát (talk) 18:22, 23 September 2020 (UTC)
I fixed one more hardware issue. Now all should be fine. — Ivan A. Krestinin (talk) 19:15, 8 October 2020 (UTC) P. S. Never buy Kingston SSD :)
Thanks so much! (OK: Kingston SSD :) Palotabarát (talk) 23:26, 8 October 2020 (UTC)

England IS a country

Every time I include the country value England in items your bot changes it to United Kingdom. England IS a country, one of four that make up the United Kingdom. Even the description of England calls it a country. For the demonym Londoners and Kentishmen, the country is BOTH England and United Kingdom. Adam Schiff (talk) 18:42, 15 September 2020 (UTC)

No. At least England can't be used as a value of P17 because it is not independant state. --Infovarius (talk) 14:29, 16 September 2020 (UTC)

Bot add claim then remove it

Hi! Recently your bot has been editing Scriptores Historiae Augustae (Q9334638), adding a claim before removing it, now five time in a row (since 11 Septembre). Could you fix it? Thanks! --Jahl de Vautban (talk) 07:07, 16 September 2020 (UTC)

Fixed. — Ivan A. Krestinin (talk) 20:15, 17 September 2020 (UTC)

Bot removing valid VIAF claims

Why is this bot removing completely valid VIAF IDs from records that I added it to? For example, on Q98690908 and Q98700075. I had to add all of them back in. --UWashPrincipalCataloger (talk) 18:27, 19 September 2020 (UTC)

@UWashPrincipalCataloger: Hi! Probably it isn't an error: in the two cases you cite, the IDs where in fact invalid because the initial "1" was missing. I don't know if in other cases the bot was effectively wrong. --Epìdosis 20:27, 19 September 2020 (UTC)
@Epìdosis: Thanks, I must not have noticed that! I usually paste the entire VIAF URL in and a script removes everything but the ID. But I've noticed if I copy just the ID and paste, the first digit is removed. It's a bug of the script I guess, and maybe those IDs were input that way and I didn't notice. Thanks for pointing this out, I'll be more careful! --UWashPrincipalCatalogerdosis 27:27, 19 September 2020 (UTC)

Summary data available in spreadsheet-loadable format?

Hi Ivan. Thanks so much for krbot2! I am interested in doing some analysis on the data in the summary table at https://www.wikidata.org/wiki/Wikidata:Database_reports/Constraint_violations/Summary. I need to load the table into Microsoft Excel. Does the bot generate a spreadsheet-loadable format, such as CSV or TSV? If so, is it available anywhere? David L Martin (talk) 00:19, 26 September 2020 (UTC)

Hi David, bot does not generate some machine-readable files currently. But you can copy the page data, paste it to Excel or Open Office Calc and save to CSV. — Ivan A. Krestinin (talk) 19:12, 8 October 2020 (UTC)
Hi Ivan. Thanks for that tip. I had not known about that capability of the spreadsheet products.

Несуществующая EditGroup

Не могу отменить: [109] --Infovarius (talk) 22:59, 3 October 2020 (UTC)

Приветствую, прошу прощения, что долго не отвечал. Лучше спросить у автора этого инструмента: User talk:Pintoch. Но впрочем в данном конкретном случае проще руками отменить, там всего две правки на одной странице. — Ivan A. Krestinin (talk) 19:06, 8 October 2020 (UTC)

Изменение свойства

Добрый день! Что думаете по этому поводу? Игорь Темиров (talk) 19:17, 18 October 2020 (UTC)

Приветствую, Игорь, давно я не погружался в эту тематику, да и времени на новое погружение особо нет. Возможно если начать обсуждение в рувики, то там найдется больше интересующихся темой участников. — Ivan A. Krestinin (talk) 19:23, 7 November 2020 (UTC)

Resolving a redirect causes duplicate VIAF

https://www.wikidata.org/w/index.php?title=Q770125&type=revision&diff=1289319039&oldid=1272932489

I marked that VIAF as "Deprecated rank" because it contains not only the village in canada, but also a church in france (and at the time I marked it, also a second church in france) and now krbot resolves a redirecting VIAF causing the same VIAF to be contained twice. Please use a different way to resolve such a "conflict", maybe simple drop the redirecting VIAF. Notice: Kolja21. --Wurgl (talk) 10:34, 25 October 2020 (UTC)

Placeholder

Hello, I have reverted manually numerous edits of your bot because it removes files from the image field. These files are in svg format, is that the reason for their deletion? For example in this case. Cheers. --Rodelar (talk) 22:01, 4 November 2020 (UTC)

@Rodelar: Sorry to mingle into this discussion, but why would you want to add placeholder images to Wikidata? Do you suggest adding placeholder images to all Wikidata items without an image? Vojtěch Dostál (talk) 07:24, 5 November 2020 (UTC)
@Vojtěch Dostál: Hello, maybe it is not the most suitable archive, but it was the best way I found to visualize that a building no longer exists. This, in the context of initiatives to obtain photographs of heritage elements (WLM for example), I think is very interesting information so that, if a building no longer exists, people do not travel to that place and get an unpleasant surprise. Can you think of an archive that could be included, instead of a photograph (which, not existing, will no longer be possible)?. Cheers. --Rodelar (talk) 20:19, 5 November 2020 (UTC)
@Rodelar: I'd suggest that you use Wikidata Query Service to filter out those item which have instance of (P31): destroyed building or structure (Q19860854) or state of conservation (P5816) : demolished or destroyed (Q56556915). I don't know where you inform your photographers (on Wikipedia? in some tool?) but my proposed solution should work the same way.Vojtěch Dostál (talk) 20:33, 5 November 2020 (UTC)
Please stop adding placeholder images to Wikidata. --- Jura 22:41, 5 November 2020 (UTC)
+1, and please add dissolved, abolished or demolished date (P576) --Hannolans (talk) 08:25, 6 November 2020 (UTC)

@Jura1: @Hannolans: It would be great to dialogue to find a solution (especially if I am doing something wrong), instead of pointing at me as if I were a vandal, thank you. @Vojtěch Dostál: That information is exposed in Wikipedia articles, hence the interest in visualizing it in some way. In applications like WikiShootMe the coordinate point would continue to appear, as an element without a photo, but in that case I think I can't do anything. Since it is not possible through an image, I will use P5816. Cheers. --Rodelar (talk) 18:22, 6 November 2020 (UTC)

The solution really depends on your application. If you want a placeholder image in the Wikipedia infobox, this could be coded there. If WikiShootMe doesn't provide you with what you are looking for, you need to discuss it with its coder. If you add a placeholder to P18, this can appear in many places without actually being useful there. Obviously, we can still find and add images of buildings that have been destroyed decades ago. --- Jura 18:36, 6 November 2020 (UTC)
@Rodelar: Czech Wikipedia displays placeholders automatically whenever the image is not available there and not even in Wikidata. Eg. w:cs:Chýlava (zaniklá vesnice) - the infobox is empty, there is no placeholder image in Wikidata, but the template automatically detects that and shows the image there. When you click on the image, it gets you to the UploadWizard on Commons. This way, we were able to obtain hundreds of useful images.Vojtěch Dostál (talk) 19:10, 6 November 2020 (UTC)

Почему?

Что не так? Почему из статьи Оттокар Карлович Адеркас (Q1334696) 11 ноября 2020 ботом было удалено заявление: „код Baltisches Biographisches Lexikon Digital (P2580): 0000000112166559, перемещено в / moved to код ISNI (P213)“? --Nick Fishman (talk) 06:10, 13 November 2020 (UTC)

Добрый день, робот сделал это в соответствии с параметрами шаблона {{Autofix}} расположенного на странице Property talk:P2580. Если считаете, что эти параметры некорректны, то удалите шаблон со страницы Property talk:P2580. Но лучше сначала обсудите проблему с участником, который установил этот шаблон. — Ivan A. Krestinin (talk) 18:52, 23 November 2020 (UTC)

Problem

Hello, your robot was removed 3 times of occupation on the Wikidata of Swedish House Mafia and Axwell Ingrosso. I don't know why he do these one, but if he remove again, I will try to block him. --Sakida0 (talk) 09:45, 22 November 2020 (UTC)

Hi, occupation (P106) is used for persons only, not for group of persons. You can extract person group occupation using request has part(s) (P527)occupation (P106) if needed. — Ivan A. Krestinin (talk) 19:07, 23 November 2020 (UTC)