User talk:Pintoch

Jump to navigation Jump to search

About this board

Previous discussion was archived at User talk:Pintoch/Archive 1 on 2017-06-21.

Epìdosis (talkcontribs)

Hi! I'm trying a very little import with OpenRefine (fewer than 4 lines), but after starting it gets stuck at "0% complete". Also @Alexmar983: is having problem in starting edits. Are you noticing anything strange in these days? Thank you very much in advance!

Epìdosis (talkcontribs)

OK, in my case it has just done the edits. Anyway, if you have noticed some slowness or strangeness, let us know! Thank you anyway!

Alexmar983 (talkcontribs)

yes because in my case I cannot start the import at all. It freezes when I log in my password on Mozilla Firefox and on the Microsoft browser... all becomes kinda grey-ish and I have to reboot OpenRefine totally to do anything at all (can't even come back). That's why I asked @Epìdosis to do that little test. this problem is going on since many days and it never disappeared so I am quite worried.

Alexmar983 (talkcontribs)

BTW, since few days I can import again. maybe it was the simple reboot of the PC, no idea. I will tell you if the issue appears again.

Discostu (talkcontribs)


when trying to revert one of my batches I came to this page . It says:

The current lag is 1 day, 13:50:39 (136239 seconds) - if this is more than an hour then something is probably wrong - please notify the maintainers of the tool and try again later.

Reply to "EditGroups Lag"
Bargioni (talkcontribs)

Hi, I tried to run two batches in QuickStatements using curl, say:

curl \

-d action=import \

-d submit=1 \

-d username=Bargioni \

-d "batchname=add en labels to journals from" \

--data-raw 'token=***' \

--data-urlencode data@qs.tsv

Job #28473 and #28478 were executed, but they remain in a state of RUN, not DONE. No modifications were applied to items.

More, both jobs lack the edit group: Edit group "28473" not found.

Thx a lot.

Pintoch (talkcontribs)

Hi, this sounds like a problem between you and QuickStatements. EditGroups will not detect the edit group until at least one edit has been performed successfully on Wikidata.

Bargioni (talkcontribs)
Bargioni (talkcontribs)

Ok, it seems you refer to the single batch. If so, I have a problem with QS when accessed through curl. Sorry for the previous reply.

Reply to "Edit group not found"

Property:P5318 - Property:P5319

Jack Rabbit Slim's (talkcontribs)

Bonjour Pintoch,

je m'adresse à vous car vous semblez un spécialiste de Wikidata ;-), et surtout parce que vous avez créé et modifié ces éléments : Property:P5318 (film) Property:P5319 (personnalités)

J'ai remarqué que le site de l'Académie des César, avait eu la "brillante idée" de changer toutes les URL concernant les films et les acteurs/actrices.

Exemple avec le film De rouille et d'os :

Avant l'URL était la suivante :,film,43.html

Maintenant c'est celle-ci :

Idem pour les personnalités du cinéma, exemple avec Marion Cotillard :

Avant :,personne,2949.html

Maintenant :

Si vous avez une idée comment résoudre le problème, je vous en serais reconnaissant, car moi, mes connaissances sur Wikidata sont très très limitées.

Cordialement. Jack

Jack Rabbit Slim's (talkcontribs)

Je notifie également @Thierry Caro qui a plusieurs fois modifié ces deux éléments.

Thierry Caro (talkcontribs)

J'ai traité les films, pour commencer.

Thierry Caro (talkcontribs)

Et les gens ensuite. C'est bon maintenant.

Jack Rabbit Slim's (talkcontribs)

Un grand merci Thierry Caro ! Et désolé du dérangement.

Bien cordialement. Jack

Jheald (talkcontribs)

Hi! I'd like to use OpenRefine to help de-duplicate the recent import from The Peerage website.

In particular, at Property_talk:P4638#English_Wikipedia_articles_without_ID there are some lists of items that one might strongly suspect should have The Peerage person ID (P4638) IDs, but currently don't.

Is there any neat workflow for loading one of these batches into OpenRefine and reconciling it, ignoring the known Qid, (or, second-best, looking only for Qids above Q75000000) ?


Pintoch (talkcontribs)


Intuitively this sounds like a use case for third-party reconciliation.

I don't know if you have access to a full dump of The Peerage (with names and identifiers) - if you do, then you could potentially try to set up a reconciliation service for it, for instance using reconcile-csv. If you can only access their website, then you could perhaps write up a small reconciliation service which wraps their own search engine in the reconciliation API, but that is a bit more work.

Once you have such a service, you can just use it in place of the lobid-GND service demonstrated in the tutorial.

Let me know if it works for you!

Jheald (talkcontribs)

It's a bunch of webpages, raw html, lots of entries on each page, not much of a searh facility. The best most accessible version of names + IDs is now Wikidata, where there was a 100% upload (700,000 IDs). Of course those could be extracted reasonably easily by a SPARQL query, with their WD labels; or alternatively the catalog data could be downloaded from Mix'n'match, if the original labels are preferred -- so not hard to get a local copy of the data.

But what I was really wondering was whether there was a technique to match eg a few hundred names against Wikidata, but either excluding a specific Qid for each match (or filtering it from the suggestions post-match); or alternatively, excluding all Qids with Q-numbers lower than 75,000,000

Pintoch (talkcontribs)

Yeah I think I see the idea - I don't really have a workflow to propose on top of my head, reconciliation isn't really designed to do deduplication. It's something we could improve (it's not clear to me what it would look like but definitely worth thinking about).

Still, I think you could potentially use third-party reconciliation for that: download the 700,000 qids, ids and labels from wikidata, put that in reconcile-csv (if it can handle that load - I haven't tried), and then try to reconcile to this your lists of items which should have a The Peerage id.

Reply to "OpenRefine for de-duplication"

thanks for creating Beta Masaheft property!

Pietromarialiuzzo (talkcontribs)

Dear Pintoch,

Thank you very much for creating the property on Wikidata for us!

Reply to "thanks for creating Beta Masaheft property!"
Jheald (talkcontribs)

Hi, I was just wondering whether reconciliation is down (for everyone) at the moment, or whether it's just me?

It was working for me on Monday (13th), but today I'm getting a "502 Bad Gateway" error. (At 3pm this afternoon, and again when trying again now at 10pm).

Just wondered whether this is likely to be a server-side issue; or whether some problem has come up my end (eg firewall issue?)


Pintoch (talkcontribs)

Interesting! It is working now as far as I can tell, but in general you are right that it is really not as stable as it should.

Jheald (talkcontribs)

I don't know what I can be doing wrong then... I've tried turning Windows firewall off, tried versions 3.2 beta, 3.2 stable, 3.3 rc1, tried reconciling a different file -- still seeing the same thing.

Curiously when I added a column with a single Q-number, and asked it to reconcile that, that worked. But try a column of actual strings, and I'm getting the 502 error

Jheald (talkcontribs)

It seems I can successfully reconcile to VIAF, using the service by RDM Page, just not Wikidata.

Jheald (talkcontribs)

I tried putting in the Wikidata API address as an additional reconciliation service and got the following onscreen warning, even though the link seems to give a response when just put into a web-browser address bar:

"Error contacting recon service: error : Not Found - Error contacting recon service: timeout : timeout -"

Jheald (talkcontribs)

Now seems to be working again.

Pintoch (talkcontribs)

Oops, you were right indeed! I restarted the service this morning. I really need to find a more reliable hosting provider than toollabs.

Jheald (talkcontribs)


Esteban16 (talkcontribs)

Hi, I was just wondering how you do to add a lot of properties' data in a single edit. Is it some gadget or something else?


Pintoch (talkcontribs)
Reply to "Properties"
Arlo Barnes (talkcontribs)

Does the bot automatically recheck if the page is edited after a fail to validate regex?

Pintoch (talkcontribs)

No, my property creations are supervised. I am not checking Wikidata often as I am on Wikibreak :)

Arlo Barnes (talkcontribs)

I see. Well, apologies for disturbing you from it. I just updated the proposal to have a more correct length of description.

Reply to "wikidata:property proposal/Atlas Obscura identifier"
So9q (talkcontribs)

I would like to revert a recent batch. Can you help decrease the lag somehow?

Reply to "Editgroups lag is now 14 hours"