User talk:Djm-leighpark

From Wikidata
Jump to navigation Jump to search

This account is now abandoned. Please address any communications to User talk:DeirgeDel


Signature check[edit]

Double checking tlak link from custom signature following a mistake. User:Djm-leighpark(a)talk 05:02, 5 September 2022 (UTC)[reply]

adding site references for ENWQ[edit]

Hi Djm-leighpark,

I see you have added a lot of site references to items that have a page on the English Wikiquote. I have seen others do this too, but it is usually users who are the creators of the page on enwq. You seem to do it more methodically so as to include all existing enwq pages to WD, I think?

Do you have any thoughts about User_talk:Mike_Peel#Pi Bot for ENWQ? If so, it would be great if you would post them there. Thanks in advance, and sorry if I make no sense by using the wrong terminology. I am not very familiar with WD. Ottawahitech (talk) 15:07, 23 January 2023 (UTC)[reply]

@Ottawahitech: During the #SheSaid campaign I became aware of the usefulness of getting Wikiquote bio articles with an associated Wikidata Item. For an Wikiquote article with a "curated" article in the English Wikipedia their will be a Wikipedia Item easily identifiable from the English Wikipedia article. If a Wikiquote "honestly" links through to its rightful English Wikipedia article (and not something different) its easy to manually then go through via the English Wikipedia article to the correct Wikidata item and the make the sitelink back to the Wikiquote item. This is a sub one minute manual action. With the 500+ needed examining during the SheSaid campaign that was a daunting task. I sort of hinted I was looking for a bot ... and bots have been used on this for some other "quote" wiki's ... but in the end I developed my "lollypop.py" (not open source) which took a feed from my quarry queries of wikiquote shesaid aricles and for articles that ought to be safely linkable (essentially has a Wikidata template present) would produce a list of site links statements that could be used to feed into Wikidata (there was still a need to post check manually each one) That cleared most of the "easy" cases and meant it was possible to concentrate on cases which mere more complex. Broadly by 31 December 2022 the backlot of missing sitelinks for bios over the campaign period had been for practical purposes eliminated unless apart from some Wikiquote articles in a deletion process and the quarry queries were updated to show that. Post 1 January 2023 enWikiquote: New articles in the past two weeks is good at noting new articles that are not associated with a wikidata item and making that happen: easy if there is an English Wikipedia article and more difficult if there isn't; and if a wikidata item can't be found or reasonably created a deletion process can be initiated fairly rapidly. I've also worked out how to handle non-bio articles such as Books, media seasons, proverbs, .... in fact pretty much how to handle any Wikiquote article without a sitelink from Wikidata. But I am not into creating wikidata items for every wikiquote article; where there is in my opinion a possiblity of a lack of notability issue they have been sent to a deletion process. I think I got close to achieving that any #SheSaid 2022 article that needed to in to be in a deletion process was in that process by the end of the Campaign on 31 December 2022 so these would not be being pciked up for notability by the end of the campaign. While the quarry queries were great for looking at the campaign period or the last 2 weeks I'd be reluctant to (repeatedly) run a similar query over all Wikiquote articles. I remembered an admin there suggesting Special:UncoonectedPages and while dismissing it at the time asthe quarry queries were giving me the required information and more and much more rich information to work with. However post #SheSaid I revisted that suggestion, realised that I was only interested in mainspace articles, and that [1] indicated 1245 (from memory) Wikiquote articles not associated with Wikidata items. I determined to try and address this. I adjusted lollypop.py to accept a feed from the Special:UncoonectedPage(main article) list and it was able to produce a batch of about 400+ quickstatements that could with reasonable safety link an existing wikidata item to a wikiquote item. Those had to be manaully checked but overall that got UnconnectedPages down to under 800 which have needed to be manually sorted. Though a process of cherry picking some obvious cases and sending other cases to an appropriate deletion process where appropriate. By 14 January 2022 the number of unconnected pages was down to 350 per q:en:User:Djm-leighpark/US. I am "slogging" thought that list and while there have been a few easy wins for getting a wikidata item many cases are quite tougher. I'm gettinmg quite frammy with anyone who even gets near to suggesting it isn't a slog. I'm also aware some bad actors may have caught what I'm about and created some trivial wikidata items to support dodgy wikiquote items, and that was the reason I was doing a little CIR obscurity about my purposes to avoid possible disruption of those purposes. Unconnected pages is now down to about 225 however further reductions are possible but tough but I believe I can sustainable reduce that towards a logical zero. q:en:User:Djm-leighpark/HOWTO (should be hownottodo) gives some guidance/directions on how to use Spreadshee->Quickstatements to end up in a Wikidata boghole - these use technique what I amusingly call my "project egg" technology to use a speadsheet as a input form onto another sheet as quickstatements (via cut/past) feed to create book editions (this will be better if it becomes a project chicken consumed python) -- User:Djm-leighpark(a)talk 01:29, 24 January 2023 (UTC)[reply]

Drawing slightly close to phrasing that got be blocked but I am usually in for the hunour I am not aware of Mike Peel's BOT and how hairy it is but I I recall correctly his image on his user page is quite hairy but if there is a prescriptive way it can help I would be grateful to hear it. I am however aware of Mike & Co.s {{Cite Q}} and my project egg->Chicken is a personal related tool for loading book editions to Wikidata; any I use that mostly for Ireland railway related online books . Due to indef. blocking on enWQ I am forced to go my own way this. I have however used it to provide a wikiquote target fors stuff. See for example q:en:Special:Diff/3237866, Through the eyes of rebel women : the Young Lords : 1969-1976 (Q116287866) ... although thats not a great example apart from how it hell can anyone fail to persuade closer to Wikiquote:Votes for deletion/Minerva Solla keep I have now clue. Don't ask. Just don't ask. -- User:Djm-leighpark(a)talk 01:29, 24 January 2023 (UTC)[reply]

Your current import batch[edit]

Hi Djm,

which kind of people are you currently importing? There is quite a high incidence of duplicates.

Kind regards Chad McKittrick (talk) 12:33, 30 January 2023 (UTC)[reply]

@Chad McKittrick: (rushed response): Broady speaking I'm working through the list of articles on Wikiquote which are not connected to Wikidata items and trying to match them up. I'm down to the the "rump end" of the process and Working on articles that are not linked to an article on the english wikipedia ... (that would automatically define a precise english Wikipedia article). Some I will choose to create a wikidata item for, some I will choose to send to a deletion process, and some I will think over for a little longer. For poeple I have a very basic spreadsheet to help me do a batch via Quickstatements .. very crude but effective. It does rely on me checking Wikidata for an existing article first and while I am finding some I undoubtably am missing some others (including a quite a few American Professors yesterday .... (These where bound to be notable so I think I got on a run and forgot to do a Wikidata duplicate check on a number of occasions). Thanks for Picking this up and I apologise. Background links include [2], q:en:User:Djm-leighpark/US (incl. old versions) & q:en:https://en.wikiquote.org/wiki/User:Djm-leighpark/HOWTO. Some other stuff will related to loading of book editions (particularly accessible online) for use by Cite Q or to back a wikiquote article on a book. Will try to be more careful. Have some Japanese poets I'll be tackling next ... (possible dulicate nightmare)... Will try to be more careful. -- User:Djm-leighpark(a)talk 13:08, 30 January 2023 (UTC)[reply]
Thanks for doing this ungrateful job. But why are you not using OpenRefine? Your current process has no deduplication step. Since the import of complete authority files (e.g. Czech National Authority Database (Q13550863) by User:Vojtěch Dostál) many authors already exist here.--Chad McKittrick (talk) 13:12, 30 January 2023 (UTC)[reply]
@Chad McKittrick: Because I don't know what I am doing. And I didn't actually know what I was doing at the start of this. I did actually see open refine being mentioned somewhere four or file days ago. And if I changed course I might give up completely, in fact if I gate up two weeks ago I would have never got back into it. I've already spent to long on this. I got indef blocked on enWQ last April so I am sort of on my own. I got some some SparQL training back last August but I've not really been able to exercise that but I learned a couple of Wikidata things off the back of it. Thanks. -- User:Djm-leighpark(a)talk 16:20, 30 January 2023 (UTC)[reply]
I'm sorry to hear this. Keep going the way you prefer, people like me or McKittrick will have no problem merging the eventual duplicates.--Phaic Thǎn (talk) 18:24, 30 January 2023 (UTC)[reply]
@{u|Béarnese}}, Chad McKittrick: Thanks to both of you for the encouragement and support. January 2023 has been quite a learning curve for me on this process. While I'd previously built some skills with creating wikidata items for books and book editions and perhaps a little in relation to monuments I've transferred those in January 2023 into creating items for television series seasons and then for African subcontinent politicians and queens to support wikiquote articles and recently in the last week moved into authors/educators where I think I've been more guilty of not spotting some obvious duplicates. I'm proud of avoiding a duplicate at Kozan Ikkyō (Q61058232) where there was no English Language description but not happy with missing Ernst G. Frankel (Q116464744) / Ernst G. Frankel (Q112442811) which was sloppy as there was a suitable English Language label. I probably haven't been reviewing the loaded items carefully enough to check for incorrectly entered VIAF's and duplicate VIAF's. On the latter point I am concerningly noting Ernst G. Frankel (Q112442811) is still claiming a duplicate VIAF between itself and Q116464744 post merge. Thanks for support. -- User:Djm-leighpark(a)talk 20:04, 30 January 2023 (UTC)[reply]

@Chad McKittrick, Béarnese:: Just to say I've broadly completed by exercise of getting all (excluding e.g. those in deletion/move/merrgto processed) connected to a Wikidata item. There may be a little follow up work and the odd escapee and perhaps even the odd wikidata item improvement related to that exercise but broadly speaking I'm finished and likely to be moving to other things than Wikiquote or even Wikipedia. Thanks for the encouragement, quite uplifting. Thankyou. -- 23:19, 3 February 2023 (UTC)[reply]

Work outstanding[edit]

Key unconnected[edit]

✓ Done: This are all handled. -- User:Djm-leighpark(a)talk 22:57, 3 February 2023 (UTC)[reply]

Automated report of empty item: Q116401147[edit]

Hello, an item that you have edited (and you are the only non-bot editor) is considered empty and will be deleted in 72 hours if it doesn't improve. Your automated cleaner, Dexbot (talk) 11:57, 7 February 2023 (UTC)[reply]

Used Djm-leighpark by mistake[edit]

There was a lump of new articles on enWikiQuote and I've just used Djm-leighpark rather than DeirgeDel for the Quickstatments batch by mistake, realising that when I'd set it off. I'll try to make sure I don't do that in future. Thankyou. -- User:Djm-leighpark(a)talk 22:16, 17 March 2023 (UTC)[reply]



This account is now abandoned. Please address any communications to User talk:DeirgeDel