היי אהבה כהן,
ראי בבקשה את הערות של האדמין בדף שיחתי
Welcome to Wikidata, AhavaCohen!
Wikidata is a free knowledge base that you can edit! It can be read and edited by humans and machines alike and you can go to any item page now and add to this ever-growing database!
Need some help getting started? Here are some pages you can familiarize yourself with:
Please remember to sign your messages on talk pages by typing four tildes (~~~~); this will automatically insert your username and the date.
If you have any questions, please ask me on my talk page. If you want to try out editing, you can use the sandbox to try. Once again, welcome, and I hope you quickly feel comfortable here, and become an active editor for Wikidata.
Previous discussion was archived at User talk:AhavaCohen/Archive 1 on 2021-02-21.
All structured data from the main, Property, Lexeme, and EntitySchema namespaces is available under the Creative Commons CC0 License; text in the other namespaces is available under the Creative Commons Attribution-ShareAlike License; additional terms may apply. By using this site, you agree to the Terms of Use and Privacy Policy.
Hi Ahava! It is a pleasure to write you some months after our meetings at the IFLA Working Group sessions (BTW, how is the SIG application going?). As I said to Geagea, I will try to help you a bit in cleaning the imprecise IDs in the next days. I remain available for whichever collaboration, as always. See you soon!
Thanks so much for your help with this.
The SIG application is being worked on. The past month I've had a family emergency so right now I'm taking a back seat, but hope to get back on the horse next month or at the start of the year at the latest.
Hope you're doing well!
Well checked again Jacob Weinberg (Q19668022) and yes you are right. Only the last id is ok (987007269764805171) the rest are id's for works of him
I have checked filed no 10 which is LCCN id and it's lead to me to the to the VIAF id's.
Hi, my opinion about this removal. There are two possibilities, both of them will start with a sparql query to collect the items with properties P949 or P8189, and end up in a batch of QuickStatements like -Qnnn P949 "NLI_id" or -Qnnn P8189 "J9U_id".
To filter items from the sparql query, and obtain Qnnn, we can access the NLI/J9U catalog using the id. A lot of http connections are required. Is this good for the web server of the catalog, even at a graceful pace? Or you may publish the personal names auth records in a .mrc (compressed) file. Please, let me know your opinion.
Thx a lot.
@Bargioni:, beside the issue raised by Epìdosis, and continuance to the conversation in my talk page. Can "moreIdentifiers" tool (brilliant idea of user:Epìdosis) may run through items that have VIAF ID (P214) which National Library of Israel J9U ID (P8189) is missing and add them automatically.
If I'm not wrong about what you are asking for, moreIdentifiers is a gadget and, unfortunately, no more than this. I.e. it cannot be run as a robot or something similar.
That's pity. Do you have a way to create query of a list items with VIAF ID (P214) without National Library of Israel J9U ID (P8189). User Epìdosis tried to help with wdumps, but we have not enough technical skills to produce the query. Of course if there is any way to compare with VIAF site and keep only the id's that have J9U id, it would be the best.
I'm working on a method to clean invalid personal names in P949. Please, wait.
Good news. Invalid NLI ids were removed. Please, see https://editgroups.toolforge.org/b/QSv2/70352/
I'm working on P8189 now.
Hi! As of now, invalid NLI IDs where found through a scrape on the catalog and then removed by Bargioni; on the other hand, invalid J9U IDs are still there, and repeating the same method would be very very long (and heavy for the catalog) because in the meanwhile big imports of J9U IDs have been made, so it would be necessary to scrape not only about 85k pages (already very long), but nearly 170k pages! So I think that the best solution would be sending Bargioni a MARC dump of all the personal authority records, so that Bargioni can 1) remove from Wikidata invalid J9U IDs 2) create a Mix'n'match catalog for J9U. Would it be possible? Thanks and happy 2022!
A. I have simple solution to this. I can ask for a list of the equivalent P8189 id based on P949 from the records of the National Libary of Israel. Once we have the list we (you or user Bargioni) can remove the extra id's. What do you think?
B. regarding to the MARC dump. Already spoken to user AhavaCohen and it should be between user Bargioni and AhavaCohen. With the holidaies do'nt really know where it stands.😊
Solution A is viable in my opinion; anyway, if solution B comes first, solution A can be avoided (so less work for you). If the MARC dump can arrive in one or two weeks, I think we can of course wait for it. Thanks!
I will be requesting a marcxml file of all personal name authorities on Sunday and send them to Bargioni.