Wikidata talk:Bot requests

Latest comment: 2 years ago by MisterSynergy in topic Idea for data cleaning

Request for archiving old threads here edit

A number of these topics are quite old (eg up to 7 months). If they are not active, would it be possible to archive them? Having such a long page is confusing and, for users such as myself who will not always have access to high-speed internet, can be time-consuming and frustrating to load. --LT910001 (talk) 21:05, 10 January 2015 (UTC)Reply

Topics get archived after 6 months or if the template {{Section resolved}} is added to the section. --Pasleim (talk) 23:50, 19 January 2015 (UTC)Reply
Perhaps you could decrease that to 3 or 4 months in the interest of users without high-speed internet or mobile like myself? --LT910001 (talk) 21:06, 25 January 2015 (UTC)Reply
@LT910001: The page should be much smaller soon, I hope that helps you. Reduction of the time until automatic archiving I don't support as bot requests should be resolved and not moved to the archive without dealing with them. --Pasleim (talk) 18:51, 28 February 2015 (UTC)Reply

Question to operators of Coordinate bots edit

Hi, I'm wondering why for Q19311449 no bot has taken over the coordinates from de:Weitmars. Just in case someone can spot the reason, maybe it leads to a bug in a bot code, or in a de.wp template? Regards --Dealerofsalvation (talk) 21:09, 6 January 2016 (UTC)Reply

PS is there a tool that can list all items that have coordinates in de.wikipedia, but not in Wikidata, preferably limited to a region? --Dealerofsalvation (talk) 04:07, 7 January 2016 (UTC)Reply

@Dealerofsalvation: are you still interested in answer to you PS part? I don't have a tool, but could give you a number/list of such articles, not sure about that region part of question. --Edgars2007 (talk) 16:13, 17 March 2016 (UTC)Reply
Hi @Edgars2007:, yes, that would be fine if that's not much work for you, I'd particularly be interested in such a list for all objects in the category tree of de:Category:Regierungsbezirk Stuttgart, that's Q8172 for the geographical item resp. Q7470260 for the corresponding category. Thanks --Dealerofsalvation (talk) 12:22, 20 March 2016 (UTC)Reply
Here is the list, Dealerofsalvation. Those are articles, which have WD item, are in "Regierungsbezirk Stuttgart" category (depth: 20 subcategories), have template Coordinate and don't have coordinate location (P625). There may be some false results, but the list should be good enough. --Edgars2007 (talk) 08:43, 27 March 2016 (UTC)Reply
Thanks. I've checked about 15 samples and found no errors. It's about 530 items, so this is by far too much to update manually. I'll open a request on the "other side" of this page. --Dealerofsalvation (talk) 19:47, 27 March 2016 (UTC)Reply

Archiving redux edit

Despite the discussion above, this page is now 145,637 bytes long, and no archiving seems to have been done for some time. Can someone fix this, please? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:16, 17 March 2016 (UTC)Reply

Data import hub, data preperation instructions and import workflow for muggles edit

Hi all

Myself and NavinoEvans have been working on a bare bone as possible workflow and instructions for making importing data into Wikidata available to muggles like me. We have written instructions up to the point where people would make a request on the bot request page.

Please take a look and share your thoughts, I would like to know if the instructions are sufficient to produce a spreadsheet that please completing bot requests to find useful and usable.

https://www.wikidata.org/wiki/User:John_Cummings/Dataimporthub

https://www.wikidata.org/wiki/User:John_Cummings/wikidataimport_guide

Thanks very much

--John Cummings (talk) 15:02, 21 November 2016 (UTC)Reply

Is it possible to add these links to the top of the page? edit

Hi all

I've created some extra pages on Wikidata to try to create a clear and documented workflow for people interested in importing data into Wikidata, would it be possible to add these links to the top of the page? I tried to do it but the header is complicated and I don't understand how to do it so it doesn't break the formatting.

You may find these related resources helpful:

  Data Import Hub
  Why import data into Wikidata.
  Learn how to import data
  Bot requests
  Ask a data import question


Thanks very much
--John Cummings (talk) 10:50, 25 November 2016 (UTC)Reply

{{done} at Wikidata:Bot_requests/header2. An alternative could be Wikidata:Bot requests/Header. Nice work btw.
--- Jura 11:05, 25 November 2016 (UTC)Reply

Thanks very much Jura, is it possible to have it in the main header box to fit with the styling of the other pages? Thanks, --John Cummings (talk) 08:32, 26 November 2016 (UTC)Reply

I think the other pages should adapt this though, as the other is quite ugly and space-consuming. Sjoerd de Bruin (talk) 09:35, 26 November 2016 (UTC)Reply

Please check property talk pages before mass importing data edit

Please see my post on the above. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:50, 31 January 2017 (UTC)Reply

P172 edit

According with this discusion I ask to copy-paste the statements

Armenian Soviet Encyclopedia (Q2657718) and Karabakh War 1988–1994 (Q16392167)

from described by source (P1343) 
to ethnic group (P172)

as a source of second statement.

Only in items with ethnic group (P172) Armenians (Q79797) and described by those sources. Both sources are aviable in hy.wikisource. - Kareyac (talk) 14:26, 6 February 2017 (UTC)Reply

I think this belongs to the project page itself. Matěj Suchánek (talk) 14:41, 6 February 2017 (UTC)Reply

Reduction of time before archiving request edit

Can we reduce the time before archiving the request in order to clean the page ? Currently a request is moved to archives after 6 months. I propose to move it after 3 or 4 month.

In order to save discussion we should create a template for bot requests requiring different informations allowing a better understanding of the request and reducing the discussions. As parameters to provide I propose:

  • Date of the request:
  • Task description:
  • Original license of the data to be imported (if relevant):
  • Link to discussions justifying the request:
  • State of the request: pending/accepted by bot op/ Finished

Snipre (talk) 12:01, 21 April 2017 (UTC)Reply

  Strong support making a template for requests. Matěj Suchánek (talk) 12:15, 21 April 2017 (UTC)Reply
@Matěj Suchánek:   Done See Wikidata:Bot requests/Form. But this is working only when using the button at the top of the page. If this not working use the updated version of Wikidata: Bot requests/Header. Snipre (talk) 11:50, 22 April 2017 (UTC)Reply
I cleaned it all up. Thanks for your effort. Matěj Suchánek (talk) 12:16, 22 April 2017 (UTC)Reply

Which activities require approval for new bot operator? edit

Is it advisable to use test.wikidata.org when initially trying to use a bot?

Which activities, either on the test site or the real site, require a bot account, a bot flag, and/or prior approval:

  • reading a small number of items (2 or 3)
  • reading medium number of items ( around 200)
  • modifying one or two items with immediate manual inspection after the changes
  • modifying about ten items with immediate manual inspection after the changes

I would be using pywikibot.

Jc3s5h (talk) 22:51, 29 July 2017 (UTC)Reply

What do you mean by "reading"? Modifying less than ten items, in my view, does not require a bot flag.--Ymblanter (talk) 04:33, 1 August 2017 (UTC)Reply
By "reading" I mean using a bot to move information from Wikidata to my computer, but leaving the information in Wikidata unchanged. Thanks for responding. Jc3s5h (talk) 10:30, 1 August 2017 (UTC)Reply
No, I do not think reading 200 edits requires a bot flag either.--Ymblanter (talk) 05:20, 3 August 2017 (UTC)Reply

Subheading markup edit

Please fix the the template used by some people on this page, which currently uses malformed list markup for subheadings (example: ;{{int:Talk}}), so that it instead uses proper, accessible heading markup (example: === {{int:Talk}} ===). Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:34, 14 September 2017 (UTC)Reply

Bot requests for Google Code-in edit

It would be interesting to give Google Code-in students some bot requests, especially ones which are more about difficult item traversal, collecting data/aggregation with few or no data modifications to lower the risk. I would definitely be a mentor, and would structure the task such that we immediately unassign any newbies who havent done other coding tasks, so it is only the 10-20 very competent students who are doing the task, and they would immediately fail if they havent published their code and had it reviewed before it did any data modifications.

Please let me know if there are any old or new bot requests which might be suitable. John Vandenberg (talk) 05:09, 28 December 2017 (UTC)Reply

Could Json-ld be imported into Wikidata automatically/directly? edit

I have some web pages which contain linked data (JSON-ld). Could Jason-ld be imported into Wikidata automatically/directly?

Here are the examples:

https://heritage.lib.hkbu.edu.hk/routes/view/ids/HER-011335

https://digital.lib.hkbu.edu.hk/yunshan/ids/HYS-000010

Or I need to create a spreadsheet first?

Thank you.

I'm unaware of a tool to do this but it'd be cool to have. I'm gonna guess there isn't much demand for it. We support the other direction e.g. https://www.wikidata.org/wiki/Special:EntityData/Q42.jsonld BrokenSegue (talk) 03:20, 6 May 2021 (UTC)Reply
That would be very hard to do since how would it know to which WD items and properties to map the JSONLD terms? Vladimir Alexiev (talk) 14:56, 18 November 2021 (UTC)Reply

Idea for data cleaning edit

Once in a while, I spend my time reducing the number of short pages at Special:ShortPages. Normally, I focus on the items that have a bytesize smaller than 161, since that indicates empty items. By looking at the history of those pages, I can find most of the time a link to a wp page that still exists and is connected to WD. Then I can merge the two items. However, this task is quite time-consuming and since I came back to this task earlier this week, I noticed a huge increase in empty items. Normally, the list contained 10-200 items which are empty, but at the moment this list has over 2000 items. I wondered if there may be a bot-solution to this problem. The steps I could identify for a program that might do such task: search in the history of an item for the version that is largest (in bytesize), and then find the site links that are at that moment available. Then find the most up-to-date version of that link (following trails of possible redirections) and check if those articles are linked to a WD item. If so, merge the empty item with the found WD item, and the job is complete. Is this a feasible job, and are people willing to develop this? Q.Zanden questions? 22:12, 4 January 2022 (UTC)Reply

Unsure about this idea but it seems a lot of these blank items were created by @Exec8:. I left them a warning. We should probably just revert all their changes. I don't think your suggestion always results in correct outcomes. BrokenSegue (talk) 22:43, 4 January 2022 (UTC)Reply
Since I have observed this problem during my deletion routines, I have something similar available. However, I think the merging cannot be automated since conceptual identity of the items in question should be verified manually—way too often, items should *not* be merged.
So, if you are interested, I could try to make worklists for you (and others) to have a look at. I am not really able to work this down by myself, since I have already way too many things to do here. —MisterSynergy (talk) 22:47, 4 January 2022 (UTC)Reply
Return to the project page "Bot requests".