Shortcut: WD:RBOT

Wikidata:Bot requests

From Wikidata
Jump to navigation Jump to search
Bot requests

If you have a bot request, add a new section using the button and tell exactly what you want. To reduce the process time, first discuss the legitimacy of your request with the community in the Project chat or in the Wikiprojects's talk page. Please refer to previous discussions justifying the task in your request.

For botflag requests, see Wikidata:Requests for permissions.

Tools available to all users which can be used to accomplish the work without the need for a bot:

  1. PetScan for creating items from Wikimedia pages and/or adding same statements to items
  2. QuickStatements for creating items and/or adding different statements to items
  3. Harvest Templates for importing statements from Wikimedia projects
  4. OpenRefine to import any type of data from tabular sources
  5. WikibaseJS-cli to write shell scripts to create and edit items in batch
On this page, old discussions are archived. An overview of all archives can be found at this page's archive index. The current archive is located at 2021/10.
Filing cabinet icon.svg
SpBot archives all sections tagged with {{Section resolved|1=~~~~}} after 2 days.

Import Treccani IDs[edit]

Request date: 6 February 2019, by: Epìdosis

Task description

At the moment we have four identifiers referring to http://www.treccani.it/: Biographical Dictionary of Italian People ID (P1986), Treccani ID (P3365), Treccani's Enciclopedia Italiana ID (P4223), Treccani's Dizionario di Storia ID (P6404). Each article of these works has, in the right column "ALTRI RISULTATI PER", a link to the articles regarding the same topic in other works (e.g. Ugolino della Gherardesca (Q706003) Treccani ID (P3365) conte-ugolino, http://www.treccani.it/enciclopedia/conte-ugolino/ has links also to Enciclopedia Italiana (Treccani's Enciclopedia Italiana ID (P4223) and Dizionario di Storia (Treccani's Dizionario di Storia ID (P6404)). This cases are extremely frequent: many items have Biographical Dictionary of Italian People ID (P1986) and not Treccani ID (P3365)/Treccani's Enciclopedia Italiana ID (P4223); others have Treccani ID (P3365) and not Treccani's Enciclopedia Italiana ID (P4223); nearly no item has Treccani's Dizionario di Storia ID (P6404), recently created.

My request is: check each value of these identifiers in order obtain values for the other three identifiers through the column "ALTRI RISULTATI PER".

Discussion

Fix local dialing code (P473) wrongly inserted[edit]

Request date: 7 November 2019, by: Andyrom75

Task description

Several entities has a wrong value for the local dialing code (P473) according to the format as a regular expression (P1793) specified in it: [\d\- ]+, as clarified "excluded, such as: ,/;()+"

Typical examples of wrong values, easily identified are the following two:

  1. local dialing code (P473) that includes at the beginning the country calling code (P474)
  2. local dialing code (P473) that include at the beginning the "optional" zero
  • Case 1 can be checked looking for "+", when present, should be compared with the relevant country calling code (P474) and if matched, it should be removed
  • Case 2 can be checked looking for "(" and ")" with zeros inside. If matched it should be removed
Discussion
Request process

Cleaning of streaming media services urls[edit]

Request date: 12 December 2020, by: Swicher

I'm not sure if this is the best place to propose it but when reviewing the urls of a query with this script:

import requests
from concurrent.futures import ThreadPoolExecutor

# Checks the link of an item, if it is down then saves it in the variable "novalid"
def check_url_item(item):
    # Some sites may return error if a browser useragent is not indicated
    useragent = 'Mozilla/5.0 (X11; Linux i686) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.77'
    item_url = item["url"]["value"]
    print("Checking %s" % item_url, end="\r")
    req = requests.head(item_url, headers = {'User-Agent': useragent}, allow_redirects = True)
    if req.status_code == 404:
        print("The url %s in the element %s returned error" % (item_url, item["item"]["value"]))
        novalid.append(item)

base_query = """SELECT DISTINCT ?item ?url ?value
{
%s
  BIND(IF(ISBLANK(?dbvalue), "", ?dbvalue) AS ?value)
  BIND(REPLACE(?dbvalue, '(^.*)', ?url_format) AS ?url)
}"""
union_template = """  {{
    ?item p:{0} ?statement .
    OPTIONAL {{ ?statement ps:{0} ?dbvalue }}
    wd:{0} wdt:P1630 ?url_format.
  }}"""
properties = [
    "P2942", #Dailymotion channel
    "P6466", #Hulu movies
    "P6467", #Hulu series
]
# Items with links that return errors will be saved here
novalid = []

query = base_query % "\n  UNION\n".join([union_template.format(prop) for prop in properties])
req = requests.get('https://query.wikidata.org/sparql', params = {'format': 'json', 'query': query})
data = req.json()

# Schedule and run 25 checks concurrently while iterating over items
check_pool = ThreadPoolExecutor(max_workers=25)
result = check_pool.map(check_url_item, data["results"]["bindings"])

I have noticed that almost half are invalid. I do not know if in these cases it is better to delete or archive them but a bot should periodically perform this task since the catalogs of streaming services tend to be very changeable (probably many of these broken links are due to movies/series whose license was not renewed). Unfortunately I could only include Hulu and Dailymotion since the rest of the services have the following problems:

For those sites it is necessary to perform a more specialized check than a HEAD request (like using youtube-dl (Q28401317) for Youtube).

In the case of Hulu I have also noticed that some items can have two valid values in Hulu movie ID (P6466) and Hulu series ID (P6467) (see for example The Tower of Druaga (Q32256)) so you should take that into account when cleaning links.

Request process

Ontario public school contact info[edit]

Request date: 27 December 2020, by: Jtm-lis

Link to discussions justifying the request
Task description

https://www.wikidata.org/wiki/Wikidata:Dataset_Imports/_Ontario_public_school_contact_information

Licence of data to import (if relevant)
Discussion

request to import podcast identifiers (2021-01-03)[edit]

Request date: 3 January 2021, by: Sdkb

Link to discussions justifying the request
Task description

Several properties have recently been created (see e.g. Castbox show ID (P9005) for podcast identifiers), which are being used for the new w:Template: Podcast platform links on Wikipedia. I was told to come here to get help importing the identifiers for a bunch of podcast items.

Licence of data to import (if relevant)
Discussion

@Sdkb: Please use https://pltools.toolforge.org/harvesttemplates/ , this doesn't need to be done by a bot.Vojtěch Dostál (talk) 12:32, 15 June 2021 (UTC)[reply]

@Vojtěch Dostál: thanks for looking at this. I just looked at Harvest Templates and have no clue how it would fetch external database data on podcast identifiers. Could you advise? {{u|Sdkb}}talk 15:48, 15 June 2021 (UTC)[reply]
@Sdkb: I thought you meant to ask bots to import data from w:Template: Podcast platform links. Is that incorrect? In that case, where do you want to import the data from, and using what key to assign the identifiers to items?Vojtěch Dostál (talk) 19:24, 15 June 2021 (UTC)[reply]
@Vojtěch Dostál: Sorry for the confusion. The podcast platform links template uses data from Wikidata, so I'm looking for Wikidata to mass-import the various identifiers it includes, such as Apple podcasts ID. I'd hope at least some of those could be imported from e.g. Apple itself, but beyond that I'm not sure. {{u|Sdkb}}talk 19:32, 15 June 2021 (UTC)[reply]
@Sdkb: Hmm, I am not sure you were advised correctly by @Sic19:. This page is, to my knowledge, reserved for mostly mindless works which require bots. However, what you have in mind is a dataset import, a much more complex endeavour consisting of several steps - acquisition of CC0-licensed data, cleaning, pairing to existing entities, import etc. Your proposal is a Wikidata version of someone on Wikipedia asking for an article to be written :-) Vojtěch Dostál (talk) 19:39, 15 June 2021 (UTC)[reply]
@Vojtěch Dostál: Ah, thanks for the info. Given the backlog at w:WP:Requested articles, I'm guessing if the analogy holds my odds aren't too great of it being taken up anytime soon. But I'll look at the dataset imports page and see if there's a place to add a request. {{u|Sdkb}}talk 19:47, 15 June 2021 (UTC)[reply]
@Sdkb: If you can be a little more specific about what dataset you want imported and how it could be done I might be interested in picking it up. But generically importing any data using unknown means (how we would link the datasets) it's a bit too much to do. BrokenSegue (talk) 23:54, 15 June 2021 (UTC)[reply]
@BrokenSegue: The most significant IDs are probably Apple, Google, and Spotify, so if any of those seem to have open data, those would be the ones to import. The structures at those places are explained at Apple Podcasts podcast ID (P5842), Google Podcasts show ID (P9003), and Spotify show ID (P5916). If that's enough for you to go off of, it'd be fantastic to see those properties showing up more often at Wikidata items, and would in turn allow us to start using w:Template: Podcast platform links more widely on Wikipedia. But if the data isn't open or something, we might just be out of luck; the next step would have to be reaching out to the platforms for help (they should theoretically be eager to help us, as it gets their name onto a bunch of podcast Wikipedia pages). {{u|Sdkb}}talk 01:42, 16 June 2021 (UTC)[reply]
Request process

reference URL (P854)Holocaust.cz person ID (P9109) (2021-02-05)[edit]

Request date: 5 February 2021, by: Daniel Baránek

Task description

After intoducing Holocaust.cz person ID (P9109), reference URL (P854) in references can be replaced by this new identificator. The result of edits should be like this. It is 285,282 references. You can see all references, their reference URL (P854) value and value for Holocaust.cz person ID (P9109) here:

SELECT ?ref ?url ?id WHERE {
  ?ref prov:wasDerivedFrom [ pr:P248 wd:Q104074149 ; pr:P854 ?url ].
  BIND (REPLACE(STR(?url),"^.*/([0-9]+)[-/].*$","$1") as ?id)
  }
Try it!
Discussion


Request process

request to add identifiers from FB (2021-02-11)[edit]

Thanks to a recent import, we currently have more than >1.2 items where the only identifier is Freebase ID (P646). However, checking https://freebase.toolforge.org/ some of them have identifiers available there.

Samples:

See Wikidata:Project_chat#Freebase_(bis) for discussion.

Task description

Import ids where available. Map keys to properties if not available at Wikidata:WikiProject_Freebase/Mapping.

Discussion


Request process

request to change Belarusian language description from "спіс атыкулаў у адным з праектаў Вікімедыя" to "спіс артыкулаў у адным з праектаў Вікімедыя" in all the articles. A letter "р" was missed (2021-02-23)[edit]

Request date: 23 February 2021, by: Belarus2578

Link to discussions justifying the request

There is not discussion. There is only obvious mysprint. --Belarus2578 (talk) 05:01, 25 February 2021 (UTC)[reply]

Task description

Please, change Belarusian language description from "спіс атыкулаў у адным з праектаў Вікімедыя" to "спіс артыкулаў у адным з праектаў Вікімедыя" in all the articles. A letter "р" was missed. --Belarus2578 (talk) 06:47, 23 February 2021 (UTC)[reply]

Discussion
Pictogram voting comment.svg Comment There are over 250,000 items. --Matěj Suchánek (talk) 10:15, 13 March 2021 (UTC)[reply]
I would like to tackle this, do this still need community discussion? Ammarpad (talk) 14:07, 8 June 2021 (UTC)[reply]
@Ammarpad: I don't think you need full community discussion, but an independent confirmation by one more Belarusian speaker would be nice. @EugeneZelenko, Liashko, Хомелка: can you please confirm this? Vojtěch Dostál (talk) 12:38, 21 June 2021 (UTC)[reply]
@Vojtěch Dostál: Good. If any of them can confirm, I would do it then. Ammarpad (talk) 13:42, 21 June 2021 (UTC)[reply]
Yes, request is reasonable and new title is correct. --EugeneZelenko (talk) 14:13, 21 June 2021 (UTC)[reply]
@Ammarpad Still working on this? :) Vojtěch Dostál (talk) 10:58, 29 September 2021 (UTC)[reply]
@Vojtěch Dostál: Yes, I wrote the code but didn't submit bot request due to some other work. I will do that now. Ammarpad (talk) 12:49, 29 September 2021 (UTC)[reply]
Request process

Accepted by (Ammarpad (talk) 17:13, 29 September 2021 (UTC)) and under process
[reply]

request to uprank current existing countries (2021-04-10)[edit]

Request date: 10 April 2021, by: Bouzinac

Link to discussions justifying the request
Task description

Help clean P17 data by:

Exemple : Q2492784#P17 --> Ukraine (Q212) [which does not have any P576] + Soviet Union (Q15180) [which has a P576] ==>Ukraine (Q212) should be upranked

Discussion
Request process

Fix capitalization and grammar of Bosnian labels (2021-04-14)[edit]

Request date: 14 April 2021, by: Srđan

Link to discussions justifying the request
Task description

See: quarry:query/54093

Could you run the query once more? As it should show now a lot less then the 418824 items of April 14th. Edoderoo (talk) 15:05, 2 May 2021 (UTC)[reply]
@Edoderoo:: Sorry for the late reply. Just re-ran the query and it's sitting at 224,889 items. Definitely fewer than before, but still a lot to go. – Srđan (talk) 16:13, 8 May 2021 (UTC)[reply]
Update: Resultset (141.438 rows) Edoderoo (talk) 14:57, 21 May 2021 (UTC)[reply]
The wikidata-queries were empty, but the quarry had still some left, those are now in process. Almost finished ;-) Edoderoo (talk) 15:00, 8 June 2021 (UTC)[reply]

These are the descriptions that show be written in lowercase and slightly altered:

Srđan (talk) 08:37, 30 April 2021 (UTC)[reply]

Licence of data to import (if relevant)
Discussion
  • Here is a query: [1]. Maybe check if any bots still add more of them. --- Jura 09:55, 14 April 2021 (UTC)[reply]
Request process

Accepted by (Edoderoo (talk) 15:05, 2 May 2021 (UTC)) and under process
[reply]

Task completed

There are some items left, those are not true Wikimedia Categories, but special categories. I solved the ones with many items in them (like the category for stubs) but there are too many with just a few entries in the end. 99% of the request is done.

request to .. (2021-04-19)[edit]

Request date: 19 April 2021, by: Powell Street Festival Society

Link to discussions justifying the request

I have been tasked by the Powell Street Festival Society to upload to Wikidata a listing of Japanese-Canadian Artist information from the Japanese-Canadian Artists Directory.

Task description

I have worked through the various Wikidata steps to prepare the data to be imported. The data is in an Excel spreadsheet. It appears that I am on Step 6. I can provide a sample file with column headers to check that I have parsed the data properly.

Thank you for your attention with this request. I look forward to your response.

Regards, Michael

Licence of data to import (if relevant)
Discussion

@Powell Street Festival Society: What list of steps are you following? You don't necessarily need a bot to do this import. BrokenSegue (talk) 17:35, 19 April 2021 (UTC)[reply]

Hello BrokenSegue I am new to this process (and to Wikidata) and have been following the steps in the "Data Import Guide", I have created the "Dataset Summary" and it appears I am on Step 7: Match the data to Wikidata (Option 2: self import). I could really use some help to figure this out. I am not even sure if I am replying properly :)

Request process

September 17, 2021

Discussion

BrokenSegue if you are available, I still need help with the steps to properly upload the Excel data to Wikidata. As mentioned I am currently on step 6 of the "Data Import Guide" and not sure how to proceed. Any assistance you could provide would be greatly appreciated. Regards,

Michael

request to import data for "Cheung Chau Piu Sik Parade" (2021-05-06)[edit]

Request date: 6 May 2021, by: Hkbulibdmss

Link to discussions justifying the request
Task description

https://www.wikidata.org/wiki/Wikidata:Dataset_Imports/Cheung_Chau_Piu_Sik_Parade

Please help to import the dataset. The URL of a spreadsheet is : https://docs.google.com/spreadsheets/d/1iUVrHNsXVmn94IygtZYj0-foeUg9yvdOcwQ_V-CQbto/edit?usp=sharing

Licence of data to import (if relevant)
Discussion


Request process

request to fix parliamentary group = caucus, != party (2021-05-12)[edit]

Request date: 12 May 2021, by: Jura1

Link to discussions justifying the request
Task description
Discussion


Request process

request to automate marking preferred_rank for full dates. (2021-05-28)[edit]

Request date: 28 May 2021, by: Richard Arthur Norton (1958- )

Task description

We have year only dates and full dates for date_of_birth and date_of_death. See for instance Eliot Blackwelder (Q16785350). We need to mark the full date as "preferred rank" and add in the reason_for_preferred_rank=most complete record (Q105749746). The problem is when we have two dates of equal rank, both display in infoboxes. --RAN (talk) 04:45, 28 May 2021 (UTC)[reply]

Discussion
Request process

@Richard Arthur Norton (1958- ): What about references though? What if the less complete date has a reference and the other does not? Should we still do this? I might be able to find time to do this. BrokenSegue (talk) 05:21, 28 May 2021 (UTC)[reply]

I guess in the case where the two dates disagree we should not perform the update. BrokenSegue (talk) 05:22, 28 May 2021 (UTC)[reply]
That would be great, I haven't seen the bot in action yet, I am still plugging away by hand as I come across them. --RAN (talk) 20:20, 28 May 2021 (UTC)[reply]
No, my bot does not manipulate ranks. --Matěj Suchánek (talk) 11:52, 29 May 2021 (UTC)[reply]
  • Maybe it was someone else's. Sorry then. --- Jura 11:59, 29 May 2021 (UTC)[reply]
  • @Matěj Suchánek: I think I had this in mind. --- Jura 09:33, 30 May 2021 (UTC)[reply]
    Indeed, my bot still does that (every Wednesday). In fact, it has evolved since, it also merges (seemingly) duplicate dates (that issue with -00-00 vs. -01-01 etc.). But it does not change ranks, and it even avoids statements with non-normal rank. --Matěj Suchánek (talk) 10:26, 30 May 2021 (UTC)[reply]

@Matěj Suchánek: Are you interested in picking this task up? It does kinda overlap with the task Jura mentioned. Actually, hmm, there is some subtlety here that I can see being tricky (multiple dates with different qualifiers sometimes shouldn't be merged e.g. for start time (P580)s with a applies to part (P518)). If not I may still do it. BrokenSegue (talk) 12:40, 30 May 2021 (UTC)[reply]

Sorry, I am not right now. I guess it's easy now that we have Ranker (Q105394978), which can be driven by SPARQL. (Or maybe not that easy if the qualifier is also required, but QS can do this part.) I made up a query which can be used as basis.
What if the less complete date has a reference and the other does not? Preferred statements should always be sourced. If there is no evidence for the more precise date, it should be either removed or sourced (and then up-rank'd). --Matěj Suchánek (talk) 13:12, 30 May 2021 (UTC)[reply]
Thanks for the query; you're a SPARQL wizard. I write my bot actions self-contained in python so I don't need ranker. BrokenSegue (talk) 14:07, 30 May 2021 (UTC)[reply]
Excellent! I know there are several bots trying to fill in references for dates, but they are mostly pulling data from sources that give year-only dates. At one time I calculated that about 20% of year-only dates are off by a year because they are back calculated from the age at death in an obituary. --RAN (talk) 00:37, 1 June 2021 (UTC)[reply]
Do you know who is operating these bots? Wikibase in theory supports adding uncertainty in dates but in practice I believe the correct way to add a date with that kind of uncertainty is to use e.g. earliest date (P1319). BrokenSegue (talk) 01:31, 1 June 2021 (UTC)[reply]

request to replace qualifiers in GND ID (2021-06-07)[edit]

Request date: 7 June 2021, by: Kolja21

Link to discussions justifying the request
Task description

Please replace in GND ID stated as (P1932) with named as (P1810)

  1. GND ID (P227) delete qualifier stated as (P1932)
  2. import name of object from GND with qualifier named as (P1810)
  3. add retrieved (P813)

Scope: 5.161 qualifiers stated as (P1932), see Wikidata:Database reports/Constraint violations/P227#Properties statistics.

Pictogram voting comment.svg Comment (in German): Man könnte hinzufügen, dass man über die OpenRefine Reconciliation oder über https://d-nb.info/gnd/100045642/about/lds.ttl (gndo:preferredNameForThePerson) recht einfach und schnell die aktuelle Version abfragen kann. (User:Emu)

Example
Discussion
Request process

Accepted by (Ammarpad (talk) 14:01, 10 June 2021 (UTC)) and under process
[reply]

request to cleanup DOI only items (2021-07-04)[edit]

Request date: 4 July 2021, by: Jura1

Task description

Items like Q57554778 consist mainly of DOI: the DOI is repeated as title and label.


@Daniel Mietchen: who created some or all of them. @Trilotat: who mentioned some on Wikidata:Request_a_query#Items_with_DOI_(P356)_that_start_with_10.1023/A:_without_a_Label_or_a_title_(P1476). --- Jura 13:24, 4 July 2021 (UTC)[reply]

@Jura1: To be precise, I was looking for items without a label, but I had seen this and did some research. A web search for any of the "DOI as title" DOIs will find that they are all or almost all noted in ResearchGate publication ID (P5875) items associated with Entomologia Experimentalis et Applicata (Q15753202) journal. These items are published in (P1433) CrossRef Listing of Deleted DOIs (Q53952674).
  • Q57554778 is 10.1023/A:1003902321787 and that DOI is mentioned in ResearchGate publication ID (P5875) 226608108. That researchgate item mentions the title and article details as Q107413498.
  • I added the deleted DOI to that matched item as deprecated (as withdrawn identifier value).
  • They should be merged, but I didn't as I thought it might confuse this bot request.
In the future, I think we can add the new DOI to the bad items and then rerun SourceMD as I did with Q57030816, right? Trilotat (talk) 14:54, 4 July 2021 (UTC)[reply]
List of items: User:Jura1/DOI as label. It was done using regexp 10\..+/ for title (P1476) values. — Ivan A. Krestinin (talk) 20:15, 21 July 2021 (UTC)[reply]
Request process

request to add reference (2021-07-04)[edit]

Request date: 4 July 2021, by: Data Gamer

Link to discussions justifying the request
Task description

Hello. In all items (56 items) that have position held (P39) -> member of the House of Representatives of Cyprus (Q19801674) with qualifier parliamentary term (P2937) -> 12th Cypriot Parliament (Q107003549)

I want to add the reference to above statement:

reference URL (P854) -> http://www.parliament.cy/el/general-information/%CE%B2%CE%BF%CF%85%CE%BB%CE%B5%CF%85%CF%84%CE%B9%CE%BA%CE%AD%CF%82-%CE%B5%CE%BA%CE%BB%CE%BF%CE%B3%CE%AD%CF%82/%CE%B5%CE%BA%CE%BB%CE%BF%CE%B3%CE%AD%CF%82-30%CE%AE%CF%82-%CE%BC%CE%B1%CE%90%CE%BF%CF%85-2021

title (P1476) -> Εκλογές 30ής Μαΐου 2021 (in Greek (el) Language)

retrieved (P813) -> 2021-07-04

archive URL (P1065) -> https://web.archive.org/web/20210704152630/http://www.parliament.cy/el/general-information/%CE%B2%CE%BF%CF%85%CE%BB%CE%B5%CF%85%CF%84%CE%B9%CE%BA%CE%AD%CF%82-%CE%B5%CE%BA%CE%BB%CE%BF%CE%B3%CE%AD%CF%82/%CE%B5%CE%BA%CE%BB%CE%BF%CE%B3%CE%AD%CF%82-30%CE%AE%CF%82-%CE%BC%CE%B1%CE%90%CE%BF%CF%85-2021

archive URL (P1065) -> https://archive.is/loRfw

archive date (P2960) -> 2021-07-04

language of work or name (P407) -> Greek (Q9129)

publisher (P123) -> House of Representatives (Q1112381)

Thanks.

Licence of data to import (if relevant)
Discussion


Request process

Proliferate external-IDs from qualifiers and references to main statement (2021-07-06)[edit]

Request date: 6 July 2021, by: Vladimir Alexiev

Link to discussions justifying the request
Task description

Take a prop like ORCID: Property_talk:P496 says that 22.7% of uses are as reference, and 0.1% as qualifier.

I bet that some of those uses are not reflected as main statement.

SELECT ?itemLabel ?wdt ?wdLabel ?id { # ?ref ?wdr ?statement {
  ?wd wikibase:propertyType wikibase:ExternalId; wikibase:directClaim ?wdt; wikibase:reference ?wdr.
  ?ref ?wdr ?id.
  ?statement prov:wasDerivedFrom ?ref.
  # ?item ?prop ?statement
  # filter not exists {?item ?wdt ?id}
  SERVICE wikibase:label { bd:serviceParam wikibase:language "[AUTO_LANGUAGE],en". }
} limit 10
Try it!

Of course, sifting through all those external-IDs used as refs will be a huge task. WD times out even on a count query:

SELECT (count(*) as ?c) {
  ?wd wikibase:propertyType wikibase:ExternalId; wikibase:reference ?wdr.
  ?ref ?wdr ?id.
}
Try it!
Discussion


Request process

Request to change lexeme forms' grammatical features (2021-07-08)[edit]

Request date: 8 July 2021, by: Bennylin

Link to discussions justifying the request
Task description

How can I change grammatical features of form? (I operate bot, I just need to know the commands). I have the list of lexemes. I reckon this should be not too hard, I'm just not familiar with the command to do the changes.

Licence of data to import (if relevant)
Discussion


Request process

request to update Template:Tr langcodes counts monolingual text periodically (2021-07-14)[edit]

Request date: 14 July 2021, by: Jura1

Task description

--- Jura 09:50, 14 July 2021 (UTC)[reply]

Request process

Help Bota .. (2021-07-27)[edit]

Request date: 27 July 2021, by: Takhirgeran Umar

Link to discussions justifying the request
Task description
Licence of data to import (if relevant)
Discussion
Pictogram voting comment.svg Comment There are around 120,000 changes. --Matěj Suchánek (talk) 16:20, 6 August 2021 (UTC)[reply]
to clarify you want all items with that description replaced with that other description? Is there discussion around this? I can do it easily but no idea if this is an "Accepted" change. BrokenSegue (talk) 19:37, 15 August 2021 (UTC)[reply]
Request process

request to Change References to Qualifiers on property P2949 (2021-08-07)[edit]

Request date: 7 August 2021, by: Lesko987a

Link to discussions justifying the request
Task description
  • I was adding properties named as (P1810) and retrieved (P813) as a reference to the WikiTree person ID (P2949) identifier. User Jura1 suggested to change it from a reference to a qualifier. And I agree. It was done on almost all P2949 properties (210K out of 215K). That is a lot of changes to do and Jura suggested you can do it much faster. I would need to delete all references added and then add them as qualifiers.
Licence of data to import (if relevant)
Discussion


Request process
@Pasleim: may help.--GZWDer (talk) 17:22, 8 August 2021 (UTC)[reply]

request to add YOB and/or YOD to TP descriptions (2021-09-01)[edit]

Request date: 1 September 2021, by: Jura1

Task description

Many TP imported items have a description in the form "Peerage person ID=\d*". These were added when these items didn't include more information.

In the meantime, some of these items include date of birth (P569) and/or date of death (P570). To make it easier to identify them, the years from these dates should be added to the description.

  • Sample edit: [3].
  • Query to find items (currently 28776):
SELECT DISTINCT ?item ?itemLabel ?d
{
  hint:Query hint:optimizer "None".
  ?item wdt:P4638 [] .
  ?item (wdt:P569|wdt:P570) [] .
  ?item schema:description ?d . 
  FILTER( lang(?d) && regex (?d, "^Peerage person ID=\\d+$") ) 
  SERVICE wikibase:label { bd:serviceParam wikibase:language "[AUTO_LANGUAGE],en". }
}
Try it!

Thanks --- Jura 12:44, 1 September 2021 (UTC)[reply]

Discussion

I had a look to some of them, but the data is quite messy, due to a source that is messy too. I do not see a good reason to have this data in WikiData, the only reason it got imported is because it was there. Any effort in describing this, will not make a lesser mess. Edoderoo (talk) 20:09, 19 September 2021 (UTC)[reply]

  • I thought I was the only one who wasn't really convinced by the import. It has some merits and has gotten a bit more useful since. If we don't want to delete some of it, we should at least try to normalize its labels and descriptions. --- Jura 21:37, 19 September 2021 (UTC)[reply]
    @Jura1 I'm on it. How do you envision the new descriptions? Eg. "Peerage person ID=92310 (1856–1915)"? Vojtěch Dostál (talk) 06:13, 29 October 2021 (UTC)[reply]
    • Either this or just the years. --- Jura 12:42, 29 October 2021 (UTC)[reply]

Also adding dates to descriptions adds data duplication, causes problems when the dates are found to be wrong later and don’t really help that much for TP entries from my experience. --Emu (talk) 08:13, 29 October 2021 (UTC)[reply]

labels and descriptions are meant to duplicate information also available as statements. If you have a better strategy for TP, let's hear it. --- Jura 12:42, 29 October 2021 (UTC)[reply]
There’s a reason why dates aren’t mentioned in WD:D. As for a better strategy: Just leave it as it is. Your claim that different descriptions would “make it easier to identify them” begs the question: Why? To whom? And will those people try to add information to those items or does it just add aesthetic value? --Emu (talk) 16:54, 29 October 2021 (UTC)[reply]
Can you point me to the explanation with the reason (for English descriptions)? It's a standard element for descriptions in Dutch.
Maybe you can also explain what the TP ID should be doing instead?
It's fairly common that years (not dates) of birth/death are used to disambiguate. At least in databases .. --- Jura 18:47, 29 October 2021 (UTC)[reply]
I don’t know the full backstory of English descriptions in Wikidata. What I do know is that descriptions with YOB and/or YOD do often cause a lot of work when dates are corrected.
I’m not sure what you mean by your second question. If you refer to the descriptions Peerage person ID=: Well, they are there. There’s little use in deleting them so why bother.
And yeah, they are. You don’t need descriptions for that. --Emu (talk) 20:38, 29 October 2021 (UTC)[reply]
@Emu To whom? For example to people who work with Mix'n'Match to pair external databases to these items. I definitely see merit in having dates of birth and death in description, in absence of a better descriptor. Vojtěch Dostál (talk) 19:55, 29 October 2021 (UTC)[reply]
@Vojtěch Dostál: I don’t quite follow – M’n’M has autodescriptions that include life dates.
Don’t get me wrong, if you two are really keen on this bot job, go ahead. I just don’t see the need. --Emu (talk) 20:38, 29 October 2021 (UTC) [reply]
Request process

Parts for duos (14 September 2021)[edit]

A while back, we generated missing parts for duos. Each duo would generally have one item for each member. This finds some that lack parts. Maybe some more filtering needs to be done.

Sample items: Q6161933, Q52375494.

For a list, see Wikidata:WikiProject Q5/lists/duos.

Previous request: Wikidata:Bot_requests/Archive/2016/12#duos_without_parts. @Matěj Suchánek: --- Jura 14:56, 14 September 2021 (UTC)[reply]

@Jura1: There was one iteration in October [4] and I've scheduled one more for November. But there is a hard constraint for the bot that it must find at least one label for both new items. It can be helped by adding a more specific class items, like Special:Diff/1518556965, Special:Diff/1518552635. But this must be done by hand and the information needs to be in the Wikipedia articles.
Do you think it makes sense to create an item for "a cousins duo", like José and Francisco Díaz (Q6294250) or Mary and Molly Bell (Q6781014)? --Matěj Suchánek (talk) 10:07, 28 October 2021 (UTC)[reply]
Thanks for looking into this.
Both samples have a Wikipedia article linked to them. I think it does make sense to create an item for each individual, but that we have one for them together depends on Wikipedia.
I will try to do the remaining ones manually once the November run has gone through. --- Jura 12:39, 29 October 2021 (UTC)[reply]

Improve descriptions of items for people (22 September 2021)[edit]

Maybe you have seen Wikidata:Guide_to_James_Bakers.

It tries to enumerate and describe items for people with a fairly frequent names. It also has a todo list, indicating how many there are still to come.

A few queries on that page find these items differently.

As items grow, it could be interesting to complete the items' descriptions, e.g. with years, occupation, nationality, etc.

Maybe the same could be extended to any name found on Special:WhatLinksHere/Q22808320. Currently 352178.

This should facilitate choosing among potential items. --- Jura 21:07, 22 September 2021 (UTC)[reply]

Request to create a bot for automatically updating YouTube subscriber counts (2021-09-23)[edit]

Request date: 23 September 2021, by: Sdkb

Link to discussions justifying the request
Task description

YouTube personality infoboxes on English Wikipedia include subscriber counts, and these are constantly being edited on YouTuber pages to manually update the numbers. In this discussion, we reached consensus to move the subscriber counts to Wikidata if they can be updated here by bot. This would result in more up-to-date data, reduce the amount of editor effort needed, and declutter watchlists. Could someone write an update bot so that we can move forward with this? It seems like a great opportunity for better integration, so I hope we don't let it pass by. {{u|Sdkb}}talk 18:13, 23 September 2021 (UTC)[reply]

Licence of data to import (if relevant)
Discussion
Request process
@Sdkb: Good news. I have a bot already doing this. See for example this edit made by the bot. BrokenSegue (talk) 18:18, 23 September 2021 (UTC)[reply]
(edit conflict)Oh, wait, looking into this, it seems that it's already being partially handled by BorkedBot. I'm not seeing it always run weekly, though. BrokenSegue, could you give an update on the status/effectiveness of the bot? If there are improvements that could be made, we should do that before its data starts being used on a much wider scale. (Wikipedians don't hesitate to attack Wikidata when things go wrong, and we don't want to lose the opportunity to centralize the data here.) {{u|Sdkb}}talk 18:23, 23 September 2021 (UTC)[reply]
@Sdkb: how frequent do wikipedians want it? The bot currently runs once per item and then I run it again every year or so. I don't know if wikidata users would appreciate having it run weekly. We don't delete old data so it'd quickly clutter things up. Users could of course update it manually more frequently if it moved drastically. BrokenSegue (talk) 18:27, 23 September 2021 (UTC)[reply]
@Sdkb: oh also currently the bot doesn't save subscriber counts for channels with less than 10000 followers BrokenSegue (talk) 18:34, 23 September 2021 (UTC)[reply]
@BrokenSegue: From what I've seen at Wikipedia, the updates are frequently done by IP/new editors, and in many cases they really seem to love keeping things as up to date as possible (given that subscriber counts are a sort of currency, there's an appeal to topping them off). I think we'd want to make the updates at least weekly if not more frequent than that, since otherwise there will be pressure from those sources. Regarding cluttering up, I'm not sure if this would be unorthodox, but maybe the bot could update weekly but then also remove some portion of older data so it's not excessive? Another thing that's unorthodox but possibly helpful would be to build in some anti-vandalism functionality (since subscriber counts are a possible target), where if someone drastically changes a subscriber count, it notices and checks the latest numbers and corrects if needed. Regarding <1000, I don't think that is likely to be too much of an issue, as I assume there are only a small handful of such articles on Wikipedia, and I don't think subscriber counts would be sorely missed if left out of those (the infobox will presumably have some manual override built in when we do the conversion). {{u|Sdkb}}talk 18:53, 23 September 2021 (UTC)[reply]
Hmm, the anti-vandalism work would need some thinking. That's pretty out of scope for this bot. I could do weekly but I do wonder if I would need to go back and request permission since this is not what I said the bot would do. Maybe every two weeks would be acceptable? Personally I don't really see value in updating it more frequently for anyone. I would also worry about users changing the most recent value instead of adding a new value as we would prefer in wikidata. BrokenSegue (talk) 18:59, 23 September 2021 (UTC)[reply]
What I envision happening if the updates aren't frequent enough is lots of scenarios like this: Channel Foobar surpassed 1 million subscribers about a week ago. The creator puts out a celebration video, and some of the fans notice that the Wikipedia page still only lists the subscriber count as 990,000. The Wikidata interface is too complicated for them to figure out, so they either attempt to update it and make a mistake that messes something up, or they end up at forums like the Teahouse where experienced editors have to spend time looking into the situation and advising them to be patient and wait for the bot's next update. We saw a lot of this type of behavior for COVID infection data, and the nature of YouTube subscriber counts makes me think it'd be even more intense for that. {{u|Sdkb}}talk 19:09, 23 September 2021 (UTC)[reply]
Ok, I'll file a quick bot request to increase the frequency. BrokenSegue (talk) 19:11, 23 September 2021 (UTC)[reply]

Request to add publication numbers to NGA lighthouse ID (P3563) (2021-09-30)[edit]

Request date: 30 September 2021, by: MSGJ

Link to discussions justifying the request
Task description

There are more than 1000 violations of the formatting constraint listed at Wikidata:Database reports/Constraint violations/P3563#"Format" violations. In most cases this is because the publication number is missing from the identifier. This is a 3-digit number (110, 111, 112, 113, 114, 115 or 116). Without this number the identifier is not unique and so almost useless.

The publication number depends on the area of the world (see map). It could be deduced from coordinate location (P625) but I think this may be too difficult. Probably easier to identify each country (P17) to the relevant number. I can help to generate this list. (There will be a few exceptions, for example overseas territories of some countries, but I can track down and fix these.)

Thanks for considering this task — Martin (MSGJ · talk) 20:22, 30 September 2021 (UTC)[reply]

Example

Sunosaki Lighthouse (Q1087248) was in the list of violations. NGA lighthouse ID (P3563) was 5004 and it was missing its publication number. Its country (P17) is Japan (Q17) which means it is in area 112. So NGA lighthouse ID (P3563) is changed to 112-5004 and now the everything is good.

Discussion

Hi @MSGJ:, I would like to have the list (i.e country QID => code) . Ammarpad (talk) 14:41, 3 October 2021 (UTC)[reply]

@Ammarpad: I do not yet have a list, but this is copied from en:List of lights#United States:
  • PUB. 110 - Greenland, the East Coasts of North and South America (excluding Continental U.S.A. except the East Coast of Florida) and the West Indies
  • PUB. 111 - The West Coasts of North and South America (Excluding Continental U.S.A. and Hawaii), Australia, Tasmania, New Zealand, and the Islands of the North and South Pacific Oceans
  • PUB. 112 - Western Pacific and Indian Oceans Including the Persian Gulf and Red Sea
  • PUB. 113 - The West Coasts of Europe and Africa, the Mediterranean Sea, Black Sea and Azovskoye More (Sea of Azov)
  • PUB. 114 - British Isles, English Channel and North Sea
  • PUB. 115 - Norway, Iceland and Arctic Ocean
  • PUB. 116 - Baltic Sea with Kattegat, Belts and Sound and Gulf of Bothnia

If you could help me generate a query for the items with missing publication numbers, sorted by Country, then I will start making the list based on the above. Thanks — Martin (MSGJ · talk) 10:00, 4 October 2021 (UTC)[reply]

Actually I think I can handle this myself using QuickStatements. Thanks anyway — Martin (MSGJ · talk) 22:12, 5 October 2021 (UTC)[reply]
(Belated response) I am not sure how to write SPARQL queries, if that's what you wanted. But I created this (User:AmmarBot/P3563) with custom script. Is that what you're looking for? Ammarpad (talk) 07:33, 7 October 2021 (UTC)[reply]
Request process

request to populate items with relevant information from snwiki (2021-10-10)[edit]

Request date: 10 October 2021, by: Capmo

Task description

There are currently 1815 articles in Category:Mazita eVanhu of the Shona Wikipedia. All of them seem to have an item at Wikidata, but the only information provided at item creation was a link to the article. These items should be populated with relevant information as was done here.

Discussion


Request process

Accepted by (Edoderoo (talk) 18:11, 17 October 2021 (UTC)) and under process
Task completed, see source code. About 15 items could not get a description, because they are actually merge candidates with other items that have the same label/description. (05:28, 18 October 2021 (UTC)) The properties could have been added with PetScan easily, for descriptions we used to have descriptioner, but this tool died recently by lack of maintenance. Edoderoo (talk) 08:33, 19 October 2021 (UTC)[reply]

Great! Thank you. —capmo (talk) 09:32, 19 October 2021 (UTC)[reply]
I think that this discussion is resolved and can be archived. If you disagree, don't hesitate to replace this template with your comment. Matěj Suchánek (talk) 11:46, 28 October 2021 (UTC)[reply]

request to depreciated ethnic group only sourced with P143 (2021-10-23)[edit]

Request date: 23 October 2021, by: Fralambert

Link to discussions justifying the request
Task description
Hi, since ethnic group (P172) is a highly contencious subject, the property already mandate a source and imported from Wikimedia project (P143) is not a reliable source, it would be fine if a bot put a depreciated rank when the statement in ethnic group (P172) use only imported from Wikimedia project (P143) as a source. Also the bot could add reason for deprecation (P2241) and source known to be unreliable (Q22979588)as a qualifier. We could also only remove statement with this source, but they are likely to come back, so depreciated them would be a best. --Fralambert (talk) 15:12, 23 October 2021 (UTC)[reply]
Licence of data to import (if relevant)
Discussion
  •  Support - it would also be nice to deprecate or remove all those same statements, when there is no source at all… --Hsarrazin (talk) 15:19, 23 October 2021 (UTC)[reply]
  •  Support, this could be an ongoing task, since this is not the first time this property needs to be cleaned up: Property talk:P172. — eru [Talk] [french wiki] 15:32, 23 October 2021 (UTC)[reply]
  •  Support; I would even support a complete removal of unsourced or Wikipedia-imported claims. —MisterSynergy (talk) 15:58, 23 October 2021 (UTC)[reply]
    complete removal would only lead to re-adding of the same statements…  I've already cleaned hundreds of so-called "French" ethnic group, only to see them back after months - a lot of contributors tend to use P172 instead of P27… Hsarrazin (talk) 16:43, 23 October 2021 (UTC)[reply]
    If we were to make it, say, a daily job, then we would not accumulate larger amounts of unsourced claims anymore and the users who add these unsourced claims would also learn quickly to adapt to the new situation. —MisterSynergy (talk) 18:30, 23 October 2021 (UTC)[reply]
    I think that unsourced statement should be deleted, since sourcing of this property is mandatory. As Wikipedia-imported claim, best is to kept them as depreciated as they are the most likely to come back. Fralambert (talk) 18:51, 23 October 2021 (UTC)[reply]
  • delete them. As Help:Ranking#Deprecated_rank says, deprecation isn't an option for claims that can't be sourced. --- Jura 14:01, 25 October 2021 (UTC)[reply]
    Agree. Deprecation would cause a mess. Remove if you think that's necessary. Vojtěch Dostál (talk) 06:01, 29 October 2021 (UTC)[reply]
Request process

request to delete wrong references (2021-10-27)[edit]

Request date: 27 October 2021, by: LutiV

Link to discussions justifying the request
Task description

Please delete the wrong references from title (P1476) and genre (P136): it was used the property Archivio Storico Ricordi person ID (P8290) instead property Archivio Storico Ricordi opera ID (P8732). The list is: https://w.wiki/4DFS

Licence of data to import (if relevant)
Discussion


Request process