Wikipedia:Bots/Requests for approval/MusikBot 14


 * The following discussion is an archived debate. Please do not modify it. To request review of this BRFA, please start a new section at WT:BRFA. The result of the discussion was

MusikBot
Operator:

Time filed: 01:06, Sunday, November 18, 2018 (UTC)

Function overview: Stores the size of the new page patrol backlog over time for use in charts.

Automatic, Supervised, or Manual: Automatic

Programming language(s): Ruby

Source code available: GitHub

Links to relevant discussions (where appropriate): Wikipedia talk:New pages patrol/Reviewers (permalink)

Edit period(s): 4 times a daily

Estimated number of pages affected: 4 JSON pages that live under New pages patrol.

Namespace(s): Wikipedia

Exclusion compliant (Yes/No): No, not applicable.

Function details: For a while the hard-working folks at WT:NPP/R have been keeping track of the new pages patrol backlog by manually entering in dates and values at Wikipedia talk:New pages patrol/Reviewers/Backlog chart. This bot task automates this process, whilst offering four separate granularities -- hourly, daily, weekly and monthly. These are stored on four separate JSON pages. The data is fetched from 's database on Toolforge. This has been running smoothly in the MusikBot userspace for about a week now. See the chart template in action at User:MusikBot/NPPChart/Chart. Each dataset is pruned after a set amount of time, to prevent it from growing to be too large. This is configured at User:MusikBot/NPPChart/config (see also the documentation). The JSON pages include the following (I will also indicate their current time span):
 * New pages patrol/Backlog chart/hourly -- this is a bit of a misnomer. It actually stores data for 4-hour windows, not single-hour. Data goes back 4 days.
 * New pages patrol/Backlog chart/daily -- Data goes back 6 months.
 * New pages patrol/Backlog chart/weekly -- All available data, since August 29, 2017.
 * New pages patrol/Backlog chart/monthly -- All available data, since August 29, 2017. This is a bit redundant to the weekly granularity, but in the future the weekly dataset might be trimmed down to say, a year's worth of data, in which case the monthly dataset will be unique.

I have personally compared Nettrom's data with the PageTriage API (the official backlog size), and it matches verbatum. This is excluding redirects.

Essentially, this BRFA is about moving the datasets and the template to be subpages of New pages patrol. Specifically, we'll go with New pages patrol/Backlog chart as the template, with the JSON subpages at /hourly, /daily, /weekly and /monthly.

Discussion

 * Note there was previously an issue with the cron, which is why you see edits to the same page side by side, e.g. Special:Diff/869189547 and Special:Diff/869189734 (one minute later). This has been fixed. I can't explain why the values changed. I guess Nettrom's bot continually updates the values for that time slot. Regardless, once MusikBot does it's next run, all data start to finish is refreshed, so it should always be accurate. &mdash; MusikAnimal  talk  01:06, 18 November 2018 (UTC)
 * Good task for a bot, non-controversial (sticks to designated report pages in WP namespace), experienced bot-op, can probably be speedy approved - F ASTILY   08:01, 20 November 2018 (UTC)
 * Great task for a bot, uncontroversial, and helpful.  SQL Query me!  09:35, 20 November 2018 (UTC)
 * Here are the relevant edits: hourly, daily, and weekly. The bot appears to have worked without issues. There's also the monthly dataset, but the trial did not overlap the 1st of a month, so there are no edits to show. &mdash; MusikAnimal  talk  15:53, 27 November 2018 (UTC)
 * task approved. — xaosflux  Talk 20:31, 27 November 2018 (UTC)
 * The above discussion is preserved as an archive of the debate. Please do not modify it. To request review of this BRFA, please start a new section at WT:BRFA.