Jump to content

User talk:Magnus Manske/Archive 8

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia
Archive 5Archive 6Archive 7Archive 8Archive 9Archive 10Archive 11

Facto Post – Issue 15 – 21 August 2018

Facto Post – Issue 15 – 21 August 2018

The Editor is Charles Matthews, for ContentMine. Please leave feedback for him, on his User talk page.
To subscribe to Facto Post go to Wikipedia:Facto Post mailing list. For the ways to unsubscribe, see the footer.

Neglected diseases
Anti-parasitic drugs being distributed in Côte d'Ivoire
What's a Neglected Disease?, ScienceSource video

To grasp the nettle, there are rare diseases, there are tropical diseases and then there are "neglected diseases". Evidently a rare enough disease is likely to be neglected, but neglected disease these days means a disease not rare, but tropical, and most often infectious or parasitic. Rare diseases as a group are dominated, in contrast, by genetic diseases.

A major aspect of neglect is found in tracking drug discovery. Orphan drugs are those developed to treat rare diseases (rare enough not to have market-driven research), but there is some overlap in practice with the WHO's neglected diseases, where snakebite, a "neglected public health issue", is on the list.

From an encyclopedic point of view, lack of research also may mean lack of high-quality references: the core medical literature differs from primary research, since it operates by aggregating trials. This bibliographic deficit clearly hinders Wikipedia's mission. The ScienceSource project is currently addressing this issue, on Wikidata. Its Wikidata focus list at WD:SSFL is trying to ensure that neglect does not turn into bias in its selection of science papers.

Links

If you wish to receive no further issues of Facto Post, please remove your name from our mailing list. Alternatively, to opt out of all massmessage mailings, you may add Category:Wikipedians who opt out of message delivery to your user talk page.
Newsletter delivered by MediaWiki message delivery

MediaWiki message delivery (talk) 13:23, 21 August 2018 (UTC)

CommonsDelinker localized namespace issue

Hello.

I opened issue 33 yesterday for CommonsDelinker not handling local namespace correctly at the hindi wikipedia. I've since discovered that the issue is quite old. Issue 20 was reported in May 2017 with reference to the Bengali Wikipedia. Searching through the contributions of User:CommonsDelinker at hi-wp, I've discovered that the issue originated between 18th and 20th January 2015. The issue was not present on 18 Jan 15, but was present on 20 Jan 15. I'm guessing (looking at the commit history of Commons Delinquent) that the rewrite was probably deployed during this time, meaning that the issue has been present since the beginning! Also, clearly the issue is cross-wiki.

Kindly look into the issue ASAP. Thanks--Siddhartha Ghai (talk) 05:32, 3 September 2018 (UTC)

Hello

Hello User:Magnus_Manske ~Inked Soul~ (talkcontribs) 07:00, 15 September 2018 (UTC)

GLAMorgan is cool!

I'm impressed by this tool, which I found while researching the new metrics tool we're building for event organizers. You did a very nice job, for example, of laying out the complex file usage reports. Is there a user Help doc someplace? For example, I couldn't figure out whether there was a way to specify a date range beyond 1 month. Thanks! JMatazzoni (WMF) (talk) 17:12, 21 September 2018 (UTC)

Discussion invite

Hi! I am currently exploring some issues regarding CommonsDelinkerBot and a few other misc. things related to CAT:MISSFILE. I wanted to let you know that I have began a discussion here which you may be interested in contributing to. Cheers, Katniss May the odds be ever in your favor 18:27, 27 September 2018 (UTC)

Facto Post – Issue 16 – 30 September 2018

Facto Post – Issue 16 – 30 September 2018

The Editor is Charles Matthews, for ContentMine. Please leave feedback for him, on his User talk page.
To subscribe to Facto Post go to Wikipedia:Facto Post mailing list. For the ways to unsubscribe, see the footer.

The science publishing landscape

In an ideal world ... no, bear with your editor for just a minute ... there would be a format for scientific publishing online that was as much a standard as SI units are for the content. Likewise cataloguing publications would not be onerous, because part of the process would be to generate uniform metadata. Without claiming it could be the mythical free lunch, it might be reasonably be argued that sandwiches can be packaged much alike and have barcodes, whatever the fillings.

The best on offer, to stretch the metaphor, is the meal kit option, in the form of XML. Where scientific papers are delivered as XML downloads, you get all the ingredients ready to cook. But have to prepare the actual meal of slow food yourself. See Scholarly HTML for a recent pass at heading off XML with HTML, in other words in the native language of the Web.

The argument from real life is a traditional mixture of frictional forces, vested interests, and the classic irony of the principle of unripe time. On the other hand, discoverability actually diminishes with the prolific progress of science publishing. No, it really doesn't scale. Wikimedia as movement can do something in such cases. We know from open access, we grok the Web, we have our own horse in the HTML race, we have Wikidata and WikiJournal, and we have the chops to act.

Links

If you wish to receive no further issues of Facto Post, please remove your name from our mailing list. Alternatively, to opt out of all massmessage mailings, you may add Category:Wikipedians who opt out of message delivery to your user talk page.
Newsletter delivered by MediaWiki message delivery

MediaWiki message delivery (talk) 17:57, 30 September 2018 (UTC)

Missing Topics not work

Hi Magnus! Something wrong with excellent Missing Topics Tool - it starts but... not work correct (and display something about "Undefined variable: wrapper_tfc"). It began work bad yesterday as I remember. Try this for example. Thanks! -- Alexey Gustow (talk) 23:17, 4 October 2018 (UTC)

Thanks, fixed! --Magnus Manske (talk) 10:31, 5 October 2018 (UTC)

GLAMorous

Nabend Magnus,

das Tool listet nur noch ein Foto, mache ich irgendwas falsch? Hab es mit unterschiedlichen Benutzern versucht. --Ralf Roletschek (talk) 19:45, 4 October 2018 (UTC)

Ist das so richtig? --Magnus Manske (talk) 10:34, 5 October 2018 (UTC)
Nicht ganz, mit [x] show Details. Jetzt geht es wieder. Komisch, Hamsterschnupfen? Danke für das sehr hilfreiche Tool! --Ralf Roletschek (talk) 16:13, 5 October 2018 (UTC)

A barnstar for you!

The Original Barnstar
For your wonderful contributions -[[User:Akbarali]] ([[Talk:Akbarali|Talk]]) (talk) 02:26, 6 October 2018 (UTC)

Deep-out-of-sight doesn't work

Hey ;) Sorry to bother you, but the tool doesn't work. The interface is there but no results are displayed, for example: here. Many editors on plwiki find this tool very useful so it'd be great if the tool started working again ;-) Thanks in advance, Tufor (talk) 19:46, 10 October 2018 (UTC)

Fixed. --Magnus Manske (talk) 09:45, 15 October 2018 (UTC)

Not in other language tool

You know what would be super cool! If it was possible to order the results here

By the number of pageviews in English each received in the prior month. That would help translators prioritize which they should translate first :-) Doc James (talk · contribs · email) 15:41, 15 October 2018 (UTC)

I agree that would be cool! Sadly, getting those numbers is slow, essentially, one http request per article (per month). So, sorting thousands of pages would mean thousands of such requests. (that is the way WMF wants us to access page view stats. Not much I can do about it.) --Magnus Manske (talk) 13:40, 19 October 2018 (UTC)
We have this tool that gives you all the page-views in one click.[1] You actually need to run it twice to get it to work. Could we use that? You just need to set the depth to "2" and type in the category "RTT". Would only need to run it once a month. Doc James (talk · contribs · email)

Precious anniversay

A year ago ...
"I am not here for rules" software
... you were recipient
no. 1753 of Precious,
a prize of QAI!

--Gerda Arendt (talk) 08:06, 21 October 2018 (UTC)

Facto Post – Issue 17 – 29 October 2018

Facto Post – Issue 17 – 29 October 2018

The Editor is Charles Matthews, for ContentMine. Please leave feedback for him, on his User talk page.
To subscribe to Facto Post go to Wikipedia:Facto Post mailing list. For the ways to unsubscribe, see the footer.

Wikidata imaged

Around 2.7 million Wikidata items have an illustrative image. These files, you might say, are Wikimedia's stock images, and if the number is large, it is still only 5% or so of items that have one. All such images are taken from Wikimedia Commons, which has 50 million media files. One key issue is how to expand the stock.

Indeed, there is a tool. WD-FIST exploits the fact that each Wikipedia is differently illustrated, mostly with images from Commons but also with fair use images. An item that has sitelinks but no illustrative image can be tested to see if the linked wikis have a suitable one. This works well for a volunteer who wants to add images at a reasonable scale, and a small amount of SPARQL knowledge goes a long way in producing checklists.

Gran Teatro, Cáceres, Spain, at night

It should be noted, though, that there are currently 53 Wikidata properties that link to Commons, of which P18 for the basic image is just one. WD-FIST prompts the user to add signatures, plaques, pictures of graves and so on. There are a couple of hundred monograms, mostly of historical figures, and this query allows you to view all of them. commons:Category:Monograms and its subcategories provide rich scope for adding more.

And so it is generally. The list of properties linking to Commons does contain a few that concern video and audio files, and rather more for maps. But it contains gems such as P3451 for "nighttime view". Over 1000 of those on Wikidata, but as for so much else, there could be yet more.

Go on. Today is Wikidata's birthday. An illustrative image is always an acceptable gift, so why not add one? You can follow these easy steps: (i) log in at https://tools.wmflabs.org/widar/, (ii) paste the Petscan ID 6263583 into https://tools.wmflabs.org/fist/wdfist/ and click run, and (iii) just add cake.

Birthday logo
Links

If you wish to receive no further issues of Facto Post, please remove your name from our mailing list. Alternatively, to opt out of all massmessage mailings, you may add Category:Wikipedians who opt out of message delivery to your user talk page.
Newsletter delivered by MediaWiki message delivery

MediaWiki message delivery (talk) 15:01, 29 October 2018 (UTC)

CommonsHelper PHP error

Hi Magnus, I seem to be getting an error message when I try to use CommonsHelper. The error message that appears when I click "Get text" is:

Fatal error: Class 'Template' not found in /data/project/commonshelper/public_html/index.php on line 317

I was trying to directly transfer en:File:José Luis Villarreal Jacksonville Armada FC.png. I tried re-authorizing myself with OAuth Uploader, but that didn't fix it. Could you please take a look at this? Thanks, IagoQnsi (talk) 20:01, 23 October 2018 (UTC)

Found the same today. Please help to solve it. --Arnd (talk) 17:16, 24 October 2018 (UTC)
Same problem for transfering from sr.wiki to Commons. --Miljan Simonović (talk) 19:48, 24 October 2018 (UTC)
This should be fixed now. --Magnus Manske (talk) 10:34, 31 October 2018 (UTC)

File:Pierre Bayle.jpg

Hello. There appear to be a problem with File:Pierre Bayle.jpg and File:François Poullain de La Barre.jpg. I guess the second one should be deleted? I thought you should be informed since you are the author of the first file. Sapphorain (talk) 16:01, 15 November 2018 (UTC)

ArbCom 2018 election voter message

Hello, Magnus Manske. Voting in the 2018 Arbitration Committee elections is now open until 23.59 on Sunday, 3 December. All users who registered an account before Sunday, 28 October 2018, made at least 150 mainspace edits before Thursday, 1 November 2018 and are not currently blocked are eligible to vote. Users with alternate accounts may only vote once.

The Arbitration Committee is the panel of editors responsible for conducting the Wikipedia arbitration process. It has the authority to impose binding solutions to disputes between editors, primarily for serious conduct disputes the community has been unable to resolve. This includes the authority to impose site bans, topic bans, editing restrictions, and other measures needed to maintain our editing environment. The arbitration policy describes the Committee's roles and responsibilities in greater detail.

If you wish to participate in the 2018 election, please review the candidates and submit your choices on the voting page. MediaWiki message delivery (talk) 18:42, 19 November 2018 (UTC)

Wiki Tool prepares Infobo

Enable THIS TOOL FOR SINDHI WIKIPEDIA, So it can also create infoboxes on Sindhi Wikipedia. Thanks JogiAsad  Talk 02:45, 25 November 2018 (UTC)

Possible to delete stopped SourceMD batch?

Can one delete a stopped batch? I have no expectation of restarting it, so I'd like to "make it go away." My reason is simple, I don't want to see it in my batch list. Apologies if this is a capability that I should already know how to do. -Trilotat (talk) 14:08, 29 November 2018 (UTC)

Facto Post – Issue 18 – 30 November 2018

Facto Post – Issue 18 – 30 November 2018

The Editor is Charles Matthews, for ContentMine. Please leave feedback for him, on his User talk page.
To subscribe to Facto Post go to Wikipedia:Facto Post mailing list. For the ways to unsubscribe, see the footer.

WikiCite issue

GLAM ♥ data — what is a gallery, library, archive or museum without a catalogue? It follows that Wikidata must love librarians. Bibliography supports students and researchers in any topic, but open and machine-readable bibliographic data even more so, outside the silo. Cue the WikiCite initiative, which was meeting in conference this week, in the Bay Area of California.

Wikidata training for librarians at WikiCite 2018

In fact there is a broad scope: "Open Knowledge Maps via SPARQL" and the "Sum of All Welsh Literature", identification of research outputs, Library.Link Network and Bibframe 2.0, OSCAR and LUCINDA (who they?), OCLC and Scholia, all these co-exist on the agenda. Certainly more library science is coming Wikidata's way. That poses the question about the other direction: is more Wikimedia technology advancing on libraries? Good point.

Wikimedians generally are not aware of the tech background that can be assumed, unless they are close to current training for librarians. A baseline definition is useful here: "bash, git and OpenRefine". Compare and contrast with pywikibot, GitHub and mix'n'match. Translation: scripting for automation, version control, data set matching and wrangling in the large, are on the agenda also for contemporary library work. Certainly there is some possible common ground here. Time to understand rather more about the motivations that operate in the library sector.

Links

Account creation is now open on the ScienceSource wiki, where you can see SPARQL visualisations of text mining.

If you wish to receive no further issues of Facto Post, please remove your name from our mailing list. Alternatively, to opt out of all massmessage mailings, you may add Category:Wikipedians who opt out of message delivery to your user talk page.
Newsletter delivered by MediaWiki message delivery

MediaWiki message delivery (talk) 11:20, 30 November 2018 (UTC)

Filtr in Glamorous

Hey, firstly many thanks for such great tool. We (in Wikimedia Czech Republic) use it for the evaluation of contributors activites and creating lins from Commons to Wikipedia. Would it be possible to add filter for selected category by Uploader and Date (from-to)? That would help us to do our work even better. Now if there are files from different users in category its hard to focus just on usage of those from one user.--Juandev (talk) 15:55, 3 December 2018 (UTC)

Lista_de_pinturas_de_Silva_Porto

Hello! I tried to create a List of paintings of the portuguese painter Silva Porto. I published at the PTWP the entry pt:Lista de pinturas de Silva Porto with the {{Lista de Wikidata}}. At the Wikidata query service I managed to produce the list (bottom) but don´t know how to export the list to Wikipedia. Can you help me, please? Thank you, JMdosPerais 00:25, 5 December 2018 (UTC)

  1. Lista de pinturas de Silva Porto
  2. defaultView:Table

SELECT ?image ?itemDescription ?data_de_publicação ?coleção ?localização ?itemlabel WHERE {

 ?item wdt:P31 wd:Q3305213.
 ?item wdt:P170 wd:Q610761.

OPTIONAL { ?item wdt:P18 ?image .} OPTIONAL { ?item wdt:P577 ?data_de_publicação.} OPTIONAL { ?item wdt:P195 ?coleção.} OPTIONAL { ?item wdt:P276 ?localização.} }

I have created a basic list there, feel free to change it! --Magnus Manske (talk) 11:02, 6 December 2018 (UTC)
Thank you! JMdosPerais 14:18, 6 December 2018 (UTC) — Preceding unsigned comment added by GualdimG (talkcontribs)

Rust; WikiBase scalability; etc

- Zazpot (talk) 16:56, 15 December 2018 (UTC)

Listeria Image gallery?

Hi Magnus, I am interested to know if listeria can be used to create am image gallery based on a query - rather than a list? or perhaps there is another tool which does this? Thanks! Jason.nlw (talk) 11:06, 18 December 2018 (UTC)

Catnap

(lovely name for a tool :)

I couldn't see it on bitbucket, so I'm submitting a simple but major bug report here:

In toolforge:Catnap, when we generate output from a category it works well, but the links back to re-search Catnap from each of the results gives a 404, because there's a stray catnap.php in the anchor target. I.e.

Hopefully easy to fix. Thanks for the nifty tool! Quiddity (talk) 04:52, 17 December 2018 (UTC)

Fixed. --Magnus Manske (talk) 11:41, 20 December 2018 (UTC)

Facto Post – Issue 19 – 27 December 2018

Facto Post – Issue 19 – 27 December 2018

The Editor is Charles Matthews, for ContentMine. Please leave feedback for him, on his User talk page.
To subscribe to Facto Post go to Wikipedia:Facto Post mailing list. For the ways to unsubscribe, see the footer.

Learning from Zotero

Zotero is free software for reference management by the Center for History and New Media: see Wikipedia:Citing sources with Zotero. It is also an active user community, and has broad-based language support.

Zotero logo

Besides the handiness of Zotero's warehousing of personal citation collections, the Zotero translator underlies the citoid service, at work behind the VisualEditor. Metadata from Wikidata can be imported into Zotero; and in the other direction the zotkat tool from the University of Mannheim allows Zotero bibliographies to be exported to Wikidata, by item creation. With an extra feature to add statements, that route could lead to much development of the focus list (P5008) tagging on Wikidata, by WikiProjects.

Zotero demo video

There is also a large-scale encyclopedic dimension here. The construction of Zotero translators is one facet of Web scraping that has a strong community and open source basis. In that it resembles the less formal mix'n'match import community, and growing networks around other approaches that can integrate datasets into Wikidata, such as the use of OpenRefine.

Looking ahead, the thirtieth birthday of the World Wide Web falls in 2019, and yet the ambition to make webpages routinely readable by machines can still seem an ever-retreating mirage. Wikidata should not only be helping Wikimedia integrate its projects, an ongoing process represented by Structured Data on Commons and lexemes. It should also be acting as a catalyst to bring scraping in from the cold, with institutional strengths as well as resourceful code.

Links

Diversitech, the latest ContentMine grant application to the Wikimedia Foundation, is in its community review stage until January 2.

If you wish to receive no further issues of Facto Post, please remove your name from our mailing list. Alternatively, to opt out of all massmessage mailings, you may add Category:Wikipedians who opt out of message delivery to your user talk page.
Newsletter delivered by MediaWiki message delivery

MediaWiki message delivery (talk) 19:08, 27 December 2018 (UTC)

QuickStatements - how to question

Magnus,

I tried to use

Q70784 P6216 Q19652 P1001 Q30 P459 Q59693547
Q70784 P6216 Q19652 P1001 Q60332278 P459 Q29940705

hoping to get

copyright status
Normal rank public domain edit
applies to jurisdiction countries with 100 years pma or shorter
determination method or standard 100 years or more after author(s) death
▼ 0 reference
+ add reference
Normal rank public domain edit
applies to jurisdiction United States of America
determination method or standard published more than 95 years ago
▼ 0 reference
+ add reference
+ add value

However I got

copyright status
Normal rank public domain edit
applies to jurisdiction countries with 100 years pma or shorter
United States of America
determination method or standard 100 years or more after author(s) death
published more than 95 years ago
▼ 0 reference
+ add reference
+ add value

instead. Is there any way to get the first result using QuickStatements? --Jarekt (talk) 16:54, 7 January 2019 (UTC)

Not at the moment. It's a known issue. --Magnus Manske (talk) 11:08, 9 January 2019 (UTC)
In case the resolution of this issue is involves new syntax, one temporary "workaround" using current syntax migh be to enable handling of redirected items. Currently if QuickStatements encounters any redirected statements, it reports error. However if it was able to follow redirects than we could use existing redirects, like public domain (Q15687061) -> public domain (Q19652), so that

Q70784 P6216 Q19652 P1001 Q30 P459 Q59693547
Q70784 P6216 Q15687061 P1001 Q60332278 P459 Q29940705
could have been used to get the first option. Thare might be a better solution, but this was one of workarounds I tried. --Jarekt (talk) 17:18, 11 January 2019 (UTC)

QuickStatements - assigning a retrieval date

I've tried using

Q13220521 P846 "8284185" S248 Q1531570 S813 "+2019-01-05T00:00:00Z/9"

with and without double quotes around the date&time value, but it doesn't appear in the item's property-reference dropdown (Abacetus basilewskyi (Q13220521)). I also tried simpler formats "5 January 2019" & "2019 January 5", for good measure, but to no avail. Do you know how to fix this?   ~ Tom.Reding (talkdgaf)  21:00, 5 January 2019 (UTC)

The problem could be your use of "year precision" ("/9") with a specific day. Try "/11" for day precision. --Magnus Manske (talk) 10:42, 7 January 2019 (UTC)
Thanks, I just tried
Q13220521 P846 "8284185" S248 Q1531570 S813 "+2019-01-05T00:00:00Z/11"
and the result was the same, unfortunately.   ~ Tom.Reding (talkdgaf)  17:44, 13 January 2019 (UTC)

Distributed Game removing constraints

See this edit to FloraBase ID (P3101) that removed Wikidata item of this property (P1629), which formerly fulfilled an inverse constraint at FloraBase (Q5460267).   ~ Tom.Reding (talkdgaf)  16:07, 14 January 2019 (UTC)

"ERROR: Not logged into WiDaR!"

I can't seem to get the OAuth to stick.

I keep seeing "ERROR: Not logged into WiDaR!" when I try and import a new catalog (from https://tools.wmflabs.org/mix-n-match/import.php).

Is that related to the jsonp breaking change? — Preceding unsigned comment added by ATL EAC ENTERER (talkcontribs) 02:41, 17 January 2019 (UTC)

I found a better error:

Something went wrong with your authentication: Not authorized (bad API response [isAuthOK]: {"error":{"code":"mwoauth-invalid-authorization-invalid-user","info":"The authorization headers in your request are for a user that does not exist here","*":"See https:\/\/commons.wikimedia.org\/w\/api.php for API usage. Subscribe to the mediawiki-api-announce mailing list at <https:\/\/lists.wikimedia.org\/mailman\/listinfo\/mediawiki-api-announce> for notice of API deprecations and breaking changes."},"servedby":"mw1340"})

when I tried to log in from https://tools.wmflabs.org/tooltranslate/#tool=10&languages=en ATL EAC ENTERER (talk) 19:53, 17 January 2019 (UTC)

alphabetical order by the bot

Should´t the word "Évora" be alphabetically sorted before any word beginning with F? In another way, shouldn´t any letter with a signal be treated, for alphabetically order, as without the signal? I notice a case were de Listeria Bot places the word "Évora" below/after "Viseu", beginning with A..., for instance in w:pt:Lista de Castelos de Portugal. Greetings, JMdosPerais 22:41, 25 December 2018 (UTC)

Happy New Year! JMdosPerais 10:49, 1 January 2019 (UTC)
I´m sorry Magnus, but this "bug" is not yet solved. I´m not able to do so, and don´t know who can. Would you send this question to the right place, or the right people? Thank you. Greetings, JMdosPerais 21:56, 21 January 2019 (UTC)

CommonsDelinker making broken edits in Wikisource

Hi,

After deletion of a file as duplicate and redirect creation in Wikimedia Commons, Commons delinker made edits in plwikisource that are broken: eg. here it changes not only links to files but also links to index page (in the Index: namespace) inside the <pages ... > tag and provided as a template parameter ("strona indeksu="). Well, the index page name should be the same as the name of the DjVu file (and so names of pages in the Page: namespace - hundreds of them usually - which are also based on index/file name). If the bot cannot made the changes correctly (renaming also Index/Page pages or not changing links to them) it would be better if the bot refrained from any change in Wikisources that is related to a DjVu/PDF file rename. This is unlikely that such renames are performed correctly without assistance of a Wikisource user. And the way how CommonsDelinker makes such "fixes" now is disruptive. Ankry (talk) 13:45, 28 January 2019 (UTC)

Facto Post – Issue 20 – 31 January 2019

Facto Post – Issue 20 – 31 January 2019

The Editor is Charles Matthews, for ContentMine. Please leave feedback for him, on his User talk page.
To subscribe to Facto Post go to Wikipedia:Facto Post mailing list. For the ways to unsubscribe, see the footer.

Everything flows (and certainly data does)

Recently Jimmy Wales has made the point that computer home assistants take much of their data from Wikipedia, one way or another. So as well as getting Spotify to play Frosty the Snowman for you, they may be able to answer the question "is the Pope Catholic?" Possibly by asking for disambiguation (Coptic?).

Amazon Echo device using the Amazon Alexa service in voice search showdown with the Google rival on an Android phone

Headlines about data breaches are now familiar, but the unannounced circulation of information raises other issues. One of those is Gresham's law stated as "bad data drives out good". Wikipedia and now Wikidata have been criticised on related grounds: what if their content, unattributed, is taken to have a higher standing than Wikimedians themselves would grant it? See Wikiquote on a misattribution to Bismarck for the usual quip about "law and sausages", and why one shouldn't watch them in the making.

Wikipedia has now turned 18, so should act like as adult, as well as being treated like one. The Web itself turns 30 some time between March and November this year, per Tim Berners-Lee. If the Knowledge Graph by Google exemplifies Heraclitean Web technology gaining authority, contra GIGO, Wikimedians still have a role in its critique. But not just with the teenage skill of detecting phoniness.

There is more to beating Gresham than exposing the factoid and urban myth, where WP:V does do a great job. Placeholders must be detected, and working with Wikidata is a good way to understand how having one statement as data can blind us to replacing it by a more accurate one. An example that is important to open access is that, firstly, the term itself needs considerable unpacking, because just being able to read material online is a poor relation of "open"; and secondly, trying to get Creative Commons license information into Wikidata shows up issues with classes of license (such as CC-BY) standing for the actual license in major repositories. Detailed investigation shows that "everything flows" exacerbates the issue. But Wikidata can solve it.

Links

If you wish to receive no further issues of Facto Post, please remove your name from our mailing list. Alternatively, to opt out of all massmessage mailings, you may add Category:Wikipedians who opt out of message delivery to your user talk page.
Newsletter delivered by MediaWiki message delivery

MediaWiki message delivery (talk) 10:53, 31 January 2019 (UTC)

8th ISCB Wikipedia Competition: a reminder

Listeriabot

Listeriabot has gone 500 Internal Server Error for the past day and a half. Already reported it to the talk page of Listeria, but it does not seem to be used much. Thanks, --Folengo (talk) 19:46, 16 February 2019 (UTC)

Facto Post – Issue 21 – 28 February 2019

Facto Post – Issue 21 – 28 February 2019

The Editor is Charles Matthews, for ContentMine. Please leave feedback for him, on his User talk page.
To subscribe to Facto Post go to Wikipedia:Facto Post mailing list. For the ways to unsubscribe, see the footer.

What is a systematic review?

Systematic reviews are basic building blocks of evidence-based medicine, surveys of existing literature devoted typically to a definite question that aim to bring out scientific conclusions. They are principled in a way Wikipedians can appreciate, taking a critical view of their sources.

PRISMA flow diagram for a systematic review

Ben Goldacre in 2014 wrote (link below) "[...] : the "information architecture" of evidence based medicine (if you can tolerate such a phrase) is a chaotic, ad hoc, poorly connected ecosystem of legacy projects. In some respects the whole show is still run on paper, like it's the 19th century." Is there a Wikidatan in the house? Wouldn't some machine-readable content that is structured data help?

File:Schittny, Facing East, 2011, Legacy Projects.jpg
2011 photograph by Bernard Schittny of the "Legacy Projects" group

Most likely it would, but the arcana of systematic reviews and how they add value would still need formal handling. The PRISMA standard dates from 2009, with an update started in 2018. The concerns there include the corpus of papers used: how selected and filtered? Now that Wikidata has a 20.9 million item bibliography, one can at least pose questions. Each systematic review is a tagging opportunity for a bibliography. Could that tagging be reproduced by a query, in principle? Can it even be second-guessed by a query (i.e. simulated by a protocol which translates into SPARQL)? Homing in on the arcana, do the inclusion and filtering criteria translate into metadata? At some level they must, but are these metadata explicitly expressed in the articles themselves? The answer to that is surely "no" at this point, but can TDM find them? Again "no", right now. Automatic identification doesn't just happen.

Actually these questions lack originality. It should be noted though that WP:MEDRS, the reliable sources guideline used here for health information, hinges on the assumption that the usefully systematic reviews of biomedical literature can be recognised. Its nutshell summary, normally the part of a guideline with the highest density of common sense, allows literature reviews in general validity, but WP:MEDASSESS qualifies that indication heavily. Process wonkery about systematic reviews definitely has merit.

Links

If you wish to receive no further issues of Facto Post, please remove your name from our mailing list. Alternatively, to opt out of all massmessage mailings, you may add Category:Wikipedians who opt out of message delivery to your user talk page.
Newsletter delivered by MediaWiki message delivery

MediaWiki message delivery (talk) 10:02, 28 February 2019 (UTC)

Listeriabot not listing all the item that SPARQL does

Hi Magnus.

I've written a simple SPARQL query that lists 60 results in Wikidata Query Service. However, when I add it to this list it only prints the first 3 elements. Do you know what I'm doing wrong? Thanks in advance.

YoaR (talk) 07:58, 12 March 2019 (UTC)

You used "links=local", that limits the rows to items that have an article in your local wikipedia. I have removed that line, it now shows all results. Check out the documentation if you want redlinks etc. --Magnus Manske (talk) 14:01, 12 March 2019 (UTC)

/missingtopics/

tools.wmflabs.org/missingtopics/, after some warnings, the error:

Fatal error: Uncaught exception 'Exception' in /data/project/magnustools/public_html/php/ToolforgeCommon.php:241

What is going wrong? --FriedhelmW (talk) 20:05, 24 March 2019 (UTC)

GLAMorous broken?

GLAMorous has not been working for me the last couple months. Is it broken, or am I doing something wrong? --Lasunncty (talk) 09:53, 13 March 2019 (UTC)

Works for me. Example? --Magnus Manske (talk) 17:18, 13 March 2019 (UTC)
I'm trying to get a list of things that need work, so I did this search which returned lots of errors. I reduced the depth to 5 and still got errors. I then reduced it to 1 and got 0 results. I know that is not correct because I can go to the first image in the first subcategory and see that it is indeed used in Wikipedia. Just now I tried depth 2 and it seems to work, so I guess it is not completely broken. --Lasunncty (talk) 09:04, 15 March 2019 (UTC)
I'm also running in some errors (using boh Firefox and Chrome on Windows7), when I do this query everything above "Details (top 1000 images)" looks OK, but below I only see the thumbnails of the 1000 images, so without the detailed file usage reports I used to see there. Best, --OlafJanssen (talk) 12:53, 22 March 2019 (UTC)


Hi, for me GLAmorous broken too. Please see my used images. With best regards, -- George Chernilevsky talk 20:10, 24 March 2019 (UTC)