User:Ɛs Huvər/sandbox

Blah blah blah

Another paragraph

More text

Call details

 * Date: 2021-02-09
 * Topic: Mapping the Scottish Reformation Project
 * Presenters: Michelle (Mikki) Brock and Christopher (Chris) Langley

Presentation materials

 * Blog and project information: Mapping the Scottish Reformation Blog – A database of the Scottish clergy, 1560 to 1689
 * Map interface
 * Twitter

Presentation

 * Intro to the Scottish Reformation
 * Trying to trace the careers of religious figures in Scotland in the 16th & 17th centuries
 * Examples of ministers who had several relocations and moved after short periods of time, and ministers who were relatively sedentary and had long postings in their parishes. Characteristics of mobility vs. characteristics of longevity in their career.
 * Ministers and their family had outsized religious, community, and intellectual weight. Big archival footprint. Useful for historians and genealogy
 * How to trace the careers of these important individuals?
 * Create one of the largest databases of Protestant thinkers in history
 * Chart growth, movement, and networks of these individuals
 * “Biggish” data to reimagine what we understand about the Scottish Reformation
 * Starts with: the inadequacy of sources they had about “the clergy as cohort.” Difficulty with the source material, bits of information scattered across manuscript records. Printed materials often hold errors.
 * For any digital humanities project, think about the specific questions you have and what data you need to answer those questions
 * Manuscript records are also inaccessible (geographically) and often in terms of readability for those not familiar with the paleography. Incredibly difficult to find the information you want.
 * Rethinking how to present historical documents and the data within them
 * Test case/proof of concept to focus on the Synod of Lothian and Tweeddale, 10,000 pages of manuscript, gathering 874 full clerical journeys across 116 parishes, including “significant events” for 500, as well as 450 clerical wives.
 * Hope to write the clerical wives back into history as community figures within the Scottish Reformation
 * Chris on the website, map interface
 * 5 map views, explore different aspects of the subjects lives with different views
 * Multiple search and filter options.
 * Different ways to visualize the ministry: see individual detail along with the aggregate view.
 * Start to answer questions like “what does the average ministry look like?”
 * ZOOM button coming soon!
 * Journeys page:
 * Visualize travels over someone’s life.
 * Search by minister or residence, date filters
 * What patterns will emerge? Regional patterns?
 * Education
 * Rich data available in popups, will be downloadable in next version.
 * Can work out kilometers traveled
 * More aggregations coming in 1.1 release as well
 * Wikidata’s Role in the Project
 * “Late to the party but started dancing quickly.” - Chris
 * Scalability, capturing complex information quite easily
 * Using qualifiers heavily and multilayered references. Need data that can be aggregated into an academic style reference.
 * Example from slides: John Oswald
 * Wikidata allows prototyping data models really easily, using the query service
 * Export process
 * WQS export to JSON. Queries built around the ways they’ve structured data.
 * Some assumptions that their data is obscure enough within Wikidata that it’s easily identifiable as theirs.
 * Manually check for auto-generated errors (occasionally get character errors)
 * Website run on static snapshot, not a live feed. PM benefits: better editorial control, can batch snapshots at the end of a project stage. Though website doesn’t grow automatically with voluntary work.


 * Next steps
 * Expand to all of Scotland, 50,000 more manuscript pages
 * Additional help from post-doc, training undergraduates as well to enter Wikidata items

Questions

 * JL: Curious, did these folks in the Scottish reformation adopt or write a confession? Is there a 'paper trail' for that kind of thing in the archives?
 * Written record varies widely by minister. Working on how to work with all this qualitative data and connect to other data sources
 * V: How did you train the team in wikidata query service? Is this something you are covering later? I am also interested in more details about working with JSON
 * Help from Univ Edinburgh Wikimedian in Residence, had worked on the Scottish witchcraft
 * Had tutorials available, have begun working with QuickStatements
 * Trial and error to learn how to use the Query Service. Lots of help with SPARQL from community members
 * MS: Did you try using Wikidata ShapeEx and Entity schemas to get quality?
 * No, bespoke. Had worked with OpenRefine, but moved to QuickStatements
 * KM: How did you first learn of wikidata/base, and who comprises the project team?
 * Mikki was a fellow at Univ. Edinburgh. Went to a talk about the Scottish witches project. Had been gathering info in google sheets, so the talk was inspiring in terms of “what to do” with all this data. Quickly became immersed, scouring YouTube for tutorials.
 * Expanding advisory board to include more Wikidata and tech expertise, in addition to Scottish historians and Scottish National Records
 * Metadata Librarian at Washington and Lee, have added tech team
 * JV: I'm curious about the editorial workflow. How to you manually inspect changes? Some diff tool?
 * Looking at errors on the maps, find them within the visualizations
 * Getting feedback from advisory board historians, users on Twitter
 * V: What are you using for your website CMS - example Webflow or Wordpress?
 * Wordpress, JS fetch to JSON, eventually a bespoke CMS
 * MS: Is the source code on GITHUB if yes add some GITHUB topics so we can find things like this easy?
 * Eventually, yes. Blogging about technical aspects of MSR as well.

Call details

 * Date: 2020-09-22
 * Topic: Lightning talks on Wikidata projects
 * Presenters: Rod Page (University of Glasgow), Nicole Coleman and Jeremy Nelson (Stanford Libraries), Steve Baskauf (Jean & Alexander Heard Libraries, Vanderbilt University), Heidi Lerner (Stanford University Libraries), Crystal Clements (University of Washington Libraries)

Presentation materials

 * Agenda
 * Slides and Other Materials
 * Rod Page
 * Slides
 * Demo - ALEC (a simple Wikidata viewer)
 * Code to populate Wikidata
 * Nicole Coleman and Jeremy Nelson
 * Slides
 * Demo - Stanford ETD Abstract Similarity
 * Demo - Stanford FAST Assignment Biology ETDs
 * Steve Baskauf
 * Slides
 * Crystal Clements
 * Slides
 * Project Page
 * Project Dashboard

Call details

 * Date: 2021-01-26
 * Topic: Wikibase, the Integrated Authority File (GND) project at German National Library, and the WikiLibrary Manifesto
 * Presenters: Georgina Burnett (Wikimedia Deutschland), Barbara Fischer (DNB)

Presentation materials

 * Agenda
 * Slides

Wikibase - Georgina Burnett (Wikimedia Deutschland)

 * Wikibase from a Wikidata perspective:
 * Software underlying Wikidata
 * Developed considering needs of Wikidata community in mind
 * Lots of functionalities appreciated in Wikidata carry over to Wikibase instances
 * More recently, this has changed--Wikibase has become a product in its own right, with dedicated staff developing it
 * Benefits, largely overlapping with Wikidata:
 * rich UI
 * allows data models specific to institutional needs
 * SPARQL query service
 * Data can be exported in a number of formats
 * When might I look to Wikibase?
 * Need for multilingual access
 * (vs Wikidata) if data collection is beyond the scope of Wikidata, either in size or degree of specialisation
 * Vision: linking Wikibase instances together
 * Current ecosystem at a glance: see pie chart in slides
 * Lots of GLAM institutions
 * Resources from the presentation
 * Wikibase mailing list and telegram group
 * WikibaseHQ on twitter
 * Wikibase community user group

Questions/discussions

 * interesting that such a large portion of WB instances are GLAM, since GLAM = only 25-30% of wikipedia articles!
 * interested in more info about the scope of wikidata - if specialised/granular info is stored in sep wikibase instances, can that metadata be integrated into a global search?
 * Georgina will get back to group on:
 * Pie chart N
 * Feasibility of federation, given highly specialised nature of many Wikibase instances
 * there are still/currently challenges populating wikibase instances with properties - what is the current status of efforts to make it easier to copy wikidata properties over to a wikibase instance? are there tools being developed?
 * Yes. Also see: 2020 wikibase roadmap
 * For technical questions, contact Georgina (georgina.burnett AT wikimedia.de). She will find answers and get back to you!

GND project at the German National Library (DNB) - Barbara Fischer (DNB)

 * GND (integrated authority files) - 300 properties and their qualifiers
 * objectives/exploring/questions:
 * how to enable synchronization, between primary/main copy in CBS and the wikibase instance
 * how to empower users - make it easier for them to add new items/datasets to GND
 * observations so far - notes for improvement:
 * tech/infra: still too complex for new adopters to get WB instance installed via Docker,
 * WB is not yet a complete software architecture -- bc it was developed from mediawiki to serve needs of wikidata, this is still a limitation. need a more modern, internally-consistent architecture
 * functionality: gaps bt the specific needs of institutions, and the solutions from WB so far:
 * bulk import - DNB is trying to import 9 million items, and also the properties
 * export - needs to meet specific needs of institution
 * very aware of: being so dependent on WB development, and the direction/decisions of its core functions, but without a real say in this process/priorities
 * these organizations who are using WB lack the IT personnel/expertise to use/support this software -- so feels like they are relying on a very small pool of “experts”/contracting from the same small group of folks!
 * as clients/institutions still have a diff mindset from some of the smaller groups that are WB users. hard to fit in to that system, and causes org to wonder if we can be quick enough to keep up with the community?
 * WikiLibrary Manifesto
 * 2 parts bc they wanted to make a commitment to implementing specific actions/not just high-level ideas https://www.wikimedia.de/the-wikilibrary-manifesto/
 * 2 parts:
 * principles
 * measures

Questions/discussions

 * will bibframe ontology play a part in your transition to RDA ontology as well?
 * No, at least in the beginning stages.
 * is the mapping that DNB has created shareable?
 * yes! and it will be shared. covid has delayed the work
 * is the new UI for editor an implementation of ShEx?
 * Will be designed to be as interactive as possible (lots of questions given to user to help mould/define the item), so that people who have little context for current rules can still use it
 * As things stand now--there’s a lot of PDF reading involved to describe items
 * (vs Wikidata) Wikidata tries to make items as rich as possible--lots of context. With authority files, the focus is more on getting enough information to clearly distinguish different entities.
 * at the most recent SWIB conference - Bernhard Krabina argued for Semantic Wikimedia as a better fit for many library use cases than wikibase - has GND discussed this?
 * yes, we are monitoring this discussion, but focused on WB for now
 * IT infra and small pool of staffing available for these kinds of projects - what kind of paths are there?


 * there are already many people with expertise in wikidata and associated skills/tools - SPARQL queries, and the tools available to wikidata - there is a large volunteer community around this, and many contractors available. But when seeking staffing for WB project specifically, very few applicants! have also heard similar stories from other orgs trying to implement WB.
 * WMDE has set up a page to document contractors with WB expertise -- currently so few people; it feels like an exclusive club!
 * does the lack of staffing/expertise lead more orgs to just try things on their own, and so potentially = greater inconsistency, lack of shared standards/processes?
 * Lots of bootstrapping with whatever existing technical knowledge, by independent orgs putting together their own WB instances.
 * As part of effort to lower barriers to entry, WMDE recently brought on technical writer (Dan Shick; presented during November 2020 LD4 call on community documentation)

Call details

 * Date: 2021-01-12
 * Topic: Using QuickStatements for bulk uploading archived website data to Wikidata
 * Presenter: Peter Chan
 * Link to original agenda with link to recording: https://docs.google.com/document/d/14KHUNQoTo5eX6iclLJdVSnOyPLNyPYQA4V854Mc7iY4/edit

Presentation materials

 * Working with archived websites doc
 * QuickStatements slides

Publish Archived Website Metadata as LOD

 * Collections of archived website at Stanford in SearchWorks catalog
 * Still not linked data
 * https://searchworks.stanford.edu/view/mk656nf8485
 * Collection level item in Wikidata: https://www.wikidata.org/wiki/Q98908071
 * Item for website in Wikidata: https://www.wikidata.org/wiki/Q99851990
 * Modeling issues
 * Archive URL is qualifier and subproperty of URL
 * Has constraints on formatting as well
 * When use collection property has constraint suggestion of location--physical location not virtual or Internet
 * Options might be allowing a virtual location with collection or adding virtual collection as a property

Discussion

 * K1: this approach makes sense within Wikidata's current use of "archive URL," but it makes me feel like "archive URL" should be a top-level property in Wikidata (as opposed to a qualifier as it is now), and you could add collection as a qualifier to a given archive URL
 * Website not in collection of web archive, but archive URL in collection
 * D1 agrees
 * E1: I agree with that too. It’s the archived website that’s being cataloged; not the website itself.
 * K1: Maybe URL of item should be web archived URL and instance of something like archived website
 * G1: asking for more granularity with instance? "archived website" verse "live website" both are websites though
 * K1: I'm still a little unsure of that it would make sense to say that the website is in the collection of the fugitive us agencies web archive. I feel like it's not that the website is in the collection of the web archive except the archive URL is in the collection of the web. Item that gets created or if a property could be added to a website that is the archive URL. is there a difference in terms of wiki data to say something as the instance of a live website for the archive website and I know this is something that my colleagues at the frick
 * M1 and S1, et al. at Frick worked on BIBFRAME model for web archives that might help share a data model for Wikidata
 * M1: When we were thinking about creating the profile for web archives, it became clear that in order to create the distinctions that we needed regarding live sites versus archived sites (in our records we link to both when available), we had to parse them out into two separate instances--in part that was because the software didn't allow us to distinguish between the two with things like collection date or or the actual URL itself, but with Wikidata if you create an item or something related to the live site, it's different from the archive site in that the archive site does have that collection that its associated with--you can include the dates in terms of crawls, etc. that’s a bit more granular that obviously doesn't apply to a live site.  Those were the distinctions that we were making. Making the distinction between the two and then potentially linking them as pointing to the same resource would help alleviate confusion about where the archive site fits within a collection
 * E1: why didn’t you use archives at property for the web archives?
 * https://www.wikidata.org/wiki/Property:P485
 * Not sure that the way that you have it modeled necessarily makes any sense to use it because archives at usually connects a person to an institution
 * it might be interesting to use if you're collecting particular agencies’ websites and add that information to the agency's wikidata page to say that archives are at Stanford's web archives--connects people who are interested in the particular entity that you are describing to give people an alternative point of access into those web archives
 * Peter had considered, but wasn’t sure if could be used in this context
 * J1: In my experience, Wikidata property constraints get adjusted fairly often, and warnings may therefore appear and disappear from time to time. Usually (and in the best case) this is due to community feedback on how properties are being, and should be, used. But it can be a bit bewildering to folks who edit Wikidata repeatedly.
 * (Also in my experience, if an unprecedented Wikidata warning shows up that seems completely unreasonably, it’s often gone within a day or two after the constraint that raised it gets reverted.)
 * H1: this is a really interesting query! What was the predicate for "main subject" and does "Federal Government" have a subject entity/linked to an LCSH or other controlled vocabulary? (Apologies if I missed that and it's already in the article)
 * Predicate: P921
 * H1: I do see this in the article "Linked the collection to 3 main subjects which have identifiers from authoritative agencies such as the Library of Congress" so that answers that last question
 * D2: Is your plan to upload all of Stanford’s archived websites to Wikidata? This would be a great presentation to show Archive-It users if they are hosting any virtual conferences. Perhaps Society of American Archivists (SAA) also?
 * A1: suggests property collection creator, instance of digital collection (Q60474998), instance of web archive (Q30047053)
 * would just use instance of: web archive; “archived website with the title in the label”--would just say “archived website” and use “title” property to record the title
 * E1: Also +1 to A1’s comment to add the instance of web archive to help with retrieval
 * suggest the label is very odd: "archived website with the title in the label". I would just say "archived website" and use the "title" property to record the title

Introduction to QuickStatements

 * See slides

Helpful Resources About QuickStatements
Demo
 * Spreadsheet

Questions
Share your tips and tricks for using QuickStatements here
 * S2: How did you convert your data to the QS syntax (V1 & CSV )?
 * CSV syntax was easy with this data--simply changed headers of the columns
 * V1 more involved. For working with large datasets could try this tool: https://ash-dev.toolforge.org/wdutils/csv2quickstatements.php
 * K1: can quickstatements be used to bulk update/change existing data in wikidata or is it mostly for creating new items/properties?
 * Can do both
 * Add a statement with V1 Command syntax: https://www.wikidata.org/wiki/Help:QuickStatements#Add_simple_statement
 * Create an  item with V1 Command syntax: https://www.wikidata.org/wiki/Help:QuickStatements#Item_creation
 * Add a statement with CSV syntax:
 * P1: Is there any way quick statements will warn you if you are creating a duplicate?
 * Best to first reconcile your data with Wikidata
 * Creating new Wikidata items with OpenRefine and Quickstatements
 * Reconciliation with Open Refine Affinity Group Call
 * 2020-10-22
 * Use “run” to load statements. “Run in background” may create duplicates: https://www.wikidata.org/wiki/Help:QuickStatements#Using_QuickStatements_version_2_in_batch_mode
 * If you do realize you created a duplicate, you can merge them
 * Existing statements with an exact match (property and value) will not be added again; however additional references might be added to the statement.