Documenting the Now develops tools and builds community practices that support the ethical collection, use, and preservation of social media content.
I’ve been talking with the folks at MassMine (I’m on their Advisory Board) about tools that can gather information off the web and I was pointed to the Documenting the Now project that is based at the University of Maryland and the University of Virginia with support from Mellon. DocNow have developed tools and services around documenting the “now” using social media. DocNow itself is an “appraisal” tool for twitter archiving. They then have a great catalog of twitter archives they and others have gathered which looks like it would be great for teaching.
MassMine is at present a command-line tool that can gather different types of social media. They are building a web interface version that will make it easier to use and they are planning to connect it to Voyant so you can analyze results in Voyant. I’m looking forward to something easier to use than Python libraries.
Speaking of which, I found a TAGS (Twitter Archiving Google Sheet) which is a plug-in for Google Sheets that can scrape smaller amounts of Twitter. Another accessible tool is Octoparse that is designed to scrape different database driven web sites. It is commercial, but has a 14 day trial.
One of the impressive features of Documenting the Now project is that they are thinking about the ethics of scraping. They have a Social Labels set for people to indicate how data should be handled.
Part of what is interesting in the essay is how Salvador documents the different views about what such simulations were good for. SimRefinery was not a accurate simulation that would cover the complexity of the chemical engineering of a refinery; so what was it good for. Chevron apparently wanted something to help the staff who weren’t engineers to understand some of the connectiveness of a refinery – how problems in one area could impact others. Will Wright, the genius behind Maxis, didn’t think serious simulations were possible or something they wanted to do. He saw SimCity as a caricature that was fun. At best it might give people a “mental model” of the issues around city management. It was for that reason that MBS was a spin-off designed to contract with businesses that felt serious simulations were feasible and useful.
The Computer Literacy Project, on the other hand, is what a bunch of producers and civil servants at the BBC thought would be the best way to educate the nation about computing. I admit that it is a bit elitist to suggest we should laud this group of people for teaching the masses what they were incapable of seeking out on their own. But I can’t help but think they got it right. Lots of people first learned about computing using a BBC Micro, and many of these people went on to become successful software developers or game designers.
I’ve just discovered Two-Bit History (0b10), a series of long and thorough blog essays on the history of computing by Sinclair Target. One essay is on Codecademy vs. The BBC Micro. The essay gives the background of the BBC Computer Literacy Project that led the BBC to commission as suitable microcomputer, the BBC Micro. He uses this history to then compare the way the BBC literacy project taught a nation (the UK) computing to the way the Codeacademy does now. The BBC project comes out better as it doesn’t drop immediately into drop into programming without explaining, something the Codecademy does.
I should add that the early 1980s was a period when many constituencies developed their own computer systems, not just the BBC. In Ontario the Ministry of Education launched a process that led to the ICON which was used in Ontario schools in the mid to late 1980s.
Like many, I learned to program multimedia in HyperCard. I even ended up teaching it to faculty and teachers at the University of Toronto. It was a great starting development environment with a mix of graphical tools, hypertext tools and a verbose programming language. It’s only (and major) flaw was that it wasn’t designed to create networked information. HyperCard Stacks has to be passed around on disks. The web made possible a networked hypertext environment that solved the distribution problems of the 1980s. One wonders why Apple (or someone else) doesn’t bring it back in an updated and networked form. I guess that is what the Internet Archive is doing.
The goal of Lincs is to create a shared linked data store that humanities projects can draw on and contribute to. This would let us link our digital resources in ways that create new intellectual connections and that allow us to reason about linked data.
Kieji Amano deserves a lot of credit for putting together the largest Replaying Japan programme ever. The folks at the Ritsumeikan Center for Games Studies should also be thanked for organizing the facilities for both conferences. They have established themselves as leaders in Japan in the field.
I gave two papers:
“The End of Pachinko” (given with Amano) looked at the decline of pachinko and traditional forms of gambling in the face of the legalization of casinos. It looked at different types of ends, like the ends of machines.
“Work Culture in Early Japanese Game Development” (with Amano, Okabe, Ly and Whistance-Smith) used text analysis of Szczepaniak’ series of interviews, the Untold History of Japanese Game Developers, as a starting point to look at themes like stress and gender.
The quality of the papers in both conferences was very high. I expect this of DiGRA, but it was great to see that Replaying Japan, which is more inclusive, it getting better and better. I was particularly impressed by some of the papers by our Japanese colleagues like a paper delivered by Kobayashi on the “Early History of Hobbyist Production Filed of Video Games and its Effect on Game Industries in Japan.” This was rich with historical evidence. Another great one was “Researching AI technologies in 80’s Japanese Game Industry” delivered by Miyake who is involved in some very interesting preservation projects.
The question I want to explore today is this: what do we do about distant reading, now that we know that Franco Moretti, the man who coined the phrase “distant reading,” and who remains its most famous exemplar, is among the men named as a result of the #MeToo movement.
Lauren Klein has posted an important blog entry on Distant Reading after Moretti. This essay is based on a talk delivered at the 2018 MLA convention for a panel on Varieties of Digital Humanities. Klein asks about distant reading and whether it shelters sexual harassment in some way. She asks us to put not just the persons, but the structures of distant reading and the digital humanities under investigation. She suggests that it is “not a coincidence that distant reading does not deal well with gender, or with sexuality, or with race.” One might go further and ask if the same isn’t true of the digital humanities in general or the humanities, for that matter. Klein then suggests some thing we can do about it:
We need more accessible corpora that better represent the varieties of human experience.
We need to question our models and ask about what is assumed or hidden.