Our CAUT/ACPPU Bulletin has a front-page story about the Lakehead Faculty Association grievance against Lakehead for moving everyone onto gmail. While the arbitrator decided to dismiss the grievance, I doubt this will go far – I suspect all sorts of bodies from Faculty Associations to student unions will start pressuring their univeristies to protect our privacy better. Here is a quote from the story.
In his decision, arbitrator Joseph Carrier acknowledged the university exposed its academic staff to greater danger because “…the likelihood of such incursions by U.S. authority into a private e-mail system (Lakehead’s own former system) was marginal compared to what might occur in the presence of the Google system.”
He also commented favourably on the opinion of Stephen Schulhofer, the faculty association’s expert witness and the Robert B. McKay Professor of Law at New York University.
“I am satisfied Professor Schulhofer’s opinion was valid and more than adequate to confirm that e-mail originating within Canada and coming within the jurisdiction of U.S. authorities would be open to surveillance by agencies of that country and, but for safeguards here, would expose the author to potential consequences of the U.S. antiterrorism legislation,” Carrier’s decision states.
Yet, the collective agreement language does not prevent the employer from endangering the privacy of LUFA members because the agreement does not specify the obligation to ensure “absolute privacy to faculty members,” Carrier argues. (Page A1, “Arbitrator Dismisses Google Grievance”, Vol. 56, No. 6, June 2009)
Stephen Ramsay has created a xtra normal movie of our Untitled #4 (PDF) dialogue on Writing as Programming as Writing. xtr normal lets you create a movie from text using text-to-speech with virtual characters – an interesting variant on visualization and a different way of representing a text. I want it known that I look nothing like the character he chose for me.
Update: The dialogue is now up on YouTube in two parts. See Part I and Part II.
Screen shot from inside the Lost Museum
I just stumbled on the Lost Museum web site which is about the American Museum that Phineas Taylor Barnum opened in New York in 1841. Until 1865 when it burned, it offered everything from stuffed animals, wax figures, voyeuristic exhibits and strange documents.
What is interesting is that they (the American Social History Project) created a 3D model of the museum and you can explore it through an interactive interface that lets you pan, zoom and connect to the archives. It has a somewhat 1990s feel as if created by Voyager in Quicktime VR. There is a very good reflective essay by Joshua Brown, “History and the Web, From the Illustrated Newspaper to Cyberspace: Visual Technologies and Interaction in the Nineteenth and Twenty-First Centuries” where he talks about that period when new media designers were trying to emulate Myst,
Emulating 3-D games such as Myst and Doom, The Lost Museum took users on a seamless, self-contained narrative journey, preserving (as best as we could given the technical limitations of online technology) an illusion of visiting another place in another time. On the other hand, The Lost Museum transcended commercial 3-D explorations’ vapid content while also rejecting the fragmentation of data-bases and the derivative hybridity or scrapbook orientation of most multi-media.
That was our intention. We quickly learned, however, that we had fallen into a pattern that is seemingly intrinsic to the spatial interactive game approach. Instead of expanding the historical imagination of users and promoting their active inquiry, we had actually limited the choices open to them, in particular curtailing their ability to make informational linkages and to draw their own conclusions. In short, the narrative outcomes were preordained, confirming only the predominance of designers over users—as demonstrated by ‘test’ audiences of teachers and students who gleefully clicked on different 3-D exhibits but professed utter bewilderment about the significance of what they found. (On the coercive power of the multi-media designer, see Cubitt 2000, pp. 167–168.)
None the less, I find the virtual space compelling, if only because it too is of its time. Every museum is both about something and an exhibit itself of the concerns of its age. Just as the American Museum was in its eclectic entertainment a work of the 19th century, so the Lost Museum is also one of the best of a particular type of experiment.
centerNet met with a representative from Google Book Search, Jon Orwant, about how Google could support the humanities. I believe there are four levels of collaboration.
- Content Curation Interface: We could partner to make possible the careful cleaning and encoding of the books scanned. In most cases the quality of the OCRed text is still poor. It would be nice to have a social layer that allowed people to sign out texts voluntarily to clean them out. We could also help with the selection of editions that are scanned.
- Collections Research Interface: Google could make it possible to build tools that let users create research study collections that are subsets of Google Books that can be studied. For this we need access to an API so research portals can access collections not just individual texts. Google will want assurance that those who have access don’t abuse it.
- Social Research Tools Interface: We need a way to run tools against texts and collections. We need an API so that tools can be plugged in that can then access texts and collections. Again there is an issue of access. Perhaps Open Social could become a standard for tool plug-ins.
- Republication Interface: We need a way to be able create study sites for research groups or courses that make some subset of texts and tools available for a specific purpose.
In all these cases it is clear that Google doesn’t want to read applications, correct lost of texts, or build tools. For that matter none of us know what tools should be written. They see themselves doing smart engineering that creates a platform that enables others who might build layers (research tools, collections portals, and so on) which might be used by others.
John spoke to the centerNet meeting at DH 2009. The motto of Google is to organize the world’s information and make it accessible and useful. The crawl, index, and search the web. One can index and search the world’s books, but it is hard to crawl books (or newpapers or movies.)
There are about 120 millions works in the world and 165 million manifestations. They have an agreement in principle with the publishers that has still not been ruled on. (I think I have that right.) If it is approved in court then Google will be able to some cool things:
- Authors/publishers will be able to opt in or out.
- If authors/publishers opt in then Google could sell their book if they are still under copyright. They have algorithmic pricing to figure out what to charge.
- They could give universities access to the full text of collections of out of date works for a license.
- They could create a terminal at every library that has every book that is out of copyright.
- They could create a “research corpus” that could be used released for experimentation under a creative commons license. This could be used in contests like T-REX.
John gave some fascinating examples of things his intern has been doing from within the firewall.
I’ve been meaning to blog about the Final Report of the Tools for Data-Driven Scholarship Workshop. This workshop was organized by the Center for History and New Media at George Mason and the Maryland Institute for Technology and the Humanities in October of 2008 and they have put up the final report with a number of sensible recommendations. The report summarizes the issues around tool development, the need for reward systems, and it discussed the idea of an “invisible college” of scholars/tool developers who would exchange ideas and support. They distilled the problems down to:
1. Tools need to work better with other tools.
2. Tools need to connect better with content and use that content in a more robust way.
3. Tools need better mechanisms for being found by the scholars who need them. They are not currently finding their audience(s).
They acknowledge that “There may be intellectual and even practical value in reinvention-in ‘recreating the wheel.’” This is a tack we need to take seriously since tool development in the humanities has been going on since the 70s (or earlier if you count Busa’s work). Perhaps the reinvention in the humanities is like reinterpretation – a sign of life not a problem.
I just finished my conference report on Interacting With Immersive Worlds 2009 which was held at Brock. Keynotes included Espen Aarseth and Janet Murray, both of whom where interesting on narrative and games. Kevin Kee and I gave a dialogue on serious games which was well received.
Screenshot of Prezi
Prezi – The zooming presentation editor is a neat PowerPoint alternative. You build a presentation that is one large map, then you script a tour that zooms in and out. This avoids the problem Tufte points out of fragmented discourse. Users can zoom in and out. Prezi is also presented as a service where you craft and store the presentations online rather than through local software.
Screenshot of Exemplar
Springer has an interesting tool that lets you search for a pattern and see its distribution. When you search for a term it allows you to see distribution over time in the upper left, then distribution over disciplinary categories. It also shows a KWIC (Keyword in Context.) See Springer Exemplar: Search Results for Interactivity. The design is clean and easy to explore, but the content seems to be only recent materials in Springer journals.
The Dictionary of Words in the Wild has made it to over 5000 words thanks to its many contributors.
We also made a smooth transition over to the University of Alberta and a new domain name: lexigraphi.ca.