Cybersyn: Before the Coup, Chile Tried to Find the Right Software for Socialism

Image of Cybersyn Opsroom

In New York for my last f2f meeting of the MLA Committee on Information Technology I got a New York Times with an intriguing article about a Chilean management system, Cybersyn, titled Before the Coup, Chile Tried to Find the Right Software for Socialism.

Cybersyn was born in July 1971 when Fernando Flores, then a 28-year-old government technocrat, sent a letter to Mr. Beer seeking his help in organizing Mr. Allende’s economy by applying cybernetic concepts. Mr. Beer was excited by the prospect of being able to test his ideas.

He wanted to use the telex communications system – a network of teletypewriters – to gather data from factories on variables like daily output, energy use and labor “in real time,” and then use a computer to filter out the important pieces of economic information the government needed to make decisions.

Cybersyn was apparently semi-functional before the coup that overthrew Allende’s government and it was used to help manage around the small-business and truckers strike in 1972. I don’t think the Opsroom pictured above was ever fully operational, but visualization screens were important even if at the time they were hand-drawn slides that were projected rather than computer generated visualizations (see http://varnelis.net/blog/kazys/project_cybersyn on the chairs of the Opsroom.) Beer and the Chileans wanted Cybersyn to help them implement an alternative socialist economy that was managed in real time rather than “free” and chaotic or planned in the heavy handed way of most socialist economies of the time.

Rooting around, I found a good article about Cybersyn and the English visionary designer Stafford Beer from 2003 in the Guardian by Andy Beckett, Santiago Dreaming. It turns out that Beer gave the Massey Lectures in 1971 and they have been reprinted by Anansi as Designing Freedom. He also moved part-time to Toronto in the 80s where his last partner, Dr. Allenna Leonard of Metaphorum still resides. He died in 2002.

Another interesting thread is Fernando Flores who was the political lead of Cybersyn and the person that recruited Beer for the project. After the coup, Flores went to the US and got a Ph.D. in Computer Science collaborating with Terry Winograd, and being influenced by Maturana, also Chilean. That’s right – the Flores of Understanding Computers and Cognition. He is now back in Chile as a senator and supports various projects there.

The common thread is that Beer, Flores and Maturana all seem interested in viable systems in different spheres. They were applying cybernetics.

Dreyfus: Alchemy and Artificial Intelligence

Willard in Humanist pointed us towards an interesting RAND Paper by Hubert L. Drefus from 1965, Alchemy and Artificial Intelligence which suggests that artificial intelligence research is like alchemy – initial success has led to it being oversold when the fundamental paradigm is wrong.

Alchemists were so successful in distilling quicksilver from what seemed to be diret, that after several hundred years of fruitless effort to convert lead into gold they still refused to believe that on the chemical level one cannot transmute metals. To avoid the fate of the alchemists, it is time we asked where we stand. Now, before we invest more time and money on the information-processing level, we should ask whether the protocols of human subjects suggest that comptuer language is appropriate for analyzing human behaviour. Is an exhaustive analysis of human intelligent behavior into discrete and determinate operations possible? Is an approximate analysis of human intelligent behavior in such digital terms probable? The answer to both these questions seems to be, “No.”

In this paper Dreyfus leverages the lack of progress after people like H. A. Simon in 1957 predicted the extraordinary. Dreyfus does more than make fun of the hype, he uses it to question what AI research might achieve at all and to think about intelligence.

Now that we are 50 years after Simon’s predictions things are more complicated. We do have chess playing machines that are better players than humans. (Drefus points out how the early machines being hyped were really stupid chess players.) We do have machines that can recognize complex patterns and recognize speech. We do have better machine translation. It may be going slowly, but research is moving forward. Perhaps the paradigm of the mind as a machine is wrong, but thinking about it that way and trying to model intelligent behaviour is getting results. What then do we make of the alchemical insult. Is it too easy to call magical thinking those projects that are ambitious and make the mistake of predicting success? Having recently read Siegried Zielinski’s Deep Time of the Media, I’m finding myself more sympathetic of magical projects that promise to transmute data into intelligence. Impossible … probably, but that is no reason not to try.

To paraphrase the third of (recently died) Arthur C. Clarke’s three laws of prediction:

“Any sufficiently magical proposal should be indistinguishable from research.”

This obviously applies to grant proposals.

Harvard and Open Access

Peter Suber in Open Access News has reproduced the text of the motion that the Faculty of Arts and Science at Harvard passed requiring faculty to deposit a copy of their articles with the university.

The Faculty of Arts and Sciences of Harvard University is committed to disseminating the fruits of its research and scholarship as widely as possible. In keeping with that commitment, the Faculty adopts the following policy: Each Faculty member grants to the President and Fellows of Harvard College permission to make available his or her scholarly articles and to exercise the copyright in those articles.

According to another post by Peter Suber, Harvard is the first North American university to adopt an open access policy. He calls it a “permission mandate” (granting permission to the university to make research open) rather than a “deposit mandate.” It has the virtue that the university takes responsibility for maintaining the access, not the faculty member.

More on this can be found here (another Suber post) and here (Chronicle of Higher Ed.).

Ian Hacking: analogue bodies and digital minds

The Cartesian vision fulfilled: analogue bodies and digital minds is an essay in Interdisciplinary Science Reviews (2005, v. 30, n. 2) by Ian Hacking that first argues that despite the dislike for the Cartesian mind-body split in philosophy there is a degree to which Western culture is acting as if the body was analogue and the mind digital. Our metaphors, our representations, our sciences are Cartesian. Medicine treats the body as a messy mechanism, cognitive science treats the mind like a computer. Here is the abstract:

Current intellectual wisdom, abetted by philosophers of all stripes, teaches that the Cartesian philosophy is both wrong and dead. This wisdom will be overtaken by events. Present and future technologies – ranging from organ transplants to information coding – will increasingly make us revert to Descartes’s picture of two absolutely distinct types of domains, the mental and the physical, which nevertheless constantly interact. We as humans are constituted in both domains, and also must inhabit them. This is less a matter of facts – for what a person is, is never simply a matter of fact – than of how we will come to conceive of ourselves in the light of the facts that will press in upon us.

What is impressive and distracting about the essay (and what makes it accessible) is that he takes us on a tour of contemporary media culture from Japanese entertainment robots, manga, to Stelarc. It is only at the end that he makes his second move, which is to declare, without giving us a similar tour, that the representation of the mind as digital is “dated”.

Minds, on the other hand, we represent as information processors. And in this age we represent the processing of information by sequences of binary digital operations. Here I am less confident of the metaphor, which I find a bit dated. (p. 164)

He concludes by talking about Antonio Damasio’s theory which is that, “A human being is a neurologically nested triad of mind, brain and body.” (p. 165) The science that is showing the importance of the body to emotion and emotion to mind “leaves the digital mind in the dust.” (p. 165) Hardly. I find it hard to believe that science will give up on trying to formally model the mind as a method for testing hypotheses and understanding.

Barry Allen on the Tool: Artifice and Design

On Friday I heard Barry Allen talk about tools. His talk touched on points he makes in his forthcoming book, Artifice and Design: Art and Technology in Human Experience. He defined a tool as having two distinctive qualities:

  1.  Artifactual economy. A tool is an artifact and it is part of an economy of tools. Tools are made by other tools, they are controlled by other tools, and often operate on tools. This is what characterizes modern technology – a technical culture of machines driving machines making machines and interacting with machines.
  2. Functionless functionality. By this Allen, I think, meant that tools do not have a single function or purpose – that they are often used for unanticipated functions, but are still functional. The general purpose computer might be the paradigm of a tool with no fixed function that is therefore adaptable to all sorts of functions.

After the talk Barry and I talked about software as tools. He would say that a book or a movie is an artifact, but not a tool. At what point does a digital artifact go from being data to being a tool? When it is executable? Is a web page information or a tool?

Barry also made an interesting point about first-order and second-order machines. First-order machines are “devices that extend human capacities by exploiting a mechanical advantage.” Second-order machines are like factories, “an assembly of first-order machines, coupled to produce a multiplying effect. Exploits the economic equivalent of a mechanical advantage.” These second-order machines are the factories of tools that make our culture more than tool using, but an economy of technology. Allen warned that it is these second-order tools that so easily turned to waste.

Zonbu: cares about the planet too

Image of Zonbu BoxZonbu is a environmental personal computer with some interesting features. It runs a version of Linux and comes with bundled applications. You buy it with a monthly plan that gives you off-site storage and maintenance. It has no hard-drive, just a flash card for local storage. All of this means it is extremely energy efficient (consumes as much as a light bulb) and that it is easy to run. They also promise to take it back and disassemble it for recycling.As interesting as the green aspect of Zonbu is, I’m also struck by their service model. You buy it for $99 (without keyboard or monitor) and then pay $13 a month or more for the storage and support. You don’t get root access and they manage the computer for you. It comes with all the basic applications. As some commentators have put it – the Zonbu makes for a good second home computer for the family (at least those who don’t want to run PC games.)

Guy sent me this after reading my Blog Action Day grumbling.

As It Happens, Privacy, and the Mechanical Turk

As It Happens on CBC Radio just played a good double segment on “Google Eyes”. The first part looked at the Amazon Mechanical Turk task looking for Steve Fossett’s plane on satellite images. The second part looked at privacy issues around street level imaging from outfits like Google.

Mechanical Turk (Artificial Artificial Intelligence) is a project where people can contribute to tasks that need many human eyes like looking at thousands of satellite images for a missing plane. It reminds me of the SETI@home project which lets users install a screen saver that uses your unused processing cycles for SETI signal processing. SETI@home is not part of a generalized project, BOINC that, like the Mechanical Turk, has a process for people to post tasks for others to work on.

The Privacy Commissioner of Canada announced yesterday that she has written both Google and Immersive Media (who developed the Street View technology used by Google) “to seek further information and assurances that Canadians’ privacy rights will be safeguarded if their technology is deployed in Canada.” The issue is that,

While satellite photos, online maps and street level photography have found useful commercial and consumer applications, it remains important that individual privacy rights are considered and respected during the development and implementation of these new technologies.

This is a growing concern among privacy advocates as a number of companies have considered integrating street level photography in their online mapping technologies.

In street level photography the images are, in some cases, being captured using high-resolution video cameras affixed to vehicles as they proceed along city streets.

Google, according to the commission on the radio, has not replied to the August 9th letter.

Richard Rorty Dies at 75 – New York Times

Richard Rorty, Image ofThe New York Times has an obituary for Richard Rorty by Patricia Cohen, Richard Rorty, Philosopher, Dies at 75 (June 11, 2007). Rorty, when I was a Haverford, was presented to us as a philosopher looking to the American tradition of James and Dewey to reconcile Continental and Anglo-American philosophy.

Gary Madison has an essay online, Coping with Nietzsche’s Legacy: Rorty, Derrida, Gadamer that nicely positions Rorty in postmodern philosophy.

Robotic age poses ethical dilemma

Roboethics ImageThe BBC has a story about roboethics, Robotic age poses ethical dilemma, triggered by a South Korean initiative to develop a Robot Ethics Charter as part of a focus on robotics as a growth area.

In the past, robots were considered just a useful tool in the manufacturing industry. But it is gradually embedded in human life by cleaning homes, protecting them from thieves and providing education. Nowadays robots are also used to rescue people at accident spots such as fires.

This year, various robots are to be introduced: a robot that teaches English and sings songs to children, a robot that guides people at the post office and a robot designed to save people at disaster areas. (Korea.net, Robots, cars, batteries hold key to future growth)

Poking around I found this Painter Robot from Yahoh. (Sounds like Yahoo to me.) The BBC story also mentions the Roboethics.org – Official Roboethics website which has issued a Roboethics Roadmap.

Roboethics is the ethics applied to Robotics, guiding the design, construction and use of the robots.
In this site you may find: birth and history of Roboethics; all the information concerning the development of the concept of a human-centered Roboethics; the events which have marked the update of the original proposal; the international projects on Roboethics; the EURON Roboethics Roadmap; the activity of the IEEE-RAS Technical Committee on Roboethics.

Thanks to Daryl for this link.

Web Mining for Research

Web Mining for Research is a white paper I’ve just written to get my ideas down about how we should be using the Web as evidence not just for social science research, but in the humanities. Digital humanities is more than studying old wine in new digital bottles – the challenge is to do humanities research using the digital as evidence. For me the challenge is how to rethink philosophy now that we can mine concepts in their sites, to paraphrase Ian Hacking.