Marriage is a form of social glue: it creates connection, often between groups (circles of friends, family, etc.) that where not connected before. In that sense, our amorous entanglements can play a similar role as Granovetter’s weak ties, that is, they may establish bridging connections between groups that can be quite dissimilar concerning the habitual social parameters (class, ethnicity, religion, etc.) thereby effectively contributing to social cohesion. This article from German news magazine Die Zeit argues that marriage used to be an important means for social mobility:

“One in two men used to marry below their social status and one in two women above.”

The fact that secretaries married their bosses and nurses the doctors at their hospital contributed greatly to the mixing of milieus and even played a role in producing a more even distribution of riches. What I found really interresting though is that this seems to be changing and that online dating is one of the reasons.

Dating websites establish what could be described as a regime of data based perception. Profiles contain a lot of information about a person, some of which allow for an easy and quick categorization on the scale of social status: job and income, level of education, cultural preferences, choice of language, etc. make it easy to compare one’s own social position to that of the millions of profiles that we have access to. According to the cited article, online dating significantly reduces interaction between social strata. Both women and men now strive to find mates that have a similar educational and socioeconomic background and the data offered by most dating sites allows filtering out potential candidates – down to a very fine degree – and still get access to hundreds of profiles. The haphazard component of interaction in physical space where attraction might emerge in unexpected ways is greatly reduced when every single person from the pool of available mates is a profile competing in a handful of very large markets (= dating sites) that are based on specific data constructs and algorithms that allow to project that data according to one’s own wishes.

One could say that in a sense, digital space implies a shifting of perceptual categories: the physical attributes of a person are compressed into pictures and a small number of descriptors (weight, height, body shape) while information that would traditionally have to be uncovered in a conversation (job, education, etc.) is magnified and pushed to the front. With economic insecurity and status angst on the rise, this shift perfectly fits a public increasingly eager to include rational arguments when it comes to choosing a mate.

As always, the chain link that interests me most is the designer of such information systems that, in this case as least, becomes a full fledged social engineer whose choices might have larger consequences for social cohesion than he or she would probably be willing to admit…

Two things currently stand out in my life: a) I’m working on an article on the relationship between mathematical network analysis and the humanities, and b) continental Europe is finally discovering Facebook. The fact that A is highly stimulating (some of the stuff I’m reading is just very decent scholarship, especially Mathématiques et Sciences humaines [mostly French, some English] is a source of wonder) and B quite annoying (no, I don’t miss kindergarten) is of little importance here; there is, however, a connection between the two things that I would like to explore a little bit here.

Part of the research that I’m looking into is what has been called “The New Science of Networks” (NSN), a field founded mostly by physicists and mathematicians that started to quantitatively analyze very big networks belonging to very different domains (networks of acquaintance, the Internet, food networks, brain connectivity, movie actor networks, disease spread, etc.). Sociologists have worked with mathematical analysis and network concepts from at least the 1930ies but because of the limits of available data, the networks studied rarely went beyond hundreds of nodes. NSN however studies networks with millions of nodes and tries to come up with representations of structure, dynamics and growth that are not just used to make sense of empirical data but also to build simulations and come up with models that are independent of specific domains of application.

Very large data sets have only become available in recent history: social network data used to be based on either observation or surveys and thus inherently limited. Since the arrival of digital networking, a lot more data has been produced because many forms of communication or interaction leave analyzable traces. From newsgroups to trackback networks on blogs, very simple crawler programs suffice to produce matrices that include millions of nodes and can be played around with indefinitely, from all kinds of angles. Social network sites like Facebook or MySpace are probably the best example for data pools just waiting to be analyzed by network scientists (and marketers, but that’s a different story). This brings me to a naive question: what is a social network?

The problem of creating data sets for quantitative analysis in the social sciences is always twofold: a) what do I formalize, i.e. what are the variables I want to measure? b) how do I produce my data? The question is that of building a representation. Do my categories represent the defining traits of the system I wish to study? Do my measuring instruments truly capture the categories I decided on? In short: what to measure and how to measure it, categories and machinery. The results of mathematical analysis (which is not necessarily statistical in nature) will only begin to make sense if formalization and data collection were done with sufficient care. So, again, what is a social network?

Facebook (pars pro toto for the whole category qua currently most annoying of the bunch) allows me to add “friends” to my “network”. By doing so, I am “digitally mapping out the relationships I already have”, as Mark Zuckerberg recently explained. So I am, indeed, creating a data model of my social network. Fifty million people are doing the same, so the result is a digital representation of the social connectivity of an important part of the Internet-connected world. From a social science research perspective, we could now ask whether Facebook’s social network (as database) is a good model of the social network (as social structure) it supposedly maps. This does, of course, depend on what somebody would want to study but if you ask yourself, whether Facebook is an accurate map of your social connections, you’ll probably say no. For the moment, the formalization and data collection that apply when people use a social networking site does not capture the whole gamut of our daily social interactions (work, institutions, groceries, etc.) and does not include many of the people that play important roles in our lives. This does not mean that Facebook would not be an interesting data set to explore quantitatively; but it means that there still is an important distinction between the formal model (data and algorithm, what? and how?) of “social network” produced by this type of information system and the reality of daily social experience.

So what’s my point? Facebook is not a research tool for the social sciences and nobody cares whether the digital maps of our social networks are accurate or not. Facebook’s data model was not created to represent a social system but to produce a social system. Unlike the descriptive models of science, computer models are performative in a very materialist sense. As Baudrillard argues, the question is no longer whether the map adequately represents the territory, but in which way the map is becoming the new territory. The data model in Facebook is a model in the sense that it orients rather than represents. The “machinery” is not there to measure but to produce a set of possibilities for action. The social network (as database) is set to change the way our social network (as social structure) works – to produce reality rather than map it. But much as we can criticize data models in research for not being adequate to the phenomena they try to describe, we can examine data models, algorithms and interfaces of information systems and decide whether they are adequate for the task at hand. In science, “adequate” can only be defined in connection to the research question. In design and engineering there needs to be a defined goal in order to make such a judgment. Does the system achieve what I set out to achieve? And what is the goal, really?

When looking at Facebook and what the people around me do with it, the question of what “the politics of systems” could mean becomes a little clearer: how does the system affect people’s social network (as social structure) by allowing them to build a social network (as database)? What’s the (implicit?) goal that informs the system’s design?

Social networking systems are in their infancy and both technology and uses will probably evolve rapidly. For the moment, at least, what Facebook seems to be doing is quite simply to sociodigitize as many forms of interaction as possible; to render the implicit explicit by formalizing it into data and algorithms. But beware merry people of The Book of Faces! For in a database “identity” and “consumer profile” are one and the same thing. And that might just be the design goal…

I have admired the work of Geoffrey Bowker and Susan Leigh Star for quite a while, especially their co-authored book Sorting Things Out is a major step towards understanding how systems of classification structure fields of perception and, consequently, action. The study of advanced technology is intrinsically related to information handling (in the largest sense, ranging from human cognition to information science): building categories, models, languages, and metaphors is a major part of designing information systems and with the ongoing infiltration of society by IT, the process of formalization (i.e. the construction of analytical categories that translate our messy world into manageable symbolic representations) has become a major difficulty in many software projects that concern human work settings. Ontology is ontology indeed but very often “reality as phenomenon” does resist being turned into “reality as model” – our social world is too complex and incoherent to fit into tidy data models. The incongruity between the two explains why there are so many competing classifications, models, and theories in the humanities and social sciences: no single explanation can claim to adequately cover even a small section of the cultural world. Our research is necessarily cumulative and tentative.

The categories and models used to build information systems are only propositions too, but they are certainly not (only) descriptive in nature. There is peculiar performativity to information structures that are part of software because they do not only affect people on the level of “ideas have impacts”. A scientific theory has to be understood, at least in part, in order to make a difference. When PageRank, which is basically a theory on the production of relevancy, became an implemented algorithm, there was no need for people to understand how it worked in order for it to become effective. Information technology relies on the reliable but brainless causality of the natural world to in-form the cultural world.

Why am I writing about this? The University of Vienna (my first alma mater) is organizing a workshop [german] on search engines before Google. And “before” should be read as “before digital technology” (think “library catalogue”). This is a very good idea because instead of obsessing about the “effects” that IT might have (or not) on “society” I believe we should take a step back and look at the categories, models, and theories that our information technologies are based on. And as a first step that means going back in time and researching the intellectual genealogy that is behind these nasty algorithms. The abstract I sent in (four days late, shame on me) proposes to look at early developments in bibliometrics that lead to the development of impact analysis, which is the main inspiration for PageRank.

The proposal is part of this project on mathematics and the humanities that I’m fantasizing about, but that’s a story for another day.

Since MySpace and Facebook have become such a big hype, lot of text has been dedicated to social networking. For people like myself whose social drive is not very developed, the attraction of “hey, dude, I love you so much!!!” is pretty difficult to parse into a familiar frame of reference, but apparently there’s something to all that cuddling online. Being alone has to be learned after all. I somehow can’t shake the feeling that people are going to get bored with all the poking eventually…

Independently form that, there is something really interesting about Facebook and that is, of course, Facebook Platform, the API that allows third party developers to write plug-in like applications for the system. Some of them are really impressive (socialistics and the touchgraph app come to mind), others are not. What I find fascinating about the whole thing is that in a certain sense, the social network (the actual “connections” between people – yes, the quotes are not a mistake) becomes an infrastructure that specific can applications “run” on. For the moment, this idea has not yet been pushed all that far, but it is pretty easy to imagine where this could go (from filesharing to virtual yard sale, from identity management to marketing nirvana). In a sense, “special interest” social networks (like LinkedIn who’s currently scrambling to develop their own platform) could plug onto Facebook and instead of having many accounts for different systems you’ve got your Facebook ID (FB Passport) and load the app for a specific function. If the computer is a Universal Machine, the Internet the Universal Network, Facebook Platform might just become what sociologists since Durkheim have been talking about: the universal incarnation of sociality. Very practical indeed – when Latour tells us that the social is not explaining anything but is, in fact, that what has to be explained, we can simply say: Facebook. That’s the Social.

That’s of course far around the corner and futurism is rarely time well spent – but still, actor-network theory is becoming more intelligible by the day. Heterogeneous Associations? Well, you just have to look at the Facebook interface and it’s all there, from relationship status to media preferences – just click on Le Fabuleux Destin d’Amélie Poulain on you profile page (come on, I know it’s there) and there’s the list of all the other people whose cool facade betrays a secret romantic. This is a case of mediation and it’s both technical and symbolic, part Facebook, part Amélie, part postmodern emptiness and longing for simpler times. Heterogeneous, quoi.

A Facebook Platform thought to its end could mediate on many additional levels, take part in producing the social through many other types of attachment, when it will no longer be a social network application but a social network infrastructure. At least Actor-Network theory will be a lot easier to teach then…

I have recently been thinking quite a lot about what it means to be “critical”. At a lot of the conferences I go to, the term is used a lot but somehow it remains intuitively unintelligible to me. The dictionary says that a critical person would be “inclined to judge severely and find fault” and a critical reading “characterized by careful, exact evaluation and judgment”. I cannot shake the impression that a lot of the debate about the political and ethical dimension of information systems is neither careful, nor exact. Especially when it comes to analyzing the deeds of big commercial actors like Google, there has been a pointed shift from complete apathy to hysteria. People like Siva Vaidhyanathan, whose talk about the “googlization of everything” I heard at the New Network Theory Conference, are, in my view, riding a wave of “critical” outrage that seemingly tries to compensate for the long years of relative silence about issues of power and control in information search, filtering, and structuration. But instead of being careful and exact – apparently the basis of both critical thought and scholarly pursuit – many of the newly appointed Emile Zolas are lumping together all sorts of different arguments in order to make their case. In Vaidhyanathan’s case for example, Google is bad because its search algorithms work too well and the book search not well enough.

Don’t get me wrong, I’m not saying that we should let the emerging giants of the Web era off the hook. I fully agree with many points Jeffrey Chester recently made in The Nation – despite the sensationalist title of that article. What I deplore is a critical reflex that is not concerned with being careful and exact. If we do not adhere, as scholars, to these basic principles, our discourse loses the basis of its justification and we are doing a disservice to both the political cause of fighting for pluralism of opinion in the information landscape and the academic cause of furthering understanding. Our “being critical” should not lead to obsession with the question of whether Google (or other companies for that matter) are “good” or “bad” but to an obsession about the more fundamental issues that link these strange systems that serve us the Web as a digestible meal to matters of political and economic domination. I’ve been reading a lot recently about how Google is invading our privacy but very little about the actual social function of privacy, seen as a historical achievement, and how the very idea could and should be translated into the information age where every action leaves a footprint of data waiting to be mined. We still seem to be in a “1984” mindset that, in my view, is thoroughly misleading when it comes to understanding the matters at hand. If we phrase the challenges posed by Google in purely moral terms we might miss the ethical dimension of the problem – ethics understood as the “art of conduct” that is.

This might sound strange, but under the digital condition the protection of privacy faces many of the same problems as the enforcement of copyright, because they both concern the problem of controlling flows of data. And whether we like it or not, both technical and legal solutions to protecting privacy might end up looking quite similar to the DRM systems we rightfully criticize. It is in that sense that the malleability of digital technology throws us back to the fundamentals of ethics: how do we want to live? What do we want our societies to look like? What makes for a good life? And how do we update the answers to those questions to our current technological and legal situation? Simply put: I would like to read more about why privacy is fundamentally important to democracy and how protection of that right could work when everything we do online is prone to be algorithmically analyzed. Chastising Google sometimes look to me like actually arguing on the same level as the company’s corporate motto: “don’t be evil” – please?

We don’t need Google to repent their sins. We need well-argumented laws that clearly define our rights to the data we produce, patch up the ways around such laws (EULAs come to mind) and think about technical means (encryption based?) that translate them onto the system level. Less morals and more ethics that is.

Oliver Ertzscheid’s blog recently had an interesting post (French) pointing to a couple of articles and comments on The Facebook, among which an article at the LA Times’ entitled “The Facebook Revolution“. One paragraph in there really stands out:

Boiled down, it goes like this: Humans get their information from two places — from mainstream media or some other centralized organization such as a church, and from their network of family, friends, neighbors and colleagues. We’ve already digitized the first. Almost every news organization has a website now. What Zuckerberg is trying to do with Facebook is digitize the second.

This quote very much reminds me of some of the issues discussed in the “Digital Formations” volume edited by Robert Latham and Saskia Sassen in 2005. In their introduction (available online) they coin the (unpronounceable and therefore probably doomed) term “sociodigitization” by distinguishing it from “digitization”:

The qualifier “socio” is added to distinguish from the process of content conversion, the broader process whereby activities and their histories in a social domain are drawn up into the digital codes, databases, images, and text that constitute the substance of a digital formation. As the various chapters below show, such drawing up can be a function of deliberate planning and reflexive ordering or of contingent and discrete interactions and activities. In this respect as well, sociodigitization differs from digitization: what is rendered in digital form is not only information and artifacts but also logics of social organization, interaction, and space as discussed above.

Facebook, then, is quite plainly an example for the explicit (socio-)digitization of social relations that were mediated quite differently in the past. The “network of family, friends, neighbors and colleagues” that is now recreated inside of the system has of course been relying on technical (and digital) means of communication and interaction for quite a while, and these media did play a role in shaping the relations they helped sustain. There is no need to cite McLuhan to understand that relating to distant friends and family by mail or telephone will influence the way these relations are lived and how they evolve. Being rather stable dispositifs, the specific logics of individual media (their affordances) were largely covered up by habitualization (cf. Berger & Luckmann1967, p.53); it is the high speed of software development on the Web that makes the “rendering of logics of social organization, interaction, and space” so much more visible. In that sense, what started out as media theory is quickly becoming software theory or the theory of ICT. There is, of course, a strong affiliation with Lawrence Lessig’s thoughts about computer code (now in v. 2.0) and its ability to function as both constraint and incentive, shaping human behavior in a fashion comparable to law, morals, and the market.

The important matter seems to be the understanding of how sociodigitization proceeds in the context of the current explosion of Web-based software applications that is set to (re)mediate a great number of everyday practices. While media theory in the tradition of McLuhan has strived to identify the invariant core, the ontological essence of individual media, such an endeavor seems futile when it comes to software, whose prime caracteristic is malleability. This forces us to concentrate the analysis of “system properties” (i.e. the specific and local logic of sociodigitization) on individual platforms or, at best, categories of applications. When looking at Facebook, this means analyzing the actual forms the process of digitization leads to as well as the technical and cultural methods involved. How do I build and grow my network? What are the forms of interaction the system proposes? Who controls data structure, visibility, and perpetuity? What are the possibilities for building associations and what types of public do they give rise to?

In the context of my own work, I ask myself how we can formulate the cultural, ethical, and political dimension of systems like Facebook as matters of design, and not only on a descriptive level, but on the level of design methodology and guidelines. The critical analysis of social network sites and the cultural phenomena that emerge around them is, of course, essential but shouldn’t there be more debate of how such systems should work? What would a social network look like that is explicitly build on the grounds of a political theory of democracy? Is such a think even thinkable?

I have been working, for a couple of month now, on what has been called “network theory” – a rather strange amalgam of social theory, applied mathematics and studies on ICT. What has interested me most in that area is the epistemological heterogeneity of the network concept and the difficulties that come with it. Quite obviously, a cable based computer network, an empirically established social network and the mathematical exploration of dendrite connections in worm brains are not one and the same thing. The buzz around a possible “new science of networks” (Duncan J. Watts) suggests, however, that there is enough common ground between a great number of very different phenomena to justify the use of similar (mathematical) tools and concepts to map them as networks. The question of whether of these things (the Internet, the spreading of disease, ecosystems, etc.) “are” networks or not, seems of less importance than the question of whether network models produce interesting new perspectives on the areas they are being applied to. And this is indeed the case.

One important, albeit often overlooked, aspect of any mathematical modeling is the question of formalization: the mapping of entities from the “real” world onto variables and back again, a process that necessarily implies selection and reduction of complexity. This is a first layer of ambiguity and methodological difficulty. A second one has been noted even more rarely, and it concerns software. Let me explain: the goal of network mapping, especially when applied to the humanities, is indeed to produce a map: a representation of numerical relations that is more intuitively readable that a matrix. Although graph (or network) theory does not need to produce a graphical representation as its result, such representations are highly powerful means to communicate complex relationships in a way that works well with the human capacity for visual understanding. These graphs, however, are not drawn by hand but generally modeled by computer software, e.g. programs like InFlow, Pajek, different tools for social network analysis, or a plethora of open source network visualization libraries. It may be a trivial task to visualize a network of five or ten nodes, but the positioning of 50 or more nodes and their connections is quite a daunting task and there are different conceptual and algorithmic solutions to the problem. Some tool use automatic clustering methods that lump nodes together and allow users to explore a network structure as hierarchical system where lower levels only fold up by zooming in on them. Parabolic projection is another method for reducing the number of nodes to draw for a given viewport. Three-dimensional projections represent yet another way to handle big numbers of nodes and connections. Behind these basic questions lurk matters of spatial distribution, i.e. the algorithms that try to make a compromise between accurate representation and visual coherence. Interface design adds yet another layer, in the sense that certain operations are made available to users, while others are not: zooming, dragging, repositioning of nodes, manual clustering, etc.

The point I’m trying to make is the following: the way we map networks is in large part channeled by the software we use and these tools are therefore not mere accessories to research but indeed epistemological agents that participate in the production of knowledge and participate in shaping research results. For the humanities, this is, in a sense, a new situation: while research methods based on mathematics are nothing new (sociometrics, etc.), the new research tools that a network science brings with it (other examples come to mind, e.g. data-mining) might imply a conceptual rift where part of the methodology gets blackboxed into a piece of software. This is not necessarily a problem but something that has to be discusses, examined, and understood.

Most of the political potentialities of automated surveillance depend on two elements. The debate has generally concentrated on the first: data acquisition. But “digital wiretapping”, in whatever technical shape it may come, is only one part of the issue. Making sense of the data collected is the more complicated, albeit often neglected, half of the equation. Data mining technologies have certainly much advanced over the last couple of years but the commercial applications are not necessarily adapted to the demands that government agencies might have. This article makes the claim that scientist have developed a software that can “help detect terrorists before they strike”. It reads:

Computer and behavioral scientists at the University at Buffalo are developing automated systems that track faces, voices, bodies and other biometrics against scientifically tested behavioral indicators to provide a numerical score of the likelihood that an individual may be about to commit a terrorist act.

Although I’m quite skeptical about the actual performance of such software in the field (the real world is pretty messy after all) it shows the direction things are heading. The actual piece of software comes pretty close to what Virilio has described as “machine de vision” (vision machine) – a device that not only records reality (camera) but also interprets it (the man on the subway platform walking around nervously is not just a heap of pixels but a potential terrorist). Virilio talks about the “delegation of perception” and this is perhaps the most interesting aspect of the increasing technologizing of control: part of the process of decision-making (and therefore part of the responsibility) is transferred to algorithms and questions of professional ethics become matters of optimizing parameters and debugging.