Matthew Fuller on Tue, 21 Sep 1999 02:17:47 +0200 (CEST) |
[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]
<nettime> BREAK THE LAW OF INFORMATION |
BREAK THE LAW OF INFORMATION notes on search engines and 'Natural Selection'1 Matthew Fuller BACK UP There is no permanent back-up device for the internet. A number of institutions have made one-off synchronous copies of the web which claim to be entire for the duration of the period it takes their spiders to reach the furthest corners of the web.2 There are also systems which actively monitor every email or newsgroup message sent on behalf of the odder fans of political action and dissolution3. Both of these are partial of course. As what Gregory Ulmer has usefully called a pre-broken system, there is no centralised back-up. Copies of elements of the net are backed up across thousands of storage devices off and across itself. Were the whole lot to go down, putting back together the results of such disgraceful degradation would be an insane task. Inevitably, some large areas of the web would slide back into place with the efficiency of a file amongst filing cabinets. Others however, made on the fly, left without a double, would go missing. Still more would have their links tangled, lost - a mess of stringy, gloopy, gone data. Natural Selection4 is in a sense the result of this crash before it happens. It is the web, picked up in pieces by people who never made it, putting it back together in a way that makes more sense. It is obvious that the British Airports Authority site would 'want' to link to information about people getting killed by anti-immigration police5. It is obvious that someone looking for information on fertility treatments would want to find a site mixing the histories of different acts of racial 'improvement' into a corporate brochure. Just as it is obvious that anyone looking for traces of overtly racialised political activism on the net is going to want their browser to become infested by a fast-breeding Jamaican virus in the guise of a triple-consciousnessed poet6. BLACK BOX A search engine is a black box. That is to say, the relationship between its input and output is masked. The tension between making the a search engine usable, predictable and 'refinable' and the commercial necessity of maintaining its constituent algorithms and processes as proprietary information provides one of the key contexts for this project. Piracy of any electronic technology is usually carried out by reverse engineering: feeding it inputs, monitoring its output until a tabulation of its workings can be drawn up. The black box is not actually cracked open and scrutinised, but is fed line after line until it can be simulated. What is true of the search engine is also true of the user. When you use HotBot for instance your search-string and all the links you make from the results to your enquiry are logged and recorded along with your domain and IP number. This information is useful for many reasons and becomes increasingly so over time and repeated use. It allows the more accurate targeting of search results to specific enquiry entries (according to frequency of choice) and advertising, (according say to proportional interest from certain domains). Any impression that engines are there to provide a useful service financed by incidental advertising is of course pure 'company history', as Yahoo calls it. Just as the engine helpfully provides a field in which you can type in your search-string, the engine is also opening up an aperture through which the user can be interrogated. There are always at least two black boxes in this equation. The interface is where they meet and engineer each other. Natural Selection sits on top of any of the mainstream search engines, filtering their results and, if certain strings are fed to it, slipping in a result from the bottom of the deck. It is in producing itself as an interface to an interface between: the targeted engine, its interface on one side and the user's manifestation to the engine as a search string (with a history of others) and an internet location along with the need of the user to access relevant links on the other that Natural Selection is able to operate as a double-blinded bait and switch. It is this context, of the specific rules of formation, of the grammars of information put into play by various search engines that Natural Selection operates in and which this story hopes to open up a little. INFORMATION TRANSFERENCE Shannon7: The amount of information in a message is in proportion to its unpredictability. Of course, this definition of information defines it in strict relation to the context in which it is received. It is strictly relativist, (and provides one of the crucial problems for designers of search engines attempting to produce a machine-enforced dialogism, where the information derived from the net is the result of an ongoing process of discrimination - agents for instance). This has important implications for the construction of search engines. The search engine is absolutely unable to treat a word or any collection of symbols entered into it in a contextualised manner, there are ways of refining or narrowing down the search for sure, but the core of what it has to act upon is the string of characters that it has just been requested to find matches or correspondences to. One response to this has been to produce a situation where the user has the responsibility to do the thinking 'for' the engine. Yahoo for instance prides itself on its 'Ontology' - it's super-mainstream view of the world. Users too, as members of a conformist society are expected to have internalised this ontology, but also to be able to use it reflexively. That is to say, that the conformism of the user can be operated upon, tweaked, at the same time as it is being used in order to get to the required link. This is not to say that this internalised conformism has an open architecture in the case of either the user or the engine, but rather that what Natural Selection does is in a sense to make manifest the production of this double consciousness and open it to some form of reconfiguration. To great or lesser extents, non-directory-based systems avoid this conformism, largely as is examined later, to the degree that they attempt to ascribe, involve, or invent semantic meaning in the field which they operate on. To the degree at which it occurs, the refusal of making straightforward sense by the engine becomes the best tool by whose help the most secret components of the net can be opened.8 Here, the occurrence of information in Shannon's sense, veering towards its upper limit of utter randomness, is limited solely by the congruence of search strings - making a permanently locked on to its self-same narcissism an impossibility. At the same time though, their avoidance of overt taxonomisation of the world - their heightened level of information in this sense - means that the development of a search-consciousness, perhaps a temporary micro-subject thrown together by the agglomeration of strings, results and guesses into a briefly memorised sensorium, remains trapped at the level of articulation of that loaded bane of software culture - intuition. A more sustained, distributed, inquisitive or conflictual sensorium demands more difficult and supple ways to cohere. THE COPY JUMPED One other definition for the amount of information in a sentence, object, process and so on is the length of the set of instructions - the program - necessary to carry out in order to make a copy or model of it9. A game of cricket has more information than a game of knuckles. A cup has more information than a plain sheet of paper. By this definition at least, the internet is significantly high in information. To create a set of instructions to copy it would be just short of impossible. What this project does is, in a sense, just that. It turns the internet into a copy of itself. This denaturalised selection treats the net as a found object: mangy paint-dripped goat; soup can; planet. The instructions for its copying are found in every line of code that makes it up. However massified the net becomes, however much 'diversity' is attached Natural Selection will always add one more bit, one last line, that, instead of adding information to the net will always force the cognisance of its own grounding incompleteness. Thus at once, the copy will never be made - the thing has to be done for real. CONDENSATION The search string, the series of symbols tapped into the search field on the search engine, like the password for the hacker draws all the protocols and machines in onto itself as an act of condensation. In common with elements within indexes, search terms are words or numbers disengaged from direct social or linguistic involvement. The string of characters constructs itself as an aperture to alignment with a potentially infinite amount of interlocutions, the key around which a thousand locks ensconce their levers. For the second or so that it is typed, all of this potential is under pressure. From the moment that the search command is given, the moment, the string and its ramifications begins to extrude. An arbitrary wad of bits heats temporarily up into a larval plasticity that has it tonguing for the contradictory, the useless and the lost. What Natural Selection does is to maintain this sense of plasticity. When the search string is entered, it is under a pressure that is not let up as the results file up by percentages onto the page. Rather than each result providing a smooth transition from risk, from information, into assuredness it becomes evident that every link is articulated in many ways. The switch-tongue is in constant spasm. It is in this state that a politics of the switches and of the string can be made. FORMS One of the ways in which the world wide web has been found to be most 'useful' is in the ability to construct forms which can be filled out online. This again is the creation of a reading aperture into the user. Feeding into databases they "traverse and cancel the", already questionable "public/private distinction"10 by a literal interrelation. A question implies an incomplete picture of reality which an answer completes. In this case however, ticking checkboxes implies an addition to reality - an accumulatory combination of positives which individually and together combine to produce form in numbers. In "The Temple of Confessions" and their related techno-ethnographic internet work11 Guillermo Gomez-Pena and Roberto Suifentes use this use this shape-making facility of forms to entice the self-production of fear and desire about Chicanos and Mexicans in US art and net audiences. The confessions of fear and desire pour into a database where they form a switch tongue writhing in a perpetual glottal roar of confusion, hate and fantastical need. In comparison, Natural Selection takes this database composed of fear and desire as already pre-existing - as the entire net. One of the modes in which elements of Natural Selection operates is a reverse of what Gomez Pena calls the 'Dantean model of multiculturalism'12 in which the flaneur from the North descends to the South to sample the strictly localised flavour. Here, euro-american technoculture is exoticised and taken for the mumbo-jumbo status it hungers for but can never attain. The accumulatory combination of positives produces a weight in data, a gravitational pull that turns conceptions of what constitutes marginality inside out. ENTER A TOURIST. LEAVE A LOCAL Web work is endless. Why do something as text when you can do it as sound? Why do it as mere sound when you can do it as a realtime post-raster 3D walkthrough intentional community with badword filtering and ambient billboards? If the net is already a copy, then to double it demands a constitutional heterogeneity that is nothing if not a mess. The founding statements of Romanticism demanded a radical incompleteness, a cursed restlessness against the apparent foreclosures of industrialisation. This is at once the condition of work for all those who put bread on the table by shovelling content and code onto servers, and in a reversioned materialist form realised in the widely popularised conceptual device of the rhizome. Such refusal of finitude is both what forms an incomprehensible glimpse over the abyss for regulation and for those in need of the reassurance of 'critical understanding', as well as the excuse for extending overtime into your sleep. It is inherent to the texture and dynamics of the web and forms the swamp on which Natural Selection is grounded. This mess sprawls out in several ways: in terms of distribution across the net and position within different server-regimes; in terms of producing work for machine readerships; in terms of working many technico-aesthetic approaches simultaneously; and clearly, in terms of authorship. Much has been made of the potential on the web for small groups of people or individuals to develop equal visibility or presence as major companies or institutions, often to usurp vertical hold over their public image. This was especially the case in the early to mid nineties. Later as the web-design market restructured along rationalised lines and the corporate influx onto the net became an increasingly normalised procedure and new forms of software and increasing production times raised barriers to making ostensibly corporate sites, various groups developed ways of destabilising the apparent authorship of sites. Many of these remained strictly on the level of swapping zeros for ones and standing back to admire the - inevitably invisible - confusion and collect the letters from the lawyers. Others however knotted themselves more intimately and uncomfortably into the political and technical fabric of representation and communication13. When dealing with such a diffuse dynamic as racialisation, pressing attention solely on a specific target is not necessary. Being able to deal with it in as many ways and as many places it exists is. Inevitably this means relinquishing the appearance of 'clarity' essential to validation in terms of art in order to turn more than one kind of purism into more matter for mongrelisation. Doubling the web demands that glib coffee-table web design becomes an appearance along with default-grey file directories, that rainbow coloured, <blink>-heavy home-pages get bitten along with customer service oriented purchase opportunity provision included product information displays. At the same time as one mode of operation within Natural Selection is to produce a destabilisation of the apparent reliability of racialised sites by hooking their familiar routines into heritage free-fall, it is equally imperative that other dynamics come through. There are around twenty sites that sit directly after the first folderload in the mongrel.org.uk URL. They come from people working in different countries, different political, narrative and interactive angles. Many kinds of personal histories are in operation alongside approaches to users and interrelationships with elements of the web that they link themselves into or cut themselves off from. The Natural Selection front end, the doctored search engine is a specific technical and cultured intervention, a hack, into 'a culture that has no culture' - information science - in order to tease out what it represses and what it produces. Making this break into the net produces a surplus of possibility - we could do anything with it - that allows others to take it on as an opportunity. This surplus produced by a technical and conceptual break - what might happen if you could begin to sense the politics of algorithms whilst you used them? - allows a multitude of aesthetics, experiences, voices, flows, positions and lives to operate interconnected and in parallel without making them conform to any particular model of operation. Different kinds of energies, different speeds are also capable of mobilisation in a process of work that is not overdetermined by a finalised shape: energy from hate, from derision, communication, stupidity; programming energy; research. For a long while this meant that the some of it never climbed out of Beta, but it also means that it produces an extended period of experimentation that will potentially at least never be quite over. This drawn out, not quite programmatic dispersal of working over time that is inherent in all kinds of work on the web is also echoed in the distribution of the component sites across it. Natural Selection is composed of a many sites. All of the separate sites are accessible via the search engine, as they are via any other search engine. Only a proportion of them are pushed to be indexed on the ostensible front-end as 'Star Sites'. Forming a constellation of sites on and off the mongrel domain, with varying thicknesses of cross-linking it never appears as a whole, there is no complete index, no climbing up the URL to find the answer. Some of it may never get seen by users, but will be constantly read by crawlers and agents. In any work on the internet, machines both hard and soft are the primary audience. BRIGHT LATTICES OF LIGHT UNFOLDING AGAINST THE COLOURLESS VOID In one software millionaire's mansion, the walls are covered not with paintings but with flatscreen monitors linked to a voice activated database of images. Any time a conversation occurs in some part of the house the screens change to reflect it. The millionaire coos to his baby son whilst feeding him bottle and dictating a letter into an intercom phone. Some cracked up old madonna and child appears, a ball game, gouache cowboys, pulmonary cancer on ultrasound, then three-dee NASDAQ flows. A family snap taken on the lawn. Mummy herself enters to usher in a group of business associates and take baby away. Courbet's studio appears. The naked model is clad in a seasonally updated business suit. Something transcendental from National Geographic. Talk moves on, the computer reads every word uttered, flashing up a picture corresponding to any one of thousands of keywords from the database, everyone effects not to notice. The web sites produced as part of Natural Selection involve the mobilisation of words that have become crystallised forms of racialisation. Typing a search string into the Natural Selection search engine either allows you access to the net as per usual - or, should you enter one of several thousand keywords - clicking on an apparently transparent search result drops you into its double. Imagining into the lexical space of the fear, desire, hate, the will to control and define that composes racialisation allows the registration of a shape composed in crisis by language. The accumulation of shifting populations of terms of abuse, of jargons, of political or scientific code, of euphemisms, agglomerate into conceptual and signifying phase forms. Networks of pot-holes and caverns under the homogenised surface of mediatised language are opened up for re-articulation and engineering. This occurs not just as an incursion into enemy mind-space by a punitive band of artists with the oxygen-tent souls of angels but as an response to senses of what might be done and as the public development of models that might be taken further. .sitUATE Two models that feed into Natural Selection are conjoined by use of the same string of characters to mark kinds of poesis, making, that are attractive by both their livid intensity and their attempts at critical technics. For Guy Debord, writing before the formation of the Situationist International (SI), the construction of situations was "the concrete construction of momentary ambiences of life and their transformation into a superior passional quality".14 Situations were to be something produced in the city, in real space. For Donna Haraway, writing about feminist scientific practice, situated knowledge is equally "rooted in yearning, not an abstract philosophical foundation".15 Tensions exists between the two uses of the same sign. Debord insists on a practice that provokes the release of "primary sentiments".16 Haraway, involved in writing which still has to contend with determinist scientisms whose agenda is very much the 'discovery' (and political use) of such compulsive drives proposes instead the practices that are "suspicious, implicated, knowing, ignorant, worried and hopeful".17 Writing later, after the foundation of the SI, and in a more complex manner about what is intended by situations, there is what might be taken as an almost literal description of certain aspects of Natural Selection: "We are not going to limit ourselves... ...with environments of mechanistically provoked surprises."18 Clearly, entering a search string and finding that it doesn't take you to the hoped-for homepage of the European Skinhead Army but to some pervert nonsense looking too much like the real thing to be quite reassuringly wrong enough for comfort is exactly that. You're tricked by a few lines of code. What can clearly be seen though is that in both the preparation and execution of this project is that, "...setting up, on the basis of more or less clearly recognised desires, a temporary field of activity favourable to those desires"19 is not quite so simple - this could be the description of a business plan. Situated knowledge is for Haraway something equally particular, but complicated by embodiment, by "sometimes painful" processes of building relationships, structures and accountability into the process. What unites both forms, the 'construction of situations' and 'situated knowledge', however is the urgency of and the possibility for establishing concrete freedoms via genuine experiment. Correctly understanding the 'merely empirical experimentalism'20 of the academic avant garde to be rigidifying and pre-determinate of outcome in the manner of heavily disciplined scientific procedure21, Debord criticises experimental work which seeks only to operate within pre-established conceptual categories. Haraway lives in an altogether trickier space, which (although it is at least partly in the grip of an academic feminism which has to make the most tortuous excuses before it feels permitted to crack a joke, but feels quite at liberty to punish the world with nauseatingly didactic allegorical paintism as book covers) seeks to establish intermeshing and conflictual fields of lived experience, scientific method, political mobilisation, diffusion of knowledges in order to go beyond and change them. The making of alliances, something which is inherent to the composition of Mongrel, demands concrete practices of learning and accountability in order to make the connections that it desires. For, "If they are not so bound, connection and coalition disintegrate into orgies of moralism."22 - something profoundly familiar to post-situationist shufflers. It is this thickness of connection in the work - a situatedness of lived relations instead of identity - rather than any access to primal realities, that we believe allows the combination of machines and people to make concrete both a refusal of slavery and access to life. INSTRUCTIONS In search engines using automatic indexing to produce inverted files - the number of occurrences of a word related to the number of occurrences of all words in the file - an apparent transparency is generated by the simple procedure of simply pulling the words directly from the sites that it logs. What is important, what is notable appears simply by virtue of being logged by the crawler. Inevitably, this flat perspective produces an outcry about quality control. For information science, "It is entirely possible that vast concatenations of information resources are just not susceptible to any really effective measure of control and access, at least not by any automatic procedures applied on a large scale"23 The web challenges not only the status of the document, the corpus, but of the index. This challenge is the result of an attempt to refuse, or at least avoid the burden of classification built into the architecture of the web. The Uniform Resource Locator, (URL) of the World Wide Web is an attempt to, "point to any document (or any other type of resource) in the universe of information"24 without a priori specification of the resource within any system of classification. Within this schema, anything that is digitizable can be stored, networked and located. The architecture of the WWW is an attempt to route round classification, or at least provide a situation were many such interpretative formations can operate without universalisation. This is embedded not just in the device of the URL but also in details such as the simple geek idealism of physical and logical styles, (where, in the latter, the site designer can specify that the individually determined settings of the user's browser can determine the way to display certain forms of text emphasis). The WWW attempts to produce a 'preconceptual'25 technology which in a similar way to a computer's memory storing both the data it uses and the programs the data is used with produces a field within which concepts coexist alongside the rules to which the field is subjected. Elsewhere there has been discussion of how particular modes of use of HTML have tended to lock down the potential of the network into an idealised form of paper-based media26. What is hoped to move towards here is a consideration of how the rules of formation that operate in the construction of search engines produce many particular overt or anonymous ways of making and using the web. It is not suggested that these are secret rules for the confabulation of the world beyond the magic gate internet into something that it is not, nor are they necessarily merely illusions, misinformation, bad habits of thought, but a series of perceptual devices which are themselves reliant on and interwoven with others - and which are as such open to reinvention. THE EATING Within computer science there is argument as to the verification of the accuracy of both hardware and software. Donald McKenzie whose work has focused on the sociology of 'proof' suggests that there are three identifiable positions in the debate: "...the formal, mechanised verification of programs and hardware designs; the denial that verification confers certainty akin to that conferred by proof in mathematics; and the advocacy of human, rather than machine, proof"27. There is a triangulation of tension here between the relative status of verification and proof and between mathematics performed by humans and by machines. Software and hardware are relied upon for matters, not just of life and death, but for financial and military operations. Inevitably then, the need to determine the correct expression of an accurate result - and within software, concatenations of millions of them - has produced, if not agreement as to what constitutes accuracy, at least variable standards of proof. (The Ministry of Defence for instance distinguishes between a 'Formal Proof' and a 'Rigorous Argument' in its specifications for the assessment of mission critical software28). These arguments are played out at a 'lower' level in the construction of search engines. However, the accuracy of results returned are impossible to check because of the proprietary closure of search engines' programming. They are also complicated by their interface between social and technical protocols. For users, getting sense out of search engines is empirically based, deductive rather than inductive. For search engines getting sense, or at least data, out of the web is based around three major strategies. Directories specifically trade on their particular take on the web. Yahoo! for instance gathers URLs, largely from people submitting their own web sites for classification, but also from their crawler. Each site is then classified by an employee who places the site into one of several thousand sub-categories. What started out as the ironically named 'Yet Another Hierarchically Ordered Ontology', but, being too revealing, swiftly became just another yelping logo.gif of capitalised joy, is able in this way to bunch results together into classes. Because the system refers to digital documents rather than books on a shelf it also allows the possibility of having sites stored in the cross-hairs of several different classifications. However, because they both sell themselves on and operate by a claim to be able to at once abstract the essence of all things and at the same time conform as accurately as possible to the expectations of their statistically idealised user, the terms on which they classify are rendered solely by a normalised idea of 'subject'. Just as search engines can only operate given the pre-standardisation of language and spelling already put into place over several centuries by printed documents, directories assume a conceptual normativity. The 'subject' or 'topic' of a web site is what it is really talking about, its essence. It is not classified by whatever conscious or unconscious machine propels its production; by its use of bgcolor #A5C679; by that which it makes allegorical reference to; or by that which forms the centre of its coming into existence but which it is unable to name. Constructing a directory is of course complicated by the existence of contradictory or completely incommensurable schemas, but this becomes a negligible problem after it is understood that the clear asymmetry between such systems and the world ensures that the only moment any directory becomes completely stable is also the moment that everything dies. One pact between terminal equilibrium and the sustained recursion of the-same-but-different is formulated as the theory of quality: of establishing a mechanism by which certain levels of scientific discipline; relevance; erotic intensity; rarity; legal, commercial or other forms of trustworthiness; timeliness and so on can be reliably guaranteed. Perhaps as simple would be to have search engines organised not by exactitude of match by subject, but by degree of absolutely determined irrelevance. If directories promise nothing less than the ultimate victory of common sense over difference, the two broad types of semantic network in use to produce search engines rely on far less explicit ways of codifying the web. Before establishing what these are, it is worth illustrating how a 'standard' search engine works. The Boolean approach to information retrieval is the most widely used. Links to documents are returned only if they contain exactly the same words as your query. Even if a site is related in some way, it would never be retrieved unless it contains the user's search string. In order to get better results from this, most search engines finesse things a little. The results from the inverted file are modified by factors such as: where the search term appears in the item - whether the string of characters appears in specially tagged areas such as titles or head, or whether it appears early in the body text of the html document; how close the search words are together, or whether they form an 'exact phrase'; the frequency of occurrence of search terms in the inverted file; whether the search terms appear as keywords in the metatags in the html file's head; how frequently the terms appear in relation to the document's length (greater frequency indicating a probable greater relevance). These techniques gradually move over into semantic analysis. For instance selection by the frequency of occurrence of words in general in which very uncommon words get heavier weight. (Although this is probably slightly unnecessary as their occurrence in the inverted file, coupled with its realisation in the search term is already 'unusual' enough). The two key methods of semantic analysis of the web are hierarchical- based on ordered word relationships defined by a database, or self-organising - based on the build up of evidence of search strings finding their correlates in practice. The most simple method of constructing a semantic network is to use a precodified arrangement of terms and their correlates - using a document like an expanded thesaurus to widen and clarify search possibilities29 by aiding query formulation through synonym relations between words and hierarchical and other relations between concepts. Whilst they come into being on the basis of automatic word sense disambiguation running the I, me, she, he, it, you, them, that, on things - the syntactical machine that allows and disallows conjugation with what word, with what it means - top-down, 'Hard' semantic networks aggravate polysemy in the sense that all 'Dictionaries advertise semantic discrepancies',30 which allow other diachronic and synchronic meanings, other takes to leak through. However, hard semantic networks are like dictionaries again in that even if effectively and constantly enriched by lexicographers, they still only really compose logs of past signifiance. At the same time, whilst their hierarchical organisation allows for an extremely coherent co-ordination of words from a single perspective they are unable to negotiate the variable status of words as they are used. Allegory, irony, metaphor, repurposing, the rich meshwork of language that mitigates against its petrification all go like woodcutters amongst the trees in the database, whilst utterly incommensurable figurations of language such as sacredness throw up errors, not just of computation, but of essence. One way of attempting to avoid as much disjuncture between the semantic model of the information referenced by the search engine and that of its users in the way commonly imposed by hierarchically organised systems is to assume that the words that are most commonly found together are actually related in some way. This is a technique that makes best use of machine rather than human interpretation of data. In Latent Semantic Indexing31 a set of documents is represented by a matrix whose entries are measures of frequencies of occurrences of terms in each document. The greater the frequency of co-occurrence, the greater the relevance. This matrix is then approximated by a sum of rank-one matrices determined using a singular value or related decomposition. This allows for great compression of the data and fast, relatively robust and encompassing referencing of sites. However, what is found to be the most related in terms of quantity doesn't necessarily match for each user, or each search. Latent Semantic Indexing finds a way of working the inverted file mined with homonyms, synonyms and routes to a thousand pointless choice trees that plays to the data processing strengths of computers and thus avoids the implicitly ontological nature of directories. However, these techniques are always caught in a self-organised conformism to the homeostatic pressures of what is statistically deemed, by frequency of use and by frequency of co-occurrence, to be the most relevant. Searching for the most relevant, the most related - simply rendered as the most linked to is another technique adopted by a variety of search engines in compiling their results from the inverted field. AltaVista, Excite, Infoseek, Inktomi32 rate sites higher in their list of search results if a substantial number of sites link to it. Google, usefully makes this process visible and useable, allowing backlinking from its results. Related to citation analysis - the scheme whereby the relative importance of scientific papers is accorded by the frequency of their being cited by other scientific papers - is another approach adopted in different ways both by the Web Stalker33 and by the 'Clever'34 project at IBM: "We have developed a new kind of search engine that exploits one of the Web's most valuable resources - its myriad hyperlinks. By analysing these connections, our system automatically locates two types of pages: authorities and hubs. The former are deemed to be the best sources of information on a particular topic: the latter are collections of links to those locations."35 'Authorities' are sites that are linked to many times, 'Hubs' are sites that contain many links.36 Despite their claims, Clever has no way of actually defining a topic other than by resorting to imposing hierarchical taxonomies. What it does do is - by defining a limited set of properties: links-to and links-from, as a way of organising an interpretation of the topological form of the web - use some of ways the web is actually put together to develop a representation of the emergent, unpredictable interplay of multipolar gravitational tendencies within the hyperlinked structure of the web. Some of these may be sufficiently closely clustered to be interpretable from specific perspectives as a topic. Other sites which might be highly pertinent from other perspectives might not link at all. What Clever has usefully done though is to distinguish the structure of the web from its content and use it as an organisational principle. What they collapse in their descriptions of the project is the distinction between the hierarchical pretence to objectivity and what they actually represent which is an aggregate of subjective decisions by the thousands of people making sites to incorporate hyperlinks. Here, 'objectivity is less about realism than it is about intersubjectivity'37 Self-organisation as a phrase may sniff like anarchist democratics, but it is in the choice of qualities, terms, devices, elements to register as the evidence for, or as the most consequential part of, the self-organisation of this data that the political choice - closed to scrutiny or reinvention - as to what is important on the web or about the web is made. It is a symptom of a self-reinforcing 'misplaced concreteness' that the mode of interpretation of processes becomes the means by which they are molded. Go to search engines of either sort and type in Jamaica. You'll get information on holidays coming up first -along with banner adverts for Sandals resorts. Type in Africa and you'll predominantly get wildlife and famine. Self-organisation of data is organised on the basis of what 'self' is determined to be important. This, at present, is something put into place by the demographics of internet use. How data is interpreted and processed, how (whether it is hierarchically ordered, or ordered according to emergent compositions of specified elements) the grammar by which the strings of symbols are sorted and ordered - what is possible to come into composition with what - particularly in a situation in which most of this work is done by a closed culture of proprietary specialists, has immense importance. Finding ways in which this process can be opened up to speculation and organisation by other forms of 'self' even more. Blah, click, Blah, click, Blah During the first batch of journalistic excitement over the web as a phenomenon there were many articles of the 'around the world in sixty minutes' variety. Someone sits at their desktop and links their way round the web. Despite the barest relation of geography to the act of calling up HTML files on a computer it did speak of the enthusiasm generated by a device that could incorporate the most distant elements within one interface. Most of these articles were written at a time when there were no search engines, (the only thing that came close at this stage were, sometimes vast, lists of ordered and annotated links) so the 'travel' had to be done by linking from one site to another by links that had been constructed on a site-by-site basis. The fascination generated was that solely by virtue of this meshwork of links, and even in the early stages of this technology, a path could be constructed through extremely distant resources. The distance in this case was supplied by geography. The distance now which provides the most amazement is that of subjectivity. ASCII is text in its mobile form. The standardised procedure of writing as a sequence of claim-argument-conclusion becomes extraordinarily distributed, messed with. The consequentiality of sequence becomes, not unimportant, but subject to a richer field of influence. This offer of suprise or detail or resource, this at least potential, richness produces the condition of the link, how it is arrived at, what it articulates. Safeways Establishing gateways between different elements of networks is an essential part of communications practice. From the development of interactive interfaces to mainframes, to the subnetwork of Interface Message Processor nodes that enabled the establishment of ARPANET and on to the creation of every hard and soft device and protocol, the history of network computing is that of increasing the circle of inclusion of hardware, software and data-types. How that connection is managed, what protocols channel it are the focus of antagonism in debates that see themselves as being primarily technically oriented. However, of course, although the crucible that these conflicts come to heat in is that of technics they are doped with many other traces. What interfaces allow, what they bring to bear, what elements they move into proximity or remove from the chance of coming into composition, what assumptions they make about the formations that move through and with them, what patternings they imprint into the way in which things are ordered or establish terms of composition, all exist and establish conditions at the logical, semantic, syntactic, social, economic, aesthetic and political levels. Just as, within an economy of increasingly immaterial labour, and within a logistical context of just in time delivery the gateway to the factory is where much of the inward and outward processing of information and materials is now done; or where the interface to companies has been stripped down to a telephone number connected to a database by a voice (a voice that, even though itself coming into being through a stripping away of commerce from the street, will become increasingly redundant as sales carried over the internet increase), the point of access becomes crucial on the nets. Search Engines work on the basis that they can turn any site into something only one click away from their search results, almost a subsidiary section of themselves. The development of portal sites is a way of turning the internet from a distributed network into a centralised one. Hierarchies such as the Name Space, or its categorical inverse on Usenet were relatively late in interpelating the meshwork, allowing every object on the net to have a unique address accessible from a central root-node. Node power is intensified the higher you climb. This remains practical and unproblematic, so long as the system is open, with location and structure visible and interrogable, at every point. However there is an increased tendency for the visibility to be locked up. (On a small though crucial scale, some shows of net-based art have blocked the display of URLs from the link status messages at the bottom of the browser window and in the location bar. On a larger basis though, compelled by the same drive to increase dependency of users on portals, browsers are being developed that tie in to specific search engines and built-in results to enquiries on 'key' topics. (Equally there is the potential once the industry 'matures', with the formation of effective cartels, for limits to be placed on the speed of accurate linkage to maintain multiple retutns to the portal within one search session). The possibility of capture of such root nodes becomes not just a matter of control but, on a temporary basis, also one of decontrol. Slipping the wrong signal into the centre of a hub - such as when, in early July, Ivan Novkovic famously spiked TV coverage of the Germany v Yugoslavia football match with a call for demonstrations against Milosevic in his home town of Leskovac - offers immense possibilities for cutting away the appearance of order. At the same time, it is pretty much the same tactic used by sites that spam search engines to climb up result hierarchies by including reams of invisible keywords, (the same colour as the background for instance) or have multiple editions of a site with different metatags and spam content. network of stoppages Control over the way in which links occur, the way in which they are interpreted, has become a key issue for web design. A list of ways in which it is suggested that sites should use links echoes the approach which is becoming widespread: "Guidelines for effective linking - Links cannot create content, but they can ruin it - Links should reinforce your message, not replace it - Most links should point into your site, not away from it. - Most links should appear as footnotes, away from the main text - Links to outside sites should open a new browser window - Every link is a maintenance issue, link sparingly if at all"38 Each web site is to become its own portal, feeding only into its own. Every site an intranet. Whilst on the web we are urged at once that 'Everything must be published!' any form of external conjunction must be done with immense prophylactic care. (The Clever project's mapping of clusters of pro-abortion and anti-abortion web sites showing their close self-interrelation and extremely rare cross-linking provides a good example of this). Of course this reads as pretty much the exact inverse of most attempts to invest links on the web with various kinds of poetics. Read it through again, but kind of backwards. Make a web site that does exactly the opposite of what the list instructs - top art mate. Guaranteed. It is at the same time the very limited nature of the link within Hypertext Transfer Protocol that makes both takes on its use possible. The link in normal use is extremely narrow in terms of the information it allows about itself or in the transition from one anchor to another - both allowing the actuation of the link to result in a suprising conjunction, or one that is completely expected. A link in a web site places it in a material not automatically allusive, or even referential, conjunctive relation to other documents on the web. Steven Johnson sharply places its particular role "The link is the first significant new form of punctuation to emerge in centuries.'39 Being, within HTTP at least, simply a form of punctuation frees the link up from some of the portentous baggage strapped to it. Much has been eleborated on the basic theme of hypertextual linkages as at once a form of augmentation of text but also of readerly poaching of inventive power over the text. However it is clear that the technology has additional possibilities. Rather than hypertextual ceding of control to the reader there is the chance to establish the power of contextualisation, the power to follow and establish in advance any argument or use of the text that strays. To establish that the author has already 'gathered' that these allusions or threads of argument could be made. No association by the reader need worry about going unprecedented within the matrix of this neurotic urge to completion. It is clear from the gluttonous expansivity of the leavings of some writers on the web that hypertext can become as richly detailed and engulfing as Piranesi's fantasy prisons. However, if we start with an acknowledgement of the simple way in which links are formed and the very restricted nature of their movement from one anchor to another rather than an imposition of a literarily inspired compulsion to recapitulate particular models of thought as fact, things have the chance to free up a little. In most browsers there is a pop-up menu which allows links or other resources to be opened, saved, forwarded and so on. This itself is a start at making some form of expanded range of modes of use available - including of course expanded terms of readership, composition, re-use and storage familiar from hypertextual theory. Whilst links are only one of the range of objects that the menu operates on, it might be possible for instance to expand this range of uses to include some expanded sense of what links might be. Presently though, what the list of constraints on 'good' linking above suggests is that as the degree to which meaning can be attached to or implemented by a link - at its simple level of conjunctive punctuation - how much it brings together, to what extent it vears towards fragmentary 'disassociation' or togetherness between both anchors depends on the degree of framing of the link. The link works along a paradigm of expectation from shock to the already anticipated. This paradigm is drawn along the line of indexicality. The common use of Frames to index sites perhaps sits at one end of the paradigm: with every subsequent link being kept within the tutelage of the primary indexical frameset. Somewhere near is the indexical link from a to b, internally within documents, or to external sources. These unilateral associations however immediately become complicated when there is a difference in any of the various ways the sites are formulated. This can be in terms of design; of technical level (which browsers of plug-ins they demand); on ideological grounds, or at least those of public appearance: for instance the politely racist conservative Southern League provides, for the benefit of users linking from the openly Neo-Nazi Stormfront, a disclaimer of its white suprematism in favour of just wanting things the way God ordained down South. At other times you can see something similar when cozy simple homepages showing pictures of a family's house, the people in the house and snaps of what they look like and like doing include portentious links to 'the search engines we recommend'. There are links operating on the basis of providing routes through sites; hypertextual uses of links based on allusion or the choice of progressive pathways through multilinear naratives; malign links; antagonistic links; supportive links; links on par; links to gain status by association; psychopathic links - you can be anything, the link could be to everywhere; links traded between sites solely to boost the appearance of advertising revenue in annual reports. All of these are, because of the limited nature of the link, based on a dialogue between the interpretation of the two conjoined elements These are rather 'Hot' uses of the link though. Their use may equally be blasé, shut off. The relatively withdrawn process of the search hijacked by Natural Selection is just as much run on indifference - in the same way that link actuation is a quick way of avoiding downloading immense gifs or applets. One use of links that acknowledges this, but also punningly overcomes it, is the use of links to create innuendo. (Steven Johnson describes this particularly well with regard to Suck) The perpetually tripped up slapstick of meaning also forms the basis by which much of the work in Natural Selection operates40. Somewhere further towards the latter along the expectation-suprise paradigm, it works on an aesthetic of addition, the piling up of multiple threads of implication, double-entendre and self-undercutting that forces the user, if they can be arsed, not into a smooth plateau of inference and lovingly fondled argument iterated in abundance, but something thicker, something that is more prone to, whilst attempting to shoot itself in the foot, blow off its head41. Information transference interrupted Earlier, the possibility of using software or types of use of software or of searches could be developed as a form of exploratory laboratory illness was suggested. Transference, the displacement of affect from a repressed condition onto the artificial figure of the analyst, synthetic stress produced as a means to operate at a more abstract level on the underlying problem. "The transference thus creates an intermediate region from illness to real life through which the transition from one to the other is made. The new condition has taken over all the features of the illness; but it represents an artificial illness which is at every point accessible to our intervention."42 As a technique it encourages the same investment of emotion and involvement that allowed Joseph Weizenbaum's program ELIZA43, a sentence generator that famously mimicked the responses of a nondirectional psychotherapist, to convince users that they were effectively undergoing an initial psychiatric interview. Transference is marked by an absence. It can be a substitute for an experience, a 'something' which cannot be expressed, that blocks comprehension, the analyst investigates what is inexpressible - allows a latent conflict to become a current one under experimental control that can be interrogated. This in its 'positive' form is the very 'unsayable' living truth that Weizenbaum later mobilises against instrumental reason, the mental strictures of a technocracy which claims that, "The only reason we have not yet succeeded in formalising every aspect of the real world is that we have been lacking a sufficiently powerful logical calculus."44, as an ineffable human quality also moves in the opposite direction as the very means of locking people into rules, habits, obedience. A recursive phenomenon, transference was the mechanism by which Freud hoped to make of psychoanalysis a science.45 At the same time, the possibility of effecting transference was reliant on the authority vested in psychoanalysis under the sign of Science. ELIZA's ability to effect transference to itself equally relied on the idea that it was "ŠClothed in the magic mantle of Science and all of Science's well-known powers may be attributed to it."46 Once properly in the hands of the analyst, whose attention, whose favour is so much desired the supposed difference between ordinary suggestion and analytic suggestion is the calculable and controllable version of the latter. The patient can be induced, by its virtue, to specify their desire by exact phrase, by name of person, to constrain within certain languages or dates, to search only for certain pre-ordained data-types, to associate from any of the specified words. The pen on the pad turns the dial of the mental radio being tuned back in to the Voice of Oedipus. Whilst the analyst is therefore supposed to be able to remain outside the scene that their role creates, using this device in order to deny and ward off the fragmentary and incomprehensible fullness of the world through an identification with the Master, with the Law. His brain is become as an interview room. A single bulb shines down on two chairs across a bite-marked table. As Freud acknowledges in 'Analysis Terminable and Interminable' transference was an idealisation, a fictional tool that provides a way of working through messy life. It is the resistances, skids of meaning, clarifications, pun-loops that provide transference with its opening chance, but also its interminability, that we can take as something more, as the material for composition. Information transference here becomes a kind of laboratory sickness in which conceptual, disciplinary, juridical, cultural, political and other 'norms' or expectations of meaning and linkage - the switch tongue bound, made orderly - can also be adopted. Each search, but also each piece of software can usefully be approached as a transferral, a synthetic agglomeration of knowing, sensing and doing. When you drive a car your mind fills out to the space occupied by the vehicle. You sense its shape in order to manoeuvre safely, or however. Equally, conceptual proprioception can be elaborated in order to negotiate the multiple levels of meaning making in a search. The user mobilises and becomes infested with, composed through, flocks of sensoria, a billion symptoms, neurosis at the service of knowledge, or a simple sticky or slimy crawl of the switch-tongue at the point of slipping away into babble and the learning of language 1 http://www.mongrel.org.uk 2 There are also archive of sections of the net - the search engine DejaNews is working towards an archive of all usenet posts going back to its inception in 1979. 3 see info on the Echelon system http://caq.com/CAQ/CAQBackIssues.htm 4 http://www.mongrel.org.uk. This project (Coordinated for Mongrel by Harwood and MF) essentially presents itself as a straight search engine. When however any of several thousand words which are either directly racist or which have racialised connotations are entered as search strings, the Natural Selection search engine returns innocent looking results (those from whichever search engine it is currently sitting on top of) that when the link is actuated drop the user into one of the over thirty other sites on and off the mongrel domain. Collaborators in producing the work include: Hakim Bey, Byju, Critical Art Ensemble, Stewart Home and Daniel Waugh, Mervin Jarman, Richard Pierre-Davis, Dimela Yekwai, and others 5 http://www.mongrel.org.uk/BAA a site produced by Mervin Jarman 6 http://www.mongrel.org.uk/Venus a site produced in collaboration with Dimela Yekwai 7 Claude E. Shannon, A Mathematical Theory of Communication, The Bell System Technical Journal, Vol. 27 p379-423, 1948 8 Sigmund Freud, 'Transference', Introductory Lectures on Psychanalysis, the Penguin Freud Library vol 1. eds. Angela Richards and James Strachey, trans. James Strachey, Penguin, London, 1991, p. 482 9 G. L.J Chaitin, Algorithmic Information Theory, Cambridge University Press, Cambridge, 1987 10 Mark Poster, The Second Media Age, Polity Press, Cambridge 1995 p.86 11 The Temple of Confessions website is at: http://www.sfgate.com/foundry/pochanostra.html 12 Guillermo Gomez-Pena, interviewed by Mildred Thompson in Glenn Hapcott ed. 'Interventions and Provocations', State University of New York Press, New York 1998, p6. 13 McSpotlight's (http://www.mcspotlight.org) most successful element was its use of frames to directly comment on the official McDonald's web-site. Irational (http://www.irational.org), renegade teletubbies fans and other groups such as the Barbie Liberation Front produced ostensibly 'real' sites on behalf of the companies and culture products they loved the most. Whilst even honest to goodness fan-sites are often compelled to mark themselves as 'unofficial', some catalogue-sites have been known to stage 'hacks' of themselves in an effort to gain press-coverage. 14 Guy Debord, 'Report on the Construction of Situations and on the International Situationist Tendency's Conditions of Organisationa and Action', in Ken Knabb ed. and trans. 'Situationist International Anthology', Bureau of Public Secrets, Berkeley, 1989, p.17 15 Donna Haraway, Modest_Witness@Second_Millenium. FemaleMan©_Meets_OncoMouse, Routledge, London 1997, p.199 16 Debord, Ibid, p23 17 Haraway, Ibid,p3 18 Situationist International, Preliminary Problems in Constructing a Situation, in Ken Knabb ed. and trans. 'Situationist International Anthology', Bureau of Public Secrets, Berkeley, 1989, 19 Situationist International, Ibid 20 Situationist International, Ibid 21 The couple which have become charmed suitors in the institionalised narrative of 'electronic art'. 22 Haraway, Ibid, p199 23 Frederick W. Lancaster, Indexing and Abstracting in Theory and Practice, 2nd Edition, London Library Association, London 1998, p.313 24 Tim Berners Lee, www: Past, Present and Future, Computer 29, October 1996, vol.29 no.10, p.69-77 25 Michel Foucault, The Archaeology of Knowledge, trans, A.M. Sheridan-Smith, Tavistock, London 1972, p.62 26 see A Means of Mutation, notes on I/O/D 4: The Web Stalker 27 Donald McKenzie, Negotiating Arithmetic, Constructing Proof: the sociology pf mathematics and information technology, Social Studies of Science 23 1993 37-65 p.56 28 McKenzie p.57 29 Visual Thesaurus, a good example of a visualisation of this process can be seen in all its ludicrously rigid beauty at: http://www.thinkmap.com This project is an interface to the WordNet lexical database at http://www.cogsci.princeton.edu/~wn/. WordNet operates basically by constructing hierarchically ordered sets of synonyms. see Christiane Fellbaum ed. WordNet: An Electronic Lexical Database, MIT Press, 1998 30 Walter J. Ong, Orality and Literacy, the technologizing of the word, Routledge, London 1982, p.46, 31 Latent Semantic Indexing, see: http://superbook.bellcore.com/~remde/lsi/ 32 source: http://www.searchenginewatch.com - a recommended site for a basic introduction to search engines 33 http://www.backspace.org/iod The Web Stalker by contrast, avoids these problems by its representation of the network as a flat aggregate of links and nodes composing itself on the screen rather than retranslating it into an ordered list. The Web Stalker does not assign importance or greater degree of relevance to sites, merely relaying the greater number of links to them by increasing luminosity of the node. Clever is more sophisticated however in terms of its mapping simply by the analytical tools it is able to bring to bear on the information. 34 http://www.almaden.ibm.com/cs/k53/clever.html 35 Hypersearching the Web, The Clever Project, Scientific American, June 1999 or at: http://www.sciam.com/1999/0699issue/0699raghavan.html 36 J.Kleinberg, Authoritative Sources in a Hyperlinked Environment, in Proceedings of ACM-SIAM Symposium on Discrete Algorithms, 1998 37 Haraway p.198 38 Patrick J. Lynch, Sarah Horton, Imprudent Linking Leaves a Tangled Web, Computer, July 1997, Vol. 30 No. 7, p.115-117 39 Steven Johnson, Interface Culture: how new technology transforms the way we create and communicate, HarperEdge, San Francisco, 1997, p.110 40 see the work by Critical Art Ensemble: http://www.mongrel.org.uk/Biocom, or by Stewart Home and Daniel Waugh, http://www.mongrel.org.uk/Skrewed 41 see for instance: http://www..mongrel.org.uk/NSA or http://www.mongrel.org.uk/Agent 42 Sigmund Freud, Remembering, Respeating and Working Through, in The Standard Edition of the Complete Psychological Works of Sigmund Freud, ed. James Strachey, vol. 11 p.155 43 Weizenbaum, Ibid 44 John McCarthy, cited in Weizenbaum, Computer Power and Human Reason, from judgement to calculation, Pelican, London 1984, p.203 45 Isabelle Stengers, Black Boxes, or Psychoanalysis a Science? in Power and Invention, situating science, University of Minnesota Press, Minneapolis, 1997 46 Weizenbaum, p.191 # distributed via <nettime>: no commercial use without permission # <nettime> is a moderated mailing list for net criticism, # collaborative text filtering and cultural politics of the nets # more info: majordomo@bbs.thing.net and "info nettime-l" in the msg body # archive: http://www.nettime.org contact: nettime@bbs.thing.net