Patrice Riemens on Mon, 26 May 2014 15:53:16 +0200 (CEST) |
[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]
<nettime> Ippolita Collective, In the Facebook Aquarium Part Two, |
Ippolita Collective, In the Facebook Aquarium, Part Two The Libertarian World Domination Project: Hacking, Social Network(s), Activism and Institutional Politics Social networks as seen through the anarcho-capitalist lens - or the management of sociality through Big Data. (section 4, continued) Profiling marks the pledge for automatic, instant freedom: contextualised ads, research into the users' sentiments so that everyone gets a personalised, 'bespoken' ad, so as to promptly realize a click-thru sale, followed by the disposal of the purchase as soon as possible in order to ... order a new one. We, the users, are in this all 'suspects' of sorts, whose most intimate details must be known so as to satisfy our compulsive craving for ever new, yet instantly obsolete objects. The fake issue of confidentiality is also regularly being bandied about, but the issue itself seems only to arise for real once confidentiality has been breached. This usually goes together with shocked, shocked rants about the shameful and pushy immorality of a system that divides people into categories. And high-octane paranoid (conspiracy) theories are rife in the era of Big Data. But the real, much more concrete, and also much more concerning issue is about the individuals themselves, not the amorphous mass we are all part of. On the one hand, people want to be profiled, and on the other, whatever we do in order to avoid profiling, our digital footprint sticks to us like glue: no way we can opt out once enlisted in the army of the data-feeding-data-suppliers, aka /prosumers/ (all at once producers and consumers). For some time already, a massive debate has been raging about the gross abuse of the /data mining/ that takes place to make profiling possible [31]. New lines of digital discrimination are being created, related to the degree of access: which researchers, which institutions, which groups have actually both the means and the opportunity to put these data to use? What are the rules, what are the limits - and: who decides? Here is not the place to go into all these issues in detail [32]. Let's stick to our main point. This is not about going against progress (and its promise of a brilliant future), nor to escape into ludism, or into its exact opposite: crypto(graphy). To hide serves no purpose, neither does the refusal to make concessions to the present (order of things). What should be done is to get a clear understanding of Big Data and profiling, in so far as they are part and parcel of concrete strategies to arrive at a society shaped by anarcho-capitalism, the ideology according to which everyone is 'free' to ransack everybody else. We do not dig this 'utopia', for sure. We rather would call it a dystopia of control and auto(self)-control. (Yet) We are, bit by bit, and very swiftly nonetheless, going from a world endowed with signification, rich as it is with relationships we are developing for our own benefit, to one that finds significance only through relationships (pre-)determined by machines. It looks like as if we no longer need neither theories nor practices that are grounded in personal belief and validated by personal experience. The status of knowledge is thereby transformed, now that data are supposed to speak for themselves. Knowledge suddenly becomes self-evident and impose itself as a certainty. Statistical correlations determine existing links between things and direct relationships between people. We do no longer shape a discourse: the data have taken over the floor. This is the pipe dream of a society ruled by data, where the role of the human subject has been blanked out to all practical purposes. All the human remnant needs to do is to obediently accept to be 'freed' (of everything), including from the possibility to choose and to desire. Give us ever more powerful machines, hand over all your data, be transparent with the machines, and we will be able to foretell the (radiant) future - the future of the market, of course. We fly above the world, we observe it from the outside, we see oceans of data, expanding at a vertiginous rate, only to be swamped by tsunamis of social crazes, as sudden as fleeting, occupying all available space before making room for the next upstart. Mass sentiment can be analyzed, and the aggregate opinion is easy to distill by way of /sentiment analysis/ and /opinion mining/ [33]. All this while we, as enthusiast and willing victims, love to be 'free' consumers: generalised, global recording is the price to be paid if we want to be truly 'free' to choose. The algorithm will tell us what we really want: it already advises us on which book to buy on Amazon.com, it edits our searches on Google, it suggests us which just-out film we should see, and it tells us which music best suits our taste. It is an algorithm which points out our potential friends on Google+ and on LinkedIn, and also those subscribers we might want to follow on Twitter. Algorithms are paying attention in our stead, and encourage us to socialise the right way. (Before soon) It will be no longer necessary to desire whatever what, since the algorithm will see and foresee for us. Which will be the equivalent of seeing with 'the eye of God', who is able to read the future in the crystal ball in which the information deluge rages. Open your heart, let your body be cut and formatted in usable chunks, speak out your mind, tell us where you are (now), what you are doing (there) and who is your current company. Don't think, say it all, now, and you will obtain all you desire, even without knowing yet what you actually desire. Inexpressible vertigo (in the literal sense of what 'cannot be expressed'), infantile enthusiasm (in the original sense of 'infans': the one who doesn't talk yet), mystical ecstasy in front of the Matrix uncoiling under our very eyes. The words and the pictures referring to Big Data often take on a religious tone, and that a bit too often to be just fortuitous. The fetish dangling behind the knowledge society of the Big Data belongs to a populist, techno-fascist religiousness, since once a sufficient quantity of data has been gathered, any hypothesis can be proven. Just as with the Bible, the Q'uran, the Torah, or any sacred book, the scope for interpretation is endless. And it is precisely because Big data embrace incredibly more stuff, that each and every assumption can be held up and proven. Statistics vouch for everything, but prove nothing, they are allegedly scientific proofs of highly ideologic presupposition. (to be continued) Next time: real and digital life explained ... -------------- [31] Dino Pedreschi et al., "Big data mining, fairness and privacy. A vision statement towards an interdisciplinary roadmap of research" ?Privacy Observatory/ 2011 http://www.kdnuggets.com/2011/10/big-data-mining-fairness-privacy.html No wonder mainstream publications like /The Economist/ call for more transparency for the sake of security, confusing thereby security with extension of control and surveillance - perhaps deliberately so. http://www.economist.com/node/15579717 "The data deluge: Businesses, governments and society are only starting to tap its vast potential" [32] For a good approach to this issue see Dana Boyd & Kate Crawford: "Six Provocations for Big Data", /A Decade in Internet Time: Symposium on the Dynamics of the Internet and Society/, September 2011 http://papers.ssrn.com/sol3/papers.cfm?abstract_id=1926431 [33] The last-born device for the automated analysis of sentiments and opinions by way of semantic systems (at time of publication of the original book in Italian -transl) was the Sentics Computing: http://sentic.net ("Helping machines to learn, leverage, love.") WKP entry on Sentiment Analysis: http://en.wikipedia.org/wiki/Sentiment_analysis (for recent developments and a lot sources) ----------------------------- Translated by Patrice Riemens This translation project is supported and facilitated by: The Institute of Network Cultures, Amsterdam University of Applied Sciences (http://networkcultures.org/wpmu/portal/) The Antenna Foundation, Nijmegen (http://www.antenna.nl - Dutch site) (http://www.antenna.nl/indexeng.html - english site under construction) Casa Nostra, Vogogna-Ossola, Italy # distributed via <nettime>: no commercial use without permission # <nettime> is a moderated mailing list for net criticism, # collaborative text filtering and cultural politics of the nets # more info: http://mx.kein.org/mailman/listinfo/nettime-l # archive: http://www.nettime.org contact: [email protected]