Posted Thursday, 25 Jul 2019 by Antonio Cerella

āThink Privacyā
Public Service Announcements by the Privacy Gift Shop
©Adam Harvey 2016
Adam HarveyĢżis an award-winning artist and researcher based in Berlin. His work has been widely covered in such publications as the New York Times, CNN and the Huffington Post, and has also been cited by critical theorists such as GrĆ©goire Chamayou and Derek Gregory. Harveyās work explores the societal impacts of networked data analysis technologies with a focus on computer vision, digital imaging technologies, and counter surveillance. In particular, Harvey is interested in the question of how fashion can be used to address āthe rise of surveillance, the power of those who surveil, and the growing need to exert more control over privacyā (Harvey 2012).
To quickly summarize his career thus far, Harveyās projects include CV Dazzle, camouflage from computer vision; Stealth Wear, his āAnti-Droneā fashion range, and the Privacy Gift Shop, an online marketplace for counter-surveillance art. For his Stealth Wear collection (2012), Harvey fabricated a set of clothing inspired by traditional Islamic dress from silver-plated fabric that reflects thermal radiation, enabling the wearer to avert overhead thermal surveillance. The metal-plated fibres reflect and diffuse the thermal radiation emitted by a body, thereby reducing the wearerās thermal signature under observation by a long wave infrared camera (Harvey 2012). This range of anti-drone clothing is available for purchase online, alongside other works of counter-surveillance art, at Harveyās Privacy Gift Shop (. In the years since it has been produced, the Stealth Wear collection has been the subject of features in the New York Times (Wortham 2013), Washington Post (Priest 2013) and Der Speigel (Backovic 2013).
If a range of artists such as Trevor Paglen, Mahwish Christy and the collaborative art installation #NotABugSplat ( undoubtedly address similar questions of the ethico-political status of drone warfare, surveillance and data collection, what arguably distinguishes Adam Harveyās work in this field is its specifically non-military, or domestic, focus. GrĆ©goire Chamayou (2015: 203-4) provides this description of his work in Drone Theory: āOne of the questions that arises is whether societies that have, for the time being, failed to rule out the use of this type of technology in wars waged on the other side of the world will eventually realize, perhaps with a jolt, that this technology is designed to be used on them too, and whether they will mobilize themselves to block its use. For it is important for them to become aware that a future of video surveillance with armed drones awaits us if we donāt prevent it. As a last resort, there is always the possibility of purchasing anti-drone clothing, such as that invented by the artist Adam Harvey. It is made from a special metallic fabric that renders the body practically invisible to dronesā thermal imaging camerasā.
In his interview with Antonio Cerella, Adam Harvey discusses the conceptual origins and reception of his work, its relationship to critical theory on drones, surveillance and biopolitics by figures like Chamayou and Gregory, the growing militarization of the domestic sphere and, more widely, how art can be used to ā playfully ā resist or challenge what he calls the fatalism of living under the gaze of domestic drone surveillance.
Antonio Cerella (AC): In recent years, your work has become increasingly influential in theoretical and philosophical circles ā GrĆ©goire Chamayou refers to your anti-drone camouflage in hisĢżDrone Theory (Chamayou 2015).ĢżTo what extent are you aware of, or interested in, contemporary theoretical debates about biopolitics, drone theory, surveillance and security studies? Do they inform your work at all?
Adam Harvey (AH): Theoretical debates around surveillance are of great interest to me, but Iāve intentionally unhinged myself from thinking through the more historical part of these debates because I think the current state of surveillance and security is mostly unprecedented.
For example, Benthamās Panopticon is of limited value in thinking about how social media is designed to accumulate data thatās used to train neural networks which are eventually licensed to government surveillance programs. And the Orwellian concept of Big Brother seems to anesthetize more complex discussions about computer vision being used to infer sexuality, criminality, or oneās psychological state.
As an artist, I see an opportunity for imaging new ways to represent the current state of surveillance, if it can even be called that. We are being observed by drones in many ways that we canāt know. And my project Stealth Wear (Figure 1) was about asking what are the limits of perceptibility ā how do you know how weāre being looked at?

Figure 1 Stealth Wear: āAnti-Droneā Burqa
Photo: ©Adam Harvey 2013
Materials: Burqa with silver-plated rip-stop nylon exterior and black silk interior.
Credits: Engineered by Adam Harvey. Designed by Adam Harvey and Johanna Bloomfield.
Model: Tate
One of the biggest challenges in discussing topics related to surveillance technologies is to avoid becoming fatalistic. My work aims to imagine new and expressive ways of adapting to this environment. In my own work, there is an object attached to it that is a ādiscussion-generatorā that people can easily refer to. From my perspective, itās important to use wording that is appealing and provocative, and to frame the work so that there is an overlap with different circles, because the most important thing is what happens next, once the project is activated.
Only after working on these projects have I discovered, and enjoyed, becoming involved in actual contemporary discussions about drones, which Iām grateful to be included in. For me, experimental art projects become a gateway to other forms of related knowledge.
AC: An important theme informing much of the research on drones is the question of what happens when military drones come āhomeā and begin to be deployed by domestic law enforcement agencies, immigration and border patrol services and so on. According to Chamayou, your work addresses the kind of society which has finally realized that the drones they are using in other parts of the world are designed to be used on them too ā that a future of civilian surveillance by armed drones is here. For example, it has recently been reported that the Metropolitan Police in London are now using drones equipped with thermal imaging technologies for counter-terrorism purposes and, as Arthur Bradley and Oliver Davis discuss in their contributions to Security Dialogue, last year in Dallas the police killed a suspect by using a robot bomb (see also Graham 2016). Could you say something about how your work addresses this āmilitarizationā of the domestic sphere? Does it challenge that process? Or perhaps even reinforce it by encouraging civilians to adopt processes of camouflage?
AH: I think the reality is that dense urban areas have become military sites. Itās not a pleasant thought, but itās how adversaries have adapted to asymmetrical military power, by advancing their mission into vulnerable non-militarized areas. The response from national security agencies has been to infiltrate communications, and to create deterrence through surveillance accountability. In New York City, where I previously lived, armed soldiers, bomb-sniffing dogs, and cell phone interceptors are a regular part of life. In 2012, when I began research and development on Stealth Wear, it seemed clear that drones were an inevitable next step once the economics worked out.
In my work, I encourage people to engage with, but essentially resist militarization, by adopting camouflaging techniques to modulate their visibility and hopefully mitigate the negativity of this militarization. The big challenge is to confront the preexisting narratives of privacy-paranoia.
So, the tension is: ādo I accept this garment at the risk of being labelled paranoidā or ādo you embrace it as a fashion-forward concept?ā I tell people that you may only wear a tuxedo once or twice a year, but itās still a great garment. So, if your rationale is that youāre okay with this garment that you wear once a year, then maybe itās okay that you wear a camouflaging garment once a year if you want to evade thermal surveillance. But I guess itās not that simple because itās still tied up with the dangerous idea of āwhy do you want to hide?ā, instead of thinking about it as an act of protection or defense. In some ways, thatās what fashion is: a kind of defense against categorization. Fashion gives you control over how people interpret you.
AC: What do you mean by protection against categories and categorization? It seems that this is related to your current work as well.
AH: Fashion gives you the ability to role-play and to shape-shift between social classes. By appearing in different garments, you can convince people that you belong to a different social class. There are different types of camouflage and one of them is social camouflage where I try to blend in to a crowd of really smart people when Iām not smart by wearing smart clothes. Or the opposite, that you dress down as an undercover agent wearing a T-shirt, carrying a newspaper, with a baseball hat on. These are equally valid forms of camouflage that arenāt seen as military, but can be still used in military ways: covert agents deploy a very basic street fashion to camouflage them. Fashion is an integral part of deception, both for civilian and military appearance.
AC: Why is the idea of camouflage so important to your work? In your work, it sometimes seems as if we canāt really stop or block surveillance as a society ā the best we can do is hide ourselves as individuals or even run away.ĢżIs this the only solution to the kind of categorization youāre talking about?
AH: I agree that we canāt stop surveillance. I think this would be an unattainable goal and this is where surveillance fatalism re-emerges. Itās not possible to opt out or hide completely. Framing it this way guarantees failure. However, there is still a lot of room to be expressive and successful in defeating, blocking, or reducing certain kinds of surveillance. In a way, both fashion and camouflage are about staying one season ahead of the latest trends.
There is an interesting story about the word ācamouflageā. In the early 1900s, Theodore Roosevelt considered it as a āform of effeminate cowardice, a mere defensive strategy [that] all but announced an unmanly desire to hide instead of fightā (Elias 2016). Throughout the early 20th century camouflage quickly proved useful and by the end of World War II it had finally evolved into sign of humanityās increasing intelligence ā of our ability to outwit our opponents, outmaneuvering their sensory or perceptual strategies.
I find it strange that we donāt see the same connection happening with privacy right now. In the beginning of this century, there was a lot of negativity around the words āhidingā and āprivacyā. And now in 2017, weāre beginning to see the same situation unfold: privacy is increasingly becoming a sign of intelligence.
AC: So, youāre putting forward a reverse argument. Youāre saying that in the last two centuries, camouflage has been used for military purposes, but we can use it for civilian purposes in order to get some privacy back?
AH: Right. Iām also saying that embracing military-strategy surplus is not taboo. I think a lot of people arenāt comfortable engaging with military culture. But, if it can be reformulated in a more fashionable way, then people will be more willing to engage with topics such as thermal surveillance or facial recognition. When I started working on CV Dazzle in 2010, face detection was mostly linked to national security, not consumer products. Subverting it meant also meant subverting national security. Similarly, in 2012-2013, the primary use of thermal drone surveillance was to locate foreign enemy combatants. Now both are a daily part of the UKās domestic surveillance program (Davis 2017)
AC: In your project CV Dazzle (Figure 2), you show how some very lo-tech fashion modifications that anyone can make can be used to overcome hi-tech surveillance technologies. You can even buy some of the technologies you develop on your website the Privacy Gift Shop (Figure 3). Another theme from our recent work in Security Dialogue is the extent to which the ever-increasing penetration of drones into the domestic sphere might also create the possibility of new forms of civil resistance. To what extent do projects like CV Dazzle describe a politics of āeveryday resistanceā?

Figure 2: CV Dazzle Look 3
Photo: Adam Harvey
Model: Jude
Hair: Pia Vivas
Art Direction: DIS Magazine
AH: Thatās a great way to look at it. When you put clothes on in the morning, when you style yourself to go out, you need to consider the new environments of mass surveillance weāve created and arenāt easily escapable. Machine vision is a reality: youāre being observed and analyzed and will be continually analyzed for the next decades whenever that information is stored on a database. How do you reflect that in the way that you appear throughout time? Well itās not possible to see into the future, but it is possible to see into the present, to a certain degree. A lot of the important information on whatās happening in the computer vision industry is buried in long technical papers, but as you become more aware, the decisions you make in the morning ā everyday resistance ā will begin to affect the way that you dress. You start dressing for a machine-readable world.
AC: Can you say a bit more the relationship between everyday resistance and the idea of play? It seems you are talking about something slightly different from the militarization of the private sphere. Itās not about militarizing yourself, itās more about a playful mode of resistance.
AH: Yeah, I like to avoid thinking of these projects as resistance. In fact, theyāre the opposite. Where resistance costs you, these projects are designed to afford new opportunities. I mean, theyāre politically-motivated and serve the same purpose, but I absolutely want to avoid creating more resistance for people.
AC:Ģż How, if at all, do you see the political debate around drones and surveillance changing in the next few years? Do you envisage any kind of concerted attempt by governments to challenge the monopolies of the big tech companies? Is the idea ofĢżāa human right to our own dataā, for example, possible or desirable?
AH: I think weāre only at the beginning of learning what it means to be digitized at varying resolutions. Take the capabilities of cameras now (a lot of my research is on visual information) ā itās astonishing how much you can learn through aggregate information with neural network classifiers. For example, with a 6 x 7 pixel face image, you can do facial recognition on a group of 50 people with 95% accuracy. At 100 x 100 pixels, you have enough information to recognize hundreds of thousands of individuals. But even with 11 x 11 pixels, itās now possible to resynthesize a low-resolution image into a higher-resolution image, meaning 11 x 11 pixels is nearly enough for some facial recognition applications.
I can imagine an EU-centric push towards a human right on protection of personal visual data. Itās astonishing how much biometric information can be obtained remotely without oneās consent. For example, itās possible to extract your heart rate and skin conditions and iris data from over a hundred meters away. This information could eventually be augmented by multispectral sensors that can measure body temperature (thermal) or vein patterns (infrared). But first we need to develop a better understanding of what exactly we mean when we say āfaceā or ābiometricā. Do we mean to include psychological inferences and heart irregularities along with facial recognition?
AC: Who has the upper hand at this moment in these new forms of visual recognition technologies? The private companies or the government? Or are they working together for security purposes?
AH: They are both working together. I think the corporations have the upper hand of talent, but not funding or hardware. The NSA will always be years ahead of corporations with their computational capabilities. But this could be changing with regards to quantum computing. You can see evidence of collaboration in the Snowden documents, which reveal that the NSA received a license for facial recognition software called PittPatt from Google. But PittPatt was a start-up from Carnegie Mellon. And Carnegie Mellon receives significant funding from the Department of Defense, DARPA, and IARPA. Depending on how you look it, itās one ecosystem.
AC: This is very much related to the idea of the fungibility of technology. The government only funds those projects that are fungible for military purposes. When a technology is invented, it has to be functional for security purposes, otherwise they would never invest any money in it.
So, anyway, you are currently in London for the opening of a new project. Could you tell us a little about this and your future plans?
AH: Recently, Iāve come to the realization that visual machine communications are an integral part of how the new world is operating. I used to approach computer vision and surveillance from the perspective of counter-surveillance and avoidance. But now this limits my ability to communicate. Iāve now changed my approach and Iām embracing these technologies, because as I understand that theyāre here to stay. Iām thinking about how they can be applied to helping human rights groups, how they can be used by groups like Forensic Architecture.ĢżTo ignore the technological capabilities that enable you to keep protesting surveillance is the wrong strategy moving forward. I now believe visual machine communications to be an essential way of understanding the world and communicating information.
The work I am presenting here in London is called MegaPixels*.*ĢżItās a real time facial recognition query into the databases that are used to train facial recognition software. Most people donāt know that they might be in these databases because the databases are created from social media images. They are created from photos posted online. The installation currently uses MegaFace, which is the largest publicly available facial recognition training dataset. When you walk up to the installation youāll see the highest confidence matches from the dataset, and there is a small chance that you may find yourself. Out of the billions of people in the world, this dataset has 672,000 people. By chance, at least 2 people during the exhibition were positively identified in the dataset.
AC: Can I ask you a very difficult final question? I donāt know if you are familiar with the work of the French philosopher Michel Foucault, but I wanted to raise what I think might be his response to the situation you are describing, namely, that we are very easy to profile because we are so easily identifiable through these new media technologies, social media, Facebook, Google, etc. Is this because of our behaviour, or is it because the system compels us to behave in a specific way which can be recognized? I think Foucault would say the latter: itās not because we are always very similar to ourselves but because the system provides us with very detailed, structured pathways which we need to follow to behave as ourselves. We have to invent ourselves by means of the system and that is the reason why we are recognizable. So, in a way, the only way out would be to unplug the system.
AH: This is reality, and you can certainly see it with facial recognition and facial analysis. People become recognizable as the identities they bought into. Faceception ( is one company that clearly shows this feedback loop. Itās a filter bubble effect in the physical world. We can become trapped in the datasets of our former selves.
References
Backovic, L (2012) āHier kommt der Drohnen-Schutzanzugā. Spiegel online, 17th January. (accessed 24.05.18).
Elias, Ann (2016) āCamouflage and its impact on Australia in WWII: An Art Historianās Perspectiveā. Salus Journal. Vol 4, No. 1.
Chamayou, GrƩgoire (2015) Drone Theory tr. Janet Lloyd, London: Penguin.
Davis, Caroline (2017) āLondon could get 50m armed police base to tackle terrorismā, The Guardian, 11 September. Accessed 01.11.17. Available at:
Graham, DA (2016) āThe Dallas shooting and the advent of police killer robotsā, The AtlanticĢż (accessed 01.11.17).
Harvey, A (2013) (accessed 24.05.18).
Priest (2013) āGovernment surveillance spurs Americans to fight backā, Washington Post, August 14. (accessed 24.05.18).
Wortham, J (2013) āStealth Wear aims to make a statementā, New York Times, June 29. (accessed 24.05.18).
This interview is a part of theĢżSecurity Dialogue special compilation āDronelandā edited by Arthur Bradley and Antonio Cerella. An to this compilation can also be found on this blog, and two peer reviewed full-length articles:Ģż**by Arthur Bradley andĢż**by Oliver Davis were published in the 2019ĢżĢż50(4) ofĢżSecurity Dialogue.