We can't find the internet
Attempting to reconnect
Something went wrong!
Hang in there while we get back on track
-
Short answer: Viewed unconventionally, digital surveillance can be read as a social mirror—revealing and shaping power relations, norms, and identities rather than just collecting data. It performs classification and persuasion, not only observation.
-
Key terms
- Surveillance — systematic collection and analysis of people’s behavior or data.
- Panopticon — a metaphor for power through possible observation (Foucault).
- Sociotechnical — systems that combine social practices and technology.
- Algorithmic governance — rules enforced or suggested by automated systems.
-
How it works
- Sensors, logs, and platforms collect signals (clicks, location, biometrics).
- Algorithms categorize and predict behavior, producing profiles.
- Institutions act on profiles (targeting, exclusion, reward), shaping futures.
- Social norms shift as people anticipate and adapt to monitoring.
- Visibility becomes a resource: some groups are hyper-visible; others erased.
-
Simple example
- A city’s CCTV + facial recognition not only deters crime but redirects policing toward certain neighborhoods, reinforcing social hierarchies.
-
Pitfalls or nuances
- Data ≠ truth: classifications embed biases.
- Power asymmetry: those who design systems decide categories.
- Resistance and evasion change system effects.
-
Next questions to explore
- Who benefits from specific surveillance categories?
- How can design redistribute power or enable accountability?
-
Further reading / references
- Discipline and Punish — Michel Foucault (background on panopticon).
- “The Rise of Big Data Policing” — NYU Law Review search query (try: “Big Data Policing NYU Law Review Katrin Nowotny” if link unavailable).