Tag Archives: big data

Code and the City workshop videos: Session 4

If you missed any of the videos from the first three sessions, they are here: Session 1, Session 2 and Session 3.

Session 4: Cities, knowledge classification and ontology

Cities and context: The codification of small areas through geodemographic classification
Alex Singleton, Geography, University of Liverpool

Abstract
Geodemographic classifications group small area geography into categories based on shared population and built environment characteristics. This process of “codification” aims to create a common language for the description of salient internal structure of places, and by extension, enable their comparison across geographic contexts. The typological study of areas is not a new phenomenon, and contemporary geodemographics emerged from research conducted in the 1970s that aimed at providing a new method of targeting deprivation relief funding within the city of Liverpool. This city level model was later extended for the national context, and became the antecedent of contemporary geodemographic classification. This paper explores the origins of geodemographics, to first illustrate that the coding of areas is not just a contemporary practice; and then extends this discussion to consider how methodological choices influence classification structure. Being open with such methods is argued as being essential for classifications to engender greater social responsibility.

The city and the Feudal Internet: Examining institutional materialities
Paul Dourish, Informatics, UC Irvine

Abstract
In “Seeing like a City,” Marianne Valverde turns to urban regulation to counter some of James Scott’s arguments about the homogenizing gaze of high modern statehood. Cities, she notes, are highly regulated, but without the panoptic order that Scott suggests. They operate instead as a splintered patchwork of regulatory boundaries – postal codes, tax assessment districts, business improvement zones, school catchment areas, zoning blocks, sanitation districts, and similar divisions that don’t quite line up. Arguments about online experience and the consequences of the Internet have a similar air to Scott’s analysis of statehood – they posit a world of consistent, compliant, and compatible information systems, in which the free flow of information and the homogenizing gaze of the digital erases boundaries (both for good and ill).

In fact, the organization of the Internet — that is, of our technologically- and historically-specific internet –is one of boundaries, barriers, and fiefdoms. We have erected all sorts of internal barriers to the free flow of information for a range of reasons, including the desire for autonomy and the extraction of tolls and rents. In this talk I want to explore some aspects of the historical specificity of our Internet and consider what this has to tell us about the ways that we talk about code and the city.

Semantic cities: Coded geopolitics and rise of the semantic web
Heather Ford and Mark Graham, Oxford Internet Institute, University of Oxford

Abstract
In 2012, Google rolled out a service called Knowledge Graph which would enable users to have their search query resolved without having to navigate to other websites. So, instead of just presenting users with a diverse list of possible answers to any query, Google selects and frames data about cities, countries and millions of other objects sourced from sites including Wikipedia, the CIA World Factbook and Freebase under its own banner.

For many, this heralded Google’s eventual recognition of the benefits of the Semantic Web: an idea and ideal that the Web could be made more efficient and interconnected when websites share a common framework that would allow data to be shared and reused across application, enterprise, community, and geographic boundaries. This move towards the Semantic Web can be starkly seen in the ways that Wikipedia, as one of the foundations for Google’s Knowledge Graph, has begun to make significant epistemic changes. With a Google funded project called WikiData, Wikipedia has begun to use Semantic Web principles to centralise ‘factual’ data across all language versions of the encyclopaedia. For instance, this would mean that the population of a city need only be altered once in WikiData rather in all places where it occurs in Wikipedia’s 285 language versions.

For Google, these efficiencies provide a faster experience for users who will stay on their website rather than navigating away. For Wikipedia, such efficiencies promise to centralise the updating process so that data are consistent and so that smaller language Wikipedias can obtain automated assistance in translating essential data for articles more rapidly.

This paper seeks to critically interrogate these changes in the digital architectures and infrastructures of our increasingly augmented cities. What shifts in power result from these changes in digital infrastructures? How are semantic standardisations increasingly encoded into our urban environments and experiences? And what space remains for digital counter-narratives, conflict, and contention?

To tackle those questions, we trace data about two cities as they travel through Google’s algorithms and the Semantic Web platforms of Wikidata and Wikipedia. In each case, we seek to understand how particular reflections of the city are made visible or invisible and how particular publics are given voice or silenced. Doing so leads us to ultimately reflect on how these new alignments of code and content shape how cities are presented, experienced, and brought into being.

Code and the City workshop videos: Session 3

If you missed our first and second sessions of the Code and the City workshop video, the embedded links will lead you to them. And now is time for Session 3!

Session 3: Locative/social media

Digital social interactions in the city: Reflecting on location-based social media
Luigina Ciolfi, Human-Centred Computing, Sheffield Hallam University
Gabriela Avram, University of Limerick

Abstract
Location-based social media increasingly mediates social and interpersonal interactions in urban settings. Such practices become coded in software representing both the log and content of social interactions and the location to which they relate. Therefore a digital “cloud” of social interactions becomes embedded into the physical reality of the city, of its neighbourhoods, public places, cafés, transportation hubs and any other location identified by social media users (by user-initiated “check-ins” or by the content that they generate, such as photographs) and by the tools they use (for example, through automatic geo-tagging). Two sets of issues to be investigated are emerging: firstly referring to how such localised interactions are populating the algorithms and infrastructures provided by the software: how are the platform of location-based social media framing people’s perceptions and identifications of locations? How is code both facilitating and representing a set of social interactions relating to various spatial configurations? A second set of issues regards the re-materialisation of such cloud of interactions in the physical world: could it be made somehow perceivable and/or tangible in the physical world by the way in which certain environments are designed?

Overall, could new approaches to urban planning and environmental design become concerned with accommodating and facilitating these social interactions as they do so by supporting in-presence, analogue ones?

This paper will attempt to define and discuss these issues drawing both from interaction design and human-computer interaction literature on physical/digital interactions and from two preliminary empirical studies of location-based social media use in two cities.

Feeling place in the city: strange ontologies, Foursquare and location-based social media
Leighton Evans, National University of Ireland Maynooth

Abstract
Certain instances of the use of location-based social media in cities can result in deep understandings of novel locations. The contributions of other users and the information pushed to users when in particular locales can help users rapidly attune themselves to places and achieve an understanding of the place. The use of a computational device and location-based social networking to achieve this understanding indicates an alteration in the achievement of placehood using computational technology. Practices and methods of understanding place can, in some situations, be delegated to the device and application. This paper explores how the moment that place is appreciated as place (that is, as a meaningful existential locale) can be reconciled with the delegation of the epistemologies of placehood to a computational device and location-based social media application. Drawing on data from an ethnographic study of Foursquare users, the phenomenological appreciation of place is understood as co-constituent between the device, application and the mood of the user. Code and computational devices are contextualised as a constant foregrounding presence in the city, and the engagement of the user, device, code and data in understanding place is a moment of revealing that is co-constituent of all these elements. This exploratory paper engages Peter Sloterdijk’s theory of spheres as a framework to understand how these four elements interact, and how that interaction of elements can orient a user to a revealing of the city that can be understood as a phenomenological revealing of place.

Cultural curation and urban Interfaces: Locative media as experimental platforms for cultural data
Nanna Verhoeff, Department of Media and Culture Studies, Utrecht University

Abstract
My contribution is concerned with the way in which urban interfaces are used for access to cultural collections – whether institutionally embedded, or bottom-up, participatory collections. Designed in code and exploring affordances of new location-based and/or mobile technologies for urban space-making, these interfaces are thought to be powerful tools for ideals of participatory urban culture. I propose to approach these “projects” as curatorial machines, as urban experimental laboratories for cultural data. This entails a threefold perspective, on curation, on code, and on principles of creative (sometimes artistic or playful) experimentation.

For this, we may remind ourselves of the curatorial project of museal and archival institutions, of preserving, and “caring” for the object, as well as creating new contexts for the object and providing access for an urban public – a field which is very much in transition as a result of current ambitions for new public engagement and ideals of participation, pervasive in all socio-economic and political regions of contemporary culture. Simultaneously we witness the current interest in the principles of data curation as the care for, interaction with, interpretation and visualisation of digital data, as the datafication and codification of culture invades all corners of urban life. Design of interfaces is central in how we can access, work with, and make meaning with digital culture. Departing from the concept of dispositif in the analysis of interfaces, I propose to bring together the fact that the interfaces are coded and designed, to (playfully) experiment with their affordances.

In my approach to this intersection of datafication of, and the proliferation of interfaces for “culture”, I aim to develop heuristic tools for critical evaluation of this phenomenon, broadly bracketed as [urban interfaces] as interfaces of cultural curation.

A Window, a message, or a medium? Learning about cities from Instagram
Lev Manovich, Computer Science, The Graduate Center, City University of New York

Abstract
Over last few years, tens of thousands of researchers in social computing and computational social sciences started to use available data from social networks and media sharing services (such as Twitter, Foursquare and Instagram) created by users of mobile platforms. The research uses techniques from statistics, machine learning, and visualization, among others, to analyze all kinds of patterns contained in this data and also (less frequently) propose new models for understanding the social. The examples include analysis of information propagation in Twitter, predicting popularity of photos on Flickr, proposing new sets of city neighborhoods using Foursquare users check-ins, and understanding connections between musical genres using listening data from Echonest.

In my talk I will address a fundamental question we face in doing this research: what exactly are we learning when analyzing can social media data? Is it a window into real-world social and cultural behaviors, a reflection of lifestyles of particular demographics who use mobile platforms and particular network services, or only an artifact of mobile apps? In other words – is social media a “message” or a “medium”?

I will discuss this question using three recent projects from my lab (softwarestudies.com). The projects use large sets of Instagram images and accompanying data together with data science and visualization tools. Phototrails.net (2013) analyzes 2.3 million photos from 13 global cities to investigate how different kinds of events are represented in these photos. The project also investigates if the universal affordances of Instagram app (same interface and same set of filters available to all users) result in universal digital visual language. Selfiecity.net (2014) analyzes the distinct artifact of mobile platforms – selfies. We compare thousands of selfies to see if cultural specificity of different places and cultural is preserved in this genre. Finally, our third project compares Instagram photos taken by visitors in a few major modern art museums, asking if photographs of famous works of the art differ depending on what these artworks are and where they are situated.

Nathaniel Tkacz – Dashboards and Data Signals

On Wednesday 8th October 2014, Nathaniel Tkacz visited the Programmable City Project and delivered a seminar on “Dashboards and Data Signals”. Nathaniel is Assistant Professor in CIM at The University of Warwick. He is currently Principal Investigator for the ESRC-funded project ‘Interrogating the Dashboard’. He is author or editor of Wikipedia and The Politics of Openness, The MoneyLab Reader (forthcoming 2015), Digital Light (forthcoming 2014) and Critical Point of View: A Wikipedia Reader.

On dashboards and data signals, Nathaniel summarised his discussion as: Screen interfaces that aggregate and visualise flows of data, namely, dashboards, are greatly increasing in number and function. They coincide with the heightened interest in information visualisation and massive claims about the transformative power of big data. David Cameron has a bespoke dashboard, as does US Federal Reserve Chair, Janet Yellen. For the rest of us, there are a range of dashboard apps available from Apple or Google’s respective markets. The control screens one would expect to see in a Bloomberg Terminal, flight control tower or security operation – that is, in strategic and logistical spaces – is today becoming generalised and individualised. As is true of all interfaces, a dashboard is a relation. It is a relation of control, to be sure, and one that equally reflects a desire for control among the ‘data deluge’ as much as it does its achievement. But a closer look at the dashboard reveals much more than the proliferation of control. It can tell us, for example, about the changing nature of indicators, the everyday experience of data-driven life, and emerging forms of rationality, and it is these things that I will explore in this presentation.

New Paper: Towards Critical Data Studies: Charting and Unpacking Data Assemblages and Their Work

Rob Kitchin and Tracey Lauriault have just published the second Programmable City Working Paper – Towards Critical Data Studies: Charting and Unpacking Data Assemblages and Their Work. It is a pre-print of a chapter written for the book, Geoweb and Big Data, edited by Joe Eckert, Andy Shears and Jim Thatcher, to be published by University of Nebraska Press.

Abstract
The growth of big data and the development of digital data infrastructures raises numerous questions about the nature of data, how they are being produced, organized, analyzed and employed, and how best to make sense of them and the work they do. Critical data studies endeavours to answer such questions. This paper sets out a vision for critical data studies, building on the initial provocations of Dalton and Thatcher (2014). It is divided into three sections. The first details the recent step change in the production and employment of data and how data and databases are being reconceptualised. The second forwards the notion of a data assemblage that encompasses all of the technological, political, social and economic apparatuses and elements that constitutes and frames the generation, circulation and deployment of data. Drawing on the ideas of Michel Foucault and Ian Hacking it is posited that one way to enact critical data studies is to chart and unpack data assemblages. The third starts to unpack some the ways that data assemblages do work in the world with respect to dataveillance and the erosion of privacy, profiling and social sorting, anticipatory governance, and secondary uses and control creep. The paper concludes by arguing for greater conceptual work and empirical research to underpin and flesh out critical data studies.

Key words
big data, critical data studies, data assemblages, data infrastructures, civil liberties

Post in 'Big Data, Big Questions' series on LSE Impact blog

Rob Kitchin provides the first interview in a new series on the LSE Impact blog entitled ‘Big Data, Big Questions’ curated by Mark Carrigan.  The post concerns the philosophy of data science, the nature of ‘big data’,  the opportunities and challenges presented for scholarship with its growing influence, the hype and hubris surrounding its advent, and the distinction between data-driven science and empiricism.  You can read more over at the LSE Impact blog.