Author Archives: Rob Kitchin

Towards geographies of and produced by data brokers

Today, Rob Kitchin participated in a panel session on spatialized information economies at the Association of American Geographers in Chicago, organized by Jeremy Crampton and Agnieszka Leszczynski.  Below is his script for an intervention titled ‘Towards geographies of and produced by data brokers’.

There have long been spatialized information economies – ever since maps, gazetteers and almanacs have been created and traded.  There’s also a well established century old history of political polling and spatialised market research and data services.  With the development of digital data from the 1950s on, the markets for spatial data and information have steadily diversified in products and exploded in volume of trade, with the growth of new market sectors for creating and processing spatial data such as GIS and CAD, and new spatial info products such as geodemographics.  This is particularly the case in the era of big data, where there is now a deluge of diverse types of continuously produced georeferenced data (principally through GPS and zip code), including digital CCTV, clickstream, online and store transactions, CRM, sensors and scanners, social media, wearables, IoT, and so on.

The data produced from these sources has become a highly valuable commodity and they have led to the rapid growth of a set of data brokers (sometimes called data aggregators, consolidators or re-sellers) who trade in a number of multi-billion dollar data markets.  Data brokers capture, gather together and repackage data for rent (for one time use or use under licensing conditions) or re-sale.  By assembling data from a variety of sources data brokers construct a vast relational data infrastructure.  For example, Epsilon is reputed to own data on 300 million company loyalty card members worldwide.  Acxiom is reputed to have constructed a databank concerning 500 million active consumers worldwide, with about 1,500 data points per person, and claim to be able to provide a ‘360-degree view’ on consumers (meshing off-line, online and mobile data).  It also manages separately customer databases for, or works with, 47 of the Fortune 100 companies.  Datalogix claim to store data relating to over a trillion dollars worth of offline purchases.  Other data broker and analysis companies include Alliance Data Systems, eBureau, ChoicePoint, Corelogic, Equifax, Experian, Facebook, ID Analytics, Infogroup, Innovis, Intelius, Recorded Future, Seisint and TransUnion.

Each company tends to specialize in different types of data and data products and services.  Products include:

  • lists of potential customers/clients who meet certain criteria and consumer and place profiles
  • search and background checks
  • derived data products wherein brokers have added value through integration and analytics
  • data analysis products that are used to micro-target advertising and marketing campaigns (by social characteristics and/or by location), assess credit worthiness and socially and spatially sort individuals, provide tracing services, predictive modelling as to what individuals might do under different circumstances and in different places, or how much risk a person constitutes, and supply detailed business analytics.

The worry of some, including Edith Ramirez, the chairperson of the Federal Trade Commission (FTC) in the US, is that such firms practice a form of ‘data determinism’ in which individuals are not profiled and judged just on the basis of what they have done, but on the prediction of what they might do in the future using algorithms that are far from perfect, which may hold in-built biases relating to race, ethnicity, gender and sexuality, and yet are black-boxed and lack meaningful oversight and remediate procedures.  Moreover, they employ the data for purposes for which they were never generated and data are hoarded as a speculative measure that they may have future value, breaking data minimization rules that stipulate that only data of defined value should be retained.  And given the volume of sensitive personal records they are a prime target for criminals intent on conducting identity theft fraud.

Interestingly, given the volumes and diversity of personal and place-based data that data brokers and analysis companies possess, and how their products are used to socially and spatially sort and target individuals and households, there has been remarkably little critical attention paid to their operations.  Indeed, there is a dearth of academic and media analysis about such companies and the implications of their work and products.  This is in part because the industry is relatively low-profile and secretive, not wishing to draw public attention to and undermine public trust in their assets and activities, which in turn might lead to public campaigns for transparency, accountability and regulation.  Moreover, data brokers are generally unregulated and are not required by law to provide individuals access to the data held about them, nor are they obliged to correct errors relating to those individuals.  As such, there is a pressing need for us to conduct research on both the geographies of the data brokerage industry (in terms of where they are located and for what reasons) and the geographies produced by that industry; to map out their associated spatial informational economies.  At present, we have little detailed understanding of either, which is why it is difficult – for me at least – to answer list of questions posed by Agnieszka and Jeremy for this session (which are listed below).

Questions of labor, legal frameworks, and privacy

  • what is the legal status of geolocational privacy rights; privacy and national security (eg PCLOB); and legal rulings (eg Jones v. USA, Riley v. California)?
  • what is the landscape of legal geographies around spatial information/the spatialization of content?
  • are laborers in the spatialized information economy experiencing increasing “control” over them? If so, does this control reflect itself in everyday working conditions? Are these working conditions exploitive?
  • what is the status of regulatory and/or oversight over labor in this sector of the economy?
  • what are the (global) geographies of divisions of labour in the spatial information economy?

Questions of innovation, investment and technology

  • what characterizes geospatial information/product lifecycles and growth trajectories?
  • what is the context of geospatial product development?
  • on the consumer side, how are geoweb technological innovations marketed?
  • are there geographical clusters of innovation and if so what is giving rise to their concentration (eg., proximity to other capital, deregulated conditions, tax incentives?)
  • how is venture capital invested in the spatial info economy, and what are the sources of investment?

Questions of (cyber)security, surveillance, and cyberwarfare

  • in what ways is geolocation underwriting and increasingly central to the surveillance activities and practices of the securities agencies?
  • how are the decentralized, global geographies of data (deterritorialized collection and flow, reterritorialized storage and analysis) complicating the cybersecurity/cyberwarfare equation?

New paper: Big data and official statistics

A new paper by Rob Kitchin – Big data and official statistics: Opportunities, challenges and risks – has been published on SSRN as Programmable City Working Paper 9.  The abstract runs thus:

The development of big data is set to be a significant disruptive innovation in the production of official statistics offering a range of opportunities, challenges and risks to the work of national statistical institutions (NSIs).  This paper provides a synoptic overview of these issues in detail, mapping out the various pros and cons of big data for producing official statistics, examining the work to date by NSIs in formulating a strategic and operational response to big data, and plotting some suggestions with respect to on-going change management needed to address the use of big data for official statistics.

The argument is pretty much distilled into this table:big data and official statistics

Download PDF

 

The Impact of the Data Revolution on Official Statistics: Opportunities, Challenges and Risks

NTTS 2015This morning Rob Kitchin presented a keynote talk at the New Techniques and Technologies for Statistics conference in Brussels.  The presentation examined the potential impact of the unfolding data revolution – big data, open and linked data, data infrastructures, and new data analytics – on the production of official statistics and the work of national statistical institutions.  The slides that accompanied the talk are below.

 

Two 3 year postdoc posts on The Programmable City project

We’re recruiting!  We are seeking two postdoctoral researchers to work on the Programmable City project to (1) unpack data assemblages, and (2) examine big data industry and smart cities.
These posts will complement the existing research team.

Post 1: Unpacking a data assemblage, including the associated technological stack
Adopting a critical data studies approach, this researcher will examine in depth, and compare and contrast, two data assemblages, mostly likely operating in the public sector.  The aim is to gain a detailed conceptual and empirical understanding from a social sciences perspective of how data infrastructures are on the one hand technically assembled (through a technical stack composed of hardware, networks, software, algorithms, data, interfaces) and operated using a set of technical and social practices, and how their production and operation is socially, politically, legally and economically framed.  The empirical research will consist primarily of ethnographic work within organisations, along with in-depth interviews with key actors and stakeholders.

Post 2:  Big data industry and smart cities
This researcher will, on the one hand, examine in depth the development of big data industries in Dublin and Boston, examining the big data ecosystem (including big data analytics and data brokers) and its associated discursive regime, and on the other, examine how big data are being mobilised and deployed within smart city initiatives as part of big data assemblages.  The aim is to gain a detailed conceptual and empirical understanding of the development and use of big data and big data analytics within the private and public sector to complement existing project work on open data, and how big data are being deployed in practice in cities and any associated consequences.  The empirical research will consist primarily of in-depth interviews with key actors and stakeholders and one and two case studies of urban big data initiatives.

There will be some latitude to re-jig these projects in negotiation with the principal investigator in order to fit the interests, expertise and experience of the appointed researchers.

Full details can be found on the Maynooth University vacancies page (labelled
Post Doctoral Researcher x 2 – NIRSA).

Video: ProgCity at Smart City Expo and Congress

In November 2014 members of the Programmable City team visited the Smart City Expo and Congress in Barcelona.  The organisers have now posted up videos of all of the sessions on their YouTube channel.  Together they make interesting viewing for anyone interested in understanding what is happening with regards to creating smart cities.  Rob Kitchin and Gavin McArdle presented a paper at the Congress (below) entitled, ‘Dublin Dashboard: Open and real-time data and visualizations for citizens, government and companies’.

Book launch: The Data Revolution and others

Mark Boyle, Chris Brunsdon & Rob Kitchin invite you to a BOOK LAUNCH Thursday 26th, February 2015, 4.30pm, Maynooth University Bookshop, North Campus

Using the story of the “West and the world” as its backdrop, this book provides for beginning students a clear and concise introduction to Human Geography, including its key concepts, seminal thinkers and their theories, contemporary debates, and celebrated case studies.

“An excellent textbook for introductory courses in Human Geography.” Prof. Patricia Wood, York University, Toronto

“In this textbook, Mark Boyle combines his broad and deep understanding of the discipline of Human Geography with his great passion and enthusiasm for education and teaching.”  Prof. Guy Baeten, Lund University

This is an excellent and student-friendly text from two of the world leaders in spatial analysis. It shows clearly why the open source software R is not just an alternative to commercial GIS, it may actually be the better choice for mapping, analysis and for replicable research. Prof.  Richard Harris, Bristol University

This is a vital primer to what is ‘Big’ about geocomputation: new data, innovative methods of analysis, new geographic information technologies and, above all, an over-arching rethink of how we represent geography. Prof. Paul Longley, UCL

The Data Revolution provides a synoptic and critical analysis of big data, open data, and data infrastructures.

“Anyone who wants to obtain a critical, conceptually honed and analytically refined perspective on new forms of data should read this book.”  David Beer, University of York

Funny, engaging, fast-paced and hugely enjoyable … a unique combination of comedy, both gentle and black, and Grand Guignol murder and mayhem.”  Michael Russell, author of The City of Strangers

ALL WELCOME