What is Big Data, Anyway?
Lynn Cherny, Visiting Knight Chair
JMM, SoC, University of Miami
@arnicas / lynn@cherny.com
Sept 25, 2015
http://ghostweather.slides.com/lynncherny/what-is-big-data-anyway/live
This was a google search, but also recommended: http://bigdatapix.tumblr.com/
use the
downarrow!
Genres
Men in front of walls of big data.
In suits.
Men in front of wall-sized networks.
Genres
Word clouds
Elephants
http://demo.relato.io/oreilly
Big data is like teenage sex: Everyone talks about it, nobody really knows how to do it, everyone thinks everyone else is doing it, so everyone claims they are doing it.
--Dan Ariely of Duke University
Big data is a broad term for data sets so large or complex that traditional data processing applications are inadequate. Challenges include analysis, capture, data curation , search, sharing , storage, transfer, visualization, and information privacy .
"Big data is distributed"
Mike Driscoll: https://www.quora.com/How-much-data-is-Big-Data
The "3 V's" of Big Data
- Volume - the amount of data generated by organizations or individuals or websites.
- Velocity - the frequency at which data is generated, captured and shared.
- Variety - More than rows and columns and numbers; often unstructured text, video, audio.
sometimes also
- Veracity - Can you trust it, is it reliable? Evaluating that is hard.
http://www.ibmbigdatahub.com/infographic/four-vs-big-data
"The Cloud"
- 1999: Salesforce, "Software as a Service" (SAAS) via web
- 2002: Amazon Web Services (including Mechanical Turk)
- 2006: AWS EC2 (cloud-based computing clusters)
AWS Mechanical Turk
Distributed Human Labor
EC2 Instances
Amazon AWS "in plain English"
There are tutorials, but they're kind of tough.
A server issue in Virginia is affecting most of the northeast, disrupting the infrastructure for many popular products and services including Netflix, Product Hunt, Medium, SocialFlow, Buffer, GroupMe, Pocket, Viber Amazon Echo and more.
It’s certainly not the first time AWS has taken much of the Internet out with it. In 2013, AWS suffered a similar outage that took services like Instagram, Airbnb and Vine offline. According to Buzzfeed, that’s a loss of about $1,100 per second for Amazon.
Tools in the Ecosystem:
"Hadoop" and Map/Reduce
Apparently still relevant...
Map/Reduce
MAP
REDUCE
The nightmare that is Java Hadoop...
This is "hello world" word count.
An alternative to Hadoop, Spark with Python
Essentially, the actual work in Big Data is still hard.
- Most Hadoop users complain bitterly about the ops and time delays on their work.
- Recent tools like Spark, Impala, etc. look better, but ops (setup, management of the systems) is a giant pain.
- Just "moving data around" takes a long time and some serious engineering expertise. (E.g., "The time required to move the data from Amazon S3 to HDFS was about 1 hour and 45 minutes." -- 1 billion reddit comments)
Giant Data Sets Are Around
Of many types...
Applications that use them, too! Via APIs.
"Mashups"
combining web data apis in new applications
GDELT
Now With Historical Book Archives too
Big Query Demo of GDELT
Crossfilter / dc.js dash I built...
Visualizing Big Data
Aggregation and Approximation
(Sometimes true for medium data, too!)
Paris Rainfall Data (323 years * 12 months)
Visualization of daily Wikipedia edits created by IBM. At multiple terabytes in size, the text and images of Wikipedia are an example of big data.
The Internet Archive's Map of Book Subjects (8M books, 5K Subjects linked to 2.6M images on flickr) by Mario Klingemann
Crossfilter: Fast Multi-Dimensional Filtering for Coordinated Views
Demo by Mike Bostock
Relevant Overviews of Vis for Big Data
- Bin-Summarize-Smooth: A Framework for Visualizing Large Data (Hadley Wickham)
- imMems: Real-Time Visual Querying of Large Data (Stanford - now UW - Vis Group)
- "Why Exploring Big Data is Hard and What We Can Do About It", Danyel Fisher's talk at OpenVisConf 2015
Critical Discussion
Big Data and whole data are not the same. Without taking into account the sample of a data set, the size of the data set is meaningless. For example, a researcher may seek to understand the topical frequency of tweets, yet if Twitter removes all tweets that contain problematic words or content – such as references to pornography or spam – from the stream, the topical frequency would be inaccurate. Regardless of the number of tweets, it is not a representative sample as the data is skewed from the beginning.
d. boyd and K. Crawford, "Critical Questions for Big Data"
... four quantitatively adept social scientists reported that Google’s flu-tracking service not only wildly overestimated the number of flu cases in the United States in the 2012-13 flu season — a well-known miss — but has also consistently overshot in the last few years. Google Flu Trends’ estimate for the 2011-12 flu season was more than 50 percent higher than the cases reported by the Centers for Disease Control and Prevention. ...Their technical criticism of Google Flu Trends is that it is not using a broader array of data analysis tools. Indeed, their analysis shows that combining Google Flu Trends with C.D.C. data, and applying a few tweaking techniques, works best.
Big Data Hubris? Or "not invented here"...
"The Anxieties of Big Data": Surveillance and the Surveilled
And while there is an enormous structural power asymmetry between the surveillers and surveilled, neither are those with the greatest power free from being haunted by a very particular kind of data anxiety: that no matter how much data they have, it is always incomplete, and the sheer volume can overwhelm the critical signals in a fog of possible correlations.
Spurious Correlations
"Squeaky Dolphin" and GCHQ's anxiety
What is Data Science?
(and how is it related?)
"The Unreasonable Effectiveness of Data"
by Halevy, Norvig, Pereira at Google in 2009
[ok, but for some problems]
The first lesson of Web-scale learning is to use available large-scale data rather than hoping for annotated data that isn’t available. For instance, we find that useful semantic relationships can be automatically learned from the statistics of search queries and the corresponding results-- or from the accumulated evidence of Web-based text patterns and formatted tables-- in both cases without needing any manually annotated data.
More Data Means...
- It can be noisy (noisier)
- Noisy is more realistic
- With the right models (and infrastructure), more data substitutes for annotated training data
Some good, recent articles about data science jobs
- What It's Like to Be on the Data Science Job Market, by Trey Causey
- Crushed It! Landing a Data Science Job, by Erin Shellman
- Doing Data Science At Twitter, by Robert Chang
Type A: Analysis
This type is primarily concerned with making sense of data or working with it in a fairly static way. The Type A Data Scientist is very similar to a statistician (and may be one) but knows all the practical details of working with data that aren’t taught in the statistics curriculum: data cleaning, methods for dealing with very large data sets, visualization, deep knowledge of a particular domain, writing well about data, and so on.
Type B: Builders
Type B Data Scientists share some statistical background with Type A, but they are also very strong coders and may be trained software engineers. The Type B Data Scientist is mainly interested in using data “in production.” They build models which interact with users, often serving recommendations (products, people you may know, ads, movies, search results).
A Day in the Life of a Data Scientist
Predictive Analytics - a Hot Topic in Data Science
- Fraud Detection
- Cross-selling (customers who bought this bought that)
- Clinical support systems: Patients likely to develop diabetes, etc.
- Customer Retention
An area of data mining that deals with extracting information from data and using it to predict trends and behavior patterns.
"Metadata"
President Obama has emphasized that the NSA is “not looking at content.” “[T]his is just metadata,” Senator Feinstein told reporters.
We were wrong. We found that phone metadata is unambiguously sensitive, even in a small population and over a short time window. We were able to infer medical conditions, firearm ownership, and more, using solely phone metadata.
Uber Affairs,
4Square Checkins, Your secrets aren't safe.
Uber had just told all its users that if they were having an affair, it knew about it. Rides to Planned Parenthood? Regular rides to a cancer hospital? Interviews at a rival company? Uber knows about them, too.
Data Journalism is struggling, but game...
Applicants should have evidence of distinction in research and teaching in humanities computing and large-scale quantitative or big data analysis of cultural archives. Applicants should have expertise in algorithmic development, data culture, data mining, and quantitative analysis of visual or textual materials. Specific domain application could be in any number of areas, from historical text archives to game databases. Applicants should be able to teach courses on topics related to humanities computing and big data for one or more of the following departments: English, Science and Technology Studies, Cinema and Digital Media.
A few takeaways...
The tools are still painful. So are the jobs.
Applications of "big data" modeling are sometimes invasively accurate and sometimes embarrassingly, dangerously off-base.
But none of this is going away.
Everyone should learn some SQL.
Thanks!
What is Big Data, Anyway?
By Lynn Cherny
What is Big Data, Anyway?
A presentation for non-technical folks in Communications and Journalism about some recent (and historical) trends in "Big Data."
- 10,362