SlideShare a Scribd company logo
1 of 35
Download to read offline
1	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
High	
  Performance	
  Python	
  on	
  
Apache	
  Spark	
  
Wes	
  McKinney	
  @wesmckinn	
  
Spark	
  Summit	
  West	
  -­‐-­‐	
  June	
  7,	
  2016	
  
2	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Me	
  
•  Data	
  Science	
  Tools	
  at	
  Cloudera	
  
•  Serial	
  creator	
  of	
  structured	
  data	
  tools	
  /	
  user	
  interfaces	
  
•  Wrote	
  bestseller	
  Python	
  for	
  Data	
  Analysis	
  2012	
  
• 	
  Working	
  on	
  expanded	
  and	
  revised	
  2nd	
  edi-on,	
  coming	
  2017	
  
•  Open	
  source	
  projects	
  
•  Python	
  {pandas,	
  Ibis,	
  statsmodels}	
  
•  Apache	
  {Arrow,	
  Parquet,	
  Kudu	
  (incubaUng)}	
  
•  Focused	
  on	
  C++,	
  Python,	
  and	
  Hybrid	
  projects	
  
3	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Agenda	
  
•  Why	
  care	
  about	
  Python?	
  
	
  
•  What	
  does	
  “high	
  performance	
  Python”	
  even	
  mean?	
  	
  
	
  
•  A	
  modern	
  approach	
  to	
  Python	
  data	
  so]ware	
  
	
  
•  Spark	
  and	
  Python:	
  performance	
  analysis	
  and	
  development	
  direcUons	
  
4	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
•  Accessible,	
  “swiss	
  army	
  knife”	
  programming	
  language	
  
	
  
•  Highly	
  producUve	
  for	
  so]ware	
  engineering	
  and	
  data	
  science	
  alike	
  
	
  
•  Has	
  excelled	
  as	
  the	
  agile	
  “orchestraUon”	
  or	
  “glue”	
  layer	
  for	
  applicaUon	
  
business	
  logic	
  
	
  
•  Easy	
  to	
  interface	
  with	
  C	
  /	
  C++	
  /	
  Fortran	
  code.	
  Well-­‐designed	
  Python	
  C	
  API	
  
	
  
Why	
  care	
  about	
  (C)Python?	
  
5	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Defining	
  “High	
  Performance	
  Python”	
  
•  The	
  end-­‐user	
  workflow	
  involves	
  primarily	
  Python	
  programming;	
  programs	
  can	
  
be	
  invoked	
  with	
  “python	
  app_entry_point.py	
  ...”	
  
	
  
•  The	
  so]ware	
  uses	
  system	
  resources	
  within	
  an	
  acceptable	
  factor	
  of	
  an	
  
equivalent	
  program	
  developed	
  completely	
  in	
  Java	
  or	
  C++	
  
•  Preferably	
  1-­‐5x	
  slower,	
  not	
  20-­‐50x	
  
	
  
•  The	
  so]ware	
  is	
  suitable	
  for	
  interacUve	
  /	
  exploratory	
  compuUng	
  on	
  modestly	
  
large	
  data	
  sets	
  (=	
  gigabytes)	
  on	
  a	
  single	
  node	
  
6	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Building	
  fast	
  Python	
  so]ware	
  
means	
  embracing	
  certain	
  
limitaUons	
  
7	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Having	
  a	
  healthy	
  relaUonship	
  with	
  the	
  interpreter	
  
•  The	
  Python	
  interpreter	
  itself	
  is	
  “slow”,	
  as	
  compared	
  with	
  hand-­‐coded	
  C	
  or	
  Java	
  
•  Each	
  line	
  of	
  Python	
  code	
  may	
  feature	
  mulUple	
  internal	
  C	
  API	
  calls,	
  
temporary	
  data	
  structures,	
  etc.	
  
	
  
•  Python	
  built-­‐in	
  data	
  structures	
  (numbers,	
  strings,	
  tuples,	
  lists,	
  dicts,	
  etc.)	
  have	
  
significant	
  memory	
  and	
  performance	
  use	
  overhead	
  
	
  
•  Threads	
  performing	
  concurrent	
  CPU	
  or	
  IO	
  work	
  must	
  take	
  care	
  not	
  to	
  block	
  
other	
  threads	
  
8	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Mantras	
  for	
  great	
  success	
  
•  Key	
  quesUon	
  1:	
  Am	
  I	
  making	
  the	
  Python	
  interpreter	
  do	
  a	
  lot	
  of	
  work?	
  
	
  
•  Key	
  quesUon	
  2:	
  Am	
  I	
  blocking	
  other	
  interpreted	
  code	
  from	
  execu-ng?	
  
	
  
•  Key	
  quesUon	
  3:	
  Am	
  I	
  handling	
  data	
  (memory)	
  in	
  a	
  “good”	
  way?	
  
9	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Toy	
  example:	
  interpreted	
  vs.	
  compiled	
  code	
  
10	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Toy	
  example:	
  interpreted	
  vs.	
  compiled	
  code	
  
	
  
Cython: 78x faster than
interpreted
11	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Toy	
  example:	
  interpreted	
  vs.	
  compiled	
  code	
  
NumPy
Creating a full 80MB temporary array +
PyArray_Sum is only 35% slower than a
fully inlined Cython ( C ) function
Interesting: ndarray.sum by itself is almost
2x faster than the hand-coded Cython
function...
12	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Submarines	
  and	
  Icebergs:	
  metaphors	
  for	
  fast	
  Python	
  so]ware	
  
13	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
SubopUmal	
  control	
  flow	
  
Elsewhere Data
Python
code
Python data
structures
Pure Python
computation
Python data
structures
Pure Python
computation
Python data
structures
Data
Deserialization Serialization
Time for a coffee
break...
14	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Beler	
  control	
  flow	
  
Extension
code
(C / C++)
Native
data
Python
code
C Func
Native
data
Native
data
C Func
Python
app logic
Python
app logic
Users only see this!
Zoom zoom!
(if the extension code is good)
15	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
But	
  it’s	
  much	
  easier	
  to	
  write	
  100%	
  Python!	
  
•  Building	
  hybrid	
  C/C++	
  and	
  Python	
  systems	
  adds	
  a	
  lot	
  of	
  complexity	
  to	
  the	
  
engineering	
  process	
  
•  (but	
  it’s	
  o]en	
  worth	
  it)	
  
•  See:	
  Cython,	
  SWIG,	
  Boost.Python,	
  Pybind11,	
  and	
  other	
  “hybrid”	
  so]ware	
  
creaUon	
  tools	
  
	
  
•  BONUS:	
  Python	
  programs	
  can	
  orchestrate	
  mulU-­‐threaded	
  /	
  concurrent	
  systems	
  
wrilen	
  in	
  C/C++	
  (no	
  Python	
  C	
  API	
  needed)	
  
•  The	
  GIL	
  only	
  comes	
  in	
  when	
  you	
  need	
  to	
  “bubble	
  up”	
  data	
  or	
  control	
  flow	
  
(e.g.	
  Python	
  callbacks)	
  into	
  the	
  Python	
  interpreter	
  
16	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
A	
  story	
  of	
  reading	
  a	
  CSV	
  file	
  
f	
  =	
  get_stream(...)	
  
df	
  =	
  pandas.read_csv(f,	
  **csv_options)	
  
while more_data():
buffer = f.read()
parse_bytes(buffer)
df = type_infer_columns()
internally, pseudocode
Concerns
Uses PyString_FromStringAndSize, must
hold GIL for this
Synchronous or asynchronous with IO?
Type infer in parallel?
Data structures used?
17	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
It’s	
  All	
  About	
  the	
  Benjamins	
  (Data	
  Structures)	
  
•  The	
  hard	
  currency	
  of	
  data	
  so]ware	
  is:	
  in-­‐memory	
  data	
  structures	
  
•  How	
  costly	
  are	
  they	
  to	
  send	
  and	
  receive?	
  
•  How	
  costly	
  to	
  manipulate	
  and	
  munge	
  in-­‐memory?	
  
•  How	
  difficult	
  is	
  it	
  to	
  add	
  new	
  proprietary	
  computaUon	
  logic?	
  
	
  
•  In	
  Python:	
  NumPy	
  established	
  a	
  gold	
  standard	
  for	
  interoperable	
  array	
  data	
  
•  pandas	
  is	
  built	
  on	
  NumPy,	
  and	
  made	
  it	
  easy	
  to	
  “plug	
  in”	
  to	
  the	
  ecosystem	
  
•  (but	
  there	
  are	
  plenty	
  of	
  warts	
  sUll)	
  
18	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
What’s	
  this	
  have	
  to	
  do	
  with	
  Spark?	
  
•  Some	
  known	
  performance	
  issues	
  in	
  PySpark	
  
•  IO	
  throughput	
  
•  Python	
  to	
  Spark	
  
•  Spark	
  to	
  Python	
  (or	
  Python	
  extension	
  code)	
  
•  Running	
  interpreted	
  Python	
  code	
  on	
  RDDs	
  /	
  Spark	
  DataFrames	
  
•  Lambda	
  mappers	
  /	
  reducers	
  (rdd.map(...))	
  
•  Spark	
  SQL	
  UDFs	
  (registerFuncUon(...))	
  
	
  
19	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Spark	
  IO	
  throughput	
  to/from	
  Python	
  
1.15 MB/s in
9.82 MB/s out
Spark 1.6.1 running on
localhost
76 MB pandas.DataFrame
20	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Spark	
  IO	
  throughput	
  to/from	
  Python	
  
Unofficial improved
toPandas
25.6 MB/s out
21	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Compared	
  with	
  HiveServer2	
  Thri]	
  RPC	
  fetch	
  
Impala 2.5 + Parquet
file on localhost
ibis + impyla
41.46 MB/s read
hs2client (C++ / Python)
90.8 MB/s
Task benchmarked: Thrift TFetchResultsReq + deserialization + conversion to
pandas.DataFrame
22	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Back	
  of	
  envelope	
  comp	
  w/	
  file	
  formats	
  
Feather: 1105 MB/s write
CSV (pandas): 6.2 MB/s write
Feather: 2414 MB/s read
CSV (pandas): 51.9 MB/s read
disclaimer: warm NVMe / OS file cache
23	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Aside:	
  CSVs	
  can	
  be	
  fast	
  
See: https://github.com/wiseio/paratext
24	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
How	
  Python	
  lambdas	
  work	
  in	
  PySpark	
  
Spark
RDD
Python task
worker pool
Python worker
Python worker
Python worker
Python worker
Python worker
Data stream +
pickled PyFunction
See: spark/api/python/PythonRDD.scala
python/pyspark/worker.py
The inner loop of RDD.map
map(f,	
  iterator)	
  
25	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
How	
  Python	
  lambdas	
  perform	
  
NumPy array-oriented operations are
about 100x faster… but that’s not the
whole story
Disclaimer: this isn’t a remotely “fair” comparison, but it helps illustrate the
real pitfalls associated with introducing serialization and RPC/IPC into a
computational process
26	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
How	
  Python	
  lambdas	
  perform	
  
8 cores
1 core
Lessons learned: Python data analytics should
not be based around scalar object iteration
27	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Asides	
  /	
  counterpoints	
  
•  Spark<-­‐>Python	
  IO	
  may	
  not	
  be	
  important	
  -­‐-­‐	
  can	
  leave	
  all	
  of	
  the	
  data	
  remote	
  
•  Spark	
  DataFrame	
  operaUons	
  have	
  reduced	
  the	
  need	
  for	
  many	
  types	
  of	
  Lambda	
  
funcUons	
  
•  Can	
  use	
  binary	
  file	
  formats	
  as	
  an	
  alternate	
  IO	
  interface	
  
•  Parquet	
  (Python	
  support	
  soon	
  via	
  apache/parquet-­‐cpp)	
  
•  Avro	
  (see	
  cavro,	
  fastavro,	
  pyavroc)	
  
•  ORC	
  (needs	
  a	
  Python	
  champion)	
  
•  ...	
  
28	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Apache	
  
Arrow	
  
http://arrow.apache.org
Some slides from Strata-HW talk w/
Jacques Nadeau
29	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Apache	
  Arrow	
  in	
  a	
  Slide	
  
•  New	
  Top-­‐level	
  Apache	
  So]ware	
  FoundaUon	
  project	
  
• 	
  hlp://arrow.apache.org	
  
	
  
•  Focused	
  on	
  Columnar	
  In-­‐Memory	
  AnalyUcs	
  
1.  10-­‐100x	
  speedup	
  on	
  many	
  workloads	
  
2.  Common	
  data	
  layer	
  enables	
  companies	
  to	
  choose	
  best	
  of	
  
breed	
  systems	
  	
  
3.  Designed	
  to	
  work	
  with	
  any	
  programming	
  language	
  
4.  Support	
  for	
  both	
  relaUonal	
  and	
  complex	
  data	
  as-­‐is	
  
	
  
•  Oriented	
  at	
  collaboraUon	
  amongst	
  other	
  OSS	
  projects	
  
Calcite
Cassandra
Deeplearning4j
Drill
Hadoop
HBase
Ibis
Impala
Kudu
Pandas
Parquet
Phoenix
Spark
Storm
R
30	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
High	
  Performance	
  Sharing	
  &	
  Interchange	
  
Today With Arrow
•  Each system has its own internal
memory format
•  70-80% CPU wasted on serialization
and deserialization
•  Similar functionality implemented in
multiple projects
•  All systems utilize the same memory
format
•  No overhead for cross-system
communication
•  Projects can share functionality (eg,
Parquet-to-Arrow reader)
31	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Arrow	
  and	
  PySpark	
  
•  Build	
  a	
  C	
  API	
  level	
  data	
  protocol	
  to	
  move	
  data	
  between	
  Spark	
  and	
  Python	
  
•  Either	
  	
  
•  (Fast)	
  Convert	
  Arrow	
  to/from	
  pandas.DataFrame	
  
•  (Faster)	
  Perform	
  naUve	
  analyUcs	
  on	
  Arrow	
  data	
  in-­‐memory	
  
•  Use	
  Arrow	
  
•  For	
  efficiently	
  handling	
  nested	
  Spark	
  SQL	
  data	
  in-­‐memory	
  
•  IO:	
  pandas/NumPy	
  data	
  push/pull	
  
•  Lambda/UDF	
  evaluaUon	
  
	
  
32	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
• Problem:	
  fast,	
  language-­‐
agnosUc	
  binary	
  data	
  frame	
  
file	
  format	
  
• Creators:	
  Wes	
  McKinney	
  
(Python)	
  and	
  Hadley	
  
Wickham	
  (R)	
  
• Read	
  speeds	
  close	
  to	
  disk	
  IO	
  
performance	
  
Arrow	
  in	
  acUon:	
  Feather	
  File	
  Format	
  for	
  Python	
  and	
  R	
  
33	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
More	
  on	
  Feather	
  
array 0
array 1
array 2
...
array n - 1
METADATA
Feather File
libfeather
C++ library
Rcpp
Cython
R data.frame
pandas DataFrame
34	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Summary	
  
•  It’s	
  essenUal	
  to	
  improve	
  Spark’s	
  low-­‐level	
  data	
  interoperability	
  with	
  the	
  Python	
  
data	
  ecosystem	
  
	
  
•  I’m	
  personally	
  excited	
  to	
  work	
  with	
  the	
  Spark	
  +	
  Arrow	
  +	
  PyData	
  +	
  other	
  
communiUes	
  to	
  help	
  make	
  this	
  a	
  reality	
  
	
  
35	
  ©	
  Cloudera,	
  Inc.	
  All	
  rights	
  reserved.	
  
Thank	
  you	
  
Wes	
  McKinney	
  @wesmckinn	
  
Views	
  are	
  my	
  own	
  

More Related Content

What's hot

The Rise of ZStandard: Apache Spark/Parquet/ORC/Avro
The Rise of ZStandard: Apache Spark/Parquet/ORC/AvroThe Rise of ZStandard: Apache Spark/Parquet/ORC/Avro
The Rise of ZStandard: Apache Spark/Parquet/ORC/Avro
Databricks
 
Best Practices for Enabling Speculative Execution on Large Scale Platforms
Best Practices for Enabling Speculative Execution on Large Scale PlatformsBest Practices for Enabling Speculative Execution on Large Scale Platforms
Best Practices for Enabling Speculative Execution on Large Scale Platforms
Databricks
 
MySQL Performance Schema in 20 Minutes
 MySQL Performance Schema in 20 Minutes MySQL Performance Schema in 20 Minutes
MySQL Performance Schema in 20 Minutes
Sveta Smirnova
 
Tuning SQL for Oracle Exadata: The Good, The Bad, and The Ugly Tuning SQL fo...
 Tuning SQL for Oracle Exadata: The Good, The Bad, and The Ugly Tuning SQL fo... Tuning SQL for Oracle Exadata: The Good, The Bad, and The Ugly Tuning SQL fo...
Tuning SQL for Oracle Exadata: The Good, The Bad, and The Ugly Tuning SQL fo...
Enkitec
 
Radical Speed for SQL Queries on Databricks: Photon Under the Hood
Radical Speed for SQL Queries on Databricks: Photon Under the HoodRadical Speed for SQL Queries on Databricks: Photon Under the Hood
Radical Speed for SQL Queries on Databricks: Photon Under the Hood
Databricks
 
Magnet Shuffle Service: Push-based Shuffle at LinkedIn
Magnet Shuffle Service: Push-based Shuffle at LinkedInMagnet Shuffle Service: Push-based Shuffle at LinkedIn
Magnet Shuffle Service: Push-based Shuffle at LinkedIn
Databricks
 

What's hot (20)

Datadog: a Real-Time Metrics Database for One Quadrillion Points/Day
Datadog: a Real-Time Metrics Database for One Quadrillion Points/DayDatadog: a Real-Time Metrics Database for One Quadrillion Points/Day
Datadog: a Real-Time Metrics Database for One Quadrillion Points/Day
 
The Rise of ZStandard: Apache Spark/Parquet/ORC/Avro
The Rise of ZStandard: Apache Spark/Parquet/ORC/AvroThe Rise of ZStandard: Apache Spark/Parquet/ORC/Avro
The Rise of ZStandard: Apache Spark/Parquet/ORC/Avro
 
Advanced RAC troubleshooting: Network
Advanced RAC troubleshooting: NetworkAdvanced RAC troubleshooting: Network
Advanced RAC troubleshooting: Network
 
Understanding Oracle RAC 11g Release 2 Internals
Understanding Oracle RAC 11g Release 2 InternalsUnderstanding Oracle RAC 11g Release 2 Internals
Understanding Oracle RAC 11g Release 2 Internals
 
Performant Streaming in Production: Preventing Common Pitfalls when Productio...
Performant Streaming in Production: Preventing Common Pitfalls when Productio...Performant Streaming in Production: Preventing Common Pitfalls when Productio...
Performant Streaming in Production: Preventing Common Pitfalls when Productio...
 
Best Practices for Enabling Speculative Execution on Large Scale Platforms
Best Practices for Enabling Speculative Execution on Large Scale PlatformsBest Practices for Enabling Speculative Execution on Large Scale Platforms
Best Practices for Enabling Speculative Execution on Large Scale Platforms
 
Road to NODES - Handling Neo4j Data with Apache Hop
Road to NODES - Handling Neo4j Data with Apache HopRoad to NODES - Handling Neo4j Data with Apache Hop
Road to NODES - Handling Neo4j Data with Apache Hop
 
Low Level CPU Performance Profiling Examples
Low Level CPU Performance Profiling ExamplesLow Level CPU Performance Profiling Examples
Low Level CPU Performance Profiling Examples
 
MySQL Performance Schema in 20 Minutes
 MySQL Performance Schema in 20 Minutes MySQL Performance Schema in 20 Minutes
MySQL Performance Schema in 20 Minutes
 
Spark and Object Stores —What You Need to Know: Spark Summit East talk by Ste...
Spark and Object Stores —What You Need to Know: Spark Summit East talk by Ste...Spark and Object Stores —What You Need to Know: Spark Summit East talk by Ste...
Spark and Object Stores —What You Need to Know: Spark Summit East talk by Ste...
 
Getting Started with Splunk Enterprise - Demo
Getting Started with Splunk Enterprise - DemoGetting Started with Splunk Enterprise - Demo
Getting Started with Splunk Enterprise - Demo
 
Tuning SQL for Oracle Exadata: The Good, The Bad, and The Ugly Tuning SQL fo...
 Tuning SQL for Oracle Exadata: The Good, The Bad, and The Ugly Tuning SQL fo... Tuning SQL for Oracle Exadata: The Good, The Bad, and The Ugly Tuning SQL fo...
Tuning SQL for Oracle Exadata: The Good, The Bad, and The Ugly Tuning SQL fo...
 
Radical Speed for SQL Queries on Databricks: Photon Under the Hood
Radical Speed for SQL Queries on Databricks: Photon Under the HoodRadical Speed for SQL Queries on Databricks: Photon Under the Hood
Radical Speed for SQL Queries on Databricks: Photon Under the Hood
 
DOAG Oracle Database Vault
DOAG Oracle Database VaultDOAG Oracle Database Vault
DOAG Oracle Database Vault
 
Clone Oracle Databases In Minutes Without Risk Using Enterprise Manager 13c
Clone Oracle Databases In Minutes Without Risk Using Enterprise Manager 13cClone Oracle Databases In Minutes Without Risk Using Enterprise Manager 13c
Clone Oracle Databases In Minutes Without Risk Using Enterprise Manager 13c
 
Grounding Conversational AI in a Knowledge Base
Grounding Conversational AI in a Knowledge BaseGrounding Conversational AI in a Knowledge Base
Grounding Conversational AI in a Knowledge Base
 
Adaptive Query Execution: Speeding Up Spark SQL at Runtime
Adaptive Query Execution: Speeding Up Spark SQL at RuntimeAdaptive Query Execution: Speeding Up Spark SQL at Runtime
Adaptive Query Execution: Speeding Up Spark SQL at Runtime
 
Real-time Stream Processing with Apache Flink
Real-time Stream Processing with Apache FlinkReal-time Stream Processing with Apache Flink
Real-time Stream Processing with Apache Flink
 
Magnet Shuffle Service: Push-based Shuffle at LinkedIn
Magnet Shuffle Service: Push-based Shuffle at LinkedInMagnet Shuffle Service: Push-based Shuffle at LinkedIn
Magnet Shuffle Service: Push-based Shuffle at LinkedIn
 
7th Class Certificate
7th Class Certificate7th Class Certificate
7th Class Certificate
 

Similar to High Performance Python on Apache Spark

28March2024-Codeless-Generative-AI-Pipelines
28March2024-Codeless-Generative-AI-Pipelines28March2024-Codeless-Generative-AI-Pipelines
28March2024-Codeless-Generative-AI-Pipelines
Timothy Spann
 

Similar to High Performance Python on Apache Spark (20)

Enabling Python to be a Better Big Data Citizen
Enabling Python to be a Better Big Data CitizenEnabling Python to be a Better Big Data Citizen
Enabling Python to be a Better Big Data Citizen
 
Next-generation Python Big Data Tools, powered by Apache Arrow
Next-generation Python Big Data Tools, powered by Apache ArrowNext-generation Python Big Data Tools, powered by Apache Arrow
Next-generation Python Big Data Tools, powered by Apache Arrow
 
Python Data Ecosystem: Thoughts on Building for the Future
Python Data Ecosystem: Thoughts on Building for the FuturePython Data Ecosystem: Thoughts on Building for the Future
Python Data Ecosystem: Thoughts on Building for the Future
 
Apache Arrow and Python: The latest
Apache Arrow and Python: The latestApache Arrow and Python: The latest
Apache Arrow and Python: The latest
 
An Incomplete Data Tools Landscape for Hackers in 2015
An Incomplete Data Tools Landscape for Hackers in 2015An Incomplete Data Tools Landscape for Hackers in 2015
An Incomplete Data Tools Landscape for Hackers in 2015
 
PyData: The Next Generation
PyData: The Next GenerationPyData: The Next Generation
PyData: The Next Generation
 
Improving data interoperability in Python and R
Improving data interoperability in Python and RImproving data interoperability in Python and R
Improving data interoperability in Python and R
 
Improving Data Interoperability for Python and R
Improving Data Interoperability for Python and RImproving Data Interoperability for Python and R
Improving Data Interoperability for Python and R
 
PyData: The Next Generation | Data Day Texas 2015
PyData: The Next Generation | Data Day Texas 2015PyData: The Next Generation | Data Day Texas 2015
PyData: The Next Generation | Data Day Texas 2015
 
Apache Kudu Fast Analytics on Fast Data (Hadoop / Spark Conference Japan 2016...
Apache Kudu Fast Analytics on Fast Data (Hadoop / Spark Conference Japan 2016...Apache Kudu Fast Analytics on Fast Data (Hadoop / Spark Conference Japan 2016...
Apache Kudu Fast Analytics on Fast Data (Hadoop / Spark Conference Japan 2016...
 
PyData Boston 2013
PyData Boston 2013PyData Boston 2013
PyData Boston 2013
 
28March2024-Codeless-Generative-AI-Pipelines
28March2024-Codeless-Generative-AI-Pipelines28March2024-Codeless-Generative-AI-Pipelines
28March2024-Codeless-Generative-AI-Pipelines
 
Luciano Resende - Scaling Big Data Interactive Workloads across Kubernetes Cl...
Luciano Resende - Scaling Big Data Interactive Workloads across Kubernetes Cl...Luciano Resende - Scaling Big Data Interactive Workloads across Kubernetes Cl...
Luciano Resende - Scaling Big Data Interactive Workloads across Kubernetes Cl...
 
DatEngConf SF16 - Apache Kudu: Fast Analytics on Fast Data
DatEngConf SF16 - Apache Kudu: Fast Analytics on Fast DataDatEngConf SF16 - Apache Kudu: Fast Analytics on Fast Data
DatEngConf SF16 - Apache Kudu: Fast Analytics on Fast Data
 
IoT Edge Data Processing with NVidia Jetson Nano oct 3 2019
IoT  Edge Data Processing with NVidia Jetson Nano oct 3 2019IoT  Edge Data Processing with NVidia Jetson Nano oct 3 2019
IoT Edge Data Processing with NVidia Jetson Nano oct 3 2019
 
TensorFlow meetup: Keras - Pytorch - TensorFlow.js
TensorFlow meetup: Keras - Pytorch - TensorFlow.jsTensorFlow meetup: Keras - Pytorch - TensorFlow.js
TensorFlow meetup: Keras - Pytorch - TensorFlow.js
 
Ibis: operating the Python data ecosystem at Hadoop scale by Wes McKinney
Ibis: operating the Python data ecosystem at Hadoop scale by Wes McKinneyIbis: operating the Python data ecosystem at Hadoop scale by Wes McKinney
Ibis: operating the Python data ecosystem at Hadoop scale by Wes McKinney
 
Building a Hadoop Data Warehouse with Impala
Building a Hadoop Data Warehouse with ImpalaBuilding a Hadoop Data Warehouse with Impala
Building a Hadoop Data Warehouse with Impala
 
Accelerate and Scale Big Data Analytics with Disaggregated Compute and Storage
Accelerate and Scale Big Data Analytics with Disaggregated Compute and StorageAccelerate and Scale Big Data Analytics with Disaggregated Compute and Storage
Accelerate and Scale Big Data Analytics with Disaggregated Compute and Storage
 
Building a Hadoop Data Warehouse with Impala
Building a Hadoop Data Warehouse with ImpalaBuilding a Hadoop Data Warehouse with Impala
Building a Hadoop Data Warehouse with Impala
 

More from Wes McKinney

More from Wes McKinney (20)

The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
 
Solving Enterprise Data Challenges with Apache Arrow
Solving Enterprise Data Challenges with Apache ArrowSolving Enterprise Data Challenges with Apache Arrow
Solving Enterprise Data Challenges with Apache Arrow
 
Apache Arrow: Open Source Standard Becomes an Enterprise Necessity
Apache Arrow: Open Source Standard Becomes an Enterprise NecessityApache Arrow: Open Source Standard Becomes an Enterprise Necessity
Apache Arrow: Open Source Standard Becomes an Enterprise Necessity
 
Apache Arrow: High Performance Columnar Data Framework
Apache Arrow: High Performance Columnar Data FrameworkApache Arrow: High Performance Columnar Data Framework
Apache Arrow: High Performance Columnar Data Framework
 
New Directions for Apache Arrow
New Directions for Apache ArrowNew Directions for Apache Arrow
New Directions for Apache Arrow
 
Apache Arrow Flight: A New Gold Standard for Data Transport
Apache Arrow Flight: A New Gold Standard for Data TransportApache Arrow Flight: A New Gold Standard for Data Transport
Apache Arrow Flight: A New Gold Standard for Data Transport
 
ACM TechTalks : Apache Arrow and the Future of Data Frames
ACM TechTalks : Apache Arrow and the Future of Data FramesACM TechTalks : Apache Arrow and the Future of Data Frames
ACM TechTalks : Apache Arrow and the Future of Data Frames
 
Apache Arrow: Present and Future @ ScaledML 2020
Apache Arrow: Present and Future @ ScaledML 2020Apache Arrow: Present and Future @ ScaledML 2020
Apache Arrow: Present and Future @ ScaledML 2020
 
PyCon Colombia 2020 Python for Data Analysis: Past, Present, and Future
PyCon Colombia 2020 Python for Data Analysis: Past, Present, and Future PyCon Colombia 2020 Python for Data Analysis: Past, Present, and Future
PyCon Colombia 2020 Python for Data Analysis: Past, Present, and Future
 
Apache Arrow: Leveling Up the Analytics Stack
Apache Arrow: Leveling Up the Analytics StackApache Arrow: Leveling Up the Analytics Stack
Apache Arrow: Leveling Up the Analytics Stack
 
Apache Arrow Workshop at VLDB 2019 / BOSS Session
Apache Arrow Workshop at VLDB 2019 / BOSS SessionApache Arrow Workshop at VLDB 2019 / BOSS Session
Apache Arrow Workshop at VLDB 2019 / BOSS Session
 
Apache Arrow: Leveling Up the Data Science Stack
Apache Arrow: Leveling Up the Data Science StackApache Arrow: Leveling Up the Data Science Stack
Apache Arrow: Leveling Up the Data Science Stack
 
Ursa Labs and Apache Arrow in 2019
Ursa Labs and Apache Arrow in 2019Ursa Labs and Apache Arrow in 2019
Ursa Labs and Apache Arrow in 2019
 
PyCon.DE / PyData Karlsruhe keynote: "Looking backward, looking forward"
PyCon.DE / PyData Karlsruhe keynote: "Looking backward, looking forward"PyCon.DE / PyData Karlsruhe keynote: "Looking backward, looking forward"
PyCon.DE / PyData Karlsruhe keynote: "Looking backward, looking forward"
 
Apache Arrow at DataEngConf Barcelona 2018
Apache Arrow at DataEngConf Barcelona 2018Apache Arrow at DataEngConf Barcelona 2018
Apache Arrow at DataEngConf Barcelona 2018
 
Apache Arrow: Cross-language Development Platform for In-memory Data
Apache Arrow: Cross-language Development Platform for In-memory DataApache Arrow: Cross-language Development Platform for In-memory Data
Apache Arrow: Cross-language Development Platform for In-memory Data
 
Apache Arrow -- Cross-language development platform for in-memory data
Apache Arrow -- Cross-language development platform for in-memory dataApache Arrow -- Cross-language development platform for in-memory data
Apache Arrow -- Cross-language development platform for in-memory data
 
Shared Infrastructure for Data Science
Shared Infrastructure for Data ScienceShared Infrastructure for Data Science
Shared Infrastructure for Data Science
 
Data Science Without Borders (JupyterCon 2017)
Data Science Without Borders (JupyterCon 2017)Data Science Without Borders (JupyterCon 2017)
Data Science Without Borders (JupyterCon 2017)
 
Memory Interoperability in Analytics and Machine Learning
Memory Interoperability in Analytics and Machine LearningMemory Interoperability in Analytics and Machine Learning
Memory Interoperability in Analytics and Machine Learning
 

Recently uploaded

Easier, Faster, and More Powerful – Alles Neu macht der Mai -Wir durchleuchte...
Easier, Faster, and More Powerful – Alles Neu macht der Mai -Wir durchleuchte...Easier, Faster, and More Powerful – Alles Neu macht der Mai -Wir durchleuchte...
Easier, Faster, and More Powerful – Alles Neu macht der Mai -Wir durchleuchte...
panagenda
 
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
TrustArc
 
CORS (Kitworks Team Study 양다윗 발표자료 240510)
CORS (Kitworks Team Study 양다윗 발표자료 240510)CORS (Kitworks Team Study 양다윗 발표자료 240510)
CORS (Kitworks Team Study 양다윗 발표자료 240510)
Wonjun Hwang
 

Recently uploaded (20)

Easier, Faster, and More Powerful – Alles Neu macht der Mai -Wir durchleuchte...
Easier, Faster, and More Powerful – Alles Neu macht der Mai -Wir durchleuchte...Easier, Faster, and More Powerful – Alles Neu macht der Mai -Wir durchleuchte...
Easier, Faster, and More Powerful – Alles Neu macht der Mai -Wir durchleuchte...
 
ChatGPT and Beyond - Elevating DevOps Productivity
ChatGPT and Beyond - Elevating DevOps ProductivityChatGPT and Beyond - Elevating DevOps Productivity
ChatGPT and Beyond - Elevating DevOps Productivity
 
ERP Contender Series: Acumatica vs. Sage Intacct
ERP Contender Series: Acumatica vs. Sage IntacctERP Contender Series: Acumatica vs. Sage Intacct
ERP Contender Series: Acumatica vs. Sage Intacct
 
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
TrustArc Webinar - Unified Trust Center for Privacy, Security, Compliance, an...
 
AI mind or machine power point presentation
AI mind or machine power point presentationAI mind or machine power point presentation
AI mind or machine power point presentation
 
JavaScript Usage Statistics 2024 - The Ultimate Guide
JavaScript Usage Statistics 2024 - The Ultimate GuideJavaScript Usage Statistics 2024 - The Ultimate Guide
JavaScript Usage Statistics 2024 - The Ultimate Guide
 
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)
AI+A11Y 11MAY2024 HYDERBAD GAAD 2024 - HelloA11Y (11 May 2024)
 
JohnPollard-hybrid-app-RailsConf2024.pptx
JohnPollard-hybrid-app-RailsConf2024.pptxJohnPollard-hybrid-app-RailsConf2024.pptx
JohnPollard-hybrid-app-RailsConf2024.pptx
 
How to Check GPS Location with a Live Tracker in Pakistan
How to Check GPS Location with a Live Tracker in PakistanHow to Check GPS Location with a Live Tracker in Pakistan
How to Check GPS Location with a Live Tracker in Pakistan
 
UiPath manufacturing technology benefits and AI overview
UiPath manufacturing technology benefits and AI overviewUiPath manufacturing technology benefits and AI overview
UiPath manufacturing technology benefits and AI overview
 
Design and Development of a Provenance Capture Platform for Data Science
Design and Development of a Provenance Capture Platform for Data ScienceDesign and Development of a Provenance Capture Platform for Data Science
Design and Development of a Provenance Capture Platform for Data Science
 
CORS (Kitworks Team Study 양다윗 발표자료 240510)
CORS (Kitworks Team Study 양다윗 발표자료 240510)CORS (Kitworks Team Study 양다윗 발표자료 240510)
CORS (Kitworks Team Study 양다윗 발표자료 240510)
 
Design Guidelines for Passkeys 2024.pptx
Design Guidelines for Passkeys 2024.pptxDesign Guidelines for Passkeys 2024.pptx
Design Guidelines for Passkeys 2024.pptx
 
Oauth 2.0 Introduction and Flows with MuleSoft
Oauth 2.0 Introduction and Flows with MuleSoftOauth 2.0 Introduction and Flows with MuleSoft
Oauth 2.0 Introduction and Flows with MuleSoft
 
Introduction to use of FHIR Documents in ABDM
Introduction to use of FHIR Documents in ABDMIntroduction to use of FHIR Documents in ABDM
Introduction to use of FHIR Documents in ABDM
 
الأمن السيبراني - ما لا يسع للمستخدم جهله
الأمن السيبراني - ما لا يسع للمستخدم جهلهالأمن السيبراني - ما لا يسع للمستخدم جهله
الأمن السيبراني - ما لا يسع للمستخدم جهله
 
How we scaled to 80K users by doing nothing!.pdf
How we scaled to 80K users by doing nothing!.pdfHow we scaled to 80K users by doing nothing!.pdf
How we scaled to 80K users by doing nothing!.pdf
 
State of the Smart Building Startup Landscape 2024!
State of the Smart Building Startup Landscape 2024!State of the Smart Building Startup Landscape 2024!
State of the Smart Building Startup Landscape 2024!
 
Microsoft CSP Briefing Pre-Engagement - Questionnaire
Microsoft CSP Briefing Pre-Engagement - QuestionnaireMicrosoft CSP Briefing Pre-Engagement - Questionnaire
Microsoft CSP Briefing Pre-Engagement - Questionnaire
 
Event-Driven Architecture Masterclass: Integrating Distributed Data Stores Ac...
Event-Driven Architecture Masterclass: Integrating Distributed Data Stores Ac...Event-Driven Architecture Masterclass: Integrating Distributed Data Stores Ac...
Event-Driven Architecture Masterclass: Integrating Distributed Data Stores Ac...
 

High Performance Python on Apache Spark

  • 1. 1  ©  Cloudera,  Inc.  All  rights  reserved.   High  Performance  Python  on   Apache  Spark   Wes  McKinney  @wesmckinn   Spark  Summit  West  -­‐-­‐  June  7,  2016  
  • 2. 2  ©  Cloudera,  Inc.  All  rights  reserved.   Me   •  Data  Science  Tools  at  Cloudera   •  Serial  creator  of  structured  data  tools  /  user  interfaces   •  Wrote  bestseller  Python  for  Data  Analysis  2012   •   Working  on  expanded  and  revised  2nd  edi-on,  coming  2017   •  Open  source  projects   •  Python  {pandas,  Ibis,  statsmodels}   •  Apache  {Arrow,  Parquet,  Kudu  (incubaUng)}   •  Focused  on  C++,  Python,  and  Hybrid  projects  
  • 3. 3  ©  Cloudera,  Inc.  All  rights  reserved.   Agenda   •  Why  care  about  Python?     •  What  does  “high  performance  Python”  even  mean?       •  A  modern  approach  to  Python  data  so]ware     •  Spark  and  Python:  performance  analysis  and  development  direcUons  
  • 4. 4  ©  Cloudera,  Inc.  All  rights  reserved.   •  Accessible,  “swiss  army  knife”  programming  language     •  Highly  producUve  for  so]ware  engineering  and  data  science  alike     •  Has  excelled  as  the  agile  “orchestraUon”  or  “glue”  layer  for  applicaUon   business  logic     •  Easy  to  interface  with  C  /  C++  /  Fortran  code.  Well-­‐designed  Python  C  API     Why  care  about  (C)Python?  
  • 5. 5  ©  Cloudera,  Inc.  All  rights  reserved.   Defining  “High  Performance  Python”   •  The  end-­‐user  workflow  involves  primarily  Python  programming;  programs  can   be  invoked  with  “python  app_entry_point.py  ...”     •  The  so]ware  uses  system  resources  within  an  acceptable  factor  of  an   equivalent  program  developed  completely  in  Java  or  C++   •  Preferably  1-­‐5x  slower,  not  20-­‐50x     •  The  so]ware  is  suitable  for  interacUve  /  exploratory  compuUng  on  modestly   large  data  sets  (=  gigabytes)  on  a  single  node  
  • 6. 6  ©  Cloudera,  Inc.  All  rights  reserved.   Building  fast  Python  so]ware   means  embracing  certain   limitaUons  
  • 7. 7  ©  Cloudera,  Inc.  All  rights  reserved.   Having  a  healthy  relaUonship  with  the  interpreter   •  The  Python  interpreter  itself  is  “slow”,  as  compared  with  hand-­‐coded  C  or  Java   •  Each  line  of  Python  code  may  feature  mulUple  internal  C  API  calls,   temporary  data  structures,  etc.     •  Python  built-­‐in  data  structures  (numbers,  strings,  tuples,  lists,  dicts,  etc.)  have   significant  memory  and  performance  use  overhead     •  Threads  performing  concurrent  CPU  or  IO  work  must  take  care  not  to  block   other  threads  
  • 8. 8  ©  Cloudera,  Inc.  All  rights  reserved.   Mantras  for  great  success   •  Key  quesUon  1:  Am  I  making  the  Python  interpreter  do  a  lot  of  work?     •  Key  quesUon  2:  Am  I  blocking  other  interpreted  code  from  execu-ng?     •  Key  quesUon  3:  Am  I  handling  data  (memory)  in  a  “good”  way?  
  • 9. 9  ©  Cloudera,  Inc.  All  rights  reserved.   Toy  example:  interpreted  vs.  compiled  code  
  • 10. 10  ©  Cloudera,  Inc.  All  rights  reserved.   Toy  example:  interpreted  vs.  compiled  code     Cython: 78x faster than interpreted
  • 11. 11  ©  Cloudera,  Inc.  All  rights  reserved.   Toy  example:  interpreted  vs.  compiled  code   NumPy Creating a full 80MB temporary array + PyArray_Sum is only 35% slower than a fully inlined Cython ( C ) function Interesting: ndarray.sum by itself is almost 2x faster than the hand-coded Cython function...
  • 12. 12  ©  Cloudera,  Inc.  All  rights  reserved.   Submarines  and  Icebergs:  metaphors  for  fast  Python  so]ware  
  • 13. 13  ©  Cloudera,  Inc.  All  rights  reserved.   SubopUmal  control  flow   Elsewhere Data Python code Python data structures Pure Python computation Python data structures Pure Python computation Python data structures Data Deserialization Serialization Time for a coffee break...
  • 14. 14  ©  Cloudera,  Inc.  All  rights  reserved.   Beler  control  flow   Extension code (C / C++) Native data Python code C Func Native data Native data C Func Python app logic Python app logic Users only see this! Zoom zoom! (if the extension code is good)
  • 15. 15  ©  Cloudera,  Inc.  All  rights  reserved.   But  it’s  much  easier  to  write  100%  Python!   •  Building  hybrid  C/C++  and  Python  systems  adds  a  lot  of  complexity  to  the   engineering  process   •  (but  it’s  o]en  worth  it)   •  See:  Cython,  SWIG,  Boost.Python,  Pybind11,  and  other  “hybrid”  so]ware   creaUon  tools     •  BONUS:  Python  programs  can  orchestrate  mulU-­‐threaded  /  concurrent  systems   wrilen  in  C/C++  (no  Python  C  API  needed)   •  The  GIL  only  comes  in  when  you  need  to  “bubble  up”  data  or  control  flow   (e.g.  Python  callbacks)  into  the  Python  interpreter  
  • 16. 16  ©  Cloudera,  Inc.  All  rights  reserved.   A  story  of  reading  a  CSV  file   f  =  get_stream(...)   df  =  pandas.read_csv(f,  **csv_options)   while more_data(): buffer = f.read() parse_bytes(buffer) df = type_infer_columns() internally, pseudocode Concerns Uses PyString_FromStringAndSize, must hold GIL for this Synchronous or asynchronous with IO? Type infer in parallel? Data structures used?
  • 17. 17  ©  Cloudera,  Inc.  All  rights  reserved.   It’s  All  About  the  Benjamins  (Data  Structures)   •  The  hard  currency  of  data  so]ware  is:  in-­‐memory  data  structures   •  How  costly  are  they  to  send  and  receive?   •  How  costly  to  manipulate  and  munge  in-­‐memory?   •  How  difficult  is  it  to  add  new  proprietary  computaUon  logic?     •  In  Python:  NumPy  established  a  gold  standard  for  interoperable  array  data   •  pandas  is  built  on  NumPy,  and  made  it  easy  to  “plug  in”  to  the  ecosystem   •  (but  there  are  plenty  of  warts  sUll)  
  • 18. 18  ©  Cloudera,  Inc.  All  rights  reserved.   What’s  this  have  to  do  with  Spark?   •  Some  known  performance  issues  in  PySpark   •  IO  throughput   •  Python  to  Spark   •  Spark  to  Python  (or  Python  extension  code)   •  Running  interpreted  Python  code  on  RDDs  /  Spark  DataFrames   •  Lambda  mappers  /  reducers  (rdd.map(...))   •  Spark  SQL  UDFs  (registerFuncUon(...))    
  • 19. 19  ©  Cloudera,  Inc.  All  rights  reserved.   Spark  IO  throughput  to/from  Python   1.15 MB/s in 9.82 MB/s out Spark 1.6.1 running on localhost 76 MB pandas.DataFrame
  • 20. 20  ©  Cloudera,  Inc.  All  rights  reserved.   Spark  IO  throughput  to/from  Python   Unofficial improved toPandas 25.6 MB/s out
  • 21. 21  ©  Cloudera,  Inc.  All  rights  reserved.   Compared  with  HiveServer2  Thri]  RPC  fetch   Impala 2.5 + Parquet file on localhost ibis + impyla 41.46 MB/s read hs2client (C++ / Python) 90.8 MB/s Task benchmarked: Thrift TFetchResultsReq + deserialization + conversion to pandas.DataFrame
  • 22. 22  ©  Cloudera,  Inc.  All  rights  reserved.   Back  of  envelope  comp  w/  file  formats   Feather: 1105 MB/s write CSV (pandas): 6.2 MB/s write Feather: 2414 MB/s read CSV (pandas): 51.9 MB/s read disclaimer: warm NVMe / OS file cache
  • 23. 23  ©  Cloudera,  Inc.  All  rights  reserved.   Aside:  CSVs  can  be  fast   See: https://github.com/wiseio/paratext
  • 24. 24  ©  Cloudera,  Inc.  All  rights  reserved.   How  Python  lambdas  work  in  PySpark   Spark RDD Python task worker pool Python worker Python worker Python worker Python worker Python worker Data stream + pickled PyFunction See: spark/api/python/PythonRDD.scala python/pyspark/worker.py The inner loop of RDD.map map(f,  iterator)  
  • 25. 25  ©  Cloudera,  Inc.  All  rights  reserved.   How  Python  lambdas  perform   NumPy array-oriented operations are about 100x faster… but that’s not the whole story Disclaimer: this isn’t a remotely “fair” comparison, but it helps illustrate the real pitfalls associated with introducing serialization and RPC/IPC into a computational process
  • 26. 26  ©  Cloudera,  Inc.  All  rights  reserved.   How  Python  lambdas  perform   8 cores 1 core Lessons learned: Python data analytics should not be based around scalar object iteration
  • 27. 27  ©  Cloudera,  Inc.  All  rights  reserved.   Asides  /  counterpoints   •  Spark<-­‐>Python  IO  may  not  be  important  -­‐-­‐  can  leave  all  of  the  data  remote   •  Spark  DataFrame  operaUons  have  reduced  the  need  for  many  types  of  Lambda   funcUons   •  Can  use  binary  file  formats  as  an  alternate  IO  interface   •  Parquet  (Python  support  soon  via  apache/parquet-­‐cpp)   •  Avro  (see  cavro,  fastavro,  pyavroc)   •  ORC  (needs  a  Python  champion)   •  ...  
  • 28. 28  ©  Cloudera,  Inc.  All  rights  reserved.   Apache   Arrow   http://arrow.apache.org Some slides from Strata-HW talk w/ Jacques Nadeau
  • 29. 29  ©  Cloudera,  Inc.  All  rights  reserved.   Apache  Arrow  in  a  Slide   •  New  Top-­‐level  Apache  So]ware  FoundaUon  project   •   hlp://arrow.apache.org     •  Focused  on  Columnar  In-­‐Memory  AnalyUcs   1.  10-­‐100x  speedup  on  many  workloads   2.  Common  data  layer  enables  companies  to  choose  best  of   breed  systems     3.  Designed  to  work  with  any  programming  language   4.  Support  for  both  relaUonal  and  complex  data  as-­‐is     •  Oriented  at  collaboraUon  amongst  other  OSS  projects   Calcite Cassandra Deeplearning4j Drill Hadoop HBase Ibis Impala Kudu Pandas Parquet Phoenix Spark Storm R
  • 30. 30  ©  Cloudera,  Inc.  All  rights  reserved.   High  Performance  Sharing  &  Interchange   Today With Arrow •  Each system has its own internal memory format •  70-80% CPU wasted on serialization and deserialization •  Similar functionality implemented in multiple projects •  All systems utilize the same memory format •  No overhead for cross-system communication •  Projects can share functionality (eg, Parquet-to-Arrow reader)
  • 31. 31  ©  Cloudera,  Inc.  All  rights  reserved.   Arrow  and  PySpark   •  Build  a  C  API  level  data  protocol  to  move  data  between  Spark  and  Python   •  Either     •  (Fast)  Convert  Arrow  to/from  pandas.DataFrame   •  (Faster)  Perform  naUve  analyUcs  on  Arrow  data  in-­‐memory   •  Use  Arrow   •  For  efficiently  handling  nested  Spark  SQL  data  in-­‐memory   •  IO:  pandas/NumPy  data  push/pull   •  Lambda/UDF  evaluaUon    
  • 32. 32  ©  Cloudera,  Inc.  All  rights  reserved.   • Problem:  fast,  language-­‐ agnosUc  binary  data  frame   file  format   • Creators:  Wes  McKinney   (Python)  and  Hadley   Wickham  (R)   • Read  speeds  close  to  disk  IO   performance   Arrow  in  acUon:  Feather  File  Format  for  Python  and  R  
  • 33. 33  ©  Cloudera,  Inc.  All  rights  reserved.   More  on  Feather   array 0 array 1 array 2 ... array n - 1 METADATA Feather File libfeather C++ library Rcpp Cython R data.frame pandas DataFrame
  • 34. 34  ©  Cloudera,  Inc.  All  rights  reserved.   Summary   •  It’s  essenUal  to  improve  Spark’s  low-­‐level  data  interoperability  with  the  Python   data  ecosystem     •  I’m  personally  excited  to  work  with  the  Spark  +  Arrow  +  PyData  +  other   communiUes  to  help  make  this  a  reality    
  • 35. 35  ©  Cloudera,  Inc.  All  rights  reserved.   Thank  you   Wes  McKinney  @wesmckinn   Views  are  my  own