Index
Symbols
- --jars command-line argument, Submitting applications
- ./bin/pyspark, Launching the Python console, Starting Spark
- ./bin/spark-shell, Launching the Scala console, Starting Spark
- =!= operator, Concatenating and Appending Rows (Union), Working with Booleans
- == (equal to) expression, Working with Booleans
- ` (backtick) character, Reserved Characters and Keywords
A
- Accelerated Failure Time (AFT), Survival Regression (Accelerated Failure Time)
- accumulators
- basic example, Basic Example-Basic Example
- custom, Custom Accumulators
- overview of, Distributed Shared Variables, Accumulators
- acknowledgements, Acknowledgments
- actions, Actions
- ADAM Project, Spark Packages
- advanced analytics (see machine learning and advanced analytics)
- aggregate function, aggregate
- AggregateByKey function, aggregateByKey
- AggregationBuffer, User-Defined Aggregation Functions
- aggregations
- aggregate function, aggregate
- AggregateByKey function, aggregateByKey
- aggregation functions, Aggregation Functions-Aggregating to Complex Types
- CombineByKey function, combineByKey
- countByKey, countByKey
- debugging, Slow Aggregations
- foldByKey function, foldByKey
- groupByKey, groupByKey
- grouping, Grouping-Grouping with Maps
- grouping sets, Grouping Sets-Pivot
- on complex types, Aggregating to Complex Types
- overview of, Aggregations-Aggregations
- performance tuning, Aggregations
- on RDDs, Aggregations-foldByKey
- reduceByKey, reduceByKey
- in Structured Streaming API, Aggregations
- User-Defined Aggregation Functions (UDAFs), User-Defined Aggregation Functions
- window functions ...
Get Spark: The Definitive Guide now with the O’Reilly learning platform.
O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.