About 39,800 results
Open links in new tab
  1. PySpark Overview — PySpark 4.1.0 documentation - Apache Spark

    Dec 11, 2025 · PySpark combines Python’s learnability and ease of use with the power of Apache Spark to enable processing and analysis of data at any size for everyone familiar with Python. PySpark …

  2. Spark SQL and DataFrames - Spark 4.1.0 Documentation

    Spark SQL is a Spark module for structured data processing. Unlike the basic Spark RDD API, the interfaces provided by Spark SQL provide Spark with more information about the structure of both …

  3. Structured Streaming Programming Guide - Spark 4.1.0 Documentation

    Structured Streaming is a scalable and fault-tolerant stream processing engine built on the Spark SQL engine. You can express your streaming computation the same way you would express a batch …

  4. Configuration - Spark 4.1.0 Documentation

    Spark provides three locations to configure the system: Spark properties control most application parameters and can be set by using a SparkConf object, or through Java system properties. …

  5. Overview - Spark 4.0.0 Documentation

    If you’d like to build Spark from source, visit Building Spark. Spark runs on both Windows and UNIX-like systems (e.g. Linux, Mac OS), and it should run on any platform that runs a supported version of Java.

  6. Spark Release 3.5.4 - Apache Spark

    While being a maintenance release we did still upgrade some dependencies in this release they are: [SPARK-50150]: Upgrade Jetty to 9.4.56.v20240826 [SPARK-50316]: Upgrade ORC to 1.9.5 You …

  7. Pandas API on Spark — PySpark 4.1.0 documentation

    Specify the index column in conversion from Spark DataFrame to pandas-on-Spark DataFrame Use distributed or distributed-sequence default index Handling index misalignment with distributed …

  8. Structured Streaming Programming Guide - Spark 4.1.0 Documentation

    Types of time windows Spark supports three types of time windows: tumbling (fixed), sliding and session. Tumbling windows are a series of fixed-sized, non-overlapping and contiguous time …

  9. Apache Spark™ - Unified Engine for large-scale data analytics

    Apache Spark is a multi-language engine for executing data engineering, data science, and machine learning on single-node machines or clusters.

  10. Spark Release 3.5.5 - Apache Spark

    Dependency changes While being a maintenance release we did still upgrade some dependencies in this release they are: [SPARK-50886]: Upgrade Avro to 1.11.4 You can consult JIRA for the detailed …