Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
130 commits
Select commit Hold shift + click to select a range
a8d01a5
[SPARK-54868][PYTHON][INFRA][FOLLOWUP] Add PYSPARK_TEST_TIMEOUT to ho…
gaogaotiantian Feb 24, 2026
730e239
[SPARK-55643][INFRA] Add connection timeout to JIRA client to prevent…
dongjoon-hyun Feb 24, 2026
37e4c2d
[SPARK-55646][SQL] Refactored SQLExecution.withThreadLocalCaptured to…
huanliwang-db Feb 24, 2026
2f37724
[SPARK-55649][K8S] Promote `Kubernetes(Driver|Executor)?FeatureConfig…
dongjoon-hyun Feb 24, 2026
178bcec
[SPARK-55075][K8S] Track executor pod creation errors with ExecutorFa…
parthchandra Feb 24, 2026
ec35791
[SPARK-55647][SQL] Fix `ConstantPropagation` incorrectly replacing at…
ilicmarkodb Feb 24, 2026
7d67ff3
[SPARK-55654][SQL] Enable TreePattern pruning for EliminateSubqueryAl…
yaooqinn Feb 24, 2026
1e1b363
[SPARK-55653][K8S] Support `NetworkPolicy` for Spark executor pods
dongjoon-hyun Feb 24, 2026
a5316a4
[SPARK-55655][MLLIB] Make `CountVectorizer` vocabulary deterministic …
LuciferYang Feb 24, 2026
7da1356
[SPARK-55264] Add ExecuteOutput command to Spark Connect pipelines proto
aditya-nambiar Feb 24, 2026
7529622
[SPARK-55052][SQL] Add AQEShuffleRead properties to Physical Plan Tree
erenavsarogullari Feb 24, 2026
52a9f01
[SPARK-55622][SQL][TESTS] Add test for DSV2 Tables with multi-part na…
szehon-ho Feb 24, 2026
b27eabf
[SPARK-55659][CORE] Improve `EventLogFileWriter` to log `stop` operation
dongjoon-hyun Feb 24, 2026
7fca140
[SPARK-55648][PS] Handle an unexpected keyword argument error `groupb…
ueshin Feb 24, 2026
f546447
[SPARK-55296][PS][FOLLOW-UP] Disconnect the anchor for more cases to …
ueshin Feb 24, 2026
1e09fc8
[SPARK-55600][PYTHON] Fix pandas to arrow loses row count when schema…
Yicong-Huang Feb 24, 2026
4e55f6e
[SPARK-54666][PS] Leave numeric types unchanged on `to_numeric`
devin-petersohn Feb 24, 2026
f0285c4
[SPARK-55636][CONNECT] Add detailed errors in case of deduplication o…
pranavdev022 Feb 24, 2026
8fd58ed
[SPARK-54276][BUILD] Bump Hadoop 3.4.3
pan3793 Feb 25, 2026
954bb57
[SPARK-55652][SQL] Optimize `VectorizedPlainValuesReader.readShorts()…
LuciferYang Feb 25, 2026
1d43629
[SPARK-55678][BUILD] Add daily test for Java 25
pan3793 Feb 25, 2026
564f20b
[SPARK-55677][BUILD] Upgrade `commons-cli` to 1.11.0
dongjoon-hyun Feb 25, 2026
a81fc2a
[SPARK-55670][BUILD] Add `-Dio.netty.noUnsafe=false` to enable Java 2…
pan3793 Feb 25, 2026
32c3b66
[SPARK-55678][BUILD][FOLLOWUP] Add Maven Java 25 daily CI and update …
pan3793 Feb 25, 2026
054cca1
[SPARK-55656][BUILD] Upgrade Guava to 33.5.0-jre
LuciferYang Feb 25, 2026
b293246
[SPARK-55675][PS][TESTS] Fix tests for window functions
ueshin Feb 25, 2026
8ac083c
[SPARK-55674][PYTHON] Optimize 0-column table conversion in Spark Con…
Yicong-Huang Feb 25, 2026
3e4df1c
[SPARK-55679][CORE] Fix dectecting `sun.io.serialization.extendedDebu…
pan3793 Feb 25, 2026
0e27265
[SPARK-55673][SQL] Add more tests for nested type encoder
szehon-ho Feb 25, 2026
dd22010
[SPARK-55681][SQL] Fix singleton DataType equality after deserialization
timlee0119 Feb 25, 2026
cd01c4c
[SPARK-55682][CORE][SQL] ServiceLoader returned iterator may throw `N…
pan3793 Feb 25, 2026
e5a203c
[SPARK-55684][BUILD] Upgrade sbt to `1.12.4`
sarutak Feb 25, 2026
8bd7b6e
[SPARK-55606][CONNECT] Server-side implementation of GetStatus API
terana Feb 25, 2026
f9a9c94
[SPARK-55683][SQL] Optimize `VectorizedPlainValuesReader.readUnsigned…
LuciferYang Feb 25, 2026
54ff4ea
[SPARK-55322][SQL][TESTS][FOLLOWUP] Fix `max_by and min_by with k` fa…
LuciferYang Feb 25, 2026
0444c02
[SPARK-55155] Fix SET CATALOG to use special chars and backticks in t…
yuyuant Feb 25, 2026
0a02d71
[SPARK-55601][SS] Hook StreamingSourceIdentifyingName into Microbatch…
ericm-db Feb 25, 2026
0a7908e
[SPARK-55688][BUILD] Upgrade `aircompressor` to 2.0.3
dongjoon-hyun Feb 25, 2026
d99e5a1
Revert "[SPARK-41398][SQL][FOLLOWUP] Update runtime filtering javadoc…
dongjoon-hyun Feb 25, 2026
361ee9c
[SPARK-54374][SQL][UI][FOLLOWUP] Fix SQL plan visualization display i…
yaooqinn Feb 25, 2026
38a7533
[SPARK-55694][SQL] Block constraints in CTAS/RTAS at parser level
yyanyy Feb 25, 2026
ebb6817
[SPARK-55672][PS][TESTS] Fix error message check due to StringDtype(n…
ueshin Feb 25, 2026
6bdae9d
[SPARK-55666][PYTHON][TEST] Fix flaky connect tests due to non-determ…
Yicong-Huang Feb 25, 2026
d73c6cb
[SPARK-55406][PYTHON][CONNECT][FOLLOW-UP] Add a timeout for release wait
gaogaotiantian Feb 25, 2026
b370375
[SPARK-55631][SQL] ALTER TABLE must invalidate cache for DSv2 tables
aokolnychyi Feb 25, 2026
53606f2
[SPARK-51695][SQL][DOCS][FOLLOW-UP] Clarify NULL handling semantics i…
yyanyy Feb 26, 2026
191801b
[SPARK-55665][PYTHON] Unify how workers establish connection with the…
gaogaotiantian Feb 26, 2026
3ca6c0f
[SPARK-55697][INFRA] Add `create_spark_jira.py` script
dongjoon-hyun Feb 26, 2026
e3b6b10
[SPARK-55671][PYTHON][TESTS] Add zstandard requirement to connect req…
gaogaotiantian Feb 26, 2026
e92ef55
[SPARK-55640][GEO][SQL] Propagate WKB parsing errors for Geometry and…
uros-db Feb 26, 2026
1d56813
[SPARK-55703][K8S][DOCS][INFRA] Upgrade Volcano to 1.14.1
dongjoon-hyun Feb 26, 2026
778e43c
[SPARK-55687][SQL][TEST] Fix ComputeCurrentTimeSuite - No duplicate l…
pan3793 Feb 26, 2026
361ece3
[SPARK-55704][K8S] Add `Constants.DEFAULT_PVC_ACCESS_MODE` for reuse
dongjoon-hyun Feb 26, 2026
811c3fa
[SPARK-55706][SQL][TESTS] Disable DB2 JDBC Driver tests
dongjoon-hyun Feb 26, 2026
ee467ff
[SPARK-55711][UI][FOLLOWUP] Fix Job DAG visualization display issue c…
yaooqinn Feb 26, 2026
1b99a56
[SPARK-55539][GEO][SQL] Allow casting from GeographyType to GeometryType
uros-db Feb 26, 2026
02453b1
[SPARK-55712][INFRA] Allow run benchmark with JDK 25
pan3793 Feb 26, 2026
7742e73
[SPARK-55709][K8S][TESTS] Fix `ExecutorLifecycleTestUtils.persistentV…
dongjoon-hyun Feb 26, 2026
3665506
[SPARK-55692][SQL] Fix `SupportsRuntimeFiltering` and `SupportsRuntim…
peter-toth Feb 26, 2026
baa8507
[SPARK-55710][BUILD] Use Google Mirror of Maven Central for SBT boots…
yaooqinn Feb 26, 2026
c15e36b
[SPARK-55705][PYTHON][INFRA] Upgrade PyArrow to 23
zhengruifeng Feb 26, 2026
f88addf
[SPARK-55662][PS] Implementation of idxmin Axis argument
devin-petersohn Feb 26, 2026
08a2cb9
[SPARK-55296][PS][FOLLOW-UP] Fix setitem with Series through iloc ind…
ueshin Feb 26, 2026
9e6b257
[SPARK-55700][PS] Fix handling integer keys on Series with non-intege…
ueshin Feb 26, 2026
4da26e4
[SPARK-55683][SQL][FOLLOWUP] Optimize `VectorizedPlainValuesReader.re…
LuciferYang Feb 26, 2026
6775a17
[SPARK-55702][SQL] Support filter predicate in window aggregate funct…
cloud-fan Feb 27, 2026
3b343b4
[SPARK-55713][PYTHON][TESTS] Add `LongArrowToPandasBenchmark` for lon…
zhengruifeng Feb 27, 2026
df1378e
[SPARK-55725][K8S][DOCS] Make `AbstractPodsAllocator` docs up-to-date
dongjoon-hyun Feb 27, 2026
8026f7d
[SPARK-55699][SS] Inconsistent reading of LowLatencyClock when used t…
eason-yuchen-liu Feb 27, 2026
7182a79
[SPARK-55463][PYTHON][FOLLOW-UP] Delete unnecessary `date_as_object=T…
zhengruifeng Feb 27, 2026
2996e30
[MINOR][PYTHON] Add `asv` in `dev/requirements.txt`
zhengruifeng Feb 27, 2026
fcf8636
[SPARK-55144][SS] Introduce new state format version for performant s…
HeartSaVioR Feb 27, 2026
1b1904e
[SPARK-55405][PYTHON][TESTS][FOLLOW-UP] Add macOS ARM overrides for f…
Yicong-Huang Feb 27, 2026
c26d0bc
[SPARK-55701][SS] Fix race condition in CompactibleFileStreamLog.allF…
zeruibao Feb 27, 2026
555cd38
[SPARK-55645][SQL] Add serdeName to CatalogStorageFormat
tagatac Feb 27, 2026
e52dfc5
[MINOR][DOCS] Document how to run a specified benchmark
zhengruifeng Feb 27, 2026
61c87b2
[SPARK-55743][PYTHON][TESTS] Add `python/benchmarks` in python linter
zhengruifeng Feb 27, 2026
3f64bca
[SPARK-55736][PYTHON][TESTS] Add tests for mismatched batch sizes in …
zhengruifeng Feb 27, 2026
b9ad073
[MINOR][INFRA] Update outdated comment for GHA runner
pan3793 Feb 27, 2026
3b06a65
[SPARK-55696][SQL] Add explicit error to Encoders.bean for interface …
szehon-ho Feb 27, 2026
72225dd
[SPARK-55740][TESTS] Fix npm vulnerabilities by `npm audit fix`
LuciferYang Feb 27, 2026
dc57455
[SPARK-54571][CORE][SQL] Use LZ4 safeDecompressor to mitigate perf re…
pan3793 Feb 27, 2026
a83eec8
[SPARK-55691][CONNECT] GetStatus client
terana Feb 27, 2026
1e76dd8
[SPARK-55519][SQL][TESTS] Fix V2FunctionBenchmark
pan3793 Feb 27, 2026
6a0a905
[SPARK-55739][SQL] Optimize `OnHeapColumnVector.putIntsLittleEndian/p…
LuciferYang Feb 27, 2026
3f2575d
[SPARK-55745][TESTS] Add GHA benchmark reports for JDK 25
pan3793 Feb 27, 2026
a0a092f
[SPARK-55746][SQL][TESTS] Fix unable to load custom metric object Sup…
pan3793 Feb 27, 2026
eae6ede
[SPARK-55741][INFRA] Use `ubuntu-slim` for GitHub Actions `preconditi…
dongjoon-hyun Feb 27, 2026
e2cd90b
[SPARK-55730][PYTHON] Not make timezone lower case
ueshin Feb 27, 2026
117e171
[SPARK-55510][SS] Fix deleteRange of Rocksdb state store to call chan…
zeruibao Feb 27, 2026
835689e
[SPARK-47672][SQL] Avoid double eval from filter pushDown w/ projecti…
sfc-gh-hkarau Feb 27, 2026
168e5cf
[DOCS] Clarify DataFrames in quickstart
celestehorgan Feb 27, 2026
f66c336
[SPARK-55501][SQL] Fix listagg distinct + within group order by bug
helioshe4 Feb 28, 2026
4b65223
[SPARK-55639][K8S] Support Recovery-mode K8s Executors
dongjoon-hyun Feb 28, 2026
40db51f
[SPARK-55752][UI] Upgrade D3.js to 7.9.0 and vis-timeline to 7.7.3
yaooqinn Feb 28, 2026
959cc95
[SPARK-55702][SQL][FOLLOWUP] Clean up dead error code and fix flaky w…
cloud-fan Feb 28, 2026
730d375
[SPARK-55756][SQL][TESTS] Add --DEBUG directive for golden file test …
cloud-fan Feb 28, 2026
a7990f3
[SPARK-55755][SQL][TESTS] Handle null ArithmeticException message on …
yaooqinn Feb 28, 2026
3f54ac5
Revert "[SPARK-55755][SQL][TESTS] Handle null ArithmeticException mes…
pan3793 Feb 28, 2026
81101bc
[SPARK-55757][CORE] Improve `spark.task.cpus` validation
dongjoon-hyun Feb 28, 2026
392dd8d
[SPARK-55753][UI] Upgrade Bootstrap from 4.6.2 to 5.3.8
yaooqinn Mar 1, 2026
c0e367a
[SPARK-55747][SQL] Fix NPE when accessing elements from an array that…
cloud-fan Mar 1, 2026
7c9456f
[SPARK-55699][SS][FOLLOWUP] Inconsistent reading of LowLatencyClock w…
eason-yuchen-liu Mar 1, 2026
3ef0362
[SPARK-55714][SQL] JDK might throw ArithmeticException without message
pan3793 Mar 2, 2026
56faeb4
[SPARK-55464][PYTHON] Support GeographyType in convert_numpy
fangchenli Mar 2, 2026
f612cd0
[SPARK-37711][PS] Reduce pandas describe job count from O(N) to O(1)
devin-petersohn Mar 2, 2026
1cc5916
[SPARK-55663][PYTHON] Unify __module__ for data source functions
gaogaotiantian Mar 2, 2026
b38e8ea
[SPARK-55583][PYTHON] Validate Arrow schema types in Python data source
Yicong-Huang Mar 2, 2026
88d5c8b
[SPARK-55780][UI] Replace PNG logo with SVG in Spark Web UI
yaooqinn Mar 2, 2026
77f05a9
[SPARK-55776][UI] Fix data-title to data-bs-title for timeline toolti…
yaooqinn Mar 2, 2026
983ccfb
[SPARK-55771][UI] Modernize progress bars using Bootstrap 5 Progress …
yaooqinn Mar 2, 2026
f2ff4a2
[SPARK-55764][UI] Use delegated event listener for Bootstrap 5 Toolti…
yaooqinn Mar 2, 2026
6d02e24
[SPARK-54687][SQL] Add more edge cases with generators
mikhailnik-db Mar 2, 2026
8b71111
[SPARK-55440][SQL] Types Framework - Phase 1a - Core Type System Foun…
davidm-db Mar 2, 2026
4ebdc4a
[SPARK-55794][SQL] Always alias `OuterReference`s
mihailotim-db Mar 2, 2026
f16570f
[SPARK-54807][SQL] Allow qualified names for built-in and session fun…
srielau Mar 2, 2026
a09b5c9
[SPARK-53226][SPARK-52088][CORE] Make ClosureCleaner work with Java22+
sarutak Mar 3, 2026
3a5f503
[SPARK-55721][INFRA] Group sbt build messages in github actions
gaogaotiantian Mar 3, 2026
854628f
[SPARK-55326][PYTHON][CONNECT] Release remote session when SPARK_CONN…
wbo4958 Mar 3, 2026
7b404cb
[SPARK-55621][PYTHON] Fix ambiguous and unnecessary unicode usage
gaogaotiantian Mar 3, 2026
5b6b806
[SPARK-55685][BUILD] Upgrade ORC to 2.3.0
dongjoon-hyun Mar 3, 2026
be10dc0
[SPARK-55801][PYTHON] Fix type hint of _SimpleStreamReaderWrapper.get…
gaogaotiantian Mar 3, 2026
0774ba0
[SPARK-55803][BUILD] Bump lz4-java 1.10.4 to bring back performance
pan3793 Mar 3, 2026
8fef8f1
[SPARK-55770][UI] Replace custom .btn-spark with Bootstrap 5 button v…
yaooqinn Mar 3, 2026
eae012e
[SPARK-55773][UI] Replace custom collapse/toggle system with Bootstra…
yaooqinn Mar 3, 2026
3036f86
[SPARK-55492][SQL][SS][FOLLOWUP] Clarify error message; do not use no…
Mar 3, 2026
14d7cca
[SPARK-55771][UI][FOLLOWUP] Fix progress bar height alignment in BS5 …
yaooqinn Mar 3, 2026
c2a739f
[SPARK-55789][CORE][TESTS] Add a microbenchmark for `Platform`
LuciferYang Mar 3, 2026
1323ab5
[SPARK-55800][PYTHON] Remove the unused type check for datetime.date
gaogaotiantian Mar 3, 2026
9eab4f6
[SPARK-XXXXX][SS] Add V1 to V2 offset log auto-upgrade
ericm-db Mar 3, 2026
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
3 changes: 2 additions & 1 deletion .github/workflows/benchmark.yml
Original file line number Diff line number Diff line change
Expand Up @@ -28,12 +28,13 @@ on:
default: '*'
jdk:
type: choice
description: 'JDK version: 17 or 21'
description: 'JDK version: 17, 21, or 25'
required: true
default: '17'
options:
- '17'
- '21'
- '25'
scala:
type: choice
description: 'Scala version: 2.13'
Expand Down
13 changes: 9 additions & 4 deletions .github/workflows/build_and_test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,9 @@ concurrency:
jobs:
precondition:
name: Check changes
runs-on: ubuntu-latest
# `ubuntu-slim` is lighter than `ubuntu-latest`.
# Please see https://docs.github.com/en/actions/how-tos/write-workflows/choose-where-workflows-run/choose-the-runner-for-a-job#standard-github-hosted-runners-for-public-repositories
runs-on: ubuntu-slim
env:
GITHUB_PREV_SHA: ${{ github.event.before }}
outputs:
Expand Down Expand Up @@ -1444,7 +1446,8 @@ jobs:
uses: medyagh/setup-minikube@v0.0.21
with:
kubernetes-version: "1.35.0"
# Github Action limit cpu:2, memory: 6947MB, limit to 2U6G for better resource statistic
# GitHub Actions limit 4C/16G, limit to 2C/6G for better resource statistic
# https://docs.github.com/en/actions/reference/runners/github-hosted-runners#standard-github-hosted-runners-for-public-repositories
cpus: 2
memory: 6144m
- name: Print K8S pods and nodes info
Expand All @@ -1464,7 +1467,7 @@ jobs:
elif [[ "${{ inputs.branch }}" == 'branch-4.0' ]]; then
kubectl apply -f https://raw.githubusercontent.com/volcano-sh/volcano/v1.11.0/installer/volcano-development.yaml || true
else
kubectl apply -f https://raw.githubusercontent.com/volcano-sh/volcano/v1.14.0/installer/volcano-development.yaml || true
kubectl apply -f https://raw.githubusercontent.com/volcano-sh/volcano/v1.14.1/installer/volcano-development.yaml || true
fi
eval $(minikube docker-env)
build/sbt -Phadoop-3 -Psparkr -Pkubernetes -Pvolcano -Pkubernetes-integration-tests -Dspark.kubernetes.test.volcanoMaxConcurrencyJobNum=1 -Dtest.exclude.tags=local "kubernetes-integration-tests/test"
Expand All @@ -1479,7 +1482,9 @@ jobs:
needs: [precondition]
if: fromJson(needs.precondition.outputs.required).ui == 'true'
name: Run Spark UI tests
runs-on: ubuntu-latest
# `ubuntu-slim` is lighter than `ubuntu-latest`.
# Please see https://docs.github.com/en/actions/how-tos/write-workflows/choose-where-workflows-run/choose-the-runner-for-a-job#standard-github-hosted-runners-for-public-repositories
runs-on: ubuntu-slim
timeout-minutes: 120
steps:
- uses: actions/checkout@v6
Expand Down
57 changes: 57 additions & 0 deletions .github/workflows/build_java25.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,57 @@
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
#

name: "Build / Java25 (master, Scala 2.13, Hadoop 3, JDK 25)"

on:
schedule:
- cron: '0 4 * * *'
workflow_dispatch:

jobs:
run-build:
permissions:
packages: write
name: Run
uses: ./.github/workflows/build_and_test.yml
if: github.repository == 'apache/spark'
with:
java: 25
branch: master
hadoop: hadoop3
envs: >-
{
"PYSPARK_IMAGE_TO_TEST": "python-311",
"PYTHON_TO_TEST": "python3.11",
"SKIP_MIMA": "true",
"SKIP_UNIDOC": "true",
"DEDICATED_JVM_SBT_TESTS": "org.apache.spark.sql.execution.datasources.parquet.ParquetFileFormatV1Suite,org.apache.spark.sql.execution.datasources.parquet.ParquetFileFormatV2Suite,org.apache.spark.sql.execution.datasources.orc.OrcSourceV1Suite,org.apache.spark.sql.execution.datasources.orc.OrcSourceV2Suite"
}
jobs: >-
{
"build": "true",
"pyspark": "true",
"sparkr": "true",
"tpcds-1g": "true",
"docker-integration-tests": "true",
"yarn": "true",
"k8s-integration-tests": "true",
"buf": "true",
"ui": "true"
}
35 changes: 35 additions & 0 deletions .github/workflows/build_maven_java25.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,35 @@
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
#

name: "Build / Maven (master, Scala 2.13, Hadoop 3, JDK 25)"

on:
schedule:
- cron: '0 14 * * *'
workflow_dispatch:

jobs:
run-build:
permissions:
packages: write
name: Run
uses: ./.github/workflows/maven_test.yml
if: github.repository == 'apache/spark'
with:
java: 25
3 changes: 2 additions & 1 deletion .github/workflows/python_hosted_runner_test.yml
Original file line number Diff line number Diff line change
Expand Up @@ -102,6 +102,7 @@ jobs:
SKIP_PACKAGING: true
METASPACE_SIZE: 1g
BRANCH: ${{ inputs.branch }}
PYSPARK_TEST_TIMEOUT: 450
steps:
- name: Checkout Spark repository
uses: actions/checkout@v6
Expand Down Expand Up @@ -153,7 +154,7 @@ jobs:
run: |
python${{matrix.python}} -m pip install --ignore-installed 'blinker>=1.6.2'
python${{matrix.python}} -m pip install --ignore-installed 'six==1.16.0'
python${{matrix.python}} -m pip install numpy 'pyarrow>=22.0.0' 'six==1.16.0' 'pandas==2.3.3' scipy 'plotly<6.0.0' 'mlflow>=2.8.1' coverage matplotlib openpyxl 'memory-profiler>=0.61.0' 'scikit-learn>=1.3.2' unittest-xml-reporting && \
python${{matrix.python}} -m pip install numpy 'pyarrow>=23.0.0' 'six==1.16.0' 'pandas==2.3.3' scipy 'plotly<6.0.0' 'mlflow>=2.8.1' coverage matplotlib openpyxl 'memory-profiler>=0.61.0' 'scikit-learn>=1.3.2' unittest-xml-reporting && \
python${{matrix.python}} -m pip install 'grpcio==1.76.0' 'grpcio-status==1.76.0' 'protobuf==6.33.5' 'googleapis-common-protos==1.71.0' 'zstandard==0.25.0' 'graphviz==0.20.3' && \
python${{matrix.python}} -m pip cache purge
- name: List Python packages
Expand Down
4 changes: 3 additions & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -32,12 +32,14 @@ This README file only contains basic setup instructions.
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/publish_snapshot.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/publish_snapshot.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_infra_images_cache.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_infra_images_cache.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_java21.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_java21.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_java25.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_java25.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_non_ansi.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_non_ansi.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_uds.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_uds.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_rockdb_as_ui_backend.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_rockdb_as_ui_backend.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_maven.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_maven.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_maven_java21.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_maven_java21.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_maven_java21_macos26.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_maven_java21_macos26.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_maven_java25.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_maven_java25.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_maven_java21_macos26.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_maven_java21_macos26.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_maven_java21_arm.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_maven_java21_arm.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_coverage.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_coverage.yml) |
| | [![GitHub Actions Build](https://github.com/apache/spark/actions/workflows/build_python_pypy3.10.yml/badge.svg)](https://github.com/apache/spark/actions/workflows/build_python_pypy3.10.yml) |
Expand Down
2 changes: 1 addition & 1 deletion build/sbt-launch-lib.bash
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,7 @@ acquire_sbt_jar () {
# artifacts from internal repos only.
# Ex:
# DEFAULT_ARTIFACT_REPOSITORY=https://artifacts.internal.com/libs-release/
URL1=${DEFAULT_ARTIFACT_REPOSITORY:-https://repo1.maven.org/maven2/}org/scala-sbt/sbt-launch/${SBT_VERSION}/sbt-launch-${SBT_VERSION}.jar
URL1=${DEFAULT_ARTIFACT_REPOSITORY:-https://maven-central.storage-download.googleapis.com/maven2/}org/scala-sbt/sbt-launch/${SBT_VERSION}/sbt-launch-${SBT_VERSION}.jar
JAR=build/sbt-launch-${SBT_VERSION}.jar

sbt_jar=$JAR
Expand Down
36 changes: 30 additions & 6 deletions common/utils/src/main/resources/error/error-conditions.json
Original file line number Diff line number Diff line change
Expand Up @@ -206,6 +206,12 @@
],
"sqlState" : "42K03"
},
"BEAN_ENCODER_INTERFACE_NOT_SUPPORTED" : {
"message" : [
"Bean encoder does not support interface type <className>."
],
"sqlState" : "0A000"
},
"BINARY_ARITHMETIC_OVERFLOW" : {
"message" : [
"<value1> <symbol> <value2> caused overflow. Use <functionName> to ignore overflow problem and return NULL."
Expand Down Expand Up @@ -1677,7 +1683,7 @@
},
"EVENT_TIME_MUST_BE_TOP_LEVEL_COLUMN" : {
"message" : [
"The event time column <eventExpr> must be a top-level column in the schema."
"The event time in withWatermark must be a top-level column, but '<eventExpr>' is a nested field. To use it, alias it to a top-level column in a select before withWatermark."
],
"sqlState" : "42K09"
},
Expand Down Expand Up @@ -4110,6 +4116,12 @@
},
"sqlState" : "42823"
},
"INVALID_TEMP_OBJ_QUALIFIER" : {
"message" : [
"Temporary <objectType> <objectName> cannot be qualified with <qualifier>. Temporary objects can only be qualified with SESSION or SYSTEM.SESSION."
],
"sqlState" : "42602"
},
"INVALID_TEMP_OBJ_REFERENCE" : {
"message" : [
"Cannot create the persistent object <objName> of the type <obj> because it references to the temporary object <tempObjName> of the type <tempObj>. Please make the temporary object <tempObjName> persistent, or make the persistent object <objName> temporary."
Expand Down Expand Up @@ -4180,6 +4192,12 @@
],
"sqlState" : "42000"
},
"INVALID_USAGE_OF_STAR_WITH_TABLE_IDENTIFIER_IN_COUNT" : {
"message" : [
"count(<tableName>.*) is not allowed. Use count(*) or expand the columns manually, e.g. count(col1, col2)."
],
"sqlState" : "42000"
},
"INVALID_UTF8_STRING" : {
"message" : [
"Invalid UTF8 byte sequence found in string: <str>."
Expand Down Expand Up @@ -4290,6 +4308,11 @@
"The function is invoked with DISTINCT and WITHIN GROUP but expressions <funcArg> and <orderingExpr> do not match. The WITHIN GROUP ordering expression must be picked from the function inputs."
]
},
"MISMATCH_WITH_DISTINCT_INPUT_UNSAFE_CAST" : {
"message" : [
"The function <funcName> with DISTINCT and WITHIN GROUP (ORDER BY) is not supported for <inputType> input. Explicitly cast the input to <castType> before passing it to both the function argument and ORDER BY expression."
]
},
"WITHIN_GROUP_MISSING" : {
"message" : [
"WITHIN GROUP is required for the function."
Expand Down Expand Up @@ -7754,6 +7777,12 @@
],
"sqlState" : "42601"
},
"WKB_PARSE_ERROR" : {
"message" : [
"Error parsing WKB: <parseError> at position <pos>"
],
"sqlState" : "22023"
},
"WRITE_STREAM_NOT_ALLOWED" : {
"message" : [
"`writeStream` can be called only on streaming Dataset/DataFrame."
Expand Down Expand Up @@ -7991,11 +8020,6 @@
"count(<targetString>.*) is not allowed. Please use count(*) or expand the columns manually, e.g. count(col1, col2)."
]
},
"_LEGACY_ERROR_TEMP_1030" : {
"message" : [
"Window aggregate function with filter predicate is not supported yet."
]
},
"_LEGACY_ERROR_TEMP_1031" : {
"message" : [
"It is not allowed to use a window function inside an aggregate function. Please use the inner window function in a sub-query."
Expand Down
Loading
Loading