Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions .github/labeler.yml
Original file line number Diff line number Diff line change
Expand Up @@ -30,9 +30,9 @@ BUILD:
- changed-files:
- any-glob-to-any-file: [
'dev/**/*',
'ep/build-velox/src/get_velox.sh',
'ep/build-velox/src/build_velox.sh',
'ep/build-clickhouse/src/build_clickhouse.sh'
'ep/build-velox/src/get-velox.sh',
'ep/build-velox/src/build-velox.sh',
'ep/build-clickhouse/src/build-clickhouse.sh'
]

DOCS:
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/build_bundle_package.yml
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@ name: Build bundle package
env:
ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION: true
CCACHE_DIR: "${{ github.workspace }}/.ccache"
SETUP: 'bash .github/workflows/util/setup_helper.sh'
SETUP: 'bash .github/workflows/util/setup-helper.sh'

concurrency:
group: ${{ github.repository }}-${{ github.head_ref || github.sha }}-${{ github.workflow }}
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/docker_image.yml
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ on:
- main
paths:
- '.github/workflows/docker_image.yml'
- '.github/workflows/util/install_spark_resources.sh'
- '.github/workflows/util/install-spark-resources.sh'
- 'dev/docker/Dockerfile.centos7-static-build'
- 'dev/docker/Dockerfile.centos8-static-build'
- 'dev/docker/Dockerfile.centos9-static-build'
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/scala_code_format.yml
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ concurrency:
cancel-in-progress: true

env:
SETUP: 'bash .github/workflows/util/setup_helper.sh'
SETUP: 'bash .github/workflows/util/setup-helper.sh'

jobs:

Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/velox_backend_arm.yml
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,7 @@ env:
ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION: true
MVN_CMD: 'mvn -ntp'
WGET_CMD: 'wget -nv'
SETUP: 'bash .github/workflows/util/setup_helper.sh'
SETUP: 'bash .github/workflows/util/setup-helper.sh'
CCACHE_DIR: "${{ github.workspace }}/.ccache"

concurrency:
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/velox_backend_cache.yml
Original file line number Diff line number Diff line change
Expand Up @@ -183,7 +183,7 @@ jobs:
# run: |
# rm -rf /opt/miniconda-for-velox/
# cd ep/build-velox/src && \
# ./get_velox.sh
# ./get-velox.sh
# cd ../build/velox_ep/
# make EXTRA_CMAKE_FLAGS="-DVELOX_ENABLE_PARQUET=ON -DVELOX_BUILD_TESTING=ON -DVELOX_BUILD_TEST_UTILS=ON"

Expand Down Expand Up @@ -221,7 +221,7 @@ jobs:
# run: |
# rm -rf /opt/miniconda-for-velox/
# cd ep/build-velox/src && \
# ./get_velox.sh
# ./get-velox.sh
# cd ../build/velox_ep/
# source /opt/rh/gcc-toolset-9/enable
# make EXTRA_CMAKE_FLAGS="-DVELOX_ENABLE_PARQUET=ON -DVELOX_BUILD_TESTING=ON -DVELOX_BUILD_TEST_UTILS=ON"
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/velox_backend_enhanced.yml
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@ env:
ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION: true
MVN_CMD: 'mvn -ntp'
WGET_CMD: 'wget -nv'
SETUP: 'bash .github/workflows/util/setup_helper.sh'
SETUP: 'bash .github/workflows/util/setup-helper.sh'
CCACHE_DIR: "${{ github.workspace }}/.ccache"

concurrency:
Expand Down Expand Up @@ -229,7 +229,7 @@ jobs:
- name: Prepare Spark Resources for Spark 3.5.5
run: |
rm -rf /opt/shims/spark35
bash .github/workflows/util/install_spark_resources.sh 3.5
bash .github/workflows/util/install-spark-resources.sh 3.5
- name: Build and Run unit test for Spark 3.5.5 (slow tests)
run: |
cd $GITHUB_WORKSPACE/
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/velox_backend_x86.yml
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,7 @@ env:
ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION: true
MVN_CMD: 'mvn -ntp'
WGET_CMD: 'wget -nv'
SETUP: 'source .github/workflows/util/setup_helper.sh'
SETUP: 'source .github/workflows/util/setup-helper.sh'
CCACHE_DIR: "${{ github.workspace }}/.ccache"
# spark.sql.ansi.enabled defaults to false.
SPARK_ANSI_SQL_MODE: false
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/velox_nightly.yml
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ on:
env:
ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION: true
CCACHE_DIR: "${{ github.workspace }}/.ccache"
SETUP: 'bash .github/workflows/util/setup_helper.sh'
SETUP: 'bash .github/workflows/util/setup-helper.sh'

concurrency:
group: ${{ github.repository }}-${{ github.head_ref || github.sha }}-${{ github.workflow }}
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/velox_weekly.yml
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ on:
env:
ACTIONS_ALLOW_USE_UNSECURE_NODE_VERSION: true
MVN_CMD: 'mvn -ntp'
SETUP: 'source .github/workflows/util/setup_helper.sh'
SETUP: 'source .github/workflows/util/setup-helper.sh'
TPCH_TEST: "env GLUTEN_IT_JVM_ARGS=-Xmx5G sbin/gluten-it.sh queries-compare
--local --preset=velox --benchmark-type=h --error-on-memleak --off-heap-size=10g -s=1.0 --threads=16 --iterations=1"
INSTALL_PREFIX: /usr/local
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ import java.util.concurrent.TimeUnit

/*
* Note: Gluten configiguration.md is automatically generated from this code.
* Make sure to run dev/gen_all_config_docs.sh after making changes to this file.
* Make sure to run dev/gen-all-config-docs.sh after making changes to this file.
*/
class VeloxConfig(conf: SQLConf) extends GlutenConfig(conf) {
import VeloxConfig._
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -79,6 +79,6 @@ class AllVeloxConfiguration extends AnyFunSuite {
AllGlutenConfiguration.verifyOrRegenerateGoldenFile(
markdown,
builder.toMarkdown,
"dev/gen_all_config_docs.sh")
"dev/gen-all-config-docs.sh")
}
}
2 changes: 1 addition & 1 deletion dev/build_arrow.sh → dev/build-arrow.sh
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ set -exu

CURRENT_DIR=$(cd "$(dirname "$BASH_SOURCE")"; pwd)
SUDO="${SUDO:-""}"
source ${CURRENT_DIR}/build_helper_functions.sh
source ${CURRENT_DIR}/build-helper-functions.sh
VELOX_ARROW_BUILD_VERSION=15.0.0
ARROW_PREFIX=$CURRENT_DIR/../ep/_ep/arrow_ep
BUILD_TYPE=Release
Expand Down
File renamed without changes.
2 changes: 1 addition & 1 deletion dev/build_libhdfs3.sh → dev/build-libhdfs3.sh
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ set -exu

CURRENT_DIR=$(cd "$(dirname "$BASH_SOURCE")"; pwd)
export SUDO=sudo
source ${CURRENT_DIR}/build_helper_functions.sh
source ${CURRENT_DIR}/build-helper-functions.sh
DEPENDENCY_DIR=${DEPENDENCY_DIR:-$CURRENT_DIR/../ep/_ep}

function build_libhdfs3 {
Expand Down
8 changes: 4 additions & 4 deletions dev/builddeps-veloxbe.sh
Original file line number Diff line number Diff line change
Expand Up @@ -210,14 +210,14 @@ function build_arrow {
get_velox && setup_dependencies
fi
cd $GLUTEN_DIR/dev
source ./build_arrow.sh
source ./build-arrow.sh
}

function build_velox {
echo "Start to build Velox"
cd $GLUTEN_DIR/ep/build-velox/src
# When BUILD_TESTS is on for gluten cpp, we need turn on VELOX_BUILD_TEST_UTILS via build_test_utils.
./build_velox.sh --enable_s3=$ENABLE_S3 --enable_gcs=$ENABLE_GCS --build_type=$BUILD_TYPE --enable_hdfs=$ENABLE_HDFS \
./build-velox.sh --enable_s3=$ENABLE_S3 --enable_gcs=$ENABLE_GCS --build_type=$BUILD_TYPE --enable_hdfs=$ENABLE_HDFS \
--enable_abfs=$ENABLE_ABFS --enable_gpu=$ENABLE_GPU --build_test_utils=$BUILD_TESTS \
--build_tests=$BUILD_VELOX_TESTS --build_benchmarks=$BUILD_VELOX_BENCHMARKS --num_threads=$NUM_THREADS \
--velox_home=$VELOX_HOME
Expand Down Expand Up @@ -263,14 +263,14 @@ function build_velox_backend {

function get_velox {
cd $GLUTEN_DIR/ep/build-velox/src
./get_velox.sh $VELOX_PARAMETER
./get-velox.sh $VELOX_PARAMETER
}

function setup_dependencies {
DEPENDENCY_DIR=${DEPENDENCY_DIR:-$CURRENT_DIR/../ep/_ep}
mkdir -p ${DEPENDENCY_DIR}

source $GLUTEN_DIR/dev/build_helper_functions.sh
source $GLUTEN_DIR/dev/build-helper-functions.sh
source ${VELOX_HOME}/scripts/setup-common.sh

echo "Start to install dependencies"
Expand Down
10 changes: 5 additions & 5 deletions dev/docker/Dockerfile.centos8-dynamic-build
Original file line number Diff line number Diff line change
Expand Up @@ -47,11 +47,11 @@ RUN set -ex; \
wget -nv https://archive.apache.org/dist/hadoop/common/hadoop-2.8.5/hadoop-2.8.5.tar.gz -P /opt/; \
git clone --depth=1 https://github.com/apache/incubator-gluten /opt/gluten; \
cd /opt/gluten/.github/workflows/util/; \
./install_spark_resources.sh 3.2; \
./install_spark_resources.sh 3.3; \
./install_spark_resources.sh 3.4; \
./install_spark_resources.sh 3.5; \
./install_spark_resources.sh 3.5-scala2.13; \
./install-spark-resources.sh 3.2; \
./install-spark-resources.sh 3.3; \
./install-spark-resources.sh 3.4; \
./install-spark-resources.sh 3.5; \
./install-spark-resources.sh 3.5-scala2.13; \
if [ "$(uname -m)" = "aarch64" ]; then \
export CPU_TARGET="aarch64"; \
fi; \
Expand Down
10 changes: 5 additions & 5 deletions dev/docker/Dockerfile.centos9-dynamic-build
Original file line number Diff line number Diff line change
Expand Up @@ -45,11 +45,11 @@ RUN set -ex; \
wget -nv https://archive.apache.org/dist/hadoop/common/hadoop-2.8.5/hadoop-2.8.5.tar.gz -P /opt/; \
git clone --depth=1 https://github.com/apache/incubator-gluten /opt/gluten; \
cd /opt/gluten/.github/workflows/util/; \
./install_spark_resources.sh 3.2; \
./install_spark_resources.sh 3.3; \
./install_spark_resources.sh 3.4; \
./install_spark_resources.sh 3.5; \
./install_spark_resources.sh 3.5-scala2.13; \
./install-spark-resources.sh 3.2; \
./install-spark-resources.sh 3.3; \
./install-spark-resources.sh 3.4; \
./install-spark-resources.sh 3.5; \
./install-spark-resources.sh 3.5-scala2.13; \
if [ "$(uname -m)" = "aarch64" ]; then \
export CPU_TARGET="aarch64"; \
fi; \
Expand Down
2 changes: 1 addition & 1 deletion dev/docker/cudf/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ RUN yum install -y sudo patch maven perl && ln -sf /usr/local/bin/cmake /usr/bin
RUN git clone --depth=1 https://github.com/apache/incubator-gluten /opt/gluten

# Install spark to folder /opt
RUN cd /opt/gluten/.github/workflows/util/ && ./install_spark_resources.sh 3.4
RUN cd /opt/gluten/.github/workflows/util/ && ./install-spark-resources.sh 3.4
ENV SPARK_HOME=/opt/spark-3.4.4-bin-hadoop3
ENV PATH=$SPARK_HOME/bin:$PATH
ENV CUDA_ARCHITECTURES=70
Expand Down
File renamed without changes.
File renamed without changes.
14 changes: 7 additions & 7 deletions docs/developers/HowTo.md
Original file line number Diff line number Diff line change
Expand Up @@ -128,16 +128,16 @@ mvn test -Pspark-3.5 -Pspark-ut -Pbackends-velox -DargLine="-Dspark.test.home=/p
Please set `wildcardSuites` with a fully qualified class name. `spark.test.home` is optional to set. It is only required for some test suites to use Spark resources.
If you are specifying the `spark.test.home` arg, it should be set to either:
* The path a directory containing Spark source code, which has already been built
* Or use the `install_spark_resources.sh` script to get a directory with the necessary resource files:
* Or use the `install-spark-resources.sh` script to get a directory with the necessary resource files:
```
# Define a directory to use for the Spark files and the Spark version
export spark_dir=/tmp/spark
export spark_version=3.5

# Run the install_spark_resources.sh script
.github/workflows/util/install_spark_resources.sh ${spark_version} ${spark_dir}
# Run the install-spark-resources.sh script
.github/workflows/util/install-spark-resources.sh ${spark_version} ${spark_dir}
```
After running the `install_spark_resources.sh`, define the `spark.test.home` directory like:
After running the `install-spark-resources.sh`, define the `spark.test.home` directory like:
`-DargLine="-Dspark.test.home=${spark_dir}/shims/spark35/spark_home"` when running unit tests.

For most cases, please make sure Gluten native build is done before running a Scala/Java test.
Expand Down Expand Up @@ -186,13 +186,13 @@ Here we will explain how to run TPC-H on Velox backend with the Parquet file for
var gluten_root = "/home/gluten"
```

- Modify `${GLUTEN_HOME}/tools/workload/tpch/run_tpch/tpch_parquet.sh`.
- Modify `${GLUTEN_HOME}/tools/workload/tpch/run_tpch/tpch-parquet.sh`.
- Set `GLUTEN_JAR` correctly. Please refer to the section of [Build Gluten with Velox Backend](../get-started/Velox.md#build-gluten-with-velox-backend)
- Set `SPARK_HOME` correctly.
- Set the memory configurations appropriately.
- Execute `tpch_parquet.sh` using the below command.
- Execute `tpch-parquet.sh` using the below command.
- `cd ${GLUTEN_HOME}/tools/workload/tpch/run_tpch/`
- `./tpch_parquet.sh`
- `./tpch-parquet.sh`

# How to run TPC-DS

Expand Down
2 changes: 1 addition & 1 deletion docs/developers/clickhouse-backend-debug.md
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ parent: /developer-overview/

1. Build Gluten ClickHouse Native Lib.
```
export CMAKE_BUILD_TYPE=Release && bash ep/build-clickhouse/src/build_clickhouse.sh
export CMAKE_BUILD_TYPE=Release && bash ep/build-clickhouse/src/build-clickhouse.sh
```
libch.so will be generated in `cpp-ch/build/utils/extern-local-engine/libch.so`.

Expand Down
4 changes: 2 additions & 2 deletions docs/get-started/ClickHouse.md
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ Following softwares are required,
- cmake 3.20 or higher version
- ninja-build 1.8.2

You can run `sudo $gluten_root/ep/build-clickhouse/src/install_ubuntu.sh` to setup the requirements. We also provide a [docker file](../../cpp-ch/local-engine/docker/image/Dockerfile), you can build your own image
You can run `sudo $gluten_root/ep/build-clickhouse/src/install-ubuntu.sh` to setup the requirements. We also provide a [docker file](../../cpp-ch/local-engine/docker/image/Dockerfile), you can build your own image
```shell
cd $gluten_root/cpp-ch/local-engine/docker/image/
docker build . -t libch_builder
Expand Down Expand Up @@ -96,7 +96,7 @@ If you have setup all requirements, you can use following command to build it di

```bash
cd $gluten_root
bash ./ep/build-clickhouse/src/build_clickhouse.sh
bash ./ep/build-clickhouse/src/build-clickhouse.sh
```


Expand Down
8 changes: 4 additions & 4 deletions docs/get-started/Velox.md
Original file line number Diff line number Diff line change
Expand Up @@ -129,7 +129,7 @@ To enable this functionality, you must set the JAVA_HOME and HADOOP_HOME environ

### Build libhdfs3

If you want to run Gluten with libhdfs3.so, you need to manually compile libhdfs3 to obtain the libhdfs3.so file. We provide the script dev/build_libhdfs3.sh in Gluten to help you compile libhdfs3.so.
If you want to run Gluten with libhdfs3.so, you need to manually compile libhdfs3 to obtain the libhdfs3.so file. We provide the script dev/build-libhdfs3.sh in Gluten to help you compile libhdfs3.so.

### Build with HDFS support

Expand Down Expand Up @@ -161,7 +161,7 @@ cp /path/to/hdfs-client.xml hdfs-client.xml

One typical deployment on Spark/HDFS cluster is to enable [short-circuit reading](https://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-hdfs/ShortCircuitLocalReads.html). Short-circuit reads provide a substantial performance boost to many applications.

By default libhdfs3 does not set the default hdfs domain socket path to support HDFS short-circuit read. If this feature is required in HDFS setup, users may need to setup the domain socket path correctly by patching the libhdfs3 source code or by setting the correct config environment. In Gluten the short-circuit domain socket path is set to "/var/lib/hadoop-hdfs/dn_socket" in [build_velox.sh](https://github.com/apache/incubator-gluten/blob/main/ep/build-velox/src/build_velox.sh) So we need to make sure the folder existed and user has write access as below script.
By default libhdfs3 does not set the default hdfs domain socket path to support HDFS short-circuit read. If this feature is required in HDFS setup, users may need to setup the domain socket path correctly by patching the libhdfs3 source code or by setting the correct config environment. In Gluten the short-circuit domain socket path is set to "/var/lib/hadoop-hdfs/dn_socket" in [build-velox.sh](https://github.com/apache/incubator-gluten/blob/main/ep/build-velox/src/build-velox.sh) So we need to make sure the folder existed and user has write access as below script.

```
sudo mkdir -p /var/lib/hadoop-hdfs/
Expand Down Expand Up @@ -461,8 +461,8 @@ All TPC-H and TPC-DS queries are supported in Gluten Velox backend. You may refe

## Data preparation

The data generation scripts are [TPC-H dategen script](../../tools/workload/tpch/gen_data/parquet_dataset/tpch_datagen_parquet.sh) and
[TPC-DS dategen script](../../tools/workload/tpcds/gen_data/parquet_dataset/tpcds_datagen_parquet.sh).
The data generation scripts are [TPC-H dategen script](../../tools/workload/tpch/gen_data/parquet_dataset/tpch-dategen-parquet.sh) and
[TPC-DS dategen script](../../tools/workload/tpcds/gen_data/parquet_dataset/tpcds-dategen-parquet.sh).

The used TPC-H and TPC-DS queries are the original ones, and can be accessed from [TPC-DS queries](../../tools/gluten-it/common/src/main/resources/tpcds-queries)
and [TPC-H queries](../../tools/gluten-it/common/src/main/resources/tpch-queries).
Expand Down
2 changes: 1 addition & 1 deletion docs/get-started/VeloxGPU.md
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@ parent: Getting-Started
- **NVIDIA Drivers**: Compatible with CUDA 12.8.
- **Container Toolkit**: Install `nvidia-container-toolkit` ([guide](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html)).
- **System Reboot**: Required after driver installation.
- **Environment Setup**: Use [`start_cudf.sh`](https://github.com/apache/incubator-gluten/tree/main/dev/start_cudf.sh) for host configuration .
- **Environment Setup**: Use [`start-cudf.sh`](https://github.com/apache/incubator-gluten/tree/main/dev/start-cudf.sh) for host configuration .

---

Expand Down
2 changes: 1 addition & 1 deletion docs/get-started/build-guide.md
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ Please set them via `--`, e.g. `--build_type=Release`.
| build_arrow | Build arrow java/cpp and install the libs in local. Can turn it OFF after first build. | ON |
| spark_version | Build for specified version of Spark(3.2, 3.3, 3.4, 3.5, ALL). `ALL` means build for all versions. | ALL |

### Velox build parameters for build_velox.sh
### Velox build parameters for build-velox.sh
Please set them via `--`, e.g., `--velox_home=/YOUR/PATH`.

| Parameters | Description | Default |
Expand Down
8 changes: 4 additions & 4 deletions docs/velox-backend-support-progress.md
Original file line number Diff line number Diff line change
Expand Up @@ -96,16 +96,16 @@ In Gluten, function support is automatically generated by a script and maintaine
When running the script, the `--spark_home` arg should be set to either:
* The directory containing the Spark source code for the latest supported Spark version in Gluten, and the Spark
project must be built from source.
* Or use the `install_spark_resources.sh` script to get a directory with the necessary resource files:
* Or use the `install-spark-resources.sh` script to get a directory with the necessary resource files:
```
# Define a directory to use for the Spark files and the latest Spark version
export spark_dir=/tmp/spark
export spark_version=3.5

# Run the install_spark_resources.sh script
.github/workflows/util/install_spark_resources.sh ${spark_version} ${spark_dir}
# Run the install-spark-resources.sh script
.github/workflows/util/install-spark-resources.sh ${spark_version} ${spark_dir}
```
After running the `install_spark_resources.sh`, the `--spark_home` for the document generation script will be
After running the `install-spark-resources.sh`, the `--spark_home` for the document generation script will be
something like: `--spark_home=${spark_dir}/shims/spark35/spark_home"`

Use the following command to generate and update the support status:
Expand Down
2 changes: 1 addition & 1 deletion ep/build-clickhouse/src/package.sh
Original file line number Diff line number Diff line change
Expand Up @@ -108,7 +108,7 @@ do
done

# build libch.so
bash "${GLUTEN_SOURCE}"/ep/build-clickhouse/src/build_clickhouse.sh
bash "${GLUTEN_SOURCE}"/ep/build-clickhouse/src/build-clickhouse.sh
cp "$GLUTEN_SOURCE"/cpp-ch/build/utils/extern-local-engine/libch.so "${PACKAGE_DIR_PATH}"/libs/libch.so

# copy bin and conf
Expand Down
2 changes: 1 addition & 1 deletion ep/build-clickhouse/src/resources/bin/gluten.sh
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
set -e

export GLUTEN_HOME=$(cd -P -- "$(dirname -- "$0")/.." && pwd -P)
source ${GLUTEN_HOME}/bin/check_env.sh || exit 1
source ${GLUTEN_HOME}/bin/check-env.sh || exit 1

[[ ! -d "${GLUTEN_HOME}"/logs ]] && mkdir -p "${GLUTEN_HOME}"/logs

Expand Down
File renamed without changes.
Original file line number Diff line number Diff line change
Expand Up @@ -170,7 +170,7 @@ function setup_linux {
elif [[ "$LINUX_DISTRIBUTION" == "openEuler" ]]; then
# this is workaround for gcc-12.3.1
# https://github.com/facebookincubator/velox/blob/b263d9dd8b8910dc642d8fdb0c0adee4b2a1fb29/CMakeLists.txt#L433
sed -i "s|no-unknown-warning-option|no-unknown-warning-option -Wno-restrict|g" ../../src/build_velox.sh
sed -i "s|no-unknown-warning-option|no-unknown-warning-option -Wno-restrict|g" ../../src/build-velox.sh
case "$LINUX_VERSION_ID" in
24.03) ;;
*)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -60,7 +60,7 @@ class GlutenCoreConfig(conf: SQLConf) extends Logging {

/*
* Note: Gluten configuration.md is automatically generated from this code.
* Make sure to run dev/gen_all_config_docs.sh after making changes to this file.
* Make sure to run dev/gen-all-config-docs.sh after making changes to this file.
*/
object GlutenCoreConfig extends ConfigRegistry {
override def get: GlutenCoreConfig = {
Expand Down
Loading
Loading