Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GH-38925: [CI] Fix spelling #38926

Merged
merged 11 commits into from
Dec 10, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions ci/conan/all/conandata.yml
Original file line number Diff line number Diff line change
Expand Up @@ -45,21 +45,21 @@ sources:
patches:
"8.0.1":
- patch_file: "patches/8.0.0-0005-install-utils.patch"
patch_description: "enable utilis installation"
patch_description: "enable utils installation"
patch_type: "conan"
- patch_file: "patches/8.0.0-0006-fix-cmake.patch"
patch_description: "use cci package"
patch_type: "conan"
"8.0.0":
- patch_file: "patches/8.0.0-0005-install-utils.patch"
patch_description: "enable utilis installation"
patch_description: "enable utils installation"
patch_type: "conan"
- patch_file: "patches/8.0.0-0006-fix-cmake.patch"
patch_description: "use cci package"
patch_type: "conan"
"7.0.0":
- patch_file: "patches/7.0.0-0006-install-utils.patch"
patch_description: "enable utilis installation"
patch_description: "enable utils installation"
patch_type: "conan"
- patch_file: "patches/7.0.0-0007-fix-cmake.patch"
patch_description: "use cci package"
Expand Down
2 changes: 1 addition & 1 deletion ci/conan/all/conanfile.py
Original file line number Diff line number Diff line change
Expand Up @@ -328,7 +328,7 @@ def _with_boost(self, required=False):
return bool(self.options.with_boost)

def _with_thrift(self, required=False):
# No self.options.with_thift exists
# No self.options.with_thrift exists
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I've added the repository to my to-do list, but I'm already a number of levels upstream. I can try to do it once I clear out the rest of these splits...

(I did Google FlatBuffers)

return bool(required or self._parquet())

def _with_utf8proc(self, required=False):
Expand Down
2 changes: 1 addition & 1 deletion ci/conan/all/test_package/test_package.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ arrow::Status VectorToColumnarTable(const std::vector<struct data_row>& rows,
// Indicate the start of a new list row. This will memorise the current
// offset in the values builder.
ARROW_RETURN_NOT_OK(components_builder.Append());
// Store the actual values. The final nullptr argument tells the underyling
// Store the actual values. The final nullptr argument tells the underlying
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

// builder that all added values are valid, i.e. non-null.
ARROW_RETURN_NOT_OK(cost_components_builder.AppendValues(row.cost_components.data(),
row.cost_components.size()));
Expand Down
2 changes: 1 addition & 1 deletion ci/docker/conda-cpp.dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ FROM ${repo}:${arch}-conda
COPY ci/scripts/install_minio.sh /arrow/ci/scripts
RUN /arrow/ci/scripts/install_minio.sh latest /opt/conda

# Unless overriden use Python 3.10
# Unless overridden use Python 3.10
# Google GCS fails building with Python 3.11 at the moment.
ARG python=3.10

Expand Down
2 changes: 1 addition & 1 deletion ci/docker/java-jni-manylinux-201x.dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@
ARG base
FROM ${base}

# Install the libaries required by the Gandiva to run
# Install the libraries required by the Gandiva to run
# Use enable llvm[enable-rtti] in the vcpkg.json to avoid link problems in Gandiva
RUN vcpkg install \
--clean-after-build \
Expand Down
2 changes: 1 addition & 1 deletion ci/docker/python-wheel-manylinux.dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,7 @@ ENV CMAKE_BUILD_TYPE=${build_type} \
VCPKG_FEATURE_FLAGS="manifests"
COPY ci/vcpkg/vcpkg.json arrow/ci/vcpkg/
# cannot use the S3 feature here because while aws-sdk-cpp=1.9.160 contains
# ssl related fixies as well as we can patch the vcpkg portfile to support
# ssl related fixes as well as we can patch the vcpkg portfile to support
# arm machines it hits ARROW-15141 where we would need to fall back to 1.8.186
# but we cannot patch those portfiles since vcpkg-tool handles the checkout of
# previous versions => use bundled S3 build
Expand Down
4 changes: 2 additions & 2 deletions ci/docker/python-wheel-windows-vs2017.dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,7 @@ ENV CMAKE_BUILD_TYPE=${build_type} \
VCPKG_FEATURE_FLAGS="manifests"
COPY ci/vcpkg/vcpkg.json arrow/ci/vcpkg/
# cannot use the S3 feature here because while aws-sdk-cpp=1.9.160 contains
# ssl related fixies as well as we can patch the vcpkg portfile to support
# ssl related fixes as well as we can patch the vcpkg portfile to support
# arm machines it hits ARROW-15141 where we would need to fall back to 1.8.186
# but we cannot patch those portfiles since vcpkg-tool handles the checkout of
# previous versions => use bundled S3 build
Expand Down Expand Up @@ -97,4 +97,4 @@ RUN python -m pip install -r arrow/python/requirements-wheel-build.txt

# For debugging purposes
# RUN wget --no-check-certificate https://github.com/lucasg/Dependencies/releases/download/v1.10/Dependencies_x64_Release.zip
# RUN unzip Dependencies_x64_Release.zip -d Dependencies && setx path "%path%;C:\Depencencies"
# RUN unzip Dependencies_x64_Release.zip -d Dependencies && setx path "%path%;C:\Dependencies"
4 changes: 2 additions & 2 deletions ci/scripts/install_python.sh
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ set -eu

declare -A platforms
platforms=([windows]=Windows
[macos]=MacOSX
[macos]=macOS
[linux]=Linux)

declare -A versions
Expand All @@ -43,7 +43,7 @@ platform=${platforms[$1]}
version=$2
full_version=${versions[$2]}

if [ $platform = "MacOSX" ]; then
if [ $platform = "macOS" ]; then
echo "Downloading Python installer..."

if [ "$(uname -m)" = "arm64" ] || [ "$version" = "3.10" ] || [ "$version" = "3.11" ] || [ "$version" = "3.12" ]; then
Expand Down
2 changes: 1 addition & 1 deletion ci/scripts/install_sccache.sh
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ if [ ! -d $PREFIX ]; then
mkdir -p $PREFIX
fi

# Extract only the sccache binary into $PREFIX and ignore README and LCIENSE.
# Extract only the sccache binary into $PREFIX and ignore README and LICENSE.
# --wildcards doesn't work on busybox.
tar -xzvf $SCCACHE_ARCHIVE --strip-component=1 --directory $PREFIX --exclude="sccache*/*E*E*"
chmod u+x $PREFIX/sccache
Expand Down
2 changes: 1 addition & 1 deletion ci/scripts/integration_spark.sh
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ test_pyarrow_only=${3:-false}
# Spark branch to checkout
spark_version=${SPARK_VERSION:-master}

# Use old behavior that always dropped tiemzones.
# Use old behavior that always dropped timezones.
export PYARROW_IGNORE_TIMEZONE=1

if [ "${SPARK_VERSION:1:2}" == "2." ]; then
Expand Down
2 changes: 1 addition & 1 deletion ci/scripts/r_valgrind.sh
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,7 @@ ${R_BIN} CMD INSTALL ${INSTALL_ARGS} arrow*.tar.gz
pushd tests

# to generate suppression files run:
# ${R_BIN} --vanilla -d "valgrind --tool=memcheck --leak-check=full --track-origins=yes --gen-suppressions=all --log-file=memcheck.log" -f testtthat.supp
# ${R_BIN} --vanilla -d "valgrind --tool=memcheck --leak-check=full --track-origins=yes --gen-suppressions=all --log-file=memcheck.log" -f testthat.supp
${R_BIN} --vanilla -d "valgrind --tool=memcheck --leak-check=full --track-origins=yes --suppressions=/${1}/ci/etc/valgrind-cran.supp" -f testthat.R |& tee testthat.out

# valgrind --error-exitcode=1 should return an erroring exit code that we can catch,
Expand Down