Skip to content

Commit

Permalink
fix(rust): Restore symbols on Apple by bumping nightly version (#20563)
Browse files Browse the repository at this point in the history
  • Loading branch information
orlp authored Jan 6, 2025
1 parent ab92c65 commit e0ef7f9
Show file tree
Hide file tree
Showing 23 changed files with 34 additions and 54 deletions.
4 changes: 2 additions & 2 deletions crates/polars-arrow/src/io/ipc/read/schema.rs
Original file line number Diff line number Diff line change
Expand Up @@ -363,7 +363,7 @@ pub fn deserialize_schema(
message: &[u8],
) -> PolarsResult<(ArrowSchema, IpcSchema, Option<Metadata>)> {
let message = arrow_format::ipc::MessageRef::read_as_root(message)
.map_err(|_err| polars_err!(oos = "Unable deserialize message: {err:?}"))?;
.map_err(|err| polars_err!(oos = format!("Unable deserialize message: {err:?}")))?;

let schema = match message
.header()?
Expand Down Expand Up @@ -430,7 +430,7 @@ pub(super) fn fb_to_schema(

pub(super) fn deserialize_stream_metadata(meta: &[u8]) -> PolarsResult<StreamMetadata> {
let message = arrow_format::ipc::MessageRef::read_as_root(meta)
.map_err(|_err| polars_err!(oos = "Unable to get root as message: {err:?}"))?;
.map_err(|err| polars_err!(oos = format!("Unable to get root as message: {err:?}")))?;
let version = message.version()?;
// message header is a Schema, so read it
let header = message
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -68,7 +68,7 @@ impl RevMapping {
val.hash(&mut hb);
});
let hash = hb.finish();
(hash as u128) << 64 | (categories.total_buffer_len() as u128)
((hash as u128) << 64) | (categories.total_buffer_len() as u128)
}

pub fn build_local(categories: Utf8ViewArray) -> Self {
Expand Down
8 changes: 1 addition & 7 deletions crates/polars-expr/src/expressions/apply.rs
Original file line number Diff line number Diff line change
Expand Up @@ -326,15 +326,10 @@ impl PhysicalExpr for ApplyExpr {
self.inputs
.par_iter()
.map(f)
.map(|v| v.map(Column::from))
.collect::<PolarsResult<Vec<_>>>()
})
} else {
self.inputs
.iter()
.map(f)
.map(|v| v.map(Column::from))
.collect::<PolarsResult<Vec<_>>>()
self.inputs.iter().map(f).collect::<PolarsResult<Vec<_>>>()
}?;

if self.allow_rename {
Expand Down Expand Up @@ -550,7 +545,6 @@ fn apply_multiple_elementwise<'a>(

ac.flat_naive().into_owned()
})
.map(Column::from)
.collect::<Vec<_>>();

let input_len = c[0].len();
Expand Down
4 changes: 2 additions & 2 deletions crates/polars-expr/src/expressions/column.rs
Original file line number Diff line number Diff line change
Expand Up @@ -156,12 +156,12 @@ impl PhysicalExpr for ColumnExpr {
// in debug builds we panic so that it can be fixed when occurring
None => {
if self.name.starts_with(CSE_REPLACED) {
return self.process_cse(df, &self.schema).map(Column::from);
return self.process_cse(df, &self.schema);
}
self.process_by_linear_search(df, state, true)
},
};
self.check_external_context(out, state).map(Column::from)
self.check_external_context(out, state)
}

#[allow(clippy::ptr_arg)]
Expand Down
3 changes: 1 addition & 2 deletions crates/polars-expr/src/expressions/window.rs
Original file line number Diff line number Diff line change
Expand Up @@ -406,7 +406,7 @@ impl PhysicalExpr for WindowExpr {
let group_by_columns = self
.group_by
.iter()
.map(|e| e.evaluate(df, state).map(Column::from))
.map(|e| e.evaluate(df, state))
.collect::<PolarsResult<Vec<_>>>()?;

// if the keys are sorted
Expand Down Expand Up @@ -551,7 +551,6 @@ impl PhysicalExpr for WindowExpr {
state,
&cache_key,
)
.map(Column::from)
},
Join => {
let out_column = ac.aggregated();
Expand Down
5 changes: 1 addition & 4 deletions crates/polars-io/src/partition.rs
Original file line number Diff line number Diff line change
Expand Up @@ -39,10 +39,7 @@ fn write_partitioned_dataset_impl<W>(
where
W: WriteDataFrameToFile + Send + Sync,
{
let partition_by = partition_by
.into_iter()
.map(Into::into)
.collect::<Vec<PlSmallStr>>();
let partition_by = partition_by.into_iter().collect::<Vec<PlSmallStr>>();
// Ensure we have a single chunk as the gather will otherwise rechunk per group.
df.as_single_chunk_par();

Expand Down
4 changes: 2 additions & 2 deletions crates/polars-mem-engine/src/executors/group_by.rs
Original file line number Diff line number Diff line change
Expand Up @@ -89,7 +89,7 @@ pub(super) fn group_by_helper(
rayon::join(get_columns, get_agg)
});

columns.extend(agg_columns?.into_iter().map(Column::from));
columns.extend(agg_columns?);
DataFrame::new(columns)
}

Expand All @@ -98,7 +98,7 @@ impl GroupByExec {
let keys = self
.keys
.iter()
.map(|e| e.evaluate(&df, state).map(Column::from))
.map(|e| e.evaluate(&df, state))
.collect::<PolarsResult<_>>()?;
group_by_helper(
df,
Expand Down
4 changes: 2 additions & 2 deletions crates/polars-mem-engine/src/executors/group_by_dynamic.rs
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ impl GroupByDynamicExec {
let keys = self
.keys
.iter()
.map(|e| e.evaluate(&df, state).map(Column::from))
.map(|e| e.evaluate(&df, state))
.collect::<PolarsResult<Vec<_>>>()?;

let (mut time_key, mut keys, groups) = df.group_by_dynamic(keys, &self.options)?;
Expand Down Expand Up @@ -63,7 +63,7 @@ impl GroupByDynamicExec {
let mut columns = Vec::with_capacity(agg_columns.len() + 1 + keys.len());
columns.extend_from_slice(&keys);
columns.push(time_key);
columns.extend(agg_columns.into_iter().map(Column::from));
columns.extend(agg_columns);

DataFrame::new(columns)
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -58,9 +58,7 @@ fn compute_keys(
df: &DataFrame,
state: &ExecutionState,
) -> PolarsResult<Vec<Column>> {
keys.iter()
.map(|s| s.evaluate(df, state).map(Column::from))
.collect()
keys.iter().map(|s| s.evaluate(df, state)).collect()
}

fn run_partitions(
Expand Down Expand Up @@ -154,7 +152,6 @@ fn estimate_unique_count(keys: &[Column], mut sample_size: usize) -> PolarsResul
let keys = keys
.iter()
.map(|s| s.slice(offset, sample_size))
.map(Column::from)
.collect::<Vec<_>>();
let df = unsafe { DataFrame::new_no_checks_height_from_first(keys) };
let names = df.get_column_names().into_iter().cloned();
Expand Down Expand Up @@ -331,9 +328,7 @@ impl PartitionGroupByExec {
.zip(&df.get_columns()[self.phys_keys.len()..])
.map(|(expr, partitioned_s)| {
let agg_expr = expr.as_partitioned_aggregator().unwrap();
agg_expr
.finalize(partitioned_s.clone(), groups, state)
.map(Column::from)
agg_expr.finalize(partitioned_s.clone(), groups, state)
})
.collect();

Expand Down
4 changes: 2 additions & 2 deletions crates/polars-mem-engine/src/executors/group_by_rolling.rs
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ impl GroupByRollingExec {
let keys = self
.keys
.iter()
.map(|e| e.evaluate(&df, state).map(Column::from))
.map(|e| e.evaluate(&df, state))
.collect::<PolarsResult<Vec<_>>>()?;

let (mut time_key, mut keys, groups) = df.rolling(keys, &self.options)?;
Expand Down Expand Up @@ -85,7 +85,7 @@ impl GroupByRollingExec {
let mut columns = Vec::with_capacity(agg_columns.len() + 1 + keys.len());
columns.extend_from_slice(&keys);
columns.push(time_key);
columns.extend(agg_columns.into_iter().map(Column::from));
columns.extend(agg_columns);

DataFrame::new(columns)
}
Expand Down
5 changes: 1 addition & 4 deletions crates/polars-mem-engine/src/executors/projection_utils.rs
Original file line number Diff line number Diff line change
Expand Up @@ -340,10 +340,7 @@ pub(super) fn check_expand_literals(
}

// @scalar-opt
let selected_columns = selected_columns
.into_iter()
.map(Column::from)
.collect::<Vec<_>>();
let selected_columns = selected_columns.into_iter().collect::<Vec<_>>();

let df = unsafe { DataFrame::new_no_checks_height_from_first(selected_columns) };

Expand Down
2 changes: 1 addition & 1 deletion crates/polars-mem-engine/src/executors/stack.rs
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ impl StackExec {
// new, unique column names. It is immediately
// followed by a projection which pulls out the
// possibly mismatching column lengths.
unsafe { df.column_extend_unchecked(res.into_iter().map(Column::from)) };
unsafe { df.column_extend_unchecked(res) };
} else {
let (df_height, df_width) = df.shape();

Expand Down
4 changes: 1 addition & 3 deletions crates/polars-ops/src/series/ops/duration.rs
Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,5 @@ pub fn impl_duration(s: &[Column], time_unit: TimeUnit) -> PolarsResult<Column>
duration = (duration + weeks * multiplier * SECONDS_IN_DAY * 7)?;
}

duration
.cast(&DataType::Duration(time_unit))
.map(Column::from)
duration.cast(&DataType::Duration(time_unit))
}
Original file line number Diff line number Diff line change
Expand Up @@ -244,7 +244,7 @@ mod tests {

encode::<bool, _, _>(&mut vec, iter, 1)?;

assert_eq!(vec, vec![(2 << 1 | 1), 0b10011101u8, 0b00011101]);
assert_eq!(vec, vec![((2 << 1) | 1), 0b10011101u8, 0b00011101]);

Ok(())
}
Expand All @@ -259,7 +259,7 @@ mod tests {
1,
)?;

assert_eq!(vec, vec![(1 << 1 | 1), 0b11111111]);
assert_eq!(vec, vec![((1 << 1) | 1), 0b11111111]);
Ok(())
}

Expand All @@ -272,7 +272,7 @@ mod tests {
assert_eq!(
vec,
vec![
(2 << 1 | 1),
((2 << 1) | 1),
0b01_10_01_00,
0b00_01_01_10,
0b_00_00_00_11,
Expand All @@ -294,7 +294,7 @@ mod tests {
let expected = 0b11_10_01_00u8;

let mut expected = vec![expected; length / 4];
expected.insert(0, ((length / 8) as u8) << 1 | 1);
expected.insert(0, (((length / 8) as u8) << 1) | 1);

assert_eq!(vec, expected);
Ok(())
Expand Down
1 change: 0 additions & 1 deletion crates/polars-plan/src/dsl/function_expr/datetime.rs
Original file line number Diff line number Diff line change
Expand Up @@ -269,7 +269,6 @@ pub(super) fn time(s: &Column) -> PolarsResult<Column> {
DataType::Time => Ok(s.clone()),
dtype => polars_bail!(ComputeError: "expected Datetime or Time, got {}", dtype),
}
.map(Column::from)
}
pub(super) fn date(s: &Column) -> PolarsResult<Column> {
match s.dtype() {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -220,7 +220,7 @@ pub(super) fn datetime_ranges(
};

let to_type = DataType::List(Box::new(dtype));
out.cast(&to_type).map(Column::from)
out.cast(&to_type)
}

impl FieldsMapper<'_> {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -62,5 +62,5 @@ pub(super) fn time_ranges(
let out = temporal_ranges_impl_broadcast(start, end, range_impl, &mut builder)?;

let to_type = DataType::List(Box::new(DataType::Time));
out.cast(&to_type).map(Column::from)
out.cast(&to_type)
}
2 changes: 1 addition & 1 deletion crates/polars-plan/src/dsl/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -1516,7 +1516,7 @@ impl Expr {
.map(|ca| ca.into_column()),
}?;
if let DataType::Float32 = c.dtype() {
out.cast(&DataType::Float32).map(Column::from).map(Some)
out.cast(&DataType::Float32).map(Some)
} else {
Ok(Some(out))
}
Expand Down
1 change: 1 addition & 0 deletions crates/polars-python/src/lib.rs
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,7 @@
#![allow(non_local_definitions)]
#![allow(clippy::too_many_arguments)] // Python functions can have many arguments due to default arguments
#![allow(clippy::disallowed_types)]
#![allow(clippy::useless_conversion)] // Needed for now due to https://github.com/PyO3/pyo3/issues/4828.

#[cfg(feature = "csv")]
pub mod batched_csv;
Expand Down
6 changes: 3 additions & 3 deletions crates/polars-python/src/map/dataframe.rs
Original file line number Diff line number Diff line change
Expand Up @@ -267,9 +267,9 @@ pub fn apply_lambda_with_list_out_type<'a>(
if val.is_none() {
Ok(None)
} else {
Err(PyValueError::new_err(
"should return a Series, got a {val:?}",
))
Err(PyValueError::new_err(format!(
"should return a Series, got a {val:?}"
)))
}
},
}
Expand Down
4 changes: 2 additions & 2 deletions crates/polars-stream/src/physical_plan/lower_expr.rs
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@ use std::sync::atomic::{AtomicU64, Ordering};
use std::sync::Arc;

use polars_core::frame::DataFrame;
use polars_core::prelude::{Column, Field, InitHashMaps, PlHashMap, PlHashSet};
use polars_core::prelude::{Field, InitHashMaps, PlHashMap, PlHashSet};
use polars_core::schema::{Schema, SchemaExt};
use polars_error::PolarsResult;
use polars_expr::planner::get_expr_depth_limit;
Expand Down Expand Up @@ -283,7 +283,7 @@ fn build_fallback_node_with_ctx(
let exec_state = ExecutionState::new();
let columns = phys_exprs
.iter()
.map(|phys_expr| phys_expr.evaluate(&df, &exec_state).map(Column::from))
.map(|phys_expr| phys_expr.evaluate(&df, &exec_state))
.try_collect()?;
DataFrame::new_with_broadcast(columns)
};
Expand Down
2 changes: 1 addition & 1 deletion crates/polars-utils/src/index.rs
Original file line number Diff line number Diff line change
Expand Up @@ -218,7 +218,7 @@ impl<const CHUNK_BITS: u64> ChunkId<CHUNK_BITS> {
#[allow(clippy::unnecessary_cast)]
pub fn store(chunk: IdxSize, row: IdxSize) -> Self {
debug_assert!(chunk < !(u64::MAX << CHUNK_BITS) as IdxSize);
let swizzled = (row as u64) << CHUNK_BITS | chunk as u64;
let swizzled = ((row as u64) << CHUNK_BITS) | chunk as u64;

Self { swizzled }
}
Expand Down
2 changes: 1 addition & 1 deletion rust-toolchain.toml
Original file line number Diff line number Diff line change
@@ -1,2 +1,2 @@
[toolchain]
channel = "nightly-2024-12-19"
channel = "nightly-2025-01-05"

0 comments on commit e0ef7f9

Please sign in to comment.