Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

chore: expand the arrow version range to allow arrow v54 #616

Open
wants to merge 2 commits into
base: main
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
26 changes: 15 additions & 11 deletions Cargo.toml
Original file line number Diff line number Diff line change
Expand Up @@ -22,17 +22,21 @@ rust-version = "1.80"
version = "0.6.0"

[workspace.dependencies]
arrow = { version = ">=53, <54" }
arrow-arith = { version = ">=53, <54" }
arrow-array = { version = ">=53, <54" }
arrow-buffer = { version = ">=53, <54" }
arrow-cast = { version = ">=53, <54" }
arrow-data = { version = ">=53, <54" }
arrow-ord = { version = ">=53, <54" }
arrow-json = { version = ">=53, <54" }
arrow-select = { version = ">=53, <54" }
arrow-schema = { version = ">=53, <54" }
parquet = { version = ">=53, <54", features = ["object_store"] }
# When changing the arrow version range, also modify ffi/Cargo.toml which has
# its own arrow version ranges witeh modified features. Failure to do so will
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
# its own arrow version ranges witeh modified features. Failure to do so will
# its own arrow version ranges with modified features. Failure to do so will

# result in compilation errors as two different sets of arrow dependencies may
# be sourced
arrow = { version = ">=53, <55" }
arrow-arith = { version = ">=53, <55" }
arrow-array = { version = ">=53, <55" }
arrow-buffer = { version = ">=53, <55" }
arrow-cast = { version = ">=53, <55" }
arrow-data = { version = ">=53, <55" }
arrow-ord = { version = ">=53, <55" }
arrow-json = { version = ">=53, <55" }
arrow-select = { version = ">=53, <55" }
arrow-schema = { version = ">=53, <55" }
parquet = { version = ">=53, <55", features = ["object_store"] }
object_store = { version = ">=0.11, <0.12" }
hdfs-native-object-store = "0.12.0"
hdfs-native = "0.10.0"
Expand Down
6 changes: 1 addition & 5 deletions acceptance/src/data.rs
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,7 @@ pub fn sort_record_batch(batch: RecordBatch) -> DeltaResult<RecordBatch> {
Ok(RecordBatch::try_new(batch.schema(), columns)?)
}

// Ensure that two schema have the same field names, and dict_id/ordering.
// Ensure that two schema have the same field names, and dict_is_ordered
// We ignore:
// - data type: This is checked already in `assert_columns_match`
// - nullability: parquet marks many things as nullable that we don't in our schema
Expand All @@ -72,10 +72,6 @@ fn assert_schema_fields_match(schema: &Schema, golden: &Schema) {
schema_field.name() == golden_field.name(),
"Field names don't match"
);
assert!(
schema_field.dict_id() == golden_field.dict_id(),
"Field dict_id doesn't match"
);
assert!(
schema_field.dict_is_ordered() == golden_field.dict_is_ordered(),
"Field dict_is_ordered doesn't match"
Expand Down
6 changes: 3 additions & 3 deletions ffi/Cargo.toml
Original file line number Diff line number Diff line change
Expand Up @@ -24,13 +24,13 @@ delta_kernel = { path = "../kernel", default-features = false, features = [
delta_kernel_ffi_macros = { path = "../ffi-proc-macros", version = "0.6.0" }

# used if we use the default engine to be able to move arrow data into the c-ffi format
arrow-schema = { version = "53.0", default-features = false, features = [
arrow-schema = { version = ">=53, <55", default-features = false, features = [
"ffi",
], optional = true }
arrow-data = { version = "53.0", default-features = false, features = [
arrow-data = { version = ">=53, <55", default-features = false, features = [
"ffi",
], optional = true }
arrow-array = { version = "53.0", default-features = false, optional = true }
arrow-array = { version = ">=53, <55", default-features = false, optional = true }

[build-dependencies]
cbindgen = "0.27.0"
Expand Down
6 changes: 1 addition & 5 deletions kernel/tests/golden_tables.rs
Original file line number Diff line number Diff line change
Expand Up @@ -89,7 +89,7 @@ fn sort_record_batch(batch: RecordBatch) -> DeltaResult<RecordBatch> {
Ok(RecordBatch::try_new(batch.schema(), columns)?)
}

// Ensure that two sets of fields have the same names, and dict_id/ordering.
// Ensure that two sets of fields have the same names, and dict_is_ordered
// We ignore:
// - data type: This is checked already in `assert_columns_match`
// - nullability: parquet marks many things as nullable that we don't in our schema
Expand All @@ -103,10 +103,6 @@ fn assert_fields_match<'a>(
actual_field.name() == expected_field.name(),
"Field names don't match"
);
assert!(
actual_field.dict_id() == expected_field.dict_id(),
"Field dict_id doesn't match"
);
Comment on lines -106 to -109
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

unrelated change?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

No, arrow 54 marks dict_id as deprecated and the function will be removed in 55

assert!(
actual_field.dict_is_ordered() == expected_field.dict_is_ordered(),
"Field dict_is_ordered doesn't match"
Expand Down
Loading