-
Notifications
You must be signed in to change notification settings - Fork 194
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Count rows as a metadata only operation #1388
base: main
Are you sure you want to change the base?
Changes from all commits
731542e
c6c971e
da18837
3104a2f
c2740ea
90bca84
f7202b9
c7205b3
09f9c10
1e9da22
3ab20d4
091c0af
3cd797d
6b0924e
96cb4e9
212c83b
8bc65fa
8bb039f
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,247 @@ | ||
# Licensed to the Apache Software Foundation (ASF) under one | ||
# or more contributor license agreements. See the NOTICE file | ||
# distributed with this work for additional information | ||
# regarding copyright ownership. The ASF licenses this file | ||
# to you under the Apache License, Version 2.0 (the | ||
# "License"); you may not use this file except in compliance | ||
# with the License. You may obtain a copy of the License at | ||
# | ||
# http://www.apache.org/licenses/LICENSE-2.0 | ||
# | ||
# Unless required by applicable law or agreed to in writing, | ||
# software distributed under the License is distributed on an | ||
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY | ||
# KIND, either express or implied. See the License for the | ||
# specific language governing permissions and limitations | ||
# under the License. | ||
from abc import ABC | ||
from typing import Any, Set | ||
|
||
from pyiceberg.expressions import And, Or | ||
from pyiceberg.expressions.literals import Literal | ||
from pyiceberg.expressions.visitors import ( | ||
AlwaysFalse, | ||
AlwaysTrue, | ||
BooleanExpression, | ||
BoundBooleanExpressionVisitor, | ||
BoundPredicate, | ||
BoundTerm, | ||
Not, | ||
UnboundPredicate, | ||
visit, | ||
) | ||
from pyiceberg.partitioning import PartitionSpec | ||
from pyiceberg.schema import Schema | ||
from pyiceberg.typedef import Record | ||
from pyiceberg.types import L | ||
|
||
|
||
class ResidualVisitor(BoundBooleanExpressionVisitor[BooleanExpression], ABC): | ||
schema: Schema | ||
spec: PartitionSpec | ||
case_sensitive: bool | ||
|
||
def __init__(self, schema: Schema, spec: PartitionSpec, case_sensitive: bool, expr: BooleanExpression): | ||
self.schema = schema | ||
self.spec = spec | ||
self.case_sensitive = case_sensitive | ||
self.expr = expr | ||
|
||
def eval(self, partition_data: Record) -> BooleanExpression: | ||
self.struct = partition_data | ||
return visit(self.expr, visitor=self) | ||
|
||
def visit_true(self) -> BooleanExpression: | ||
return AlwaysTrue() | ||
|
||
def visit_false(self) -> BooleanExpression: | ||
return AlwaysFalse() | ||
|
||
def visit_not(self, child_result: BooleanExpression) -> BooleanExpression: | ||
return Not(child_result) | ||
|
||
def visit_and(self, left_result: BooleanExpression, right_result: BooleanExpression) -> BooleanExpression: | ||
return And(left_result, right_result) | ||
|
||
def visit_or(self, left_result: BooleanExpression, right_result: BooleanExpression) -> BooleanExpression: | ||
return Or(left_result, right_result) | ||
|
||
def visit_is_null(self, term: BoundTerm[L]) -> BooleanExpression: | ||
if term.eval(self.struct) is None: | ||
return AlwaysTrue() | ||
else: | ||
return AlwaysFalse() | ||
|
||
def visit_not_null(self, term: BoundTerm[L]) -> BooleanExpression: | ||
if term.eval(self.struct) is not None: | ||
return AlwaysTrue() | ||
else: | ||
return AlwaysFalse() | ||
|
||
def visit_is_nan(self, term: BoundTerm[L]) -> BooleanExpression: | ||
val = term.eval(self.struct) | ||
if val is None: | ||
return self.visit_true() | ||
else: | ||
return self.visit_false() | ||
|
||
def visit_not_nan(self, term: BoundTerm[L]) -> BooleanExpression: | ||
val = term.eval(self.struct) | ||
if val is not None: | ||
return self.visit_true() | ||
else: | ||
return self.visit_false() | ||
|
||
def visit_less_than(self, term: BoundTerm[L], literal: Literal[L]) -> BooleanExpression: | ||
if term.eval(self.struct) < literal.value: | ||
return self.visit_true() | ||
else: | ||
return self.visit_false() | ||
|
||
def visit_less_than_or_equal(self, term: BoundTerm[L], literal: Literal[L]) -> BooleanExpression: | ||
if term.eval(self.struct) <= literal.value: | ||
return self.visit_true() | ||
else: | ||
return self.visit_false() | ||
|
||
def visit_greater_than(self, term: BoundTerm[L], literal: Literal[L]) -> BooleanExpression: | ||
if term.eval(self.struct) > literal.value: | ||
return self.visit_true() | ||
else: | ||
return self.visit_false() | ||
|
||
def visit_greater_than_or_equal(self, term: BoundTerm[L], literal: Literal[L]) -> BooleanExpression: | ||
if term.eval(self.struct) >= literal.value: | ||
return self.visit_true() | ||
else: | ||
return self.visit_false() | ||
|
||
def visit_equal(self, term: BoundTerm[L], literal: Literal[L]) -> BooleanExpression: | ||
if term.eval(self.struct) == literal.value: | ||
return self.visit_true() | ||
else: | ||
return self.visit_false() | ||
|
||
def visit_not_equal(self, term: BoundTerm[L], literal: Literal[L]) -> BooleanExpression: | ||
if term.eval(self.struct) != literal.value: | ||
return self.visit_true() | ||
else: | ||
return self.visit_false() | ||
|
||
def visit_in(self, term: BoundTerm[L], literals: Set[L]) -> BooleanExpression: | ||
if term.eval(self.struct) in literals: | ||
return self.visit_true() | ||
else: | ||
return self.visit_false() | ||
|
||
def visit_not_in(self, term: BoundTerm[L], literals: Set[L]) -> BooleanExpression: | ||
if term.eval(self.struct) not in literals: | ||
return self.visit_true() | ||
else: | ||
return self.visit_false() | ||
|
||
def visit_starts_with(self, term: BoundTerm[L], literal: Literal[L]) -> BooleanExpression: | ||
eval_res = term.eval(self.struct) | ||
if eval_res is not None and str(eval_res).startswith(str(literal.value)): | ||
return AlwaysTrue() | ||
else: | ||
return AlwaysFalse() | ||
|
||
def visit_not_starts_with(self, term: BoundTerm[L], literal: Literal[L]) -> BooleanExpression: | ||
if not self.visit_starts_with(term, literal): | ||
return AlwaysTrue() | ||
else: | ||
return AlwaysFalse() | ||
|
||
def visit_bound_predicate(self, predicate: BoundPredicate[Any]) -> BooleanExpression: | ||
""" | ||
If there is no strict projection or if it evaluates to false, then return the predicate. | ||
|
||
Get the strict projection and inclusive projection of this predicate in partition data, | ||
then use them to determine whether to return the original predicate. The strict projection | ||
returns true iff the original predicate would have returned true, so the predicate can be | ||
eliminated if the strict projection evaluates to true. Similarly the inclusive projection | ||
returns false iff the original predicate would have returned false, so the predicate can | ||
also be eliminated if the inclusive projection evaluates to false. | ||
|
||
""" | ||
parts = self.spec.fields_by_source_id(predicate.term.ref().field.field_id) | ||
if parts == []: | ||
return predicate | ||
|
||
from pyiceberg.types import StructType | ||
|
||
def struct_to_schema(struct: StructType) -> Schema: | ||
return Schema(*[f for f in struct.fields]) | ||
|
||
for part in parts: | ||
strict_projection = part.transform.strict_project(part.name, predicate) | ||
strict_result = None | ||
|
||
if strict_projection is not None: | ||
bound = strict_projection.bind(struct_to_schema(self.spec.partition_type(self.schema))) | ||
if isinstance(bound, BoundPredicate): | ||
strict_result = super().visit_bound_predicate(bound) | ||
else: | ||
strict_result = bound | ||
|
||
if strict_result is not None and isinstance(strict_result, AlwaysTrue): | ||
return AlwaysTrue() | ||
|
||
inclusive_projection = part.transform.project(part.name, predicate) | ||
inclusive_result = None | ||
if inclusive_projection is not None: | ||
bound_inclusive = inclusive_projection.bind(struct_to_schema(self.spec.partition_type(self.schema))) | ||
if isinstance(bound_inclusive, BoundPredicate): | ||
# using predicate method specific to inclusive | ||
inclusive_result = super().visit_bound_predicate(bound_inclusive) | ||
else: | ||
# if the result is not a predicate, then it must be a constant like alwaysTrue or | ||
# alwaysFalse | ||
inclusive_result = bound_inclusive | ||
if inclusive_result is not None and isinstance(inclusive_result, AlwaysFalse): | ||
return AlwaysFalse() | ||
|
||
return predicate | ||
|
||
def visit_unbound_predicate(self, predicate: UnboundPredicate[L]) -> BooleanExpression: | ||
bound = predicate.bind(self.schema, case_sensitive=True) | ||
|
||
if isinstance(bound, BoundPredicate): | ||
bound_residual = self.visit_bound_predicate(predicate=bound) | ||
# if isinstance(bound_residual, BooleanExpression): | ||
if bound_residual not in (AlwaysFalse(), AlwaysTrue()): | ||
# replace inclusive original unbound predicate | ||
return predicate | ||
|
||
# use the non-predicate residual (e.g. alwaysTrue) | ||
return bound_residual | ||
|
||
# if binding didn't result in a Predicate, return the expression | ||
return bound | ||
|
||
|
||
class ResidualEvaluator(ResidualVisitor): | ||
def residual_for(self, partition_data: Record) -> BooleanExpression: | ||
return self.eval(partition_data) | ||
|
||
|
||
class UnpartitionedResidualEvaluator(ResidualEvaluator): | ||
# Finds the residuals for an Expression the partitions in the given PartitionSpec | ||
def __init__(self, schema: Schema, expr: BooleanExpression): | ||
from pyiceberg.partitioning import UNPARTITIONED_PARTITION_SPEC | ||
|
||
super().__init__(schema=schema, spec=UNPARTITIONED_PARTITION_SPEC, expr=expr, case_sensitive=False) | ||
self.expr = expr | ||
|
||
def residual_for(self, partition_data: Record) -> BooleanExpression: | ||
return self.expr | ||
|
||
|
||
def residual_evaluator_of( | ||
spec: PartitionSpec, expr: BooleanExpression, case_sensitive: bool, schema: Schema | ||
) -> ResidualEvaluator: | ||
if len(spec.fields) != 0: | ||
return ResidualEvaluator(spec=spec, expr=expr, schema=schema, case_sensitive=case_sensitive) | ||
else: | ||
return UnpartitionedResidualEvaluator(schema=schema, expr=expr) |
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -1328,6 +1328,9 @@ def filter(self: S, expr: Union[str, BooleanExpression]) -> S: | |
def with_case_sensitive(self: S, case_sensitive: bool = True) -> S: | ||
return self.update(case_sensitive=case_sensitive) | ||
|
||
@abstractmethod | ||
def count(self) -> int: ... | ||
|
||
|
||
class ScanTask(ABC): | ||
pass | ||
|
@@ -1341,19 +1344,21 @@ class FileScanTask(ScanTask): | |
delete_files: Set[DataFile] | ||
start: int | ||
length: int | ||
residual: BooleanExpression | ||
|
||
def __init__( | ||
self, | ||
data_file: DataFile, | ||
delete_files: Optional[Set[DataFile]] = None, | ||
start: Optional[int] = None, | ||
length: Optional[int] = None, | ||
residual: BooleanExpression = None | ||
) -> None: | ||
self.file = data_file | ||
self.delete_files = delete_files or set() | ||
self.start = start or 0 | ||
self.length = length or data_file.file_size_in_bytes | ||
|
||
self.residual = residual | ||
|
||
def _open_manifest( | ||
io: FileIO, | ||
|
@@ -1513,13 +1518,23 @@ def plan_files(self) -> Iterable[FileScanTask]: | |
else: | ||
raise ValueError(f"Unknown DataFileContent ({data_file.content}): {manifest_entry}") | ||
|
||
|
||
|
||
from pyiceberg.expressions.residual_evaluator import residual_evaluator_of | ||
residual_evaluator = residual_evaluator_of( | ||
spec=self.table_metadata.spec(), | ||
expr=self.row_filter, | ||
case_sensitive=self.case_sensitive, | ||
schema=self.table_metadata.schema() | ||
) | ||
return [ | ||
FileScanTask( | ||
data_entry.data_file, | ||
data_file=data_entry.data_file, | ||
delete_files=_match_deletes_to_data_file( | ||
data_entry, | ||
positional_delete_entries, | ||
), | ||
residual=residual_evaluator.residual_for(data_entry.data_file.partition) | ||
) | ||
for data_entry in data_entries | ||
] | ||
|
@@ -1594,6 +1609,29 @@ def to_ray(self) -> ray.data.dataset.Dataset: | |
|
||
return ray.data.from_arrow(self.to_arrow()) | ||
|
||
def count(self) -> int: | ||
""" | ||
Usage: calutates the total number of records in a Scan that haven't had positional deletes | ||
""" | ||
res = 0 | ||
# every task is a FileScanTask | ||
tasks = self.plan_files() | ||
|
||
for task in tasks: | ||
# task.residual is a Boolean Expression if the fiter condition is fully satisfied by the | ||
# partition value and task.delete_files represents that positional delete haven't been merged yet | ||
# hence those files have to read as a pyarrow table applying the filter and deletes | ||
if task.residual == AlwaysTrue() and not len(task.delete_files): | ||
# Every File has a metadata stat that stores the file record count | ||
res += task.file.record_count | ||
else: | ||
from pyiceberg.io.pyarrow import ArrowScan | ||
tbl = ArrowScan( | ||
self.table_metadata, self.io, self.projection(), self.row_filter, self.case_sensitive, self.limit | ||
).to_table([task]) | ||
res += len(tbl) | ||
return res | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I love this approach! My only concern is about loading too much data into memory at once, although this is loading just one file at a time, in the worst case some file could potentially be very large? Shall we define a threshold and check, for example, if
https://github.com/apache/iceberg-python/blob/main/pyiceberg/table/__init__.py#L1541-L1564 |
||
|
||
|
||
@dataclass(frozen=True) | ||
class WriteTask: | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
this count will not be accurate when there are deletes files
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi @kevinjqliu thank you for the review. I am trying to account for positional deletes, do you have a suggestion on how that can be achieved?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, this can be widely off, not just because the merge-on-read deletes, but because
plan_files
returns all the files that (might) contain relevant rows. For example, if it cannot be determined if has relevant data, it will be returned byplan_files
.I think there are two ways forward:
task.file.record_count
. We would need to extend this to also see if there are also merge-on-read deletes as Kevin already mentioned, or just fail when there are positional deletes.residual-predicate
in theFileScanTask
. When we run a query, likeday_added = 2024-12-01 and user_id = 10
, then theday_added = 2024-12-01
might be satisfied with the partitioning already. This is the case when the table is partitioned by day, and we know that all the data in the file evaluatestrue
forday_added = 2024-12-01
, then we need to open the file, and filter foruser_id = 10
. If we would leave out theuser_id = 10
, then it would beALWAYS_TRUE
, and then we know that we can just usetask.file.record_count
. This way we could very easily loop over the.plan_files()
:To get to the second step, we first have to port the
ResidualEvaluator
. The java code can be found here, including some excellent tests.There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi @Fokko I have added Residual Evaluator with Tests.
Now I am trying to create the breaking tests for count where delete has occurred and the counts should differ