-
Notifications
You must be signed in to change notification settings - Fork 3.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GH-45594: [C++][Parquet] POC: Optimize Parquet DecodeArrow in DeltaLengthByteArray #45622
base: main
Are you sure you want to change the base?
GH-45594: [C++][Parquet] POC: Optimize Parquet DecodeArrow in DeltaLengthByteArray #45622
Conversation
|
On my MacOS ( this is not the newest ) After:
Before:
|
153b065
to
aa25e2e
Compare
aa25e2e
to
4a4847f
Compare
cc @pitrou this interface is a bit ugly but I don't know whether we have better way for this. Would you mind take a look? |
Hmm, I really don't like the new Perhaps we should instead add these APIs and let Parquet use those builders? diff --git a/cpp/src/arrow/array/builder_binary.h b/cpp/src/arrow/array/builder_binary.h
index 442e4a2632..e568279508 100644
--- a/cpp/src/arrow/array/builder_binary.h
+++ b/cpp/src/arrow/array/builder_binary.h
@@ -359,6 +359,9 @@ class BaseBinaryBuilder
/// \return data pointer of the value date builder
const offset_type* offsets_data() const { return offsets_builder_.data(); }
+ TypedBufferBuilder<offset_type>* offsets_builder() { return &offsets_builder_; }
+ TypedBufferBuilder<uint8_t>* value_data_builder() { return &value_data_builder_; }
+
/// Temporary access to a value.
///
/// This pointer becomes invalid on the next modifying operation. |
Previously I've using a poc like this, maybe unsafe or other prefix can make this better?
|
I don't think adding "unsafe" would really bring anything (there is no risk of crashing, for instance). However, we should add a docstring explaining the caveats when using these methods. |
@pitrou I've address all comments here, would you mind take a look when you've spare time? |
cpp/src/parquet/decoder.cc
Outdated
RETURN_NOT_OK(offsets_builder->Reserve(num_values)); | ||
accum_length = 0; | ||
if (valid_bits == nullptr) { | ||
for (int i = 0; i < max_values; ++i) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
IIUC, VisitNullBitmapInline
can also handle the case where valid_bits == nullptr
since it uses OptionalBitBlockCounter
under the hood.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Oh you're right, would update this
cpp/src/parquet/decoder.cc
Outdated
} | ||
accum_length += length_ptr[i]; | ||
} | ||
if (ARROW_PREDICT_FALSE(accum_length > std::numeric_limits<int32_t>::max())) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Since CanFit
returned true, can this actually happen?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
No, just I just DCHECK this?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Well, for large_binary it might even exceed int32, no?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, but currently, the builder is just BinaryBuilder
, and large binary might uses multiple chunks, so it might hit the case !CanFit
. But once fit this would never accum_length > std::numeric_limits<int32_t>::max()
cpp/src/parquet/decoder.cc
Outdated
int num_values, int null_count, const uint8_t* valid_bits, | ||
int64_t valid_bits_offset, typename EncodingTraits<ByteArrayType>::Accumulator* out, | ||
int* out_num_values) { | ||
int max_values = num_values - null_count; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can you call this num_non_null_values
?
@pitrou I've tried to fix the comments, would you mind take a look again? |
if (ARROW_PREDICT_FALSE(decoder_->bytes_left() < accum_length)) { | ||
return Status::Invalid("Binary data is too short"); | ||
} | ||
RETURN_NOT_OK(out->builder->ValidateOverflow(accum_length)); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
But what happens if accum_length > chunk_space_remaining_
. Would it just fail? It should probably append a new chunk.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It would never happens, chunk_space_remaining_ >= decoder_->bytes_left() >= accum_length
, so this would not happens
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
But why is chunk_space_remaining_ >= decoder_->bytes_left()
?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ah, this is because we know that CanFit(decoder_->bytes_left())
, right? Adding a comment might help...
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Added here: c21878c
return DecodeArrowDenseFastPath(num_values, null_count, valid_bits, | ||
valid_bits_offset, out, out_num_values); | ||
} | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I wonder why we need the slow path below at all. Instead, we could just expose chunk_space_remaining_
and PushChunk
, and use them to build the output chunk by chunk as necessary.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Emmm I don't know what does this mean, when Page's buffer is larger than a chunk can handle, we need find "maximum buffer we can pushinto", and segment the arrow array to multiple sub-chunks, then switch to next batch in this case?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, basically something like this:
int64_t start = 0;
while (start < num_non_null_values) {
// Find the number of values that we can write in this chunk
int64_t chunk_data_length = 0;
for (int64_t end = start; end < num_non_null_values; ++end) {
const int64_t length = length_ptr[end];
if (ARROW_PREDICT_FALSE(length < 0)) {
return Status::Invalid("negative string delta length");
}
if (chunk_data_length + length > out->chunk_space_remaining()) {
break;
}
chunk_data_length += length;
}
// Write chunk [start, end)
...
out->PushChunk();
start = end;
}
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
So this should also find out that, given a num_non_null_values
, what's the number of num_values
in this range?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ahem... that's a good point. Ideally we would need something like VisitNullBitmapInline
, but that would stop after a certain number of non-null values...
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, seems this requires a underlying "loop" abstraction which could hint "next-value", "end", "break"...This also requires handling the all-nulls values. So I prefer using this fast-path firstly.
This abstraction might be like:
struct Stride {
int32_t num_non_null_values;
int32_t num_values;
int32_t binary_length;
};
Stride nextStride(int32_t max_length) {
if (decoder_->bytes_left() <= out->chunk_space_remaining()) {
// accumulate value lengths until max_length
// ...
return Stride {...};
}
VisitNullBitmapInline(...); // find out the fitable stride
return Stride{..};
}
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
So should I move forward with some slow stride checking or leave it a fastpath here? @pitrou
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ah! Well, I was underestimating the complexity of the clean solution, so leaving it as a fast path sounds fine.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The remaining problem, though, is that the fallback isn't tested?
Rationale for this change
See #45594
What changes are included in this PR?
Are these changes tested?
Covered by existing
Are there any user-facing changes?
no