How to get result batch size to match the set batch_size arg in read_arrow_batches #171
Replies: 2 comments 1 reply
-
|
Never mind, silly me mixed up |
Beta Was this translation helpful? Give feedback.
0 replies
-
|
Glad you could figure it out. I hope your experiment will be successful! |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello all,
I'm exploring arrow-odbc to try to replace pyodbc that's currently being used due to its speed. I'm trying to get the returned result to match the size I've set for the argument
batch_sizetoread_arrow_batchesmethod but couldn't really get the row numbers returned in a batch to match the setbatch_size. I can see from the documentation that the max number of rows upper bound can be dictated bymax_bytes_per_batchif it's lower.However I've set the
max_bytes_per_batchto a very high number, eg. 4 times (even tried 100 times just to see if it changed the outcome) the default value and I still cannot get the row numbers returned in a batch to match the setbatch_size. For example I've set thebatch_size=200000but only got 176544 rows returned in each batch. Is there a way to get the returned number of rows to exactly match thebatch_size?Thanks a lot in advance, any help is appreciated 😃
Beta Was this translation helpful? Give feedback.
All reactions