-
Notifications
You must be signed in to change notification settings - Fork 21
feat: Increase file size limit from 25GB to 50.1GB #396
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
nikita-smetanin
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi Soroush, PR looks nice, I left a few suggestions :)
|
|
||
| # Submit next part if available | ||
| if part_index < len(parts): | ||
| part_info = parts[part_index] |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Would be great to rewrite it to deduplicate this code piece with the one above. I think you can either make a for loop to submit tasks and wait on result if we have enough already, or use executor.map with buffersize to limit concurrent tasks.
|
Thanks for the PR, looks good to me! Please make sure to address this comment: |
nikita-smetanin
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for the changes! Let's ship it
src/together/filemanager.py
Outdated
| self._upload_single_part, part_info, part_data | ||
| ) | ||
| # Submit initial batch limited by max_concurrent_parts | ||
| for i in range(min(self.max_concurrent_parts, len(parts))): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'd update it to while part_index < min(self.max_concurrent_parts, len(parts)): or at least replace i with _ as it's not used
f3ba7c9 to
d104aa1
Compare
Have you read the Contributing Guidelines?
Issue #
Describe your changes
Clearly and concisely describe what's in this pull request. Include screenshots, if necessary.
Note
Increase file size limit to 50.1GB, bump multipart target part size to 250MB, add sliding-window concurrent uploads, add download timeout, and enhance file validation progress; tests updated accordingly.
TARGET_PART_SIZE_MBto250and file limitMAX_FILE_SIZE_GBto50.1._submit_parthelper; maintain max concurrency via executor; progress tracked withtqdm.request_timeout=3600to streamed downloads._check_utf8.tqdmprogress to JSONL validation loop.as_completedmocking for sliding-window logic.Written by Cursor Bugbot for commit d104aa1. This will update automatically on new commits. Configure here.