Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Tracking issue for Mac support #4

Open
pannous opened this issue Mar 5, 2023 · 3 comments
Open

Tracking issue for Mac support #4

pannous opened this issue Mar 5, 2023 · 3 comments

Comments

@pannous
Copy link

pannous commented Mar 5, 2023

M1 / M2 32GB … 128GB any hopes?

@remixer-dec
Copy link

remixer-dec commented Mar 6, 2023

No luck with this repo, "bitsandbytes" dependency is heavily relying on CUDA.
But there is a repo for cpu inference, just change the prompts to prompts[0], so it doesn't crash with max_batch_size=1.
It takes more than 10 minutes to produce output with max_gen_len=20, even GPT-J 5B took me around a minute on CPU.
I also tried to make an MPS port with gpu acceleration, it works faster, but the output is not good enough imo, not sure if it is always good on cpu or if I just got lucky on my first generation. UPDATE: the model gives good outputs with python3.10 + pytorch-nightly

@pannous
Copy link
Author

pannous commented Mar 8, 2023

thanks!

@remixer-dec
Copy link

Actually, I was wrong. After I tried my port with a higher version of python+pytorch, the outputs were as good as the cpu ones, I am happy that it worked after all!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants