Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Smaller LitGPT like configs #398

Closed
wants to merge 4 commits into from

Conversation

carmocca
Copy link
Member

@carmocca carmocca commented May 10, 2024

What does this PR do?

Alternative to #219

Weight memory usage before:

gpt-neox-like 6532864 26.131456 MB
llama1-like 602432 2.409728 MB
long-context-like 4301120 17.20448 MB
llama2-like 602432 2.409728 MB
falcon-7b-like 4359040 17.43616 MB
falcon-40b-like 1589760 6.35904 MB
codellama2-like 819520 3.27808 MB
mixtral-like 3308800 13.2352 MB

After:

gpt-neox-like 141056 0.564224 MB
llama1-like 602432 2.409728 MB
long-context-like 602432 2.409728 MB
llama2-like 602432 2.409728 MB
falcon-7b-like 77700 0.3108 MB
falcon-40b-like 398080 1.59232 MB
codellama2-like 401568 1.606272 MB
mixtral-like 257856 1.031424 MB

@carmocca carmocca self-assigned this May 10, 2024
@carmocca carmocca force-pushed the carmocca/smaller-litgpt-test-configs branch from 25989b6 to 123df4f Compare May 10, 2024 15:38
@carmocca carmocca force-pushed the carmocca/smaller-litgpt-test-configs branch from 123df4f to bba9e86 Compare May 10, 2024 15:39
@t-vi
Copy link
Collaborator

t-vi commented May 11, 2024

Thank you @carmocca ,I don't think we need to go smaller than we currently have. Anything < 500MB GPU memory should be more than fine.

@carmocca
Copy link
Member Author

Okay. We can always revisit this

@carmocca carmocca closed this May 13, 2024
@carmocca carmocca deleted the carmocca/smaller-litgpt-test-configs branch May 13, 2024 10:29
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants