Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support llama3.1:70b on compatible systems #16

Open
divbell opened this issue Sep 9, 2024 · 1 comment
Open

Support llama3.1:70b on compatible systems #16

divbell opened this issue Sep 9, 2024 · 1 comment

Comments

@divbell
Copy link

divbell commented Sep 9, 2024

I really like this tool and was curious if you could add llama3.1:70b where the system has >40GB of VRAM.

I'm running dual 3090s and can execute it, though it's slow compared to the smaller models, and I would like to benchmark it in a standardized way. I did see one result for this model on the results page though it's unclear to me how they ran it.

Thank you!

@chuangtc
Copy link
Member

Not so many people have such powerful machine as you do.
Please checkout the code, then modify the following lines to 70b

https://github.com/aidatatools/ollama-benchmark/blob/c6a5fded472370f35a158258e79f06d36313ffb1/llm_benchmark/data/benchmark1.yml#L17C15-L17C16

Then run it in development mode. Thank you.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants