Local LLM Setup & Performance (GPU / Tokens/sec) #2
Rami8612
started this conversation in
Local LLMs
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Local Model Performance Benchmarks
Running RamiBot with local models can vary significantly depending on the GPU, model size, and inference settings.
This discussion is intended to collect performance benchmarks from the community.
Please share your setup using the following format:
Hardware
Model
Inference
Example
Hardware
Model
Inference
These benchmarks help others choose the best setup for running RamiBot locally.
If you test multiple models on the same GPU, feel free to post comparisons.
Beta Was this translation helpful? Give feedback.
All reactions