• 8 Posts
  • 4 Comments
Joined 4 months ago
cake
Cake day: December 26th, 2024

help-circle







  • Though apparently I didn’t need step 6 as it started running after I downloaded it

    Hahahha. It really is a little redundant, now that you mention it. I’ll remove it from the post. Thank you!

    Good fun. Got me interested in running local LLM for the first time.

    I’m very happy to hear my post motivated you to run an LLM locally for the first time! Did you manage to run any other models? How was your experience? Let us know!

    What type of performance increase should I expect when I spin this up on my 3070 ti?

    That really depends on the model, to be completely honest. Make sure to check the model requirements. For llama3.2:2b you can expect a significant performance increase, at least.