A Child Could Run A Private Local LLM The process is less complicated than using ...

avatar

... Zoom:

  1. Install https://lmstudio.ai or similar app
  2. Select a model and download

The real bottleneck is hardware. I could only use upto 13B models. Results were worse than GPT 3.5 and slow.

#stem #leo #pob Waitio



Posted via D.Buzz



0
0
0.000
6 comments
avatar
(Edited)

Even the process of fine tuning AI has become extremely easy from where it used to be. I would not say a child could pickup the process. If you have some experience working on the terminal and have basic familiarity about AI, there are solutions already setup by experts.



Posted via D.Buzz

0
0
0.000
avatar

Ran out of RAM? Use swap files. 😂

Posted via D.Buzz

0
0
0.000
avatar
(Edited)

Just checked the screenshot, I'll take back my last comment.

He should you swap files.

Posted via D.Buzz

0
0
0.000
avatar

I gave up on local LLMs after few hours of playing around. I will have to wait till I get a better laptop. Until then I would stick with Brave's Leo or ChatGPT.

0
0
0.000
avatar

Ya when you say child, you mean an ai... like the agi with the mind of a child!

Posted via D.Buzz

0
0
0.000