A new Intel chip due in December will be able to run a generative artificial intelligence chatbot on a laptop rather than having to tap into cloud data centers for computing power, the company said on Tuesday.
You can already run LLMs on consumer hardware. They’re slow, but they definitely work. Mine is midrange from about 5 years ago and I can run a small one at a pretty reasonable speed or a medium (but “smarter”) one quite slowly at roughly 2 seconds per token.
You can already run LLMs on consumer hardware. They’re slow, but they definitely work. Mine is midrange from about 5 years ago and I can run a small one at a pretty reasonable speed or a medium (but “smarter”) one quite slowly at roughly 2 seconds per token.
deleted by creator