Run Large Language Models (LLMs) without GPU with local-llm.
With local-llm, you can run LLMs locally or on Cloud Workstations.
Join this channel to get access to perks:
/ @rishabincloud
Timestamps:
0:00 intro
0:42 key benefits of running LLMs locally
1:25 what is local-llm
3:00 installing local-llm
6:00 running a model with local-llm
8:45 outro
Google Cloud Blog post - https://cloud.google.com/blog/product...
local-llm GitHub - https://github.com/GoogleCloudPlatfor...
Resources:
Learn to Cloud - https://learntocloud.guide
The DevOps Guide - https://thedevops.guide
Support this channel:
Buymeacoffee - https://www.buymeacoffee.com/rishabin...
Fine me on GitHub - https://github.com/rishabkumar7
Connect with me:
https://rishabkumar.com
Twitter - / rishabincloud
LinkedIn - / rishabkumar7
Instagram - / rishabincloud
#llm #localllm
Watch video Run LLMs without GPUs | local-llm online without registration, duration hours minute second in high quality. This video was added by user Rishab in Cloud 29 April 2024, don't forget to share it with your friends and acquaintances, it has been viewed on our site 4,429 once and liked it 78 people.