{"html":"<iframe src=\"https://hatenablog-parts.com/embed?url=https%3A%2F%2Fimpsbl.hatenablog.jp%2Fentry%2FLMStudio_en\" title=\"LM Studio and SML on PC with 8GB RAM, No GPU - downloading models, starting server and agentic editing - Technically Impossible\" class=\"embed-card embed-blogcard\" scrolling=\"no\" frameborder=\"0\" style=\"display: block; width: 100%; height: 190px; max-width: 500px; margin: 10px 0px;\"></iframe>","description":"The emergence of llama.cpp*1 in 2023 made it possible to offload GPU processing to the CPU, making it a reality to run generative AI on a local environment. Even models that typically require a lot of VRAM can now be run on a system with as little as 8GB of RAM. The advent of LM Studio and SML (Smal\u2026","blog_title":"Technically Impossible","url":"https://impsbl.hatenablog.jp/entry/LMStudio_en","blog_url":"https://impsbl.hatenablog.jp/","width":"100%","provider_name":"Hatena Blog","provider_url":"https://hatena.blog","author_url":"https://blog.hatena.ne.jp/espio999/","published":"2025-09-02 00:00:00","type":"rich","categories":["AI","English post","IT"],"version":"1.0","author_name":"espio999","title":"LM Studio and SML on PC with 8GB RAM, No GPU - downloading models, starting server and agentic editing","image_url":"https://cdn-ak.f.st-hatena.com/images/fotolife/e/espio999/20250829/20250829225509.png","height":"190"}