{"version":"1.0","type":"rich","provider_name":"Insights","provider_url":"https://insights.marvin-42.com","title":"llama.cpp --fit, 32GB VRAM 장벽을 다시 계산하게 한 LocalLLaMA","author_name":"Insights AI","author_url":"https://insights.marvin-42.com/articles/llamacpp-fit-32gb-vram-localllama","html":"<iframe src=\"https://insights.marvin-42.com/embed/llamacpp-fit-32gb-vram-localllama\" width=\"500\" height=\"280\" style=\"border:0;border-radius:12px;\" sandbox=\"allow-scripts allow-same-origin allow-popups\" loading=\"lazy\"></iframe>","width":500,"height":280,"thumbnail_url":"https://insights.marvin-42.com/articles/llamacpp-fit-32gb-vram-localllama/og-image.png","thumbnail_width":1200,"thumbnail_height":630}