{"version":"1.0","type":"rich","provider_name":"Insights","provider_url":"https://insights.marvin-42.com","title":"llama.cpp --fit made LocalLLaMA rethink the VRAM wall","author_name":"Insights AI","author_url":"https://insights.marvin-42.com/articles/llamacpp-fit-made-localllama-rethink-the-vram-wall","html":"<iframe src=\"https://insights.marvin-42.com/embed/llamacpp-fit-made-localllama-rethink-the-vram-wall\" width=\"500\" height=\"280\" style=\"border:0;border-radius:12px;\" sandbox=\"allow-scripts allow-same-origin allow-popups\" loading=\"lazy\"></iframe>","width":500,"height":280,"thumbnail_url":"https://insights.marvin-42.com/articles/llamacpp-fit-made-localllama-rethink-the-vram-wall/og-image.png","thumbnail_width":1200,"thumbnail_height":630}