Microsoft's Shader Execution Reordering (SER) technology is delivering dramatic performance gains on modern GPUs, achieving up to 90% improvement on Intel Arc B-Series and 80% on NVIDIA Blackwell GPUs, according to TechPowerUp.
#hardware
RSS FeedTop analyst firm Gartner has predicted that the sub-$500 entry-level PC segment will vanish entirely by 2028. The firm also forecasts a 10.4% decline in worldwide PC shipments during 2026, signaling a major shift in the PC market toward higher-end, AI-capable hardware.
Soaring AI data center demand for DRAM and HBM chips is driving a global memory shortage that will push the average smartphone price 14% higher to an all-time record of $523, while eliminating sub-$100 handsets entirely.
A remarkable 13-month comparison: running frontier-level DeepSeek R1 at ~5 tokens/second cost $6,000 in early 2025. Today, you can run a significantly stronger model at the same speed on a $600 mini PC — and get 17-20 t/s with even more capable models.
NVIDIA revealed detailed specs for Vera Rubin NVL72. Each Rubin GPU delivers 50 PFLOPS inference (5x Blackwell GB200), 22 TB/s HBM4 bandwidth (2.8x Blackwell), and cuts inference cost per million tokens by 10x. Ships H2 2026.
Andrej Karpathy highlights the fundamental memory+compute trade-off challenge in LLMs: fast but small on-chip SRAM versus large but slow off-chip DRAM. He calls optimizing this the most intellectually rewarding puzzle in AI infrastructure today, pointing to NVIDIA's $4.6T market cap as proof.
A Reddit thread spotlighted LLmFit, a CLI/TUI tool for recommending runnable models per hardware profile, while commenters raised data-quality and recommendation-validity questions.
NVIDIA CEO Jensen Huang promised chips the world has never seen at GTC 2026. Industry reports point to the Feynman architecture on TSMC A16 1.6nm-class process with silicon photonics interconnects.
Researchers have built a thermodynamic computer that generates images from noise using orders of magnitude less energy than current generative AI models, offering a radically different approach to AI hardware.
Meta and NVIDIA announced a multiyear strategic AI infrastructure partnership on February 17. Meta will deploy millions of Blackwell and Rubin GPUs, plus Grace CPUs in a first-of-its-kind standalone server deployment. Analysts peg the deal at ~$50B.
NVIDIA unveiled the N1 and N1X on February 23, its first consumer SoC combining Arm CPUs with Blackwell GPU architecture for AI PCs. Dell, HP, and Lenovo laptops are expected in spring 2026, marking NVIDIA's bold entry into the PC processor market.
Startup Taalas proposes baking entire LLM weights and architecture into custom ASICs, claiming 17K+ tokens/second per user, sub-1ms latency, and 20x lower cost than cloud — all achievable within a 60-day chip production cycle.