1. Researchers from UC Santa Cruz have discovered a method to run large language models (LLMs) at a mere 13 watts without compromising performance. 2. The key to this efficiency is the elimination of matrix multiplication in LLM processing, which, when optimized, significantly boosts performance-per-watt. 3. The broader applicability of this approach to AI in general is yet to be determined.
Related Articles
- Famed gamer creates working 5 million parameter ChatGPT AI model in Minecraft, made with 439 million blocks — AI trained to hold conversations, working model runs inference in the gameabout 1 month ago
- ERC Starting Grant for Junior Professor Stefan Mönch – Towards Nearly Lossless Electrical Energy Conversionabout 2 months ago
- Aesthetic and Efficient Facades: Fraunhofer FEP Presents Advances in Building-Integrated Photovoltaics2 months ago
- OpenAI Is Readying Its GPT-5 But It May Not Be The Breakthrough Some Are Expecting3 months ago
- What Moved Markets This Week3 months ago
- InPlay Oil: High Yield, Hidden Value In Canadian Energy4 months ago
- OXLC Is Offering Investors A Rare Opportunity4 months ago
- New standards for durable electronics: The EU energy label for mobile devices5 months ago
- BTU Team Lausitz Dynamics Competes Energy-Efficiently in International Competition5 months ago
- If I Could Only Buy 2 High-Yield REITs Today5 months ago