1. OpenAI introduces 'GPT-4o mini', a cost-effective language model priced at $0.15 per 1 million input tokens and $0.60 per 1 million output tokens, significantly cheaper than previous models and 60% cheaper than GPT-3.5 Turbo; 2. The model is designed for applications requiring low cost and low latency, such as chaining or parallelizing model calls, handling large amounts of context, and real-time text responses for customer interactions; 3. The current API supports text and vision, with plans to include video and audio input/output in the future, offering a 128K token context window and up to 16K output tokens per request, and knowledge up to October 2023. The model demonstrates improved cost-efficiency for non-English text processing with a shared, upgraded tokenizer with GPT-4o. It achieves high scores on various benchmarks, outperforming competitors like 'Gemini Flash' and 'Claude Haiku'.
Related Articles
- Apple fumbles AI again: new 'LLM Siri' update for iOS 19 delayed, now expected with iOS in 20263 months ago
- The Double-Edged Sword of AI Processors: Batch Sizes, Token Rates, and the Hardware Hurdles in Large Language Model Processing3 months ago
- Chinese CPU maker Zhaoxin rolls out DeepSeek support to all processors — entire product lineup now runs DeepSeek LLMs natively3 months ago
- Government delays AI Bill3 months ago
- Researchers discover if 0.001% of AI training data misinformation the AI becomes corrupted5 months ago
- AMD Ryzen AI 9 HX 375 outperforms Intel's Core Ultra 7 258V in LLM performance — Team Red provided benchmarks show a strong lead of up to 27% in LM Studio7 months ago
- Orion Could Be OpenAI's Game-Changing AI Model7 months ago
- Dell PowerEdge XE9712: NVIDIA GB200 NVL72-based AI GPU cluster for LLM training, inference8 months ago
- A New Chip Challenges GPU Performance8 months ago
- SK hynix starts mass production of 12-layer HBM3E memory: 36GB capacity per module @ 9.6Gbps8 months ago