Lo nuevo en Article Factory y lo último en el mundo de la IA generativa

Modal Labs in Talks for Funding Round Valued at $2.5 Billion

Modal Labs in Talks for Funding Round Valued at $2.5 Billion
Modal Labs, an AI inference infrastructure startup, is in early discussions with venture capital firms about a new financing round that could value the company at roughly $2.5 billion. If completed, the round would more than double the $1.1 billion valuation reported less than five months earlier. The company’s annualized revenue run rate is about $50 million. Co‑founder and CEO Erik Bernhardsson said the conversations are general and not an active fundraising effort. Existing backers include Lux Capital and Redpoint Ventures, while General Catalyst is reportedly considering leading the round. Leer más →

Microsoft Won’t Stop Buying AI Chips From Nvidia and AMD Even After Launching Its Own, Nadella Says

Microsoft Won’t Stop Buying AI Chips From Nvidia and AMD Even After Launching Its Own, Nadella Says
Microsoft has begun deploying its home‑grown Maia 200 AI inference chip in its data centers while confirming it will continue purchasing GPUs from Nvidia and AMD. CEO Satya Nadella emphasized the company’s ongoing partnerships and the need to stay ahead in AI hardware. The Maia 200, designed for high‑performance AI model inference, will first be used by Microsoft’s Superintelligence team as they develop frontier models, and will also support OpenAI models on Azure. The move reflects a hybrid strategy of building in‑house chips while leveraging external suppliers. Leer más →

Qualcomm Unveils AI200 and AI250 Chips to Challenge Nvidia

Qualcomm Unveils AI200 and AI250 Chips to Challenge Nvidia
Qualcomm announced the AI200 and AI250 processors, built on its Hexagon neural processing technology, to target AI inference workloads and compete with Nvidia's dominance. The AI200, slated for release next year, offers 768 GB of RAM and can be combined in racks of up to 72 chips. The AI250, planned for 2027, promises a generational leap in efficiency and lower power consumption. Saudi Arabia's Public Investment Fund‑backed AI firm Humain will use both chips in its upcoming AI datacenters, signaling Qualcomm's push into the high‑performance AI market. Leer más →

Tensormesh Secures $4.5M Seed Funding to Commercialize AI Inference Cache

Tensormesh Secures $4.5M Seed Funding to Commercialize AI Inference Cache
Tensormesh, a startup emerging from stealth mode, announced a $4.5 million seed round led by Laude Ventures with additional backing from angel investor Michael Franklin. The funding will accelerate the development of a commercial product built around LMCache, an open‑source utility that can slash AI inference costs by up to tenfold. Tensormesh’s approach focuses on preserving the key‑value (KV) cache across queries, a technique that boosts efficiency for chat‑driven and agentic AI systems. The company aims to offer an out‑of‑the‑box solution that eliminates the need for extensive engineering effort, positioning itself as a cost‑saving layer for GPU‑intensive workloads. Leer más →

Cerebras Systems Raises $1.1 Billion in Series G Funding, Valued at $8.1 Billion

Cerebras Systems Raises $1.1 Billion in Series G Funding, Valued at $8.1 Billion
Cerebras Systems, the AI‑hardware startup founded in 2015, announced a $1.1 billion Series G round that values the company at $8.1 billion. The financing, co‑led by Fidelity and Atreides Management with participation from Tiger Global, Valor Equity Partners and 1789 Capital, follows a $250 million Series F round in 2021. CEO and co‑founder Andrew Feldman said the round will fund new data‑center expansions, U.S. manufacturing hubs and continued technology development. Despite an earlier plan to go public by 2025 that was delayed by regulatory review, Cerebras remains focused on scaling its AI inference services and cloud offering. Leer más →

Clarifai Launches Reasoning Engine to Accelerate AI Model Performance and Cut Costs

Clarifai Launches Reasoning Engine to Accelerate AI Model Performance and Cut Costs
Clarifai announced a new reasoning engine that promises to double inference speed and reduce costs by 40 percent. The platform combines low‑level CUDA kernel tweaks with advanced speculative decoding to extract more performance from existing GPU hardware. Independent benchmarks reported industry‑leading throughput and latency. The launch comes amid a surge in demand for AI compute, highlighted by OpenAI’s plan to spend up to $1 trillion on new data centers. Clarifai’s CEO emphasized that software and algorithmic innovations remain critical even as hardware builds out. Leer más →