Lo nuevo en Article Factory y lo último en el mundo de la IA generativa

DeepSeek Introduces Engram to Cut High‑Bandwidth Memory Needs in Large AI Models

DeepSeek Introduces Engram to Cut High‑Bandwidth Memory Needs in Large AI Models
DeepSeek, in partnership with Peking University, unveiled Engram, a new training method that separates static memory from computation in large language models. By using hashed N‑gram lookups and a context‑aware gating mechanism, Engram reduces reliance on high‑bandwidth memory (HBM), allowing models to operate efficiently on standard GPU memory while scaling parameter counts. Tests on a 27‑billion‑parameter model showed measurable gains across industry benchmarks, and the approach integrates with existing hardware solutions such as Phison’s SSD‑based accelerators and emerging CXL standards. Engram could ease pressure on costly memory hardware and stabilize DRAM price volatility. Leer más →

OpenAI Faces Growing Competitive Pressure as Rivals Accelerate AI Advances

OpenAI Faces Growing Competitive Pressure as Rivals Accelerate AI Advances
OpenAI's dominance in the generative‑AI market has waned as competitors such as Google, Anthropic and DeepSeek have introduced new models that outpace its latest release, GPT‑5. The company, led by Sam Altman, has responded with accelerated development and internal restructuring, but analysts note that its reliance on external funding and costly infrastructure deals leaves it vulnerable. While ChatGPT still draws hundreds of millions of users, the rapid growth of rival platforms and the escalating cost of AI‑related hardware have intensified a "code red" atmosphere within OpenAI. Leer más →

Nvidia CEO Warns of China’s Rapid AI Infrastructure Build as Open‑Source Models Capture 30% of Global Usage

Nvidia CEO Warns of China’s Rapid AI Infrastructure Build as Open‑Source Models Capture 30% of Global Usage
Nvidia chief executive Jensen Huang cautioned that China can construct AI data centers and even hospitals far faster than the United States, citing the country's expansive energy resources and swift construction capabilities. At the same time, a report from OpenRouter and Andreessen Horowitz shows Chinese open‑source large language models now account for roughly 30% of global AI token usage, up from just over 1% a year earlier. While Huang affirmed Nvidia’s chip technology remains ahead of China, the rapid growth of Chinese AI models and the nation’s infrastructure advantages highlight an intensifying competitive landscape. Leer más →

OpenAI’s o3 Model Wins AI Poker Tournament

OpenAI’s o3 Model Wins AI Poker Tournament
In a week‑long AI‑only poker showdown, OpenAI’s o3 model emerged victorious, out‑earning the other eight large‑language‑model competitors. The contest featured nine chatbots—including Anthropic’s Claude Sonnet 4.5, X.ai’s Grok, Google’s Gemini 2.5 Pro, Meta’s Llama 4, DeepSeek R1, Moonshot’s Kimi K2, Mistral’s Magistral, and Z.AI’s GLM 4.6—playing thousands of hands of no‑limit Texas hold ’em at $10 and $20 tables with $100,000 bankrolls each. While the bots displayed strong strategic play, they struggled with bluffing, position, and basic math, highlighting both progress and lingering gaps in AI decision‑making under uncertainty. Leer más →

Google’s 2025 Year in Search Highlights AI Chatbots, Hot Honey and Global Sports Favorites

Google’s 2025 Year in Search Highlights AI Chatbots, Hot Honey and Global Sports Favorites
Google’s annual Year in Search report shows that the top trending query in 2025 was “Gemini,” the company’s AI chatbot, followed by searches related to cricket, politics and pop culture. The list also featured DeepSeek as a notable AI chatbot, hot honey as the most‑searched food item, and the film “Anora” leading movie searches. Sports enthusiasts gravitated toward the FIFA Club World Cup and teams such as Paris Saint-Germain, while podcasts and books saw “The Charlie Kirk Show” and Colleen Hoover’s “Regretting You” at the top of their categories. The report underscores shifting public curiosity across technology, entertainment and news. Leer más →

AWS Expands Custom LLM Tools with Serverless SageMaker and Bedrock Enhancements

AWS Expands Custom LLM Tools with Serverless SageMaker and Bedrock Enhancements
Amazon Web Services introduced a suite of new capabilities aimed at simplifying the creation of custom large language models for enterprise customers. At its re:Invent conference, AWS unveiled serverless model customization in SageMaker, offering both point‑and‑click and natural‑language‑driven workflows, and announced reinforcement fine‑tuning in Bedrock. The company also launched Nova Forge, a service that builds bespoke Nova models for a fixed annual fee. These moves signal AWS’s focus on frontier AI models and could help customers differentiate their AI solutions in a market dominated by Anthropic, OpenAI, and Gemini. Leer más →

DeepSeek Unleashes Open-Source AI Models That Rival Leading U.S. Systems

DeepSeek Unleashes Open-Source AI Models That Rival Leading U.S. Systems
Chinese startup DeepSeek has released two new AI models—DeepSeek‑V3.2 and DeepSeek‑V3.2‑Speciale—under an open-source license. The models claim performance comparable to GPT‑5 and Gemini 3 Pro on long‑form reasoning, tool use, and dense problem solving while offering a 128,000‑token context window and reduced computational cost through Sparse Attention. Their launch challenges the dominance of U.S. AI firms, sparks regulatory scrutiny in Europe, and raises questions about the future of AI accessibility and geopolitics. Leer más →

DeepSeek Unveils V3.2 and V3.2‑Speciale Models to Challenge Leading AI Systems

DeepSeek Unveils V3.2 and V3.2‑Speciale Models to Challenge Leading AI Systems
DeepSeek, the Hangzhou‑based artificial‑intelligence firm, announced the release of two new reasoning‑capable models, V3.2 and V3.2‑Speciale. V3.2 is now available through the company’s app and web platform, while V3.2‑Speciale is offered via API access only. The company claims the Speciale version surpasses Google’s Gemini 3.0 Pro and OpenAI’s GPT‑5 High on benchmark tests, positioning DeepSeek as a serious contender in the global AI race. Leer más →

Chatbots Cite Russian State Media in Responses About Ukraine Conflict

Chatbots Cite Russian State Media in Responses About Ukraine Conflict
Researchers from the Institute of Strategic Dialogue examined four widely used AI chatbots and found that they frequently reference Russian state‑affiliated media and other sanctioned sources when answering questions about the war in Ukraine. The study highlights how data gaps can be exploited by disinformation networks and raises concerns about the ability of large language models to filter prohibited content, especially within the European Union. Leer más →

Microsoft Edge subtly nudges users toward Copilot on rival AI sites

Microsoft Edge subtly nudges users toward Copilot on rival AI sites
Microsoft Edge now displays a "Try Copilot" button when users visit popular AI services such as ChatGPT, Perplexity and Deepseek. The button opens the Copilot assistant in the browser sidebar, a move Microsoft says is meant to promote its own AI offering. The prompt does not appear on visits to Google’s Gemini, and a larger pop‑up appears when users navigate to the Chrome download page, urging them to stay with Edge. Some users find the nudges intrusive and have discovered how to disable the Copilot button through the browser’s settings. Leer más →

ByteDance’s Doubao Surpasses DeepSeek to Become China’s Leading AI Chatbot

ByteDance’s Doubao Surpasses DeepSeek to Become China’s Leading AI Chatbot
ByteDance’s Doubao app has overtaken DeepSeek to become the most popular AI chatbot in China, drawing tens of millions of monthly active users. Designed with a friendly avatar and deep integration with Douyin, Doubao offers text, audio, video, image generation and customizable agents. Analysts credit ByteDance’s expertise in viral app design and seamless cross‑platform sharing for Doubao’s rapid growth, while DeepSeek remains a more minimalist text‑only service. The company is now exploring extensions of Doubao into smart glasses, cars and other devices, signalling an ambition to broaden the AI assistant beyond smartphones. Leer más →

The Enduring Art of Assembly: From Classic Games to Modern AI

The Enduring Art of Assembly: From Classic Games to Modern AI
Assembly language, the low‑level code that speaks directly to a processor, has shaped both iconic video games and cutting‑edge artificial‑intelligence research. Chris Sawyer used x86 assembly to build the beloved titles RollerCoaster Tycoon and Transport Tycoon, trading modern tools for raw efficiency and personal passion. Decades later, companies such as DeepSeek and DeepMind have revisited assembly to squeeze performance out of Nvidia chips, demonstrating that even today’s AI breakthroughs can hinge on the meticulous control that only assembly provides. The story illustrates how a language once deemed archaic remains vital in today’s tech landscape. Leer más →

DeepSeek AI Chatbot Surges to Prominence Amid Global Competition

DeepSeek AI Chatbot Surges to Prominence Amid Global Competition
DeepSeek, a Chinese AI lab, has vaulted into mainstream awareness as its chatbot app climbs to the top of major app stores. Backed by High‑Flyer Capital Management, the company has released a series of models—including DeepSeek‑V2, DeepSeek‑V3, and the reasoning‑focused R1—that claim strong benchmark performance and low inference costs. The rapid rise has drawn attention from industry giants, regulators, and governments, prompting both integration into platforms like Microsoft’s Azure AI Foundry and bans on government devices in several jurisdictions. Leer más →

DeepSeek Unveils Sparse‑Attention Model to Halve API Inference Costs

DeepSeek Unveils Sparse‑Attention Model to Halve API Inference Costs
DeepSeek announced a new experimental AI model featuring Sparse Attention technology that dramatically lowers inference costs for long‑context tasks. The model, released on Hugging Face and accompanied by a research paper on GitHub, uses a lightning indexer and fine‑grained token selection to focus computational resources on the most relevant excerpts. Preliminary tests suggest API call prices can be cut by as much as 50 percent in long‑context scenarios. The open‑weight release invites third‑party validation and positions DeepSeek as a notable player in the ongoing effort to make transformer‑based AI more cost‑effective. Leer más →

Knowledge Distillation Emerges as a Core Technique for Building Smaller, Cost‑Effective AI Models

Knowledge Distillation Emerges as a Core Technique for Building Smaller, Cost‑Effective AI Models
Knowledge distillation, a method that transfers information from a large "teacher" model to a smaller "student" model, has become a fundamental tool for reducing the size and expense of AI systems. Originating from a 2015 Google paper, the technique leverages soft‑target probabilities to convey nuanced relationships between data classes, enabling compact models to retain high performance. Over the years, distillation has been applied to language models such as BERT and its distilled variant, DistilBERT, and is now offered as a service by major cloud providers. Recent developments continue to expand its utility across reasoning tasks and open‑source initiatives. Leer más →

Knowledge Distillation Emerges as a Core Technique for Building Smaller, Cost‑Effective AI Models

Knowledge Distillation Emerges as a Core Technique for Building Smaller, Cost‑Effective AI Models
Knowledge distillation, a method that transfers information from a large "teacher" model to a smaller "student" model, has become a fundamental tool for reducing the size and expense of AI systems. Originating from a 2015 Google paper, the technique leverages soft‑target probabilities to convey nuanced relationships between data classes, enabling compact models to retain high performance. Over the years, distillation has been applied to language models such as BERT and its distilled variant, DistilBERT, and is now offered as a service by major cloud providers. Recent developments continue to expand its utility across reasoning tasks and open‑source initiatives. Leer más →

Knowledge Distillation Emerges as a Core Technique for Building Smaller, Cost‑Effective AI Models

Knowledge Distillation Emerges as a Core Technique for Building Smaller, Cost‑Effective AI Models
Knowledge distillation, a method that transfers information from a large "teacher" model to a smaller "student" model, has become a fundamental tool for reducing the size and expense of AI systems. Originating from a 2015 Google paper, the technique leverages soft‑target probabilities to convey nuanced relationships between data classes, enabling compact models to retain high performance. Over the years, distillation has been applied to language models such as BERT and its distilled variant, DistilBERT, and is now offered as a service by major cloud providers. Recent developments continue to expand its utility across reasoning tasks and open‑source initiatives. Leer más →

Knowledge Distillation Emerges as a Core Technique for Building Smaller, Cost‑Effective AI Models

Knowledge Distillation Emerges as a Core Technique for Building Smaller, Cost‑Effective AI Models
Knowledge distillation, a method that transfers information from a large "teacher" model to a smaller "student" model, has become a fundamental tool for reducing the size and expense of AI systems. Originating from a 2015 Google paper, the technique leverages soft‑target probabilities to convey nuanced relationships between data classes, enabling compact models to retain high performance. Over the years, distillation has been applied to language models such as BERT and its distilled variant, DistilBERT, and is now offered as a service by major cloud providers. Recent developments continue to expand its utility across reasoning tasks and open‑source initiatives. Leer más →

Knowledge Distillation Emerges as a Core Technique for Building Smaller, Cost‑Effective AI Models

Knowledge Distillation Emerges as a Core Technique for Building Smaller, Cost‑Effective AI Models
Knowledge distillation, a method that transfers information from a large "teacher" model to a smaller "student" model, has become a fundamental tool for reducing the size and expense of AI systems. Originating from a 2015 Google paper, the technique leverages soft‑target probabilities to convey nuanced relationships between data classes, enabling compact models to retain high performance. Over the years, distillation has been applied to language models such as BERT and its distilled variant, DistilBERT, and is now offered as a service by major cloud providers. Recent developments continue to expand its utility across reasoning tasks and open‑source initiatives. Leer más →

Andreessen Horowitz Report Shows Google, Grok and Meta Narrowing Gap With ChatGPT

Andreessen Horowitz Report Shows Google, Grok and Meta Narrowing Gap With ChatGPT
A new consumer‑AI report from Andreessen Horowitz (a16z) charts the latest competitive landscape. While OpenAI’s ChatGPT remains the market leader, Google’s Gemini, AI Studio, NotebookLM and Google Labs have entered the top rankings. Meta AI and xAI’s Grok are also gaining ground, with Grok reporting roughly 20 million monthly active users and a recent 40 percent surge after its latest version. Chinese AI products continue to dominate traffic in their home market, and a wave of newcomers is reshaping the mobile app list as platforms crack down on copycat services. Leer más →