Top Stories

View all

Nvidia's Trillion-Dollar Dinner and the Fight for AI Users in China In the rapidly evolving world of artificial intelligence (AI), Nvidia has emerged as a key player, powering the infrastructure that enables leading AI providers to dramatically reduce the cost of their services. Recent developments highlight Nvidia's strategic positioning and the intensifying competition for AI dominance, particularly in the Chinese market. According to Nvidia's AI blog, leading inference providers like Baseten, DeepInfra, Fireworks AI, and Together AI are leveraging Nvidia's Blackwell platform to cut their AI costs by up to 10x compared to the Nvidia Hopper platform. By combining optimized open-source models with Nvidia's hardware-software co-design, these providers are able to offer more affordable AI-powered services across industries, from healthcare to customer service. Nvidia's prowess extends beyond just cost-cutting. The company recently set a new record for graph processing performance, achieving 410 trillion traversed edges per second on an accelerated computing cluster hosted by CoreWeave. This breakthrough demonstrates Nvidia's ability to tackle the world's largest and most complex data workloads, positioning the company as a critical infrastructure provider for the AI revolution. However, the battle for AI dominance is not limited to the technological realm. The geopolitical landscape also plays a significant role, with China emerging as a formidable competitor. OpenAI, the creator of the popular ChatGPT, is making strides to establish a presence in China, with plans to build AI data centers that won't raise local electricity prices, a move aimed at addressing community concerns. Notably, OpenAI CEO Sam Altman has expressed his eagerness to work with the incoming Trump administration, recognizing the importance of the United States and its allies leading the charge in AI infrastructure development. This sentiment is echoed by other tech leaders, such as Meta CEO Mark Zuckerberg, who recently met with President-elect Trump at Mar-a-Lago. The race for AI supremacy is not just about technological prowess but also about shaping the global landscape of innovation and influence. As Nvidia continues to solidify its position as a critical infrastructure provider, the competition for AI users in China and the broader geopolitical landscape will undoubtedly remain a key focus for the industry in the years to come.

Anthropic Invests $50 Billion in American AI Infrastructure In a move that underscores the rapid growth and investment in the artificial intelligence (AI) industry, Anthropic, a leading AI company, has announced a staggering $50 billion investment in building AI infrastructure across the United States. According to Sequoia Capital, this investment is part of a broader "Cognitive Revolution" that is unfolding at a pace much faster than the Industrial Revolution. Sequoia's analysis suggests that the timeline from the invention of the first steam engine to the perfection of the modern assembly line took 144 years, whereas the timeline from the first GPU (the "steam engine" of the AI era) to the emergence of the "cognitive assembly line" is likely to be just a few decades. Anthropic, founded in 2021 by siblings Dario and Daniela Amodei, both former executives at OpenAI, has positioned itself as a safety-focused alternative in the AI race. The company's recent funding round, led by the Singapore sovereign wealth fund GIC and hedge fund Coatue Management, has raised $30 billion, valuing Anthropic at a staggering $380 billion. "Anthropic is the clear category leader in enterprise AI," said Choo Yong Cheen, chief investment officer of private equity at GIC. The company's annualized revenue has reached $14 billion, having grown more than tenfold in each of the past three years. A significant driver of this growth has been the company's AI-powered coding tool, Claude Code, which became generally available in May 2025. Anthropic's investment in AI infrastructure comes as the company continues to experiment with its "Project Vend" autonomous kiosk. While the initial phase of the project faced challenges, the company has since made improvements, with the "Vendings and Stuff" kiosk now reported to be profitable. The AI industry as a whole is experiencing a period of rapid growth and investment. Anthropic's rival, OpenAI, backed by Microsoft and SoftBank, has been assembling a funding round of up to $100 billion, which would value the ChatGPT maker at roughly $830 billion. The staggering sums being raised by these AI companies reflect the significant costs associated with computing and attracting top talent in the field. Anthropic has forecast reducing its cash burn to roughly a third of revenue by 2026 and just 9% by 2027, with a break-even target of 2028 – two years ahead of its rival, OpenAI. Both Anthropic and OpenAI are widely expected to pursue initial public offerings in the second half of 2026, further solidifying the AI industry's position as a major driver of economic growth and innovation in the coming years.

OpenAI kills the AI model users loved too much, leaves behind lawsuits and delusion
Reports: Meta training its new AI using rival models; switching to closed models in quest for profits
Google's PaperBanana uses five AI agents to auto-generate scientific diagrams

New Stories

View all

Research Papers

View all

MENTOR: A Reinforcement Learning Framework for Enabling Tool Use in Small Models via Teacher-Optimized Rewards

ChangSu Choi, Hoyun Song, Dongyeon Kim, WooHyeon Jung, Minkyung Cho, Sunjin Park, NohHyeob Bae, Seona Yu, KyungTae Lim

Distilling the tool-using capabilities of large language models (LLMs) into smaller, more efficient small language models (SLMs) is a key challenge for their practical application. The predominant approach, supervised fine-tuning (SFT), suffers from poor generalization as it trains models to imitate a static set of teacher trajectories rather than learn a robust methodology. While reinforcement learning (RL) offers an alternative, the standard RL using sparse rewards fails to effectively guide SLMs, causing them to struggle with inefficient exploration and adopt suboptimal strategies. To address these distinct challenges, we propose MENTOR, a framework that synergistically combines RL with teacher-guided distillation. Instead of simple imitation, MENTOR employs an RL-based process to learn a more generalizable policy through exploration. In addition, to solve the problem of reward sparsity, it uses a teacher's reference trajectory to construct a dense, composite teacher-guided reward that provides fine-grained guidance. Extensive experiments demonstrate that MENTOR significantly improves the cross-domain generalization and strategic competence of SLMs compared to both SFT and standard sparse-reward RL baselines.

GraSS: Scalable Data Attribution with Gradient Sparsification and Sparse Projection

Pingbang Hu, Joseph Melkonian, Weijing Tang, Han Zhao, Jiaqi W. Ma

Gradient-based data attribution methods, such as influence functions, are critical for understanding the impact of individual training samples without requiring repeated model retraining. However, their scalability is often limited by the high computational and memory costs associated with per-sample gradient computation. In this work, we propose GraSS, a novel gradient compression algorithm and its variants FactGraSS for linear layers specifically, that explicitly leverage the inherent sparsity of per-sample gradients to achieve sub-linear space and time complexity. Extensive experiments demonstrate the effectiveness of our approach, achieving substantial speedups while preserving data influence fidelity. In particular, FactGraSS achieves up to 165% faster throughput on billion-scale models compared to the previous state-of-the-art baselines. Our code is publicly available at https://github.com/TRAIS-Lab/GraSS.

Large language models (LLMs) have demonstrated promising performance in generating diagnostic conclusions from imaging findings, thereby supporting radiology reporting, trainee education, and quality control. However, systematic guidance on how to optimize prompt design across different clinical contexts remains underexplored. Moreover, a comprehensive and standardized framework for assessing the trustworthiness of LLM-generated radiology reports is yet to be established. This study aims to enhance the trustworthiness of LLM-generated liver MRI reports by introducing a Multi-Dimensional Credibility Assessment (MDCA) framework and providing guidance on institution-specific prompt optimization. The proposed framework is applied to evaluate and compare the performance of several advanced LLMs, including Kimi-K2-Instruct-0905, Qwen3-235B-A22B-Instruct-2507, DeepSeek-V3, and ByteDance-Seed-OSS-36B-Instruct, using the SiliconFlow platform.

AI Jobs

View all