Nvidia Reaches Historic $200 Billion Revenue Milestone Driven by AI Boom In a landmark achievement, Nvidia has become the first semiconductor company to post over $200 billion in annual revenue, driven by surging demand for its AI-powered technologies. This milestone underscores the company's central role in fueling the current AI investment boom across industries. Key Highlights: - Nvidia's revenue for the fiscal year 2023 reached a record $26.9 billion, up 13% year-over-year, with the Data Center segment leading the charge at $16.2 billion in revenue. - The company's AI-focused offerings, including its Blackwell GPU architecture and Hopper AI accelerators, have been in high demand as organizations race to build out their AI infrastructure and capabilities. - Nvidia is collaborating with leading cybersecurity providers like Forescout, Palo Alto Networks, and Xage Security to bring its accelerated computing and AI technologies to operational technology (OT) and industrial control systems (ICS) environments, enhancing real-time threat detection and response across critical infrastructure. - In India, Nvidia is supporting the government's IndiaAI Mission by partnering with cloud providers like Yotta, L&T, and E2E Networks to deliver advanced AI infrastructure, including tens of thousands of Nvidia Blackwell Ultra GPUs, to fuel the country's AI transformation. - The rapid decline in large language model (LLM) token prices, driven by the release of open-source models like Llama 3.1 and hardware innovations from companies like Groq, Samba Nova, and Cerebras, is expected to further accelerate the adoption of AI applications. - Nvidia's milestone also highlights the broader push for semiconductor manufacturing onshoring, with the company and TSMC producing the first Blackwell wafer entirely on U.S. soil at TSMC's Arizona facility, a significant step toward domestic AI chip production. As the AI revolution continues to reshape industries, Nvidia's record-breaking performance underscores its position as a critical enabler of this transformative technology, driving innovation and reshaping the global semiconductor landscape.
Nous Research's NousCoder-14B is an open-source coding model landing right in the Claude Code moment
Anthropic's NousCoder-14B: Revolutionizing the Coding Landscape In the rapidly evolving world of AI-powered coding tools, Anthropic's NousCoder-14B, also known as Claude Code, has emerged as a game-changer. This open-source coding model is poised to reshape the software development landscape, as evidenced by the growing buzz and adoption among engineers in Silicon Valley. According to Boris Cherny, the head of Claude Code, the team at Anthropic built the "simplest possible thing" with the goal of anticipating the future capabilities of AI coding tools. This foresight has proven prescient, as recent months have seen a significant inflection point in the capabilities of AI coding agents. Kian Katanforoosh, the CEO of the startup Workera and an adjunct lecturer on AI at Stanford, has praised the performance of Claude Code, particularly after the launch of Anthropic's latest AI model, Claude Opus 4.5. Katanforoosh notes that the model "doesn't even feel like it's coding like a human" and has "figured out a better way" to approach coding tasks. The rapid evolution of AI coding tools has not gone unnoticed by the industry. In November 2025, Anthropic announced that Claude Code had reached $1 billion in annual recurring revenue, underscoring the growing demand for these advanced coding assistants. Anthropic has further bolstered the capabilities of Claude Code with the introduction of new desktop features that automate more of the development workflow. These include the ability to spin up development servers, display running web apps, spot and fix errors, and even merge pull requests on their own. Moreover, Anthropic has unveiled Claude Code Security, a research preview feature that leverages the latest Claude Opus 4.6 model to identify complex vulnerabilities in codebases, such as business logic flaws and access control issues, that often evade traditional static analysis tools. The success of Claude Code has not been limited to Anthropic's internal development. The company's Claude Cowork tool, which builds on Claude Code, was developed in under two weeks using the AI-powered coding assistant to write almost all the code. As the industry continues to embrace the power of AI-driven coding, Anthropic's NousCoder-14B, or Claude Code, stands out as a transformative technology that is reshaping the software development landscape. With its growing capabilities, seamless integration, and security-focused features, Claude Code is poised to become an indispensable tool for developers and organizations alike.
Distilling the tool-using capabilities of large language models (LLMs) into smaller, more efficient small language models (SLMs) is a key challenge for their practical application. The predominant approach, supervised fine-tuning (SFT), suffers from poor generalization as it trains models to imitate a static set of teacher trajectories rather than learn a robust methodology. While reinforcement learning (RL) offers an alternative, the standard RL using sparse rewards fails to effectively guide SLMs, causing them to struggle with inefficient exploration and adopt suboptimal strategies. To address these distinct challenges, we propose MENTOR, a framework that synergistically combines RL with teacher-guided distillation. Instead of simple imitation, MENTOR employs an RL-based process to learn a more generalizable policy through exploration. In addition, to solve the problem of reward sparsity, it uses a teacher's reference trajectory to construct a dense, composite teacher-guided reward that provides fine-grained guidance. Extensive experiments demonstrate that MENTOR significantly improves the cross-domain generalization and strategic competence of SLMs compared to both SFT and standard sparse-reward RL baselines.
Gradient-based data attribution methods, such as influence functions, are critical for understanding the impact of individual training samples without requiring repeated model retraining. However, their scalability is often limited by the high computational and memory costs associated with per-sample gradient computation. In this work, we propose GraSS, a novel gradient compression algorithm and its variants FactGraSS for linear layers specifically, that explicitly leverage the inherent sparsity of per-sample gradients to achieve sub-linear space and time complexity. Extensive experiments demonstrate the effectiveness of our approach, achieving substantial speedups while preserving data influence fidelity. In particular, FactGraSS achieves up to 165% faster throughput on billion-scale models compared to the previous state-of-the-art baselines. Our code is publicly available at https://github.com/TRAIS-Lab/GraSS.
Large language models (LLMs) have demonstrated promising performance in generating diagnostic conclusions from imaging findings, thereby supporting radiology reporting, trainee education, and quality control. However, systematic guidance on how to optimize prompt design across different clinical contexts remains underexplored. Moreover, a comprehensive and standardized framework for assessing the trustworthiness of LLM-generated radiology reports is yet to be established. This study aims to enhance the trustworthiness of LLM-generated liver MRI reports by introducing a Multi-Dimensional Credibility Assessment (MDCA) framework and providing guidance on institution-specific prompt optimization. The proposed framework is applied to evaluate and compare the performance of several advanced LLMs, including Kimi-K2-Instruct-0905, Qwen3-235B-A22B-Instruct-2507, DeepSeek-V3, and ByteDance-Seed-OSS-36B-Instruct, using the SiliconFlow platform.