As FM Sitharaman prepares to present the Union Budget, technology companies are urging investments in AI development, digital infrastructure, and market liquidity to accelerate adoption across sectors.
OpenAI signed a multi-year agreement with Cerebras Systems to purchase up to 750 megawatts of computing capacity, in a deal valued at over $10 billion over three years. The contract marks a deliberate move by OpenAI to reduce its dependence on NVIDIA for inference infrastructure at scale. Cerebras builds wafer-scale processors that consolidate compute onto a single silicon die, potentially reducing the inter-chip communication overhead inherent in traditional GPU clusters.
When we talk about the cost of AI infrastructure, the focus is usually on Nvidia and GPUs -- but memory is an increasingly important part of the picture.
The partnership will see the two companies exploring how World Labs’ models can work alongside Autodesk’s tools, and vice versa, starting with a focus on entertainment use cases.
Major technology companies collectively pledged approximately $700 billion in AI data center investment for 2026, representing an unprecedented level of coordinated capital expenditure on compute infrastructure. The spending is expected to significantly expand GPU supply and drive down inference costs over a multi-year horizon. In the near term, GPU allocation constraints persist as the primary bottleneck for AI development teams worldwide.
Meta announced the MTIA 300-500 series, custom AI accelerators designed to reduce dependency on NVIDIA GPUs for its inference and training workloads. The company plans mass deployment by end of 2027. Meta joins Google (TPUs), Amazon (Trainium), and Microsoft (Maia) in pursuing custom silicon to manage compute costs at scale.
As the bandwidth and power demands of AI data centers necessitate a transition from electrical to optical scaleup networking, one component has been conspicuously absent from the co-packaged optics arsenal: the laser itself. That’s no longer the case. Last month, Tower Semiconductor and Scintil Phot
The rise of generative AI has spurred demand for AI workstations that can run or train models on local hardware. Yet modern PCs have proven inadequate for this task. A typical laptop has only enough memory to load a large language model (LLM) with 8 billion to 13 billion parameters—much smaller, and
A traditional data center protects the expensive hardware inside it with a “shell” constructed from steel and concrete. Constructing a data center’s shell is inexpensive compared to the cost of the hardware and infrastructure inside it, but it’s not trivial. It takes time for engineers to consider p
CERAWeek — dubbed the Davos of energy — is where policymakers, producers, technologists and financiers gather to discuss how the world powers itself next. NVIDIA and Emerald AI unveiled at the conference last week a new way forward — treating AI factories not as static power loads but as flexible,
Microsoft announced a $10 billion investment in Japan covering AI infrastructure, cybersecurity, and workforce development — the largest single AI infrastructure commitment by a Western company in Asia. The investment targets data sovereignty concerns and positions Azure as the default AI platform for Japanese enterprise customers. Microsoft cited Japan's regulatory environment and demand for local AI capacity as primary drivers.
While browsing our website a few weeks ago, I stumbled upon “How and When the Memory Chip Shortage Will End” by Senior Editor Samuel K. Moore. His analysis focuses on the current DRAM shortage caused by AI hyperscalers’ ravenous appetite for memory, a major constraint on the speed at which large lan
Artificial intelligence harbors an enormous energy appetite. Such constant cravings are evident in the hefty carbon footprint of the data centers behind the AI boom and the steady increase over time of carbon emissions from training frontier AI models.No wonder big tech companies are warming up to n
SoftBank is uniting Japan's industrial elite to build the country's own AI foundation, trying to reduce dependence on American and Chinese models. The article Steel giants, automakers, and banks plan to build Japan's answer to US and Chinese AI dominance appeared first on The Decoder.
Surging demand for AI agents is colliding with limited compute capacity. Anthropic is struggling with outages, OpenAI announced the end of Sora, and GPU prices have jumped nearly 50 percent, according to market data. The article The AI industry is running out of compute, with outages, rationing, and
Oracle Corp. agreed to purchase as much as 2.8 gigawatts of fuel-cell power from Bloom Energy Corp. to supply data centers for artificial intelligence work.
Taiwanese stocks rose to a new record as investors returned to the pre-Iran war trading theme of chasing AI shares amid hopes for easing tensions in the Middle East.
Taiwanese stocks rose to a new record as investors returned to the pre-Iran war trading theme of chasing AI shares amid hopes for easing tensions in the Middle East.