Covering Scientific & Technical AI | Saturday, November 30, 2024

inferencing

Nvidia Releasing Open-Source Optimized Tensor RT-LLM Runtime with Commercial Foundational AI Models to Follow Later This Year

Nvidia's large-language models will become generally available later this year, the company confirmed. Organizations widely rely on Nvidia's graphics processors to write AI applications. The company has also created ...Full Article

MLCommons: MLPerf Inference Delivers Power Efficiency and Performance Gains

April 5, 2023 — MLCommons, the leading open AI engineering consortium, announced today new results from the industry-standard MLPerf Inference v3.0 and Mobile v3.0 benchmark suites, which measure the performance ...Full Article

Nvidia Dominates MLPerf Inference, Qualcomm also Shines, Where’s Everybody Else?

MLCommons today released its latest MLPerf inferencing results, with another strong showing by Nvidia accelerators inside a diverse array of systems. Roughly four years old, MLPerf still struggles to ...Full Article

AWS Boosting Performance with New Graviton3-Based Instances Now Available in Preview

Three years after unveiling the first generation of its AWS Graviton chip-powered instances in 2018, Amazon Web Services just announced that the third generation of the processors – the ...Full Article

Intel’s 30x AI Performance Aim for Xeon Sapphire Rapids CPUs May Not Solve All AI Needs: Analysts

With its upcoming Intel Sapphire Rapids CPUs, designed as the next generation of Intel Xeon CPUs after Ice Lake and slated for release in 2022, chipmaker Intel Corp. is ...Full Article

Nvidia GPUs Stay in Lead In Latest MLPerf Inference Results, but CPUs and Intel Gaining Ground

Nvidia again dominated the latest round of MLPerf inference benchmark (v 1.1) results when they were unveiled Sept. 23 (Thursday), sweeping the top spots in the closed data center ...Full Article

With $70M in New Series C Funding, Mythic AI Plans Mass Production of Its Inferencing Chips

Six months after unveiling its first M1108 Analog Matrix Processor (AMP) for AI inferencing, Mythic AI has just received a new $70 million Series C investment round to bring ...Full Article

Intel CPUs Gaining Optimized Deep Learning Inferencing from Deci in New Collaboration

Intel Corp. and deep learning startup Deci are partnering to help enterprises dramatically optimize inferencing and make their deep learning models more efficient and faster using Intel CPUs. The ...Full Article
AIwire