In this contributed article, Jonathan Friedmann, CEO & Co-Founder of Speedata, discusses a paradigm shift that is already underway. CPU’s leading role in datacenters is no longer a given, and as database workloads grow increasingly varied, market forces will point more and more in a clear direction: the development of new, dedicated, accelerated solutions. Data workloads are only going to grow more diverse and intricate from here. The same holds true for the chips that will process them.
AWS Announces General Availability of Amazon EC2 DL1 Instances
Amazon Web Services, Inc. (AWS), an Amazon.com, Inc. company (NASDAQ: AMZN), announced general availability of Amazon Elastic Compute Cloud (Amazon EC2) DL1 instances, a new instance type designed for training machine learning models. DL1 instances are powered by Gaudi accelerators from Habana Labs (an Intel company) to provide up to 40% better price performance for training machine learning models than the latest GPU-powered Amazon EC2 instances.
Building the GPU-accelerated Datacenter
[Sponsored Post] To learn about GPU-accelerated datacenters, join the upcoming live webinar from Thinkmate and PNY. You’ll hear about the future of the datacenter, why the GPU is crucial, the technology behind GPU acceleration, and what sort of options exist for different industries or types of organizations.
Yandex Finds Better Way to Train ML Models Over the Internet
A new proposal from tech giant Yandex overcomes a major hurdle in the advancement of machine learning by bringing the process to the masses, so that anyone with a home computer can help train a large neural network.
TensorRT 8 Provides Leading Enterprises Fast AI Inference Performance
NVIDIA today launched TensorRT™ 8, the eighth generation of the company’s AI software, which slashes inference time in half for language queries — enabling developers to build the world’s best-performing search engines, ad recommendations and chatbots and offer them from the cloud to the edge.
Walmart Innovates at the Retail Edge with AI
Through the combination of world-class software from Malong Technologies and a powerful processing platform Dell Technologies and NVIDIA, Walmart is able to deliver groundbreaking AI acceleration at the store edge.
Comparative Testing of GPU Servers with New NVIDIA RTX30 Video Cards in AI/ML Tasks
In early September 2020, NVIDIA debuted its second generation GeForce RTX 30 family of graphics cards, the Ampere RTX architecture. NVIDIA broke with tradition when its new generations of cards were sold more expensive than their predecessors, which means that the cost of training models has remained more or less the same.
Penguin Computing Announces OriginAI Powered by WekaIO
Penguin Computing, a division of SMART Global Holdings, Inc. (NASDAQ: SGH) and a leader in high-performance computing (HPC), artificial intelligence (AI), and enterprise data center solutions, announced that it has partnered with WekaIO™ (Weka) to provide NVIDIA GPU-Powered OriginAI, a comprehensive, end-to-end solution for data center AI that maximizes the performance and utility of high-value AI systems.
NVIDIA DGX Station A100 Offers Researchers AI Data Center-in-a-Box
NVIDIA today announced the NVIDIA DGX Station™ A100 — the world’s only petascale workgroup server. The second generation of the groundbreaking AI system, DGX Station A100 accelerates demanding machine learning and data science workloads for teams working in corporate offices, research facilities, labs or home offices everywhere.
NetApp AI and Run:AI Partner to Speed Up Data Science Initiatives
NetApp, a leading cloud data services provider has teamed up with Run:AI, a company virtualizing AI infrastructure, have teamed up to allow faster AI experimentation with full GPU utilization. The partnership allows teams to speed up AI by running many experiments in parallel, with fast access to data, utilizing limitless compute resources. Run:AI enables full GPU utilization by automating resource allocation, and NetApp® ONTAP® AI proven architecture allows every experiment to run at maximum speed by eliminating data pipeline bottlenecks.