Sign up for our newsletter and get the latest big data news and analysis.

MLCommons™ Releases MLPerf™ Inference v1.1 Results

Today, MLCommons, an open engineering consortium, released new results for MLPerf Inference v1.1, the organization’s machine learning inference performance benchmark suite. MLPerf Inference measures the performance of applying a trained machine learning model to new data for a wide variety of applications and form factors, and optionally includes system power measurement.

DataOps Dilemma: Survey Reveals Gap in the Data Supply Chain

The survey associated with this report, commission by Immuta, focused on identifying the limiting factors in the data “supply chain” as it relates to the overall DataOps methodology of the organization. DataOps itself is the more agile and automated application of data management techniques to advance data-driven outcomes, while the data supply chain represents the technological steps and human-involved processes supporting the flow of data through the organization, from its source, through transformation and integration, all the way to the point of consumption or analysis.

Best of arXiv.org for AI, Machine Learning, and Deep Learning – August 2021

In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning – from disciplines including statistics, mathematics and computer science – and provide you with a useful “best of” list for the month.

eBook: A Practical Guide to Using Third-Party Data in the Cloud

[Sponsored Post] To help you navigate a proliferating data landscape, AWS Data Exchange would like to present you with a copy of the new eBook, “A Practical Guide to Using Third-Party Data in the Cloud.” Learn how innovative teams are shifting their focus from data-driven business intelligence to accelerating insight-driven decision-making and now are turning to third-party datasets as a differentiator.

Infographic: The Fastest Supercomputers Ever Built and Who Built Them

Supercomputers are an important part of computational science, including weather forecasting, climate research, quantum mechanics, molecular modeling, and cryptanalysis, as they are able to process information more quickly than a traditional computer. The United States has historically been a leader in supercomputer development, but other countries’ technology and research have been catching up. The HP research team consulted data from supercomputer ranking project TOP500 to visualize the most powerful supercomputers in the world as well as their country and company of development.

T-Shaped Teams: A New Roadmap for AI and Big Data Adoption

CFA Institute, the global association of investment professionals, released new industry research that identifies a new organizational approach for enabling financial institutions to develop and successfully execute artificial intelligence and big data strategies.

Infographic: How to Leverage Big Data Tech in Your Construction Company

The big data market is expected to reach $99.31 billion this year, and companies that take advantage of big data analytics have reported increases in revenue, productivity and efficiency. The visual below from our friends over at BigRentz shows how construction companies can leverage big data and take advantage of new analytics technologies—from 3D modeling to material tracking and on-site safety sensors. 

Data Warehouse Costs Soar, ROI Still Not Realized

Enterprises are pouring money into data management software – to the tune of $73 billion in 2020 – but are seeing very little return on their data investments.  According to a new study out from Dremio, the SQL Lakehouse company, and produced by Wakefield Research, only 22% of the data leaders surveyed have fully realized ROI in the past two years, with most data leaders (56%) having no consistent way of measuring it. 

The Future Is Now: Why Data Is Key to Tech Research & Development

In this contributed article, IT and digital marketing specialist Natasha Lane, highlights the reasons why using data is so crucial for research and development. Using data can be the key to recognizing and solving humanity’s leading challenges in the years to come. We’re talking about everything from water shortage, climate change, the need to develop safe self-driving cars, and so on.

Yandex Finds Better Way to Train ML Models Over the Internet

A new proposal from tech giant Yandex overcomes a major hurdle in the advancement of machine learning by bringing the process to the masses, so that anyone with a home computer can help train a large neural network.