Sign up for our newsletter and get the latest big data news and analysis.

Open Source Innovations to Be Unveiled at Subsurface LIVE Winter 2021 Cloud Data Lake Conference

Dremio, the innovation leader in data lake transformation, announces the speaker lineup and full agenda for Subsurface LIVE Winter 2021, a two-day, live conference about the future of the cloud data lake industry. The virtual event takes place January 27-28, 2021, and features keynotes from senior executives of AWS, Tableau and Dremio, as well as 30+ technical sessions on the open source innovations, trends and strategies driving cloud data lake transformation and architectures.

Databricks Launches SQL Analytics to Enable Cloud Data Warehousing on Data Lakes

Databricks, the data and AI company, announced the launch of SQL Analytics, which for the first time enables data analysts to perform workloads previously meant only for a data warehouse on a data lake. This expands the traditional scope of the data lake from data science and machine learning to include all data workloads including Business Intelligence (BI) and SQL.

The Value of Data Now vs. Data Later

In this contributed article, Fluency CEO and Founder Chris Jordan discusses the inevitable extinction of Moore’s law. 90% of the world’s data has been produced over the last two years, yet companies only analyze 12% of it. With Big Data only continuing to grow, how can more innovative data storage solutions, such as the cloud, effectively respond to this level of growth?

Real-Time Analytics from Your Data Lake Teaching the Elephant to Dance

This whitepaper from Imply Data Inc. explains why delivering real-time analytics on a data lake is so hard, approaches companies have taken to accelerate their data lakes, and how they leveraged the same technology to create end-to-end real-time analytics architectures.

Real-Time Analytics from Your Data Lake Teaching the Elephant to Dance

This whitepaper from Imply Data Inc. introduces Apache Druid and explains why delivering real-time analytics on a data lake is so hard, approaches companies have taken to accelerate their data lakes, and how they leveraged the same technology to create end-to-end real-time analytics architectures.

Introducing Apache Druid

Sponsored Post Apache Druid was invented to address the lack of a data store optimized for real-time analytics. Druid combines the best of real-time streaming analytics and multidimensional OLAP with the scale-out storage and computing principles of Hadoop to deliver ad hoc, search and time-based analytics against live data with sub-second end-to-end response times. Today, […]

Introducing Apache Druid

This whitepaper provides an introduction to Apache Druid, including its evolution,
core architecture and features, and common use cases. Founded by the authors of the Apache Druid database, Imply provides a cloud-native solution that delivers real-time ingestion, interactive ad-hoc queries, and intuitive visualizations for many types of event-driven and streaming data flows.

Do You Actually Need a Data Lake?

In this contributed article, Eran Levy, Director of Marketing at Upsolver, sets out to formally define “data lake” and then goes on to ask whether your organization needs a data lake by examining 5 key indicators. Data lakes have become the cornerstone of many big data initiatives, just as they offer easier and more flexible options to scale when working with high volumes of data that’s being generated at a high velocity – such as web, sensor or app activity data.

aqfer Launches Next-Generation SI-Ready Marketing Data Lake

aqfer, a leading SaaS provider supporting a data-centric marketing architecture, announced the launch of its next generation marketing data lake platform. The solution enables systems integrators (SIs) or engineers building digital marketing platforms for managed service providers (MSPs), ad tech companies or marketing agencies to drastically reduce the time and cost associated with customizing solutions for marketers while simultaneously increasing data integration and management functionality and dramatically reducing support and operating costs.

Okera Introduces Attribute-Based Access Control for Data Lake Security and Access Management

Okera, a leading active data management company for data lake security and governance, announced the release of new attribute-based access control (ABAC) and automated business metadata tagging and policy enforcement capabilities. These new features help enterprises simplify how to manage, secure, and govern data access on data lakes at scale in an easy and automated manner.