New Data on LLM Accuracy

Juan Sequeda, Principal Scientist at data.world, recently published a research paper, “A Benchmark to Understand the Role of Knowledge Graphs on Large Language Model’s Accuracy for Question Answering on Enterprise SQL Databases.” He and his co-authors benchmarked LLM accuracy in answering questions over real business data.

Kinetica Launches Native Large Language Model for Language-to-SQL on Enterprise Data  

Kinetica, the speed layer for generative AI and real-time analytics, announced a native Large Language Model (LLM) combined with Kinetica’s innovative architecture that allows users to perform ad-hoc data analysis on real-time, structured data at speed using natural language. Unlike with public LLMs, no external API call is required and data never leaves the customer’s environment.

Kinetica Now Free Forever in Cloud Hosted Version; Accelerate the Transition to Generative AI with SQL-GPT

Kinetica, the database for time & space, announced a totally free version of Kinetica Cloud where anyone can sign-up instantly without a credit card to experience Kinetica’s generative AI capabilities to analyze real-time data. No other analytic database offers this pricing model with free storage and compute, and no expiration date.

PingCAP Empowers Open Source Community with New GitHub Data Explorer Tool

PingCAP, the provider of the advanced distributed SQL databases, announced the introduction of its new GitHub Data Explorer tool. This innovative new tool is built to help developers and open-source contributors achieve deeper insights into their GitHub activity, streamline workflows, and increase productivity.

Databricks Announces General Availability of Delta Live Tables

Databricks, the Data and AI company and pioneer of the data lakehouse paradigm, announced the general availability of Delta Live Tables (DLT), the first ETL framework to use a simple declarative approach to build reliable data pipelines and to automatically manage data infrastructure at scale. Turning SQL queries into production ETL pipelines often requires a lot of tedious, complicated operational work. By using modern software engineering practices to automate the most time consuming parts of data engineering, data engineers and analysts can concentrate on delivering data rather than on operating and maintaining pipelines.

HEAVY.AI Introduces New Capabilities that Enable Organizations to Gain Faster Access to Rapidly Growing Data Sets and Deliver Decision-Quality Information

HEAVY.AI (previously Omnisci), an innovator in advanced analytics, announced the limited release of HEAVY.AI 6.0, with general availability to follow in April. The new capabilities, such as HeavyConnect, deliver improved analytics performance and faster time to decision-quality information, allowing organizations to immediately extract value from their data. HEAVY.AI 6.0 empowers stakeholders to learn and make decisions based on data without requiring sophisticated data science skills or executing SQL commands.

Dremio Announces Open and Forever-Free Lakehouse Platform, Dremio Cloud, In Partnership with AWS

Dremio, the lakehouse company, announced the general availability of Dremio Cloud, the free data lakehouse platform and the addition of two new services: Dremio Sonar, a lakehouse engine built for SQL, and Dremio Arctic, a metadata and data management service for Apache Iceberg that provides a unique Git-like experience for the lakehouse. Dremio Sonar is now generally available, and Dremio Arctic is in public preview.

Accelerated Integration of Eventador with Cloudera – SQL Stream Builder

Within 5 months after the acquisition of Eventador, Cloudera announced that SQL Stream Builder is now being re-launched as Cloudera SQL Stream Builder. This is being done after it was fully integrated with Cloudera Data Platform’s (CDP) Shared Data Experience (SDX). This means that SQL Stream Builder can take advantage of the same unified security and governance as the rest of the platform does, using SDX.

Understanding NoSQL Databases: Document Stores

In this contributed article, Alex Williams, Writer/Researcher at Hosting Data UK, discusses document-oriented databases (also called: aggregate databases, document databases or document stores) and how they place each record, and its associative data, inside single documents. This database type is a subset of the NoSQL umbrella, which refers to the growing list of popular database management systems that use ‘non-relational’ models — i.e. databases that don’t rely on Structured Query Language, SQL.

Databricks Launches SQL Analytics to Enable Cloud Data Warehousing on Data Lakes

Databricks, the data and AI company, announced the launch of SQL Analytics, which for the first time enables data analysts to perform workloads previously meant only for a data warehouse on a data lake. This expands the traditional scope of the data lake from data science and machine learning to include all data workloads including Business Intelligence (BI) and SQL.