Sign up for our newsletter and get the latest big data news and analysis.

Video Highlights: Measuring Data Team ROI

In this video presentation from our friends over at Monte Carlo we have an all-woman panel to discuss the process of measuring data team ROI, specifically how to track and scale the impact of data teams.

Video Highlights: GPT-4 Developer Livestream

Here is Greg Brockman, President and Co-Founder of OpenAI, for a March 14, 2023 developer demo showcasing GPT-4 and some of its capabilities/limitations. Included are a number of very compelling new use case capabilities over the previous GPT-3.5 version.

Video Highlights: Copilot for R

Our video highlights selection for today is by data science industry luminary David Smith who made a presentation to the NYC Data Hackers on the topic of Copilot for R. If you haven’t come across Copilot before, it’s like an AI-based pair programmer that suggests new lines of code, and perhaps entire functions, based on context.

Video Highlights: Distributed Python with Ray

This is an introductory and hands-on guided tutorial of Ray2.0 that covers an introductory, hands-on coding tour through the core features of Ray, which provides powerful yet easy-to-use design patterns for implementing distributed systems in Python.

Data Science 101: The Data Science Process

Welcome to insideBIGDATA’s Data Science 101 channel brining you perspectives for the topics of the day in data science, machine learning, AI and deep learning. Many of the video presentations come from my lectures for my Introduction to Data Science class I teach at UCLA Extension. In today’s slide-based video presentation I discuss The Data Science Process, an overview of the steps that data scientists use solving problems with data science and machine learning technologies.

Video Highlights: Attention Is All You Need – Paper Explained

In this video presentation, Mohammad Namvarpour presents a comprehensive study on Ashish Vaswani and his coauthors’ renowned paper, “Attention Is All You Need.” This paper is a major turning point in deep learning research. The transformer architecture, which was introduced in this paper, is now used in a variety of state-of-the-art models in natural language processing and beyond. Transformers are the basis of the large language models (LLMs) we’re seeing today.

AI Under the Hood: Interactions

We asked our friends over at Interactions to do a deep dive into their technology. Mahnoosh Mehrabani, Ph.D., Interactions’ Sr. Principal Scientist shared some fascinating information about how Interactions’ Intelligent Virtual Assistants (IVAs) leverage advanced natural language understanding (NLU) models for “speech recognition” and “advanced machine learning.” The company uses NLU models to help some of today’s largest brands to understand customer speech and respond appropriately.

Video Highlights: Change Data Capture With Apache Flink

The featured video resource provided by Decodable is a webinar in which CDC experts provide an overview of CDC with Flink and Debezium. There is a growing role Change Data Capture (CDC) plays in real-time data analytics (specifically, stream processing with open source tools like Debezium and Apache Flink). CDC lets users analyze data as it’s generated by leveraging streaming from systems like Apache Kafka, Amazon Kinesis, and Azure Events Hubs to track and transport changes from one data system to another.

Video Highlights: Google Engineer on His Sentient AI Claim

Google Engineer Blake Lemoine (who worked for the company’s Responsible AI unit) joins Emily Chang of Bloomberg Technology in the video below to talk about some of the experiments he conducted that led him to believe that LaMDA (a Large Language Model) was a sentient AI, and to explain why he was placed on administrative leave and ultimately fired.

Video Highlights: Modernize your IBM Mainframe & Netezza With Databricks Lakehouse

In the video presentation below, learn from experts how to architect modern data pipelines to consolidate data from multiple IBM data sources into Databricks Lakehouse, using the state-of-the-art replication technique—Change Data Capture (CDC).