Intel Works with University of Pennsylvania in Using Privacy-Preserving AI to Identify Brain Tumors

Intel Labs and the Perelman School of Medicine at the University of Pennsylvania (Penn Medicine) are co-developing technology to enable a federation of 29 international healthcare and research institutions led by Penn Medicine to train artificial intelligence (AI) models that identify brain tumors using a privacy-preserving technique called federated learning.

Intel® Parallel Studio XE 2020: Transform Enterprise, Cloud, HPC & Artificial Intelligence with Faster Parallel Code

In this article we’ll drill down into the capabilities of Intel® Parallel Studio XE 2020, the latest release of a comprehensive, parallel programming tool suite that simplifies the creation and modernization of code. Using this newest release, software developers and architects can speed AI inferencing with support for Intel® Deep Learning Boost and Vector Neural Network Instructions (VNNI), designed to accelerate inner convolutional neural network (CNN) loops.

oneAPI: – A Unified Cross-Architecture, High Performance Programming Model Designed to Help Shape the Future of Application Development

In this article, we’ll dive into the newly announced oneAPI, a single, unified programming model that aims to simplify development across multiple architectures, such as CPUs, GPUs, FPGAs and other accelerators. The long-term journey is represented by two important first-steps – the industry initiative and the Intel beta product.

Heterogeneous Computing Programming: oneAPI and Data Parallel C++

Sponsored Post What you missed at the Intel Developer Conference, and how to catch-up today By James Reinders In the interests of full disclosure … I must admit that I became sold on DPC++ after Intel approached me (as a consultant – 3 years retired from Intel) asking if I’d help with a book on […]

2nd Generation Intel® Xeon® Platinum 9200 Processors Offer Leadership Performance, and Advance AI

Simulation, modeling, data analytics, and other workloads commonly use high performance computing (HPC) to advance research and business in many ways. However, as converged workloads involving AI grow in adoption, HPC systems must keep pace with evolving needs. 2nd Generation Intel® Xeon® Platinum processors, with built-in AI acceleration technologies, offer leadership performance to speed the most demanding HPC workloads.

How Astera Labs is Revolutionizing Semiconductor Product Development—100% in the Cloud

For any established semiconductor product developer, designing a next-generation PCIe 5.0 chipset in less than a year is no small feat. For a brand-new startup with no compute infrastructure other than laptops, however, it is a huge ask. That’s why, with time being of the essence, Astera Labs decided to take a chance on the efficiencies it would gain from a 100% cloud-based approach.

Six Platform Investments from Intel to Facilitate Running AI and HPC Workloads Together on Existing Infrastructure

Because HPC technologies today offer substantially more power and speed than their legacy predecessors, enterprises and research institutions benefit from combining AI and HPC workloads on a single system. Six platform investments from Intel will help reduce obstacles and make HPC and AI deployment even more accessible and practical.

DAOS Delivers Exascale Performance Using HPC Storage So Fast It Requires New Units of Measurement

Forget what you previously knew about high-performance storage and file systems. New I/O models for HPC such as Distributed Asynchronous Object Storage (DAOS) have been architected from the ground up to make use of new NVM technologies such as Intel® Optane™ DC Persistent Memory Modules (Intel Optane DCPMMs). With latencies measured in nanoseconds and bandwidth measured in tens of GB/s, new storage devices such as Intel DCPMMs redefine the measures used to describe high-performance nonvolatile storage.

Interview: Terry Deem and David Liu at Intel

I recently caught up with Terry Deem, Product Marketing Manager for Data Science, Machine Learning and Intel® Distribution for Python, and David Liu, Software Technical Consultant Engineer for the Intel® Distribution for Python*, both from Intel, to discuss the Intel® Distribution for Python (IDP): targeted classes of developers, use with commonly used Python packages for data science, benchmark comparisons, the solution’s use in scientific computing, and a look to the future with respect to IPD.

DarwinAI Generative Synthesis Platform and Intel Optimizations for TensorFlow Accelerate Neural Networks

DarwinAI, a Waterloo, Canada startup creating next-generation technologies for Artificial Intelligence development, announced that the company’s Generative Synthesis platform – when used with Intel technology and optimizations – generated neural networks with a 16.3X improvement in image classification inference performance. Intel shared the optimization results in a recently published solution brief.