SQLBits 2020
Machine Learning in Azure Databricks
In this session we focus on how Spark implements Machine Learning at Scale with Spark ML.
Databricks is the Swiss Army Knife of the Azure Data Analytics environment. Databricks implements Apache Spark, a robust in-memory cluster data processing framework. Spark Core supports SQL, DataFrames, Streaming, Graph Processing and Machine Learning.
In this session we focus on how Spark implements Machine Learning at Scale with Spark ML. Spark ML can do a lot, but not everything. In this session we go end-to-end, building a model in an iterative way, monitoring our improvements in MLFlow, training, testing and evaluating a Machine Learning pipeline. We will also discuss how to migrate your models to perform at scale. If you're working with Deep Learning, then let me know and we can discuss how Databricks can scale Deep Learning.
Speakers
Terry McCann's previous sessions
Docker & Kubernetes for the Data Scientist
Deployment == Return on investment. This session looks to show you how to do that for Machine Learning.
Machine Learning in Azure Synapse
There is a lot of content available on Synapse for Data Engineering, but what about Machine Learning? In this session we will look at how to integrate a SparkML model in Synapse.
Rapid Requirements: Introducing the Machine Learning Canvas
In this session, we will introduce the Advancing Analytics Machine Learning Canvas and how it can be used to capture requirements for Machine Learning Projects.
Machine Learning in Azure Databricks
In this session we focus on how Spark implements Machine Learning at Scale with Spark ML.
Deploy ML models faster with Data Science DevOps
In this session I will show you how to apply DevOps practices to speed up your development cycle and ensure that you have robust deployable models. We will focus on the Azure cloud platform in particular, however this is applicable to other platforms
Enhancing relational models with graph in SQL Server 2017
This session explores SQL Server 2017's Graph processing to better understand interconnectivity and behaviour in your data.