Azure offers a wide range of services that can be combined into a BI solution in the cloud.
This session focuses on the deeper integration of SQL Server Integration Services (SSIS) in Azure Data Factory (ADF) and the broad extensibility of Azure-SSIS Integration Runtime (IR).
In this session we'll look at ETL metadata, use it to drive process execution, and see benefits quickly emerge. I'll show how a metadata-first approach reduces complexity, enhances resilience and allows ETL processing to become self-organising.
The video is not available to view online.
We’ll take a look at how to approach making an Azure Databricks based ETL solution from start to finish. Along the way it will become clear how Azure Databricks works and we will use our SSIS knowledge to see if it can handle common use-cases
The video is not available to view online.
SQL Server 2019 expands on the Polybase feature from SQL Server 2019 by providing a robust data virtualization solution to reduce the need for ETL and data movement. Come learn how new data connectors work with sources like Oracle, MongoDB, CosmosDB, Terradata, and HDFS.
Getting better performance from your ETL
The video is not available to view online.
What data profiling is & why you should do it.
To build analytical models, we need to start by extracting, transforming, cleaning, preparing and loading the data. This session analyzes a set of scenarios that may happen during the ETL step using the Power Query in Power BI.
This session is about using the Business Intelligence Markup Language (Biml) to monitor and control your orchestration patterns. By automatically analyzing the results in ETL logs, we’ll be able to automate our staging orchestration!
How to build a Live Data Warehouse using Change Tracking capabilities
We will showcase the latest feature of SSIS 2017 such as connectors for Azure Data Lake Store (ADLS), Azure SQL Data Warehouse (SQL DW), SSIS Scale-Out at package level for the box product as well as the SSIS package execution on Azure Data Factory
Do you have large-scale SSIS platforms which experience bottlenecks due to the number of packages or volumes of data that need to be processed? Yes! Then you need to explore SSIS scale out.
A session to understand the available functionality of SSISDB Catalog
Using SQL Server Service Broker you can create near real-time data in your data warehouse for more timely analysis and reporting.
ETL Development takes time. Allen will show how PowerShell and BIML combine to make the development process seamless, consistent and reduces costly development time dramatically.
The video is not available to view online.
Azure Data Factory and SSIS are both data movement tools, but built for different purposes. In this session you will learn pros and cons of using each technology, and best practices of using each in real world scenarios.
The video is not available to view online.
Do you get involved in the creation of Business Intelligence solutions within your organisation? This session will show you the new features of SQL Server 2016 Business Intelligence components.
The video is not available to view online.
This all demo session build around a real life case study from one of the largest organizations in the world. will take you through a case of how to build a completely meta-driven SSIS solution relaying on BIML and MDS to do the hard-work for you.
See how to use Power Query to build a detail-level reporting solution
Need to automate a process in your database? Create audit tables, generate stored procedures, or any other type of manually repeatable process? In this session we'll look at how to start and the thought processes and design considerations involved.
In this session, we show how SSIS and HDInsight can work together, as well as how Hadoop jobs can be executed and how data can be transferred with SSIS between HDInsight, SQL Server, and other sources or destinations.
As data warehouse developers we deliver data to business users... but we don't deliver data as fast as we want. There are many things you can do to increase your own productivity and become a more efficient data warehouse developer.
Join me for an hour of playing with different ETL patterns by using Clustered Columnstore Indexes. Using different Hardware might lead you to different conclusions,and the size of the workload is always the paramount of your performance.Loading data first and then creating a Clustered Columnstore or creating Clustered Columnstore and than loading - join me to find the answers!
In this session we’ll look at a number of different data scenarios and explore ways of remodelling the data to optimise it for cubes and MDX. Sometimes a small ETL change can have a dramatic impact on the cube's functionality and simplicity.
I know what you're thinking, Powershell is not an ETL tool. And you're probably right. But I keep running into weird requests that were just easier to fix with Powershell. I'll show you why some things are not easy in SSIS.
Big Data is hot! And the magic word is Hadoop. But what is it? And more important: what can I do with it? In this session we will cover some basics of Hadoop, MapReduce, Pig and Hive.
In this session we will learn about the SSISDB catalog views. We will use this info plus our own info to build a monitoring solution for SSIS using SSRS.
This talk will explain the patterns that we recommend to customers when designing data warehouses. This talk will help people learning about DW for the first time and also give insight for those who wish to learn more.
More data, less time, storing everything. These ingredients mean it is becoming imperative we do not move the kitchen sink everytime we do our ETL Come to this session to find out how.
In this first of two sessions, we review the architecture of SQL Server and its BI components and deployment options for optimal performance. We'll also discuss how to optimize data warehouse load operations.
Learn all about the integration of Business Intelligence Markup Language (BIML) into BIDS Helper and see what it can do for you. BIML can help you automate the creation of SSIS packages, reduce maintenance and help code re-use across projects.
The SQL Server provides much functionality that increases the scalability and flexibility of your solution by distributing data and jobs among low-priced commodity servers.
Attunity's Ian Archibald will discuss how CDC solutions can solve problems in BI and ETL/DW projects, co-hosted with two of the industry’s highly-respected SQL Server gurus BI expert Jennifer Stirrup and ETL expert and SQL Server MVP Allan Mitchell
A new and fresh approach to datwarehousing that is agile, performant and easy to do. Data Vault is a new methodology that tries to overcome traditional problems that plagued traditional datawarehousing in the past.
Attunity will be presenting their award winning change data capture (CDC) products for SSIS. This session will discuss whatchange data capture is, why it's used, and will demonstrate how CDC processingcan be seamlessly integrated into SSIS.
Fast Track is a new reference data warehousing architecture provided by Microsoft. More than this it represents a new way of thinking about data warehousing. A Fast Track system is measured by its raw compute power - not by a DBAs ability to tune an index. Fast Track is an appliance-like solution that delivers phenomenal performance from a pre-defined, balanced configuration of CPU, memory and storage using nothing but commodity hardware. Of particular interest in a Fast Track system is the way in which the storage and SQL Server are configured. To achieve the fantastic throughput without using SSDs requires some careful configuration. This configuration is designed to make use of Sequential I/O to dramatically improve disk I/O performance. Interested? If you have a large data warehouse that's seen better days or perhaps you are about to embark on a new warehousing project then you should be! Fast Track is a great solution with a fantastic value proposition. In this one hour session we'll aim to get under the skin of Fast Track and get some answers as to how it delivers such great throughput on commodity hardware. In the process we'll aim to answer the following questions: * When might I need Fast Track? * What is Sequential I/O? * How does Sequential I/O improve performance? * What do I need to do to get Sequential I/O? * How can I monitor for Sequential I/O ? * What may I need to change in my ETL to get the benefit of sequential I/O? Still reading? I'll save you a front row seat....
Henk van der Valk from the Unisys performance lab will be my special guest for this session. He has a test system that we all would like to have at home. The ES7000 has 96 Cores and half a Terabyte of RAM yes RAM (!). We will look at various ways to push SQL Server workloads by methodically detecting and resolving bottlenecks.Together we will show you how you can apply this approach on mid-sized and even smaller systems. (We’ll even try some of them on a laptop, and of course we will remote connect into the big machine ) In this session you will: Learn how to measure what the next bottle-neck is,using perfmon and waitstats. Hear about optimization tips that will get the most out of your hardware to speed up your processsing.
A look at some of the things that make using SSIS in SQL Server 2008 a very good choice for your ETL requirements
Come to this session to see how you can create a logging, monitoring and profiling solution for your existing or new SSIS packages overcoming all usual problems that a typical solution brings: increasing package complexity, longer development times and so on. The session will propose a standard, out-of-the-box solution for all of these challenges.
Making more of SSIS in 2008
The video is not available to view online.
<<1>>