Azure Data Factory is a managed service for extracting, transforming and processing data, with ability to define pipelines and data driven workflows. in this session you will learn differences between this tools and SSIS with demos
This all demo session build around a real life case study from one of the largest organizations in the world. will take you through a case of how to build a completely meta-driven SSIS solution relaying on BIML and MDS to do the hard-work for you.
See how to use Power Query to build a detail-level reporting solution
In this session, we show how SSIS and HDInsight can work together, as well as how Hadoop jobs can be executed and how data can be transferred with SSIS between HDInsight, SQL Server, and other sources or destinations.
As data warehouse developers we deliver data to business users... but we don't deliver data as fast as we want. There are many things you can do to increase your own productivity and become a more efficient data warehouse developer.
Join me for an hour of playing with different ETL patterns by using Clustered Columnstore Indexes. Using different Hardware might lead you to different conclusions,and the size of the workload is always the paramount of your performance.Loading data first and then creating a Clustered Columnstore or creating Clustered Columnstore and than loading - join me to find the answers!
In this session, you will learn what SSIS is, what components it consists of, and how to use the SSIS catalog to track the execution of packages and how to troubleshoot packages.
In this session we’ll look at a number of different data scenarios and explore ways of remodelling the data to optimise it for cubes and MDX. Sometimes a small ETL change can have a dramatic impact on the cube's functionality and simplicity.
I know what you're thinking, Powershell is not an ETL tool. And you're probably right. But I keep running into weird requests that were just easier to fix with Powershell. I'll show you why some things are not easy in SSIS.
Big Data is hot! And the magic word is Hadoop. But what is it? And more important: what can I do with it? In this session we will cover some basics of Hadoop, MapReduce, Pig and Hive.
In this session we will learn about the SSISDB catalog views. We will use this info plus our own info to build a monitoring solution for SSIS using SSRS.
This talk will explain the patterns that we recommend to customers when designing data warehouses. This talk will help people learning about DW for the first time and also give insight for those who wish to learn more.
More data, less time, storing everything. These ingredients mean it is becoming imperative we do not move the kitchen sink everytime we do our ETL Come to this session to find out how.
In this first of two sessions, we review the architecture of SQL Server and its BI components and deployment options for optimal performance. We'll also discuss how to optimize data warehouse load operations.
Learn all about the integration of Business Intelligence Markup Language (BIML) into BIDS Helper and see what it can do for you. BIML can help you automate the creation of SSIS packages, reduce maintenance and help code re-use across projects.
The SQL Server provides much functionality that increases the scalability and flexibility of your solution by distributing data and jobs among low-priced commodity servers.
Attunity's Ian Archibald will discuss how CDC solutions can solve problems in BI and ETL/DW projects, co-hosted with two of the industry’s highly-respected SQL Server gurus BI expert Jennifer Stirrup and ETL expert and SQL Server MVP Allan Mitchell
A new and fresh approach to datwarehousing that is agile, performant and easy to do. Data Vault is a new methodology that tries to overcome traditional problems that plagued traditional datawarehousing in the past.
Attunity will be presenting their award winning change data capture (CDC) products for SSIS. This session will discuss whatchange data capture is, why it's used, and will demonstrate how CDC processingcan be seamlessly integrated into SSIS.
Fast Track is a new reference data warehousing architecture provided by Microsoft. More than this it represents a new way of thinking about data warehousing. A Fast Track system is measured by its raw compute power - not by a DBAs ability to tune an index. Fast Track is an appliance-like solution that delivers phenomenal performance from a pre-defined, balanced configuration of CPU, memory and storage using nothing but commodity hardware.
Of particular interest in a Fast Track system is the way in which the storage and SQL Server are configured. To achieve the fantastic throughput without using SSDs requires some careful configuration. This configuration is designed to make use of Sequential I/O to dramatically improve disk I/O performance.
Interested? If you have a large data warehouse that's seen better days or perhaps you are about to embark on a new warehousing project then you should be! Fast Track is a great solution with a fantastic value proposition.
In this one hour session we'll aim to get under the skin of Fast Track and get some answers as to how it delivers such great throughput on commodity hardware. In the process we'll aim to answer the following questions:
* When might I need Fast Track?
* What is Sequential I/O?
* How does Sequential I/O improve performance?
* What do I need to do to get Sequential I/O?
* How can I monitor for Sequential I/O ?
* What may I need to change in my ETL to get the benefit of sequential I/O?
Still reading? I'll save you a front row seat....
Henk van der Valk from the Unisys performance lab will be my special guest for this session. He has a test system that we all would like to have at home. The ES7000 has 96 Cores and half a Terabyte of RAM yes RAM (!). We will look at various ways to push SQL Server workloads by methodically detecting and resolving bottlenecks.Together we will show you how you can apply this approach on mid-sized and even smaller systems. (We’ll even try some of them on a laptop, and of course we will remote connect into the big machine ) In this session you will: Learn how to measure what the next bottle-neck is,using perfmon and waitstats. Hear about optimization tips that will get the most out of your hardware to speed up your processsing.
A look at some of the things that make using SSIS in SQL Server 2008 a very good choice for your ETL requirements
Come to this session to see how you can create a logging, monitoring and profiling solution for your existing or new SSIS packages overcoming all usual problems that a typical solution brings: increasing package complexity, longer development times and so on. The session will propose a standard, out-of-the-box solution for all of these challenges.