This presentation includes an overview of SQL Server 2012 & a deep dive into SQL Server 2014 columnstore indexes, including a review of the architecture, as well as the challenges, workarounds, travails, & big wins at customer sites
We have all heard about In-memory OLTP and how fast it can go, but just how do you use it? In this session we will look at how to migrate a system to use this new capability.
There are still environments which serve one single SQL Server database for both critical applications & reporting. Budgeting issues and infrastructure bottlenecks kill the application to be better managed and maintained. What could be the solution?
SQL server 2014 is one of their measure releases. Where Microsoft has done a superb job. There are so many enhancement we have in SQL Server 2014. In my session I will show you my top 5 features.
Come along to see us attempt to optimise and redesign the “Dell DVD Store” sample dot.net/SQL application from a modest 100 transactions per minute to a million transactions per minute.
Hekaton is the Greek word for 100 - the goal of In-Memory OLTP in SQL Server 2014 is to
improve query performance up to 100 times. In this session we will look under
the cover of Hekaton and the Multi Version Concurrency Control (MVCC)
In this session, we will share some of the hard lessons learned from the first large deployments in Analysis Services Tabular.
Still not sure of which is faster? Let us put this matter to rest once and for all. We will measure every single aspect of the execution of a stored procedure vs. its Ad-Hoc query counterpart, and only one will prevail!
This session will take a look at how parallel select into can be scaled to the nth degree in SQL Server such that all available hardware resources are utilised as fully as possible.
Dive deep into Columnstore Indexes and discover their advantages and some of the limitations.
Look into some of the Columnstore Indexes myths, understand more complex concepts of Batch Mode & Query Processing with another depth.
Join this session to make a deep dive into how SQL Server implements physical locking with lightweight synchronization objects like Latches and Spinlocks.
What is In-Memory OLTP? How is it architected? Where can I use in-memory technology in my solutions? Learn this and more from the Microsoft team
Learn how flash storage accelerates In-Memory OLTP performance. We will also cover limitations with SQL Server In-Memory OLTP.
In this session, you will learn the internals of SSIS and why having a deep understanding is important to solve performance problems, and how the control flow and data flow engine work.
In this session we'll look over some of the things which you should be looking at within your virtual environment to ensure that you are getting the performance out of it that you should be.
In this lunchtime session, Attunity will highlight real customer implementations of Attunity Replicate and the Microsoft BI technology stack that enabled them to achieve competitive advantage.
SQL Server 2012 brings a wide range of changes, including a key change in how SQL Server memory is managed. Earlier releases of SQL Server have two memory managers, which have now been merged into one, discover all these important changes
Do you have enough from "normal" performance tuning sessions? You can't hear anything about Indexing and Execution Plans anymore? Then you are right in this session!
In this session we'll understand the best practices of using this new technology. What exactly is it good for? When to use it and when NOT to?
Migration, Security, Administration and other tips and tricks.
Learn how to monitor Analysis Services with SQL Sentry Performance Advisor. Get tips on best practices, monitoring counters and options plus improve your understanding of how Analysis Services uses memory and where it differs from SQL Server.
Large, complex queries need memory in which to work--workspace memory--and understanding the how's, when's, and why's of this memory can help you create queries that run in seconds rather than minutes.
Sometimes some piece of T-SQL slips by, or falls out of memory.Come and revisit old favorites, and brush up on new T-SQL features and enhancements.This session is chock full of code examples, including before-and-after demos and how-to illustrations.
Virtualisation changes the way you need to monitor the performance of a virtualised instance of SQL Server. In this session I will demonstrate a balanced and well-rounded approach to performance monitoring in the virtual world along with best practices to avoid poor virtualised performance.
Fast Track is a new reference data warehousing architecture provided by Microsoft. More than this it represents a new way of thinking about data warehousing. A Fast Track system is measured by its raw compute power - not by a DBAs ability to tune an index. Fast Track is an appliance-like solution that delivers phenomenal performance from a pre-defined, balanced configuration of CPU, memory and storage using nothing but commodity hardware.
Of particular interest in a Fast Track system is the way in which the storage and SQL Server are configured. To achieve the fantastic throughput without using SSDs requires some careful configuration. This configuration is designed to make use of Sequential I/O to dramatically improve disk I/O performance.
Interested? If you have a large data warehouse that's seen better days or perhaps you are about to embark on a new warehousing project then you should be! Fast Track is a great solution with a fantastic value proposition.
In this one hour session we'll aim to get under the skin of Fast Track and get some answers as to how it delivers such great throughput on commodity hardware. In the process we'll aim to answer the following questions:
* When might I need Fast Track?
* What is Sequential I/O?
* How does Sequential I/O improve performance?
* What do I need to do to get Sequential I/O?
* How can I monitor for Sequential I/O ?
* What may I need to change in my ETL to get the benefit of sequential I/O?
Still reading? I'll save you a front row seat....
This session will reveal the basics of analyzing crash-dumps in order to troubleshoot SQL Server crashes or abnormal situations. We will also review the postmortem analysis of common (but tricky) development problems, such as connection leaks, using WinDbg and Adplus.
This session covers the new scalability features of SQL Server Reporting Services 2008 and demonstrates how to build a high performance reporting platform to deal with your enterprise reporting needs.
Henk van der Valk from the Unisys performance lab will be my special guest for this session. He has a test system that we all would like to have at home. The ES7000 has 96 Cores and half a Terabyte of RAM yes RAM (!). We will look at various ways to push SQL Server workloads by methodically detecting and resolving bottlenecks.Together we will show you how you can apply this approach on mid-sized and even smaller systems. (We’ll even try some of them on a laptop, and of course we will remote connect into the big machine ) In this session you will: Learn how to measure what the next bottle-neck is,using perfmon and waitstats. Hear about optimization tips that will get the most out of your hardware to speed up your processsing.
Learn the terminology, architecture, features and best practices that can help you understand, troubleshoot and avoid memory problems in SQL Server 2008.