Shared the stage with Kevin Kline. Paul Randal and Kimberly L. Tripp organized an excellent conference. This slide deck talks about how to design large MS SQL Server architectures with 1000s of databases that are high performance and yet easy to manage. ioMemory by Fusion-io provides performance and SQL Sentry provides an amazing interface to manage and monitor 1000s of databases.
31. Don’t forget to enter your evaluation
of this session using EventBoard!
Questions?
Thank you!
Notes de l'éditeur
Interestingly, the default configuration of the server is generally quite good. Even at very high scale there is not much additional work that can be done. The closest you get to a magic: make SQL Server go faster traceflag is 834 (http://support.microsoft.com/kb/920093 and http://msdn2.microsoft.com/en-us/library/aa366720.aspx) for Windows large-page allocations for the buffer pool.If you see a flat node, it will fill up eventually if you start doing enough work in SQL.
On heavily OLTP systems, there’s enough NIC traffic that they need their own CPU cores to process the TCP work. Use affinity mask to segregate the NIC cores.Increasing connections to ~6000 (users had think time), you’ll started seeing waits on THREADPOOLSolution: increase sp_configure ‘max worker threads’Probably don’t want to go higher than 4096Gradually increase it, default max is 980Avoid killing yourself in thread management – bottleneck is likely somewhere else
(Should be around 4:30 pm)PCI-e v1 busX4 slot: 750M/secX8 slot: 1.5GB/secX16 – fast enough, around the 3GB/secSome “v2 compliant” PCI-e bus still run at v1 speeds!
Interesting Shape, what’s causing it?
The hardware between the CPU and the physical drive is often complexDifferent topologies, depending on vendor and technologyTwo major topologies for SQL Server StorageDAS – Direct Attached StorageStandards: (SCSI), SAS, SATARAID controller in the machinePCI-X or PCI-E direct accessSAN – Storage Area Networks Standards: iSCSI or Fibre Channel (FC)Host Bus Adapters or Network Cards in the machineSwitches / Fabric access to the diskSAN & Tiered Storage ArraysSANData is explicitly placed on various disk groups which the admin must track.Moving data between tiers is manual and typically offlineGranularity is whatever the admin decides to move.Depends on the admin tracking storage hot spots and usageTiered SANArray tracks usage patterns and automatically moves data between storage tiers.Data movement is in the background and fully onlineGranularity is LUN today, moving to finer-grainedDepends on the array, tracking storage hot spots and usage
(Should be around 4:45-4:50 pm)
(Should be around 5:00 – 5:10 pm)
First image = basic AG for high availability and disaster recoverySecond image = Node and File Share Majority quorum modelThird image = Node Majority quorum model
These are the default tools.
Open up the Always On view for Instance/Group Matrix. Illustrate the thick green lines from SQL2 to SQL4.Start the job on SQL2, “Keynote workload”. The lines will begin to turn red. Discuss how the IO load on SQL2 (Fusion-IO) is backing up on SQL4 (slow disks).Start the job on SQL1, “Keynote move C/F to SQL1”. Takes a couple minutes. Nodes shut down then flip over to SQL1. The restore to create the database file is super fast. Super-low latency because of Fusion-IO. May have time to run “Keynote workload” again on the SQL1 after the flip.Start the job on SQL1, “Revert”. Takes even more time. Main point here is how long it takes because we have to restore and reinitialize database on the slow disks. Then it may not finish in time because the initialization and restore take so much longer.