Skip to main content

Column-Store index in SQL Server

Column-Store Indexes
Column-Store index is designed for improving query performance which is used for heavy amount of data like data warehouse fact tables. This type of index stores the index data in column based rather than row based. This index is introduced in version SQL server 2012; basically this is column based non-clustered index.

Why should we use Column-Store Index
Column-Store index provides a very high level of compression, typically by 10 times, to significantly reduce your data warehouse storage cost. For analytics, a column-store index offers an order of magnitude better performance than a B-Tree index.
Columns store values from the same domain and commonly have similar values, which result in high compression rates. I/O bottlenecks in your system are minimized or eliminated, and memory footprint is reduced significantly.
High compression rates improve query performance by using a smaller in-memory footprint. In turn, query performance can improve because SQL Server can perform more query and data operations in memory.
Batch execution improves query performance, typically by two to four times, by processing multiple rows together.
Queries often select only a few columns from a table, which reduces total I/O from the physical media.

1-    Use a clustered column-store index to store fact tables and large dimension tables for data warehousing workloads. This method improves query performance and data compression by up to 10 times.
2-    Use a non-clustered column-store index to perform analysis in real time on an OLTP workload.

How to create a column-store index
Using the following script you can create the column-store index in SQL Server.

--to create column-store index
CREATE COLUMNSTORE INDEX IX_OrderDetail_ProductIDOrderQty_ColumnStore
ON Sales.OrderDetail (ProductId,OrderQty);

Source of content from Microsoft: column-store index

Popular posts from this blog

Check for changes to an SQL Server table?

Problem: Suppose your team is working on the under-development project so it might be possible continuous work on the database and perform changes in Table, Stored procedure as per requirement, and daily you have to update the testing server database as per changes are done in developing server database then how it is possible to trace those changes. There are a lot of solutions for this problem which is listed below Solution 1: For SQL Server 2000, 2005 and above use the CHECKSUM command SELECTCHECKSUM_AGG(BINARY_CHECKSUM(*))FROMYour_Table_NameWITH (NOLOCK); That will return the same number each time its run as long as the table contents haven't changed. Unfortunately CHECKSUM does not work always properly to detect changes. It is only a primitive checksum and no CRC calculation. Therefore you can't use it to detect all changes, e. g. symmetrical changes result in the same CHECKSUM! Solution 2: 1.Run the following query. Before executing query replace DB_Name with your database name…

Merge and Merge join transformation in SSIS

How to drop multiple tables with common prefix in one query?

Problem: Suppose we have a situation where we need to drop those tables that have some prefixes string, so it is possible to drop those tables with a common prefix in a single query.
Solution: yes it is possible to drop all those tables that have the common prefix in a single query. Using following query you can delete all those tables that begin with a certain prefix string. In where condition just put the common prefix string in where condition (Like ‘prefix%’)
SELECT@query+=' DROP TABLE ' +QUOTENAME( +'.'+QUOTENAME(';' FROMsys.tablesASt INNERJOINsys.schemasASs ONt.[schema_id]=s.[schema_id] WHEREt.nameLIKE'MX_100%';
This query may create an issue, if a table has a foreign key relationship, you'll either need to drop them first or arrange the output to drop the tables in a certain order. If you want to monitor exactly what goes on when the query is running then use the following que…