Mastering LAG and LEAD Functions in SQL Server 2022 with the IGNORE NULLS Option

SQL Server 2022 introduced a powerful enhancement to the LAG and LEAD functions with the IGNORE NULLS option. This feature allows for more precise analysis and reporting by skipping over NULL values in data sets. In this blog, we’ll explore how to use these functions effectively using the JBDB database, and we’ll demonstrate their application with a detailed business use case.

Business Use Case: Sales Data Analysis

Imagine a retail company, JBStore, that wants to analyze its sales data to understand sales trends better. They aim to compare each month’s sales with the previous and next months, ignoring any missing data (represented by NULL values). This analysis will help identify trends and outliers, aiding in better decision-making.

Setting Up the JBDB Database

First, let’s set up the JBDB database and create a SalesData table with some sample data, including NULL values to represent months with no sales data.

-- Create JBDB database
CREATE DATABASE JBDB;
GO

-- Use the JBDB database
USE JBDB;
GO

-- Create SalesData table
CREATE TABLE SalesData (
    SalesMonth INT,
    SalesAmount INT
);

-- Insert sample data, including NULLs
INSERT INTO SalesData (SalesMonth, SalesAmount)
VALUES
    (1, 1000),
    (2, 1500),
    (3, NULL),
    (4, 1800),
    (5, NULL),
    (6, 2000);
GO

LAG and LEAD Functions: A Quick Recap

The LAG function allows you to access data from a previous row in the same result set without the use of a self-join. Similarly, the LEAD function accesses data from a subsequent row. Both functions are part of the SQL window functions family and are particularly useful in time series analysis.

Using LAG and LEAD with IGNORE NULLS

The IGNORE NULLS option is a game-changer, as it allows you to skip over NULL values, providing more meaningful results. Here’s how you can use it with the LAG and LEAD functions:

Example 1: LAG Function with IGNORE NULLS
SELECT 
    SalesMonth,
    SalesAmount,
    LAG(SalesAmount, 1) IGNORE NULLS OVER (ORDER BY SalesMonth) AS PreviousMonthSales
FROM 
    SalesData;

In this example, LAG(SalesAmount, 1) IGNORE NULLS retrieves the sales amount from the previous month, skipping over any NULL values.

Example 2: LEAD Function with IGNORE NULLS
SELECT 
    SalesMonth,
    SalesAmount,
    LEAD(SalesAmount, 1) IGNORE NULLS OVER (ORDER BY SalesMonth) AS NextMonthSales
FROM 
    SalesData;

Here, LEAD(SalesAmount, 1) IGNORE NULLS retrieves the sales amount from the next month, again skipping over NULL values.

Practical Example: Analyzing Sales Trends

Let’s combine these functions to analyze sales trends more effectively.

SELECT 
    SalesMonth,
    SalesAmount,
    LAG(SalesAmount, 1) IGNORE NULLS OVER (ORDER BY SalesMonth) AS PreviousMonthSales,
    LEAD(SalesAmount, 1) IGNORE NULLS OVER (ORDER BY SalesMonth) AS NextMonthSales
FROM 
    SalesData;

This query provides a complete view of each month’s sales, the previous month’s sales, and the next month’s sales, excluding any NULL values. This is incredibly useful for identifying patterns, such as periods of growth or decline.

Detailed Business Use Case: Data-Driven Decision Making

By utilizing the IGNORE NULLS option with LAG and LEAD functions, JBStore can:

  1. Identify Growth Periods: Detect months where sales increased significantly compared to the previous or next month.
  2. Spot Anomalies: Easily identify months with unusually high or low sales, excluding months with missing data.
  3. Trend Analysis: Understand longer-term trends by comparing sales over multiple months.

These insights can inform marketing strategies, inventory planning, and more.

Calculate Difference Between Current and Previous Month’s Sales:

SELECT SalesMonth, SalesAmount, SalesAmount - LAG(SalesAmount, 1) IGNORE NULLS OVER (ORDER BY SalesMonth) AS SalesDifference FROM SalesData;

Identify Months with Sales Decrease Compared to Previous Month:

WITH CTE AS (
    SELECT 
        SalesMonth,
        SalesAmount,
        LAG(SalesAmount, 1) IGNORE NULLS OVER (ORDER BY SalesMonth) AS PreviousMonthSales
    FROM 
        SalesData
)
SELECT 
    SalesMonth,
    SalesAmount,
    PreviousMonthSales
FROM 
    CTE
WHERE 
    SalesAmount < PreviousMonthSales;

Find the Second Previous Month’s Sales:

SELECT SalesMonth, SalesAmount, LAG(SalesAmount, 2) IGNORE NULLS OVER (ORDER BY SalesMonth) AS SecondPreviousMonthSales FROM SalesData;

Calculate the Rolling Average of the Last Two Months (Ignoring NULLs):

SELECT SalesMonth, SalesAmount, (SalesAmount + LAG(SalesAmount, 1) IGNORE NULLS OVER (ORDER BY SalesMonth)) / 2 AS RollingAverage FROM SalesData;

Compare Sales Between Current Month and Two Months Ahead:

SELECT SalesMonth, SalesAmount, LEAD(SalesAmount, 2) IGNORE NULLS OVER (ORDER BY SalesMonth) AS SalesTwoMonthsAhead FROM SalesData;

Identify Consecutive Months with Sales Increase:

WITH CTE AS ( SELECT SalesMonth, SalesAmount, LAG(SalesAmount, 1) IGNORE NULLS OVER (ORDER BY SalesMonth) AS PreviousMonthSales FROM SalesData ) SELECT SalesMonth, SalesAmount FROM CTE WHERE SalesAmount > PreviousMonthSales;

Find Months with No Sales and Their Preceding Sales Month:

SELECT SalesMonth, SalesAmount, LAG(SalesAmount, 1) IGNORE NULLS OVER (ORDER BY SalesMonth) AS PrecedingMonthSales FROM SalesData WHERE SalesAmount IS NULL;

Calculate Cumulative Sales Sum Ignoring NULLs:

SELECT 
    SalesMonth,
    SalesAmount,
    SUM(ISNULL(SalesAmount, 0)) OVER (ORDER BY SalesMonth ROWS BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW) AS CumulativeSales
FROM 
    SalesData;

Identify the First Month with Sales After a Month with NULL Sales:

SELECT SalesMonth, SalesAmount, LEAD(SalesAmount, 1) IGNORE NULLS OVER (ORDER BY SalesMonth) AS FirstNonNullSalesAfterNull FROM SalesData WHERE SalesAmount IS NULL;

    Conclusion πŸŽ‰

    The LAG and LEAD functions with the IGNORE NULLS option in SQL Server 2022 offer a more refined way to analyze data, providing more accurate and meaningful results. Whether you’re analyzing sales data, customer behavior, or any other time series data, these functions can significantly enhance your analytical capabilities.

    Happy querying! πŸš€

    For more tutorials and tips on SQL Server, including performance tuning and database management, be sure to check out our JBSWiki YouTube channel.

    Thank You,
    Vivek Janakiraman

    Disclaimer:
    The views expressed on this blog are mine alone and do not reflect the views of my company or anyone else. All postings on this blog are provided β€œAS IS” with no warranties, and confers no rights.

    Exploring the APPROX_COUNT_DISTINCT Function in SQL Server 2022

    With the release of SQL Server 2022, a range of powerful new functions has been introduced, including the APPROX_COUNT_DISTINCT function. This function provides a fast and memory-efficient way to estimate the number of unique values in a dataset, making it an invaluable tool for big data scenarios where traditional counting methods may be too slow or resource-intensive. In this blog, we will explore the APPROX_COUNT_DISTINCT function, using the JBDB database for practical demonstrations and providing a detailed business use case to illustrate its benefits. Let’s dive into the world of approximate distinct counts! πŸŽ‰


    Business Use Case: E-commerce Customer Segmentation πŸ“¦

    In an e-commerce business, understanding the diversity of customer behavior is crucial for personalized marketing and inventory management. The JBDB database contains customer transaction data, including CustomerID, ProductID, and PurchaseDate. The business aims to estimate the number of unique customers making purchases each month and the variety of products they are buying. Using the APPROX_COUNT_DISTINCT function, the company can quickly analyze this data to identify trends, optimize stock levels, and tailor marketing campaigns.


    Understanding the APPROX_COUNT_DISTINCT Function 🧠

    The APPROX_COUNT_DISTINCT function estimates the number of distinct values in a column, offering a performance-efficient alternative to the traditional COUNT(DISTINCT column) approach. It is particularly useful in large datasets where an exact count is less critical than performance and resource usage.

    Syntax:

    APPROX_COUNT_DISTINCT ( column_name )
    
    • column_name: The column from which distinct values are counted.

    Example 1: Estimating Unique Customers per Month πŸ“…

    Let’s calculate the estimated number of unique customers making purchases each month in the JBDB database.

    Setup:

    USE JBDB;
    GO
    
    CREATE TABLE CustomerTransactions (
        TransactionID INT PRIMARY KEY,
        CustomerID INT,
        ProductID INT,
        PurchaseDate DATE
    );
    
    INSERT INTO CustomerTransactions (TransactionID, CustomerID, ProductID, PurchaseDate)
    VALUES
    (1, 101, 2001, '2023-01-05'),
    (2, 102, 2002, '2023-01-10'),
    (3, 101, 2003, '2023-01-15'),
    (4, 103, 2001, '2023-02-05'),
    (5, 104, 2002, '2023-02-10'),
    (6, 102, 2004, '2023-02-15'),
    (7, 105, 2005, '2023-03-05'),
    (8, 106, 2001, '2023-03-10');
    GO

    Query to Estimate Unique Customers:

    SELECT 
        FORMAT(PurchaseDate, 'yyyy-MM') AS Month,
        APPROX_COUNT_DISTINCT(CustomerID) AS EstimatedUniqueCustomers
    FROM CustomerTransactions
    GROUP BY FORMAT(PurchaseDate, 'yyyy-MM');
    

    Output:

    MonthEstimatedUniqueCustomers
    2023-012
    2023-023
    2023-032

    This output gives an approximate count of unique customers making purchases in each month, providing quick insights into customer engagement over time.


    Example 2: Estimating Product Variety by Month πŸ“Š

    Now, let’s estimate the variety of products purchased each month to understand product diversity and demand trends.

    Query to Estimate Product Variety:

    SELECT 
        FORMAT(PurchaseDate, 'yyyy-MM') AS Month,
        APPROX_COUNT_DISTINCT(ProductID) AS EstimatedUniqueProducts
    FROM CustomerTransactions
    GROUP BY FORMAT(PurchaseDate, 'yyyy-MM');
    
    

    Output:

    MonthEstimatedUniqueProducts
    2023-013
    2023-023
    2023-032

    This data helps the business understand which months had the highest product variety, aiding in inventory and supply chain management.


    Example 3: Comparing Traditional and Approximate Counts πŸ”„

    To illustrate the efficiency of APPROX_COUNT_DISTINCT, let’s compare it with the traditional COUNT(DISTINCT column) method.

    Traditional COUNT(DISTINCT) Method:

    SELECT 
        FORMAT(PurchaseDate, 'yyyy-MM') AS Month,
        COUNT(DISTINCT CustomerID) AS ExactUniqueCustomers
    FROM CustomerTransactions
    GROUP BY FORMAT(PurchaseDate, 'yyyy-MM');
    
    

    Approximate COUNT(DISTINCT) Method:

    SELECT 
        FORMAT(PurchaseDate, 'yyyy-MM') AS Month,
        APPROX_COUNT_DISTINCT(CustomerID) AS EstimatedUniqueCustomers
    FROM CustomerTransactions
    GROUP BY FORMAT(PurchaseDate, 'yyyy-MM');
    

    Comparison:

    MonthExactUniqueCustomersEstimatedUniqueCustomers
    2023-0122
    2023-0233
    2023-0322

    The approximate method provides similar results with potentially significant performance improvements, especially in large datasets.


    Estimating Unique Products by Customer:

    • Calculate the estimated number of unique products purchased by each customer:
    SELECT 
        CustomerID,
        APPROX_COUNT_DISTINCT(ProductID) AS EstimatedUniqueProducts
    FROM CustomerTransactions
    GROUP BY CustomerID;

    Estimating Unique Purchase Dates:

    • Estimate the number of unique purchase dates in the dataset:
    SELECT 
        APPROX_COUNT_DISTINCT(PurchaseDate) AS EstimatedUniquePurchaseDates
    FROM CustomerTransactions;
    

    Regional Sales Analysis:

    • If the dataset includes a region column, estimate unique customers per region:
    SELECT 
        Region,
        APPROX_COUNT_DISTINCT(CustomerID) AS EstimatedUniqueCustomers
    FROM CustomerTransactions
    GROUP BY Region;

    Conclusion 🏁

    The APPROX_COUNT_DISTINCT function in SQL Server 2022 is a powerful tool for quickly estimating the number of distinct values in large datasets. This function is particularly useful in big data scenarios where performance and resource efficiency are crucial. By leveraging APPROX_COUNT_DISTINCT, businesses can gain rapid insights into customer behavior, product diversity, and other key metrics, enabling more informed decision-making. Whether you’re analyzing e-commerce data, customer segmentation, or product sales, this function offers a robust solution for your data analysis needs. Happy querying! πŸŽ‰

    For more tutorials and tips on SQL Server, including performance tuning and database management, be sure to check out our JBSWiki YouTube channel.

    Thank You,
    Vivek Janakiraman

    Disclaimer:
    The views expressed on this blog are mine alone and do not reflect the views of my company or anyone else. All postings on this blog are provided β€œAS IS” with no warranties, and confers no rights.

    Exploring SQL Server 2022 APPROX_PERCENTILE_DISC Function with JBDB Database

    SQL Server 2022 introduces several powerful features to enhance data analysis and performance. Among these, the APPROX_PERCENTILE_DISC function offers an efficient way to calculate discrete percentiles from large datasets. This blog will explore this function in depth, using practical examples from the JBDB database, and provide a detailed business use case to illustrate its utility. Let’s dive into the world of approximate discrete percentiles! πŸŽ‰


    Business Use Case: Analyzing Customer Satisfaction πŸ“Š

    Imagine a retail company seeking to understand customer satisfaction across different store locations. The data, stored in the JBDB database, includes satisfaction scores ranging from 1 to 5, representing customers’ overall experience. The company aims to identify key percentiles such as the median (50th percentile) and the 90th percentile to gauge typical and top-tier satisfaction levels. Using APPROX_PERCENTILE_DISC, they can efficiently compute these discrete percentiles, helping to guide strategies for improving customer experience and focusing on high-impact areas.


    Understanding the APPROX_PERCENTILE_DISC Function 🧠

    The APPROX_PERCENTILE_DISC function in SQL Server 2022 is designed to calculate approximate discrete percentiles from a sorted set of values. Unlike the continuous APPROX_PERCENTILE_CONT, this function returns the value nearest to the percentile rank, which is particularly useful for ordinal data.

    Syntax:

    APPROX_PERCENTILE_DISC ( percentile ) WITHIN GROUP ( ORDER BY column_name )
    
    • percentile: A numeric value between 0 and 1, indicating the desired percentile.
    • column_name: The column used to order the dataset before calculating the percentile.

    Example 1: Calculating Key Percentiles πŸ”

    Let’s calculate the median (50th percentile) and 90th percentile of customer satisfaction scores.

    Setup:

    USE JBDB;
    GO
    
    CREATE TABLE CustomerSatisfaction (
        CustomerID INT PRIMARY KEY,
        StoreID INT,
        SatisfactionScore INT,
        ReviewDate DATE
    );
    
    INSERT INTO CustomerSatisfaction (CustomerID, StoreID, SatisfactionScore, ReviewDate)
    VALUES
    (1, 101, 5, '2023-01-15'),
    (2, 102, 3, '2023-01-16'),
    (3, 103, 4, '2023-01-17'),
    (4, 101, 2, '2023-01-18'),
    (5, 104, 5, '2023-01-19'),
    (6, 105, 4, '2023-01-20'),
    (7, 106, 3, '2023-01-21'),
    (8, 102, 5, '2023-01-22');
    GO

    Query to Calculate 50th and 90th Percentiles:

    SELECT 
        APPROX_PERCENTILE_DISC(0.50) WITHIN GROUP (ORDER BY SatisfactionScore) AS MedianScore,
        APPROX_PERCENTILE_DISC(0.90) WITHIN GROUP (ORDER BY SatisfactionScore) AS Top10PercentScore
    FROM CustomerSatisfaction;

    Output:

    MedianScoreTop10PercentScore
    45

    This output reveals that the median satisfaction score is 4, and the top 10% of scores are 5, indicating a high level of satisfaction among the top-tier customers.


    Example 2: Store-Level Satisfaction Analysis πŸͺ

    Next, let’s analyze satisfaction scores at different store locations to identify trends and areas for improvement.

    Query for Store-Level Analysis:

    SELECT 
        StoreID,
        APPROX_PERCENTILE_DISC(0.50) WITHIN GROUP (ORDER BY SatisfactionScore) AS MedianScore,
        APPROX_PERCENTILE_DISC(0.90) WITHIN GROUP (ORDER BY SatisfactionScore) AS Top10PercentScore
    FROM CustomerSatisfaction
    GROUP BY StoreID;
    

    Output:

    StoreIDMedianScoreTop10PercentScore
    10135
    10245
    10344
    10455
    10544
    10633

    This analysis helps identify which stores are excelling in customer satisfaction and which may need targeted improvements.


    Example 3: Customer Segmentation by Satisfaction Levels πŸ“ˆ

    To further analyze the data, let’s segment customers into different satisfaction levels based on key percentiles.

    Step 1: Calculate Percentiles

    -- Calculate the 25th, 50th, and 75th percentiles
    SELECT 
        APPROX_PERCENTILE_DISC(0.25) WITHIN GROUP (ORDER BY SatisfactionScore) AS Q1,
        APPROX_PERCENTILE_DISC(0.50) WITHIN GROUP (ORDER BY SatisfactionScore) AS Q2,
        APPROX_PERCENTILE_DISC(0.75) WITHIN GROUP (ORDER BY SatisfactionScore) AS Q3
    INTO #Percentiles
    FROM CustomerSatisfaction;
    

    Step 2: Segment Customers

    -- Join with the Percentiles table to categorize customers
    SELECT 
        cs.CustomerID,
        cs.SatisfactionScore,
        CASE 
            WHEN cs.SatisfactionScore <= p.Q1 THEN 'Low'
            WHEN cs.SatisfactionScore <= p.Q2 THEN 'Medium'
            WHEN cs.SatisfactionScore <= p.Q3 THEN 'High'
            ELSE 'Very High'
        END AS SatisfactionLevel
    FROM 
        CustomerSatisfaction cs
    CROSS JOIN 
        #Percentiles p;

    Cleanup

    -- Drop the temporary table
    DROP TABLE #Percentiles;

    Explanation:

    1. Calculate Percentiles:
      • The first step calculates the 25th (Q1), 50th (Q2), and 75th (Q3) percentiles and stores them in a temporary table #Percentiles.
    2. Segment Customers:
      • The second step uses these percentile values to categorize each customer’s satisfaction score into levels: ‘Low’, ‘Medium’, ‘High’, or ‘Very High’.
    3. Cleanup:
      • Finally, the temporary table #Percentiles is dropped to clean up the session.

    Analyzing Low Satisfaction Scores:

    • Identify stores with the lowest 10th percentile satisfaction scores:
    SELECT 
        StoreID,
        APPROX_PERCENTILE_DISC(0.10) WITHIN GROUP (ORDER BY SatisfactionScore) AS Low10PercentScore
    FROM CustomerSatisfaction
    GROUP BY StoreID;
    

    Comparing Satisfaction Over Time:

    • Compare median satisfaction scores between two periods:
    SELECT 
        'Period 1' AS Period,
        APPROX_PERCENTILE_DISC(0.50) WITHIN GROUP (ORDER BY SatisfactionScore) AS MedianScore
    FROM CustomerSatisfaction
    WHERE ReviewDate BETWEEN '2023-01-15' AND '2023-01-18'
    UNION ALL
    SELECT 
        'Period 2' AS Period,
        APPROX_PERCENTILE_DISC(0.50) WITHIN GROUP (ORDER BY SatisfactionScore) AS MedianScore
    FROM CustomerSatisfaction
    WHERE ReviewDate BETWEEN '2023-01-19' AND '2023-01-22';

    3. Identifying High-Performing Stores:

    • List stores with a 90th percentile satisfaction score of 5:
    SELECT StoreID
    FROM CustomerSatisfaction
    GROUP BY StoreID
    HAVING APPROX_PERCENTILE_DISC(0.90) WITHIN GROUP (ORDER BY SatisfactionScore) = 5;

    Conclusion 🏁

    The APPROX_PERCENTILE_DISC function in SQL Server 2022 is a robust tool for efficiently estimating discrete percentiles. It offers a quick and practical solution for analyzing large datasets, making it invaluable for businesses looking to gain insights into customer behavior, product performance, and more. Whether you’re assessing customer satisfaction, analyzing sales data, or exploring other metrics, the APPROX_PERCENTILE_DISC function provides a clear and concise way to understand your data. Happy querying! πŸŽ‰

    For more tutorials and tips on SQL Server, including performance tuning and database management, be sure to check out our JBSWiki YouTube channel.

    Thank You,
    Vivek Janakiraman

    Disclaimer:
    The views expressed on this blog are mine alone and do not reflect the views of my company or anyone else. All postings on this blog are provided β€œAS IS” with no warranties, and confers no rights.