1 - active_users_count plugin

Learn how to use the active_users_count plugin to calculate the distinct count of values that appeared in a minimum number of periods in a lookback period.

Calculates distinct count of values, where each value has appeared in at least a minimum number of periods in a lookback period.

Useful for calculating distinct counts of “fans” only, while not including appearances of “non-fans”. A user is counted as a “fan” only if it was active during the lookback period. The lookback period is only used to determine whether a user is considered active (“fan”) or not. The aggregation itself doesn’t include users from the lookback window. In comparison, the sliding_window_counts aggregation is performed over a sliding window of the lookback period.

Syntax

T | evaluate active_users_count(IdColumn, TimelineColumn, Start, End, LookbackWindow, Period, ActivePeriodsCount, Bin , [dim1, dim2, …])

Parameters

NameTypeRequiredDescription
Tstring✔️The tabular input used to count active users.
IdColumnstring✔️The name of the column with ID values that represent user activity.
TimelineColumnstring✔️The name of the column that represents timeline.
Startdatetime✔️The analysis start period.
Enddatetime✔️The analysis end period.
LookbackWindowtimespan✔️The time window defining a period where user appearance is checked. The lookback period starts at ([current appearance] - [lookback window]) and ends on ([current appearance]).
Periodtimespan✔️A constant to count as single appearance (a user will be counted as active if it appears in at least distinct ActivePeriodsCount of this timespan.
ActivePeriodsCountdecimal✔️The minimal number of distinct active periods to decide if user is active. Active users are those users who appeared in at least (equal or greater than) active periods count.
Bindecimal, datetime, or timespan✔️A constant value of the analysis step period. May also be a string of week, month, or year. All periods will be the corresponding startofweek, startofmonth, orstartofyear functions.
dim1, dim2, …dynamicAn array of the dimensions columns that slice the activity metrics calculation.

Returns

Returns a table that has distinct count values for IDs that have appeared in ActivePeriodCounts in the following periods: the lookback period, each timeline period, and each existing dimensions combination.

Output table schema is:

TimelineColumndim1..dim_ndcount_values
type: as of TimelineColumn......long

Examples

Calculate weekly number of distinct users that appeared in at least three different days over a period of prior eight days. Period of analysis: July 2018.

let Start = datetime(2018-07-01);
let End = datetime(2018-07-31);
let LookbackWindow = 8d;
let Period = 1d;
let ActivePeriods = 3;
let Bin = 7d;
let T =  datatable(User:string, Timestamp:datetime)
[
    "B",      datetime(2018-06-29),
    "B",      datetime(2018-06-30),
    "A",      datetime(2018-07-02),
    "B",      datetime(2018-07-04),
    "B",      datetime(2018-07-08),
    "A",      datetime(2018-07-10),
    "A",      datetime(2018-07-14),
    "A",      datetime(2018-07-17),
    "A",      datetime(2018-07-20),
    "B",      datetime(2018-07-24)
];
T | evaluate active_users_count(User, Timestamp, Start, End, LookbackWindow, Period, ActivePeriods, Bin)

Output

Timestampdcount
2018-07-01 00:00:00.00000001
2018-07-15 00:00:00.00000001

A user is considered active if it fulfills both of the following criteria:

  • The user was seen in at least three distinct days (Period = 1d, ActivePeriods=3).
  • The user was seen in a lookback window of 8d before and including their current appearance.

In the illustration below, the only appearances that are active by this criteria are the following instances: User A on 7/20 and User B on 7/4 (see plugin results above). The appearances of User B are included for the lookback window on 7/4, but not for the Start-End time range of 6/29-30.

Graph showing active users based on the loopback window and active period specified in the query.

2 - activity_counts_metrics plugin

Learn how to use the activity_counts_metrics plugin to compare activity metrics in different time windows.

Calculates useful activity metrics for each time window compared/aggregated to all previous time windows. Metrics include: total count values, distinct count values, distinct count of new values, and aggregated distinct count. Compare this plugin to activity_metrics plugin, in which every time window is compared to its previous time window only.

Syntax

T | evaluate activity_counts_metrics(IdColumn, TimelineColumn, Start, End, Step [, Dimensions])

Parameters

NameTypeRequiredDescription
Tstring✔️The tabular input used to count activities.
IdColumnstring✔️The name of the column with ID values that represent user activity.
TimelineColumnstring✔️The name of the column that represents the timeline.
Startdatetime✔️The analysis start period.
Enddatetime✔️The analysis end period.
Stepdecimal, datetime, or timespan✔️The analysis window period. The value may also be a string of week, month, or year, in which case all periods would be startofweek, startofmonth, or startofyear.
DimensionsstringZero or more comma-separated dimensions columns that slice the activity metrics calculation.

Returns

Returns a table that has the total count values, distinct count values, distinct count of new values, and aggregated distinct count for each time window. If Dimensions are provided, then there’s another column for each dimension in the output table.

The following table describes the output table schema.

Column nameTypeDescription
TimestampSame as the provided TimelineColumn argumentThe time window start time.
countlongThe total records count in the time window and dim(s)
dcountlongThe distinct ID values count in the time window and dim(s)
new_dcountlongThe distinct ID values in the time window and dim(s) compared to all previous time windows.
aggregated_dcountlongThe total aggregated distinct ID values of dim(s) from first-time window to current (inclusive).

Examples

Daily activity counts

The next query calculates daily activity counts for the provided input table.

let start=datetime(2017-08-01);
let end=datetime(2017-08-04);
let window=1d;
let T = datatable(UserId:string, Timestamp:datetime)
[
'A', datetime(2017-08-01),
'D', datetime(2017-08-01),
'J', datetime(2017-08-01),
'B', datetime(2017-08-01),
'C', datetime(2017-08-02),
'T', datetime(2017-08-02),
'J', datetime(2017-08-02),
'H', datetime(2017-08-03),
'T', datetime(2017-08-03),
'T', datetime(2017-08-03),
'J', datetime(2017-08-03),
'B', datetime(2017-08-03),
'S', datetime(2017-08-03),
'S', datetime(2017-08-04),
];
 T
 | evaluate activity_counts_metrics(UserId, Timestamp, start, end, window)

Output

Timestampcountdcountnew_dcountaggregated_dcount
2017-08-01 00:00:00.00000004444
2017-08-02 00:00:00.00000003326
2017-08-03 00:00:00.00000006528
2017-08-04 00:00:00.00000001108

3 - activity_engagement plugin

Learn how to use the activity_engagement plugin to calculate activity engagement ratios.

Calculates activity engagement ratio based on ID column over a sliding timeline window.

The activity_engagement plugin can be used for calculating DAU/WAU/MAU (daily/weekly/monthly activities).

Syntax

T | evaluate activity_engagement(IdColumn, TimelineColumn, [Start, End,] InnerActivityWindow, OuterActivityWindow [, dim1, dim2, …])

Parameters

NameTypeRequiredDescription
Tstring✔️The tabular input used to calculate engagement.
IdCoumnstring✔️The name of the column with ID values that represent user activity.
TimelineColumnstring✔️The name of the column that represents timeline.
StartdatetimeThe analysis start period.
EnddatetimeThe analysis end period.
InnerActivityWindowtimespan✔️The inner-scope analysis window period.
OuterActivityWindowtimespan✔️The outer-scope analysis window period.
dim1, dim2, …dynamicAn array of the dimensions columns that slice the activity metrics calculation.

Returns

Returns a table that has a distinct count of ID values inside an inner-scope window, inside an outer-scope window, and the activity ratio for each inner-scope window period for each existing dimensions combination.

Output table schema is:

TimelineColumndcount_activities_innerdcount_activities_outeractivity_ratiodim1..dim_n
type: as of TimelineColumnlonglongdouble......

Examples

DAU/WAU calculation

The following example calculates DAU/WAU (Daily Active Users / Weekly Active Users ratio) over a randomly generated data.

// Generate random data of user activities
let _start = datetime(2017-01-01);
let _end = datetime(2017-01-31);
range _day from _start to _end  step 1d
| extend d = tolong((_day - _start)/1d)
| extend r = rand()+1
| extend _users=range(tolong(d*50*r), tolong(d*50*r+100*r-1), 1) 
| mv-expand id=_users to typeof(long) limit 1000000
// Calculate DAU/WAU ratio
| evaluate activity_engagement(['id'], _day, _start, _end, 1d, 7d)
| project _day, Dau_Wau=activity_ratio*100 
| render timechart 

:::image type=“content” source=“media/activity-engagement-plugin/activity-engagement-dau-wau.png” border=“false” alt-text=“Graph displaying the ratio of daily active users to weekly active users as specified in the query.”:::

DAU/MAU calculation

The following example calculates DAU/WAU (Daily Active Users / Weekly Active Users ratio) over a randomly generated data.

// Generate random data of user activities
let _start = datetime(2017-01-01);
let _end = datetime(2017-05-31);
range _day from _start to _end  step 1d
| extend d = tolong((_day - _start)/1d)
| extend r = rand()+1
| extend _users=range(tolong(d*50*r), tolong(d*50*r+100*r-1), 1) 
| mv-expand id=_users to typeof(long) limit 1000000
// Calculate DAU/MAU ratio
| evaluate activity_engagement(['id'], _day, _start, _end, 1d, 30d)
| project _day, Dau_Mau=activity_ratio*100 
| render timechart 

:::image type=“content” source=“media/activity-engagement-plugin/activity-engagement-dau-mau.png” border=“false” alt-text=“Graph displaying the ratio of daily active users to monthly active users as specified in the query.”:::

DAU/MAU calculation with additional dimensions

The following example calculates DAU/WAU (Daily Active Users / Weekly Active Users ratio) over a randomly generated data with additional dimension (mod3).

// Generate random data of user activities
let _start = datetime(2017-01-01);
let _end = datetime(2017-05-31);
range _day from _start to _end  step 1d
| extend d = tolong((_day - _start)/1d)
| extend r = rand()+1
| extend _users=range(tolong(d*50*r), tolong(d*50*r+100*r-1), 1) 
| mv-expand id=_users to typeof(long) limit 1000000
| extend mod3 = strcat("mod3=", id % 3)
// Calculate DAU/MAU ratio
| evaluate activity_engagement(['id'], _day, _start, _end, 1d, 30d, mod3)
| project _day, Dau_Mau=activity_ratio*100, mod3 
| render timechart 

:::image type=“content” source=“media/activity-engagement-plugin/activity-engagement-dau-mau-mod3.png” border=“false” alt-text=“Graph displaying the ratio of daily active users to monthly active users with modulo 3 as specified in the query.”:::

4 - activity_metrics plugin

Learn how to use the activity_metrics plugin to calculate activity metrics using the current time window compared to the previous window.

Calculates useful metrics that include distinct count values, distinct count of new values, retention rate, and churn rate. This plugin is different from activity_counts_metrics plugin in which every time window is compared to all previous time windows.

Syntax

T | evaluate activity_metrics(IdColumn, TimelineColumn, [Start, End,] Window [, dim1, dim2, …])

Parameters

NameTypeRequiredDescription
Tstring✔️The input used to calculate activity metrics.
IdCoumnstring✔️The name of the column with ID values that represent user activity.
TimelineColumnstring✔️The name of the column that represents timeline.
Startdatetime✔️The analysis start period.
Enddatetime✔️The analysis end period.
Stepdecimal, datetime, or timespan✔️The analysis window period. This value may also be a string of week, month, or year, in which case all periods will be startofweek, startofmonth, or startofyear respectively.
dim1, dim2, …dynamicAn array of the dimensions columns that slice the activity metrics calculation.

Returns

The plugin returns a table with the distinct count values, distinct count of new values, retention rate, and churn rate for each timeline period for each existing dimensions combination.

Output table schema is:

TimelineColumndcount_valuesdcount_newvaluesretention_ratechurn_ratedim1..dim_n
type: as of TimelineColumnlonglongdoubledouble......

Notes

Retention Rate Definition

Retention Rate over a period is calculated as:

where the # of customers returned during the period is defined as:

Retention Rate can vary from 0.0 to 1.0 A higher score means a larger number of returning users.

Churn Rate Definition

Churn Rate over a period is calculated as:

where the # of customer lost in the period is defined as:

Churn Rate can vary from 0.0 to 1.0 The higher score means the larger number of users are NOT returning to the service.

Churn vs. Retention Rate The churn vs. retention Rate is derived from the definition of Churn Rate and Retention Rate. The following calculation is always true:

Examples

Weekly retention rate and churn rate

The next query calculates retention and churn rate for week-over-week window.

// Generate random data of user activities
let _start = datetime(2017-01-02);
let _end = datetime(2017-05-31);
range _day from _start to _end  step 1d
| extend d = tolong((_day - _start)/1d)
| extend r = rand()+1
| extend _users=range(tolong(d*50*r), tolong(d*50*r+200*r-1), 1)
| mv-expand id=_users to typeof(long) limit 1000000
//
| evaluate activity_metrics(['id'], _day, _start, _end, 7d)
| project _day, retention_rate, churn_rate
| render timechart

Output

_dayretention_ratechurn_rate
2017-01-02 00:00:00.0000000NaNNaN
2017-01-09 00:00:00.00000000.1799100449775110.820089955022489
2017-01-16 00:00:00.00000000.7443744374437440.255625562556256
2017-01-23 00:00:00.00000000.6120967741935480.387903225806452
2017-01-30 00:00:00.00000000.6811414392059550.318858560794045
2017-02-06 00:00:00.00000000.2781456953642380.721854304635762
2017-02-13 00:00:00.00000000.2231726283048210.776827371695179
2017-02-20 00:00:00.00000000.380.62
2017-02-27 00:00:00.00000000.2955190017016450.704480998298355
2017-03-06 00:00:00.00000000.2803877703206560.719612229679344
2017-03-13 00:00:00.00000000.3606281547952890.639371845204711
2017-03-20 00:00:00.00000000.2880080280983440.711991971901656
2017-03-27 00:00:00.00000000.3061349693251530.693865030674847
2017-04-03 00:00:00.00000000.3568665377176020.643133462282398
2017-04-10 00:00:00.00000000.4950980392156860.504901960784314
2017-04-17 00:00:00.00000000.1982968369829680.801703163017032
2017-04-24 00:00:00.00000000.06188118811881190.938118811881188
2017-05-01 00:00:00.00000000.2046577275935070.795342272406493
2017-05-08 00:00:00.00000000.5173913043478260.482608695652174
2017-05-15 00:00:00.00000000.1436672967863890.856332703213611
2017-05-22 00:00:00.00000000.1991223258365330.800877674163467
2017-05-29 00:00:00.00000000.0634689922480620.936531007751938

:::image type=“content” source=“media/activity-metrics-plugin/activity-metrics-churn-and-retention.png” border=“false” alt-text=“Table showing the calculated retention and churn rates per seven days as specified in the query.”:::

Distinct values and distinct ’new’ values

The next query calculates distinct values and ’new’ values (IDs that didn’t appear in previous time window) for week-over-week window.

// Generate random data of user activities
let _start = datetime(2017-01-02);
let _end = datetime(2017-05-31);
range _day from _start to _end  step 1d
| extend d = tolong((_day - _start)/1d)
| extend r = rand()+1
| extend _users=range(tolong(d*50*r), tolong(d*50*r+200*r-1), 1)
| mv-expand id=_users to typeof(long) limit 1000000
//
| evaluate activity_metrics(['id'], _day, _start, _end, 7d)
| project _day, dcount_values, dcount_newvalues
| render timechart

Output

_daydcount_valuesdcount_newvalues
2017-01-02 00:00:00.0000000630630
2017-01-09 00:00:00.0000000738575
2017-01-16 00:00:00.00000001187841
2017-01-23 00:00:00.00000001092465
2017-01-30 00:00:00.00000001261647
2017-02-06 00:00:00.000000017441043
2017-02-13 00:00:00.00000001563432
2017-02-20 00:00:00.00000001406818
2017-02-27 00:00:00.000000019561429
2017-03-06 00:00:00.00000001593848
2017-03-13 00:00:00.000000018011423
2017-03-20 00:00:00.000000017101017
2017-03-27 00:00:00.000000017961516
2017-04-03 00:00:00.000000013811008
2017-04-10 00:00:00.000000017561162
2017-04-17 00:00:00.000000018311409
2017-04-24 00:00:00.000000018231164
2017-05-01 00:00:00.000000018111353
2017-05-08 00:00:00.000000016911246
2017-05-15 00:00:00.000000018121608
2017-05-22 00:00:00.000000017401017
2017-05-29 00:00:00.0000000960756

:::image type=“content” source=“media/activity-metrics-plugin/activity-metrics-dcount-and-dcount-newvalues.png” border=“false” alt-text=“Table showing the count of distinct values (dcount_values) and of new distinct values (dcount_newvalues) that didn’t appear in previous time window as specified in the query.”:::

5 - funnel_sequence plugin

Learn how to use the funnel_sequence plugin to learn how to calculate the distinct count of users who have taken a sequence of states, and the distribution of previous/next states that have led to/were followed by the sequence.

Calculates distinct count of users who have taken a sequence of states, and the distribution of previous/next states that have led to/were followed by the sequence. The plugin is invoked with the evaluate operator.

Syntax

T | evaluate funnel_sequence(IdColumn, TimelineColumn, Start, End, MaxSequenceStepWindow, Step, StateColumn, Sequence)

Parameters

NameTypeRequiredDescription
Tstring✔️The input tabular expression.
IdColumstring✔️The column reference representing the ID. This column must be present in T.
TimelineColumnstring✔️The column reference representing the timeline. This column must be present in T.
Startdatetime, timespan, or long✔️The analysis start period.
Enddatetime, timespan, or long✔️The analysis end period.
MaxSequenceStepWindowdatetime, timespan, or long✔️The value of the max allowed timespan between two sequential steps in the sequence.
Stepdatetime, timespan, or long✔️The analysis step period, or bin.
StateColumnstring✔️The column reference representing the state. This column must be present in T.
Sequencedynamic✔️An array with the sequence values that are looked up in StateColumn.

Returns

Returns three output tables, which are useful for constructing a sankey diagram for the analyzed sequence:

  • Table #1 - prev-sequence-next dcount

    • TimelineColumn: the analyzed time window
    • prev: the prev state (may be empty if there were any users that only had events for the searched sequence, but not any events prior to it).
    • next: the next state (may be empty if there were any users that only had events for the searched sequence, but not any events that followed it).
    • dcount: distinct count of IdColumn in time window that transitioned prev –> Sequence –> next.
    • samples: an array of IDs (from IdColumn) corresponding to the row’s sequence (a maximum of 128 IDs are returned).
  • Table #2 - prev-sequence dcount

    • TimelineColumn: the analyzed time window
    • prev: the prev state (may be empty if there were any users that only had events for the searched sequence, but not any events prior to it).
    • dcount: distinct count of IdColumn in time window that transitioned prev –> Sequence –> next.
    • samples: an array of IDs (from IdColumn) corresponding to the row’s sequence (a maximum of 128 IDs are returned).
  • Table #3 - sequence-next dcount

    • TimelineColumn: the analyzed time window
    • next: the next state (may be empty if there were any users that only had events for the searched sequence, but not any events that followed it).
    • dcount: distinct count of IdColumn in time window that transitioned prev –> Sequence –> next.
    • samples: an array of IDs (from IdColumn) corresponding to the row’s sequence (a maximum of 128 IDs are returned).

Examples

Exploring storm events

The following query looks at the table StormEvents (weather statistics for 2007) and shows which events happened before/after all Tornado events occurred in 2007.

// Looking on StormEvents statistics: 
// Q1: What happens before Tornado event?
// Q2: What happens after Tornado event?
StormEvents
| evaluate funnel_sequence(EpisodeId, StartTime, datetime(2007-01-01), datetime(2008-01-01), 1d,365d, EventType, dynamic(['Tornado']))

Result includes three tables:

  • Table #1: All possible variants of what happened before and after the sequence. For example, the second line means that there were 87 different events that had following sequence: Hail -> Tornado -> Hail
StartTimeprevnextdcount
2007-01-01 00:00:00.0000000293
2007-01-01 00:00:00.0000000HailHail87
2007-01-01 00:00:00.0000000Thunderstorm WindThunderstorm Wind77
2007-01-01 00:00:00.0000000HailThunderstorm Wind28
2007-01-01 00:00:00.0000000Hail28
2007-01-01 00:00:00.0000000Hail27
2007-01-01 00:00:00.0000000Thunderstorm Wind25
2007-01-01 00:00:00.0000000Thunderstorm WindHail24
2007-01-01 00:00:00.0000000Thunderstorm Wind24
2007-01-01 00:00:00.0000000Flash FloodFlash Flood12
2007-01-01 00:00:00.0000000Thunderstorm WindFlash Flood8
2007-01-01 00:00:00.0000000Flash Flood8
2007-01-01 00:00:00.0000000Funnel CloudThunderstorm Wind6
2007-01-01 00:00:00.0000000Funnel Cloud6
2007-01-01 00:00:00.0000000Flash Flood6
2007-01-01 00:00:00.0000000Funnel CloudFunnel Cloud6
2007-01-01 00:00:00.0000000HailFlash Flood4
2007-01-01 00:00:00.0000000Flash FloodThunderstorm Wind4
2007-01-01 00:00:00.0000000HailFunnel Cloud4
2007-01-01 00:00:00.0000000Funnel CloudHail4
2007-01-01 00:00:00.0000000Funnel Cloud4
2007-01-01 00:00:00.0000000Thunderstorm WindFunnel Cloud3
2007-01-01 00:00:00.0000000Heavy RainThunderstorm Wind2
2007-01-01 00:00:00.0000000Flash FloodFunnel Cloud2
2007-01-01 00:00:00.0000000Flash FloodHail2
2007-01-01 00:00:00.0000000Strong WindThunderstorm Wind1
2007-01-01 00:00:00.0000000Heavy RainFlash Flood1
2007-01-01 00:00:00.0000000Heavy RainHail1
2007-01-01 00:00:00.0000000HailFlood1
2007-01-01 00:00:00.0000000LightningHail1
2007-01-01 00:00:00.0000000Heavy RainLightning1
2007-01-01 00:00:00.0000000Funnel CloudHeavy Rain1
2007-01-01 00:00:00.0000000Flash FloodFlood1
2007-01-01 00:00:00.0000000FloodFlash Flood1
2007-01-01 00:00:00.0000000Heavy Rain1
2007-01-01 00:00:00.0000000Funnel CloudLightning1
2007-01-01 00:00:00.0000000LightningThunderstorm Wind1
2007-01-01 00:00:00.0000000FloodThunderstorm Wind1
2007-01-01 00:00:00.0000000HailLightning1
2007-01-01 00:00:00.0000000Lightning1
2007-01-01 00:00:00.0000000Tropical StormHurricane (Typhoon)1
2007-01-01 00:00:00.0000000Coastal Flood1
2007-01-01 00:00:00.0000000Rip Current1
2007-01-01 00:00:00.0000000Heavy Snow1
2007-01-01 00:00:00.0000000Strong Wind1
  • Table #2: shows all distinct events grouped by the previous event. For example, the second line shows that there were a total of 150 events of Hail that happened just before Tornado.
StartTimeprevdcount
2007-01-01 00:00:00.0000000331
2007-01-01 00:00:00.0000000Hail150
2007-01-01 00:00:00.0000000Thunderstorm Wind135
2007-01-01 00:00:00.0000000Flash Flood28
2007-01-01 00:00:00.0000000Funnel Cloud22
2007-01-01 00:00:00.0000000Heavy Rain5
2007-01-01 00:00:00.0000000Flood2
2007-01-01 00:00:00.0000000Lightning2
2007-01-01 00:00:00.0000000Strong Wind2
2007-01-01 00:00:00.0000000Heavy Snow1
2007-01-01 00:00:00.0000000Rip Current1
2007-01-01 00:00:00.0000000Coastal Flood1
2007-01-01 00:00:00.0000000Tropical Storm1
  • Table #3: shows all distinct events grouped by next event. For example, the second line shows that there were a total of 143 events of Hail that happened after Tornado.
StartTimenextdcount
2007-01-01 00:00:00.0000000332
2007-01-01 00:00:00.0000000Hail145
2007-01-01 00:00:00.0000000Thunderstorm Wind143
2007-01-01 00:00:00.0000000Flash Flood32
2007-01-01 00:00:00.0000000Funnel Cloud21
2007-01-01 00:00:00.0000000Lightning4
2007-01-01 00:00:00.0000000Heavy Rain2
2007-01-01 00:00:00.0000000Flood2
2007-01-01 00:00:00.0000000Hurricane (Typhoon)1

Now, let’s try to find out how the following sequence continues:
Hail -> Tornado -> Thunderstorm Wind

StormEvents
| evaluate funnel_sequence(
               EpisodeId,
               StartTime,
               datetime(2007-01-01),
               datetime(2008-01-01),
               1d,
               365d,
               EventType, 
               dynamic(['Hail', 'Tornado', 'Thunderstorm Wind'])
           )

Skipping Table #1 and Table #2, and looking at Table #3, we can conclude that sequence Hail -> Tornado -> Thunderstorm Wind in 92 events ended with this sequence, continued as Hail in 41 events, and turned back to Tornado in 14.

StartTimenextdcount
2007-01-01 00:00:00.000000092
2007-01-01 00:00:00.0000000Hail41
2007-01-01 00:00:00.0000000Tornado14
2007-01-01 00:00:00.0000000Flash Flood11
2007-01-01 00:00:00.0000000Lightning2
2007-01-01 00:00:00.0000000Heavy Rain1
2007-01-01 00:00:00.0000000Flood1

6 - funnel_sequence_completion plugin

Learn how to use the funnel_sequence_completion plugin to calculate a funnel of completed sequence steps while comparing different time periods.

Calculates a funnel of completed sequence steps while comparing different time periods. The plugin is invoked with the evaluate operator.

Syntax

T | evaluate funnel_sequence_completion(IdColumn, TimelineColumn, Start, End, BinSize, StateColumn, Sequence, MaxSequenceStepWindows)

Parameters

NameTypeRequiredDescription
Tstring✔️The input tabular expression.
IdColumstring✔️The column reference representing the ID. The column must be present in T.
TimelineColumnstring✔️The column reference representing the timeline. The column must be present in T.
Startdatetime, timespan, or long✔️The analysis start period.
Enddatetime, timespan, or long✔️The analysis end period.
BinSizedatetime, timespan, or long✔️The analysis window size. Each window is analyzed separately.
StateColumnstring✔️The column reference representing the state. The column must be present in T.
Sequencedynamic✔️An array with the sequence values that are looked up in StateColumn.
MaxSequenceStepPeriodsdynamic✔️An array with the values of the max allowed timespan between the first and last sequential steps in the sequence. Each period in the array generates a funnel analysis result.

Returns

Returns a single table useful for constructing a funnel diagram for the analyzed sequence:

  • TimelineColumn: the analyzed time window (bin), each bin in the analysis timeframe (Start to End) generates a funnel analysis separately.
  • StateColumn: the state of the sequence.
  • Period: the maximal period allowed for completing steps in the funnel sequence measured from the first step in the sequence. Each value in MaxSequenceStepPeriods generates a funnel analysis with a separate period.
  • dcount: distinct count of IdColumn in time window that transitioned from first sequence state to the value of StateColumn.

Examples

Exploring Storm Events

The following query checks the completion funnel of the sequence: Hail -> Tornado -> Thunderstorm Wind in “overall” time of 1hour, 4hours, 1day.

let _start = datetime(2007-01-01);
let _end =  datetime(2008-01-01);
let _windowSize = 365d;
let _sequence = dynamic(['Hail', 'Tornado', 'Thunderstorm Wind']);
let _periods = dynamic([1h, 4h, 1d]);
StormEvents
| evaluate funnel_sequence_completion(EpisodeId, StartTime, _start, _end, _windowSize, EventType, _sequence, _periods) 

Output

StartTimeEventTypePerioddcount
2007-01-01 00:00:00.0000000Hail01:00:002877
2007-01-01 00:00:00.0000000Tornado01:00:00208
2007-01-01 00:00:00.0000000Thunderstorm Wind01:00:0087
2007-01-01 00:00:00.0000000Hail04:00:002877
2007-01-01 00:00:00.0000000Tornado04:00:00231
2007-01-01 00:00:00.0000000Thunderstorm Wind04:00:00141
2007-01-01 00:00:00.0000000Hail1.00:00:002877
2007-01-01 00:00:00.0000000Tornado1.00:00:00244
2007-01-01 00:00:00.0000000Thunderstorm Wind1.00:00:00155

Understanding the results:
The outcome is three funnels (for periods: One hour, 4 hours, and one day). For each funnel step, a number of distinct counts of are shown. You can see that the more time is given to complete the whole sequence of Hail -> Tornado -> Thunderstorm Wind, the higher dcount value is obtained. In other words, there were more occurrences of the sequence reaching the funnel step.

7 - new_activity_metrics plugin

Learn how to use the new_activity_metrics plugin to calculate activity metrics.

Calculates useful activity metrics (distinct count values, distinct count of new values, retention rate, and churn rate) for the cohort of New Users. Each cohort of New Users (all users, which were first seen in time window) is compared to all prior cohorts. Comparison takes into account all previous time windows. For example, for records from T2 to T3, the distinct count of users will be all users in T3 who weren’t seen in both T1 and T2. The plugin is invoked with the evaluate operator.

Syntax

TabularExpression | evaluate new_activity_metrics(IdColumn, TimelineColumn, Start, End, Window [, Cohort] [, dim1, dim2, …] [, Lookback] )

Parameters

NameTypeRequiredDescription
TabularExpressionstring✔️The tabular expression for which to calculate activity metrics.
IdColumnstring✔️The name of the column with ID values that represent user activity.
TimelineColumnstring✔️The name of the column that represents the timeline.
Startscalar✔️The value of the analysis start period.
Endscalar✔️The value of the analysis end period.
Windowscalar✔️The value of the analysis window period. Can be a numeric, datetime, or timespan value, or a string that is one of week, month or year, in which case all periods will be startofweek/startofmonth/startofyear accordingly. When using startofweek, make sure start time is a Sunday, otherwise first cohort will be empty (since startofweek is considered to be a Sunday).
CohortscalarIndicates a specific cohort. If not provided, all cohorts corresponding to the analysis time window are calculated and returned.
dim1, dim2, …dynamicAn array of the dimensions columns that slice the activity metrics calculation.
LookbackstringA tabular expression with a set of IDs that belong to the ’look back’ period.

Returns

Returns a table that contains the following for each combination of ‘from’ and ’to’ timeline periods and for each existing column (dimensions) combination:

  • distinct count values
  • distinct count of new values
  • retention rate
  • churn rate

Output table schema is:

from_TimelineColumnto_TimelineColumndcount_new_valuesdcount_retained_valuesdcount_churn_valuesretention_ratechurn_ratedim1..dim_n
type: as of TimelineColumnsamelonglongdoubledoubledouble......
  • from_TimelineColumn - the cohort of new users. Metrics in this record refer to all users who were first seen in this period. The decision on first seen takes into account all previous periods in the analysis period.
  • to_TimelineColumn - the period being compared to.
  • dcount_new_values - the number of distinct users in to_TimelineColumn that weren’t seen in all periods prior to and including from_TimelineColumn.
  • dcount_retained_values - out of all new users, first seen in from_TimelineColumn, the number of distinct users that were seen in to_TimelineCoumn.
  • dcount_churn_values - out of all new users, first seen in from_TimelineColumn, the number of distinct users that weren’t seen in to_TimelineCoumn.
  • retention_rate - the percent of dcount_retained_values out of the cohort (users first seen in from_TimelineColumn).
  • churn_rate - the percent of dcount_churn_values out of the cohort (users first seen in from_TimelineColumn).

Examples

The following sample dataset shows which users seen on which days. The table was generated based on a source Users table, as follows:

Users | summarize tostring(make_set(user)) by bin(Timestamp, 1d) | order by Timestamp asc;

Output

Timestampset_user
2019-11-01 00:00:00.0000000[0,2,3,4]
2019-11-02 00:00:00.0000000[0,1,3,4,5]
2019-11-03 00:00:00.0000000[0,2,4,5]
2019-11-04 00:00:00.0000000[0,1,2,3]
2019-11-05 00:00:00.0000000[0,1,2,3,4]

The output of the plugin for the original table is the following:

let StartDate = datetime(2019-11-01 00:00:00);
let EndDate = datetime(2019-11-07 00:00:00);
Users 
| evaluate new_activity_metrics(user, Timestamp, StartDate, EndDate-1tick, 1d) 
| where from_Timestamp < datetime(2019-11-03 00:00:00.0000000)

Output

Rfrom_Timestampto_Timestampdcount_new_valuesdcount_retained_valuesdcount_churn_valuesretention_ratechurn_rate
12019-11-01 00:00:00.00000002019-11-01 00:00:00.000000044010
22019-11-01 00:00:00.00000002019-11-02 00:00:00.00000002310.750.25
32019-11-01 00:00:00.00000002019-11-03 00:00:00.00000001310.750.25
42019-11-01 00:00:00.00000002019-11-04 00:00:00.00000001310.750.25
52019-11-01 00:00:00.00000002019-11-05 00:00:00.000000014010
62019-11-01 00:00:00.00000002019-11-06 00:00:00.000000000401
72019-11-02 00:00:00.00000002019-11-02 00:00:00.000000022010
82019-11-02 00:00:00.00000002019-11-03 00:00:00.00000000110.50.5
92019-11-02 00:00:00.00000002019-11-04 00:00:00.00000000110.50.5
102019-11-02 00:00:00.00000002019-11-05 00:00:00.00000000110.50.5
112019-11-02 00:00:00.00000002019-11-06 00:00:00.000000000201

Following is an analysis of a few records from the output:

  • Record R=3, from_TimelineColumn = 2019-11-01, to_TimelineColumn = 2019-11-03:

    • The users considered for this record are all new users seen on 11/1. Since this is the first period, these are all users in that bin – [0,2,3,4]
    • dcount_new_values – the number of users on 11/3 who weren’t seen on 11/1. This includes a single user – 5.
    • dcount_retained_values – out of all new users on 11/1, how many were retained until 11/3? There are three values ([0,2,4]), while count_churn_values is one (user=3).
    • retention_rate = 0.75 – the three retained users out of the four new users who were first seen in 11/1.
  • Record R=9, from_TimelineColumn = 2019-11-02, to_TimelineColumn = 2019-11-04:

    • This record focuses on the new users who were first seen on 11/2 – users 1 and 5.
    • dcount_new_values – the number of users on 11/4 who weren’t seen through all periods T0 .. from_Timestamp. Meaning, users who are seen on 11/4 but who weren’t seen on either 11/1 or 11/2 – there are no such users.
    • dcount_retained_values – out of all new users on 11/2 ([1,5]), how many were retained until 11/4? There’s one such user ([1]), while count_churn_values is one (user 5).
    • retention_rate is 0.5 – the single user that was retained on 11/4 out of the two new ones on 11/2.

Weekly retention rate, and churn rate (single week)

The next query calculates a retention and churn rate for week-over-week window for New Users cohort (users that arrived on the first week).

// Generate random data of user activities
let _start = datetime(2017-05-01);
let _end = datetime(2017-05-31);
range Day from _start to _end step 1d
| extend d = tolong((Day - _start) / 1d)
| extend r = rand() + 1
| extend _users=range(tolong(d * 50 * r), tolong(d * 50 * r + 200 * r - 1), 1) 
| mv-expand id=_users to typeof(long) limit 1000000
// Take only the first week cohort (last parameter)
| evaluate new_activity_metrics(['id'], Day, _start, _end, 7d, _start)
| project from_Day, to_Day, retention_rate, churn_rate

Output

from_Dayto_Dayretention_ratechurn_rate
2017-05-01 00:00:00.00000002017-05-01 00:00:00.000000010
2017-05-01 00:00:00.00000002017-05-08 00:00:00.00000000.5446327683615820.455367231638418
2017-05-01 00:00:00.00000002017-05-15 00:00:00.00000000.0316384180790960.968361581920904
2017-05-01 00:00:00.00000002017-05-22 00:00:00.000000001
2017-05-01 00:00:00.00000002017-05-29 00:00:00.000000001

Weekly retention rate, and churn rate (complete matrix)

The next query calculates retention and churn rate for week-over-week window for New Users cohort. If the previous example calculated the statistics for a single week - the following query produces an NxN table for each from/to combination.

// Generate random data of user activities
let _start = datetime(2017-05-01);
let _end = datetime(2017-05-31);
range Day from _start to _end step 1d
| extend d = tolong((Day - _start) / 1d)
| extend r = rand() + 1
| extend _users=range(tolong(d * 50 * r), tolong(d * 50 * r + 200 * r - 1), 1) 
| mv-expand id=_users to typeof(long) limit 1000000
// Last parameter is omitted - 
| evaluate new_activity_metrics(['id'], Day, _start, _end, 7d)
| project from_Day, to_Day, retention_rate, churn_rate

Output

from_Dayto_Dayretention_ratechurn_rate
2017-05-01 00:00:00.00000002017-05-01 00:00:00.000000010
2017-05-01 00:00:00.00000002017-05-08 00:00:00.00000000.1903973509933770.809602649006622
2017-05-01 00:00:00.00000002017-05-15 00:00:00.000000001
2017-05-01 00:00:00.00000002017-05-22 00:00:00.000000001
2017-05-01 00:00:00.00000002017-05-29 00:00:00.000000001
2017-05-08 00:00:00.00000002017-05-08 00:00:00.000000010
2017-05-08 00:00:00.00000002017-05-15 00:00:00.00000000.4052631578947370.594736842105263
2017-05-08 00:00:00.00000002017-05-22 00:00:00.00000000.2276315789473680.772368421052632
2017-05-08 00:00:00.00000002017-05-29 00:00:00.000000001
2017-05-15 00:00:00.00000002017-05-15 00:00:00.000000010
2017-05-15 00:00:00.00000002017-05-22 00:00:00.00000000.7854889589905360.214511041009464
2017-05-15 00:00:00.00000002017-05-29 00:00:00.00000000.2376445846477390.762355415352261
2017-05-22 00:00:00.00000002017-05-22 00:00:00.000000010
2017-05-22 00:00:00.00000002017-05-29 00:00:00.00000000.6218354430379750.378164556962025
2017-05-29 00:00:00.00000002017-05-29 00:00:00.000000010

Weekly retention rate with lookback period

The following query calculates the retention rate of New Users cohort when taking into consideration lookback period: a tabular query with set of Ids that are used to define the New Users cohort (all IDs that don’t appear in this set are New Users). The query examines the retention behavior of the New Users during the analysis period.

// Generate random data of user activities
let _lookback = datetime(2017-02-01);
let _start = datetime(2017-05-01);
let _end = datetime(2017-05-31);
let _data = range Day from _lookback to _end step 1d
    | extend d = tolong((Day - _lookback) / 1d)
    | extend r = rand() + 1
    | extend _users=range(tolong(d * 50 * r), tolong(d * 50 * r + 200 * r - 1), 1) 
    | mv-expand id=_users to typeof(long) limit 1000000;
//
let lookback_data = _data | where Day < _start | project Day, id;
_data
| evaluate new_activity_metrics(id, Day, _start, _end, 7d, _start, lookback_data)
| project from_Day, to_Day, retention_rate

Output

from_Dayto_Dayretention_rate
2017-05-01 00:00:00.00000002017-05-01 00:00:00.00000001
2017-05-01 00:00:00.00000002017-05-08 00:00:00.00000000.404081632653061
2017-05-01 00:00:00.00000002017-05-15 00:00:00.00000000.257142857142857
2017-05-01 00:00:00.00000002017-05-22 00:00:00.00000000.296326530612245
2017-05-01 00:00:00.00000002017-05-29 00:00:00.00000000.0587755102040816

8 - rolling_percentile plugin

Learn how to use the rolling_percentile plugin to calculate an estimate of the rolling percentile per bin for the specified value column.

Returns an estimate for the specified percentile of the ValueColumn population in a rolling (sliding) BinsPerWindow size window per BinSize.

The plugin is invoked with the evaluate operator.

Syntax

T | evaluate rolling_percentile(ValueColumn, Percentile, IndexColumn, BinSize, BinsPerWindow [, dim1, dim2, …] )

Parameters

NameTypeRequiredDescription
Tstring✔️The input tabular expression.
ValueColumnstring✔️The name of the column used to calculate the percentiles.
Percentileint, long, or real✔️Scalar with the percentile to calculate.
IndexColumnstring✔️The name of the column over which to run the rolling window.
BinSizeint, long, real, datetime, or timespan✔️Scalar with size of the bins to apply over the IndexColumn.
BinsPerWindowint✔️The number of bins included in each window.
dim1, dim2, …stringA list of the dimensions columns to slice by.

Returns

Returns a table with a row per each bin (and combination of dimensions if specified) that has the rolling percentile of values in the window ending at the bin (inclusive). Output table schema is:

IndexColumndim1dim_nrolling_BinsPerWindow_percentile_ValueColumn_Pct

Examples

Rolling 3-day median value per day

The next query calculates a 3-day median value in daily granularity. Each row in the output represents the median value for the last 3 bins (days), including the bin itself.

let T = 
    range idx from 0 to 24 * 10 - 1 step 1
    | project Timestamp = datetime(2018-01-01) + 1h * idx, val=idx + 1
    | extend EvenOrOdd = iff(val % 2 == 0, "Even", "Odd");
T  
| evaluate rolling_percentile(val, 50, Timestamp, 1d, 3)

Output

Timestamprolling_3_percentile_val_50
2018-01-01 00:00:00.000000012
2018-01-02 00:00:00.000000024
2018-01-03 00:00:00.000000036
2018-01-04 00:00:00.000000060
2018-01-05 00:00:00.000000084
2018-01-06 00:00:00.0000000108
2018-01-07 00:00:00.0000000132
2018-01-08 00:00:00.0000000156
2018-01-09 00:00:00.0000000180
2018-01-10 00:00:00.0000000204

Rolling 3-day median value per day by dimension

Same example from above, but now also calculates the rolling window partitioned for each value of the dimension.

let T = 
    range idx from 0 to 24 * 10 - 1 step 1
    | project Timestamp = datetime(2018-01-01) + 1h * idx, val=idx + 1
    | extend EvenOrOdd = iff(val % 2 == 0, "Even", "Odd");
T  
| evaluate rolling_percentile(val, 50, Timestamp, 1d, 3, EvenOrOdd)

Output

TimestampEvenOrOddrolling_3_percentile_val_50
2018-01-01 00:00:00.0000000Even12
2018-01-02 00:00:00.0000000Even24
2018-01-03 00:00:00.0000000Even36
2018-01-04 00:00:00.0000000Even60
2018-01-05 00:00:00.0000000Even84
2018-01-06 00:00:00.0000000Even108
2018-01-07 00:00:00.0000000Even132
2018-01-08 00:00:00.0000000Even156
2018-01-09 00:00:00.0000000Even180
2018-01-10 00:00:00.0000000Even204
2018-01-01 00:00:00.0000000Odd11
2018-01-02 00:00:00.0000000Odd23
2018-01-03 00:00:00.0000000Odd35
2018-01-04 00:00:00.0000000Odd59
2018-01-05 00:00:00.0000000Odd83
2018-01-06 00:00:00.0000000Odd107
2018-01-07 00:00:00.0000000Odd131
2018-01-08 00:00:00.0000000Odd155
2018-01-09 00:00:00.0000000Odd179
2018-01-10 00:00:00.0000000Odd203

9 - rows_near plugin

Learn how to use the rows_near plugin to find rows near a specified condition.

Finds rows near a specified condition.

The plugin is invoked with the evaluate operator.

Syntax

T | evaluate rows_near(Condition, NumRows, [, RowsAfter ])

Parameters

NameTypeRequiredDescription
Tstring✔️The input tabular expression.
Conditionbool✔️Represents the condition to find rows around.
NumRowsint✔️The number of rows to find before and after the condition.
RowsAfterintWhen specified, overrides the number of rows to find after the condition.

Returns

Every row from the input that is within NumRows from a true Condition, When RowsAfter is specified, returns every row from the input that is NumRows before or RowsAfter after a true Condition.

Example

Find rows with an "Error" State, and returns 2 rows before and after the "Error" record.

datatable (Timestamp:datetime, Value:long, State:string )
[
    datetime(2021-06-01), 1, "Success",
    datetime(2021-06-02), 4, "Success",
    datetime(2021-06-03), 3, "Success",
    datetime(2021-06-04), 11, "Success",
    datetime(2021-06-05), 15, "Success",
    datetime(2021-06-06), 2, "Success",
    datetime(2021-06-07), 19, "Error",
    datetime(2021-06-08), 12, "Success",
    datetime(2021-06-09), 7, "Success",
    datetime(2021-06-10), 9, "Success",
    datetime(2021-06-11), 4, "Success",
    datetime(2021-06-12), 1, "Success",
]
| sort by Timestamp asc 
| evaluate rows_near(State == "Error", 2)

Output

TimestampValueState
2021-06-05 00:00:00.000000015Success
2021-06-06 00:00:00.00000002Success
2021-06-07 00:00:00.000000019Error
2021-06-08 00:00:00.000000012Success
2021-06-09 00:00:00.00000007Success

10 - sequence_detect plugin

Learn how to use the sequence_detect plugin to detect sequence occurrences based on provided predicates.

Detects sequence occurrences based on provided predicates. The plugin is invoked with the evaluate operator.

Syntax

T | evaluate sequence_detect (TimelineColumn, MaxSequenceStepWindow, MaxSequenceSpan, Expr1, Expr2, …, Dim1, Dim2,)

Parameters

NameTypeRequiredDescription
Tstring✔️The input tabular expression.
TimelineColumnstring✔️The column reference representing timeline, must be present in the source expression.
MaxSequenceStepWindowtimespan✔️The value of the max allowed timespan between 2 sequential steps in the sequence.
MaxSequenceSpantimespan✔️The max timespan for the sequence to complete all steps.
Expr1, Expr2, …string✔️The boolean predicate expressions defining sequence steps.
Dim1, Dim2, …string✔️The dimension expressions that are used to correlate sequences.

Returns

Returns a single table where each row in the table represents a single sequence occurrence:

  • Dim1, Dim2, …: dimension columns that were used to correlate sequences.
  • Expr1TimelineColumn, Expr2TimelineColumn, …: Columns with time values, representing the timeline of each sequence step.
  • Duration: the overall sequence time window

Examples

The following query looks at the table T to search for relevant data from a specified time period.

T | evaluate sequence_detect(datetime_column, 10m, 1h, e1 = (Col1 == 'Val'), e2 = (Col2 == 'Val2'), Dim1, Dim2)

Exploring Storm Events

The following query looks on the table StormEvents (weather statistics for 2007) and shows cases where sequence of ‘Excessive Heat’ was followed by ‘Wildfire’ within 5 days.

StormEvents
| evaluate sequence_detect(
               StartTime,
               5d,  // step max-time
               5d,  // sequence max-time
               heat=(EventType == "Excessive Heat"), 
               wildfire=(EventType == 'Wildfire'), 
               State
           )

Output

Stateheat_StartTimewildfire_StartTimeDuration
CALIFORNIA2007-05-08 00:00:00.00000002007-05-08 16:02:00.000000016:02:00
CALIFORNIA2007-05-08 00:00:00.00000002007-05-10 11:30:00.00000002.11:30:00
CALIFORNIA2007-07-04 09:00:00.00000002007-07-05 23:01:00.00000001.14:01:00
SOUTH DAKOTA2007-07-23 12:00:00.00000002007-07-27 09:00:00.00000003.21:00:00
TEXAS2007-08-10 08:00:00.00000002007-08-11 13:56:00.00000001.05:56:00
CALIFORNIA2007-08-31 08:00:00.00000002007-09-01 11:28:00.00000001.03:28:00
CALIFORNIA2007-08-31 08:00:00.00000002007-09-02 13:30:00.00000002.05:30:00
CALIFORNIA2007-09-02 12:00:00.00000002007-09-02 13:30:00.000000001:30:00

11 - session_count plugin

Learn how to use the session_count plugin to calculate the session count based on the ID column over a timeline.

Calculates the session count based on the ID column over a timeline. The plugin is invoked with the evaluate operator.

Syntax

TabularExpression | evaluate session_count(IdColumn, TimelineColumn, Start, End, Bin, LookBackWindow [, dim1, dim2, …])

Parameters

NameTypeRequiredDescription
TabularExpressionstring✔️The tabular expression that serves as input.
IdColumnstring✔️The name of the column with ID values that represents user activity.
TimelineColumnstring✔️The name of the column that represents the timeline.
Startscalar✔️The start of the analysis period.
Endscalar✔️The end of the analysis period.
Binscalar✔️The session’s analysis step period.
LookBackWindowscalar✔️The session lookback period. If the ID from IdColumn appears in a time window within LookBackWindow, the session is considered to be an existing one. If the ID doesn’t appear, then the session is considered to be new.
dim1, dim2, …stringA list of the dimensions columns that slice the session count calculation.

Returns

Returns a table that has the session count values for each timeline period and for each existing dimensions combination.

Output table schema is:

TimelineColumndim1..dim_ncount_sessions
type: as of TimelineColumn......long

Examples

For this example, the data is deterministic, and we use a table with two columns:

  • Timeline: a running number from 1 to 10,000
  • Id: ID of the user from 1 to 50

Id appears at the specific Timeline slot if it’s a divider of Timeline (Timeline % Id == 0).

An event with Id==1 will appear at any Timeline slot, an event with Id==2 at every second Timeline slot, and so on.

Here are 20 lines of the data:

let _data = range Timeline from 1 to 10000 step 1
    | extend __key = 1
    | join kind=inner (range Id from 1 to 50 step 1 | extend __key=1) on __key
    | where Timeline % Id == 0
    | project Timeline, Id;
// Look on few lines of the data
_data
| order by Timeline asc, Id asc
| take 20

Output

TimelineId
11
21
22
31
33
41
42
44
51
55
61
62
63
66
71
77
81
82
84
88

Let’s define a session in next terms: session considered to be active as long as user (Id) appears at least once at a timeframe of 100 time slots, while session look-back window is 41 time slots.

The next query shows the count of active sessions according to the above definition.

let _data = range Timeline from 1 to 9999 step 1
    | extend __key = 1
    | join kind=inner (range Id from 1 to 50 step 1 | extend __key=1) on __key
    | where Timeline % Id == 0
    | project Timeline, Id;
// End of data definition
_data
| evaluate session_count(Id, Timeline, 1, 10000, 100, 41)
| render linechart 

Example session count.

12 - sliding_window_counts plugin

Learn how to use the sliding_window_counts plugin to calculate counts and distinct counts of values in a sliding window over a lookback period.

Calculates counts and distinct count of values in a sliding window over a lookback period, using the technique described in the Perform aggregations over a sliding window example. The plugin is invoked with the evaluate operator.

Syntax

T | evaluate sliding_window_counts(IdColumn, TimelineColumn, Start, End, LookbackWindow, Bin , [dim1, dim2, …])

Parameters

NameTypeRequiredDescription
Tstring✔️The input tabular expression.
IdColumnstring✔️The name of the column with ID values that represent user activity.
TimelineColumnstring✔️The name of the column representing the timeline.
Startint, long, real, datetime, or timespan✔️The analysis start period.
Endint, long, real, datetime, or timespan✔️The analysis end period.
LookbackWindowint, long, real, datetime, or timespan✔️The lookback period. This value should be a multiple of the Bin value, otherwise the LookbackWindow will be rounded down to a multiple of the Bin value. For example, for dcount users in past 7d: LookbackWindow = 7d.
Binint, long, real, datetime, timespan, or string✔️The analysis step period. The possible string values are week, month, and year for which all periods will be startofweek, startofmonth, startofyear respectively.
dim1, dim2, …stringA list of the dimensions columns that slice the activity metrics calculation.

Returns

Returns a table that has the count and distinct count values of Ids in the lookback period, for each timeline period (by bin) and for each existing dimensions combination.

Output table schema is:

TimelineColumndim1..dim_ncountdcount
type: as of TimelineColumn......longlong

Example

Calculate counts and dcounts for users in past week, for each day in the analysis period.

let start = datetime(2017 - 08 - 01);
let end = datetime(2017 - 08 - 07); 
let lookbackWindow = 3d;  
let bin = 1d;
let T = datatable(UserId: string, Timestamp: datetime)
    [
    'Bob', datetime(2017 - 08 - 01), 
    'David', datetime(2017 - 08 - 01), 
    'David', datetime(2017 - 08 - 01), 
    'John', datetime(2017 - 08 - 01), 
    'Bob', datetime(2017 - 08 - 01), 
    'Ananda', datetime(2017 - 08 - 02),  
    'Atul', datetime(2017 - 08 - 02), 
    'John', datetime(2017 - 08 - 02), 
    'Ananda', datetime(2017 - 08 - 03), 
    'Atul', datetime(2017 - 08 - 03), 
    'Atul', datetime(2017 - 08 - 03), 
    'John', datetime(2017 - 08 - 03), 
    'Bob', datetime(2017 - 08 - 03), 
    'Betsy', datetime(2017 - 08 - 04), 
    'Bob', datetime(2017 - 08 - 05), 
];
T
| evaluate sliding_window_counts(UserId, Timestamp, start, end, lookbackWindow, bin)

Output

TimestampCountdcount
2017-08-01 00:00:00.000000053
2017-08-02 00:00:00.000000085
2017-08-03 00:00:00.0000000135
2017-08-04 00:00:00.000000095
2017-08-05 00:00:00.000000075
2017-08-06 00:00:00.000000022
2017-08-07 00:00:00.000000011

13 - User Analytics

This article describes User Analytics.

This section describes Kusto extensions (plugins) for user analytics scenarios.

ScenarioPluginDetailsUser Experience
Counting new users over timeactivity_counts_metricsReturns counts/dcounts/new counts for each time window. Each time window is compared to all previous time windowsKusto.Explorer: Report Gallery
Period-over-period: retention/churn rate and new usersactivity_metricsReturns dcount, retention/churn rate for each time window. Each time window is compared to previous time windowKusto.Explorer: Report Gallery
Users count and dcount over sliding windowsliding_window_countsFor each time window, returns count and dcount over a lookback period, in a sliding window manner
New-users cohort: retention/churn rate and new usersnew_activity_metricsCompares between cohorts of new users (all users that were first seen in time window). Each cohort is compared to all prior cohorts. Comparison takes into account all previous time windowsKusto.Explorer: Report Gallery
Active Users: distinct countsactive_users_countReturns distinct users for each time window. A user is only considered if it appears in at least X distinct periods in a specified lookback period.
User Engagement: DAU/WAU/MAUactivity_engagementCompares between an inner time window (for example, daily) and an outer (for example, weekly) for computing engagement (for example, DAU/WAU)Kusto.Explorer: Report Gallery
Sessions: count active sessionssession_countCounts sessions, where a session is defined by a time period - a user record is considered a new session, if it hasn’t been seen in the lookback period from current record
Funnels: previous and next state sequence analysisfunnel_sequenceCounts distinct users who have taken a sequence of events, and the previous or next events that led or were followed by the sequence. Useful for constructing sankey diagrams
Funnels: sequence completion analysisfunnel_sequence_completionComputes the distinct count of users that have completed a specified sequence in each time window

||||