digraph G {
0 [labelType="html" label="<br><b>WriteToDataSourceV2</b><br><br>"];
subgraph cluster1 {
isCluster="true";
label="WholeStageCodegen (7)\n \nduration: total (min, med, max (stageId: taskId))\n51 ms (1 ms, 3 ms, 39 ms (stage 9.0: task 806))";
2 [labelType="html" label="<b>Sort</b><br><br>sort time total (min, med, max (stageId: taskId))<br>0 ms (0 ms, 0 ms, 0 ms (stage 9.0: task 805))<br>peak memory total (min, med, max (stageId: taskId))<br>256.3 MiB (64.0 KiB, 64.1 MiB, 64.1 MiB (stage 9.0: task 805))<br>spill size total (min, med, max (stageId: taskId))<br>0.0 B (0.0 B, 0.0 B, 0.0 B (stage 9.0: task 805))"];
}
3 [labelType="html" label="<b>Exchange</b><br><br>shuffle records written: 7<br>shuffle write time total (min, med, max (stageId: taskId))<br>11 ms (0 ms, 0 ms, 7 ms (stage 8.0: task 643))<br>records read: 7<br>local bytes read total (min, med, max (stageId: taskId))<br>492.0 B (0.0 B, 107.0 B, 172.0 B (stage 9.0: task 807))<br>fetch wait time total (min, med, max (stageId: taskId))<br>4 ms (0 ms, 0 ms, 4 ms (stage 9.0: task 807))<br>remote bytes read total (min, med, max (stageId: taskId))<br>174.0 B (0.0 B, 0.0 B, 174.0 B (stage 9.0: task 807))<br>local blocks read: 5<br>remote blocks read: 2<br>data size total (min, med, max (stageId: taskId))<br>456.0 B (0.0 B, 0.0 B, 72.0 B (stage 8.0: task 622))<br>shuffle bytes written total (min, med, max (stageId: taskId))<br>666.0 B (0.0 B, 0.0 B, 107.0 B (stage 8.0: task 741))"];
subgraph cluster4 {
isCluster="true";
label="WholeStageCodegen (6)\n \nduration: total (min, med, max (stageId: taskId))\n571 ms (0 ms, 0 ms, 25 ms (stage 6.0: task 454))";
5 [labelType="html" label="<b>HashAggregate</b><br><br>time in aggregation build total (min, med, max (stageId: taskId))<br>38 ms (0 ms, 0 ms, 21 ms (stage 6.0: task 454))<br>peak memory total (min, med, max (stageId: taskId))<br>996.0 MiB (256.0 KiB, 256.0 KiB, 64.3 MiB (stage 6.0: task 421))<br>number of output rows: 14<br>avg hash probe bucket list iters (min, med, max (stageId: taskId)):<br>(1, 1, 1 (stage 6.0: task 421))"];
}
6 [labelType="html" label="<b>StateStoreSave</b><br><br>number of total state rows: 14<br>memory used by state total (min, med, max (stageId: taskId))<br>121.7 KiB (232.0 B, 376.0 B, 600.0 B (stage 6.0: task 421))<br>count of cache hit on states cache in provider: 1,200<br>number of output rows: 14<br>estimated size of state only on current version total (min, med, max (stageId: taskId))<br>37.3 KiB (88.0 B, 88.0 B, 312.0 B (stage 6.0: task 421))<br>time to commit changes total (min, med, max (stageId: taskId))<br>15.7 s (21 ms, 29 ms, 95 ms (stage 6.0: task 582))<br>time to remove total (min, med, max (stageId: taskId))<br>0 ms (0 ms, 0 ms, 0 ms (stage 6.0: task 406))<br>number of updated state rows: 14<br>time to update total (min, med, max (stageId: taskId))<br>575 ms (0 ms, 0 ms, 167 ms (stage 6.0: task 454))"];
subgraph cluster7 {
isCluster="true";
label="WholeStageCodegen (5)\n \nduration: total (min, med, max (stageId: taskId))\n577 ms (0 ms, 0 ms, 167 ms (stage 6.0: task 454))";
8 [labelType="html" label="<b>HashAggregate</b><br><br>time in aggregation build total (min, med, max (stageId: taskId))<br>243 ms (0 ms, 0 ms, 130 ms (stage 6.0: task 454))<br>peak memory total (min, med, max (stageId: taskId))<br>996.0 MiB (256.0 KiB, 256.0 KiB, 64.3 MiB (stage 6.0: task 421))<br>number of output rows: 14<br>avg hash probe bucket list iters (min, med, max (stageId: taskId)):<br>(1, 1, 1 (stage 6.0: task 421))"];
}
9 [labelType="html" label="<b>StateStoreRestore</b><br><br>number of output rows: 14"];
subgraph cluster10 {
isCluster="true";
label="WholeStageCodegen (4)\n \nduration: total (min, med, max (stageId: taskId))\n1.3 s (1 ms, 1 ms, 143 ms (stage 6.0: task 454))";
11 [labelType="html" label="<b>HashAggregate</b><br><br>time in aggregation build total (min, med, max (stageId: taskId))<br>118 ms (0 ms, 0 ms, 36 ms (stage 6.0: task 441))<br>peak memory total (min, med, max (stageId: taskId))<br>996.0 MiB (256.0 KiB, 256.0 KiB, 64.3 MiB (stage 6.0: task 421))<br>number of output rows: 14<br>avg hash probe bucket list iters (min, med, max (stageId: taskId)):<br>(1, 1, 1 (stage 6.0: task 421))"];
}
12 [labelType="html" label="<b>Exchange</b><br><br>shuffle records written: 9<br>shuffle write time total (min, med, max (stageId: taskId))<br>27 ms (9 ms, 17 ms, 17 ms (stage 5.0: task 403))<br>records read: 18<br>local bytes read total (min, med, max (stageId: taskId))<br>984.0 B (0.0 B, 0.0 B, 107.0 B (stage 6.0: task 547))<br>fetch wait time total (min, med, max (stageId: taskId))<br>42 ms (0 ms, 0 ms, 23 ms (stage 6.0: task 441))<br>remote bytes read total (min, med, max (stageId: taskId))<br>776.0 B (0.0 B, 0.0 B, 107.0 B (stage 6.0: task 547))<br>local blocks read: 10<br>remote blocks read: 8<br>data size total (min, med, max (stageId: taskId))<br>600.0 B (216.0 B, 384.0 B, 384.0 B (stage 5.0: task 403))<br>shuffle bytes written total (min, med, max (stageId: taskId))<br>880.0 B (320.0 B, 560.0 B, 560.0 B (stage 5.0: task 403))"];
subgraph cluster13 {
isCluster="true";
label="WholeStageCodegen (3)\n \nduration: total (min, med, max (stageId: taskId))\n102 ms (30 ms, 72 ms, 72 ms (stage 5.0: task 403))";
14 [labelType="html" label="<b>HashAggregate</b><br><br>time in aggregation build total (min, med, max (stageId: taskId))<br>78 ms (21 ms, 57 ms, 57 ms (stage 5.0: task 403))<br>peak memory total (min, med, max (stageId: taskId))<br>128.5 MiB (64.3 MiB, 64.3 MiB, 64.3 MiB (stage 5.0: task 404))<br>number of output rows: 9<br>avg hash probe bucket list iters (min, med, max (stageId: taskId)):<br>(1, 1, 1 (stage 5.0: task 404))"];
15 [labelType="html" label="<br><b>Project</b><br><br>"];
16 [labelType="html" label="<b>Filter</b><br><br>number of output rows: 42"];
17 [labelType="html" label="<b>Expand</b><br><br>number of output rows: 42"];
18 [labelType="html" label="<br><b>Project</b><br><br>"];
}
19 [labelType="html" label="<br><b>EventTimeWatermark</b><br><br>"];
subgraph cluster20 {
isCluster="true";
label="WholeStageCodegen (2)\n \nduration: total (min, med, max (stageId: taskId))\n85 ms (24 ms, 61 ms, 61 ms (stage 5.0: task 403))";
21 [labelType="html" label="<br><b>Project</b><br><br>"];
}
22 [labelType="html" label="<b>Generate</b><br><br>number of output rows: 21"];
subgraph cluster23 {
isCluster="true";
label="WholeStageCodegen (1)\n \nduration: total (min, med, max (stageId: taskId))\n89 ms (25 ms, 64 ms, 64 ms (stage 5.0: task 403))";
24 [labelType="html" label="<br><b>Project</b><br><br>"];
}
25 [labelType="html" label="<b>MicroBatchScan</b><br><br>number of output rows: 19"];
2->0;
3->2;
5->3;
6->5;
8->6;
9->8;
11->9;
12->11;
14->12;
15->14;
16->15;
17->16;
18->17;
19->18;
21->19;
22->21;
24->22;
25->24;
}
26
WriteToDataSourceV2 org.apache.spark.sql.execution.streaming.sources.MicroBatchWrite@133afe90
Sort [total_bytes#68 DESC NULLS LAST], true, 0
WholeStageCodegen (7)
Exchange rangepartitioning(total_bytes#68 DESC NULLS LAST, 200), true, [id=#214]
HashAggregate(keys=[window#69-T3000ms, hostname#12], functions=[sum(cast(bytes#47 as double))])
WholeStageCodegen (6)
StateStoreSave [window#69-T3000ms, hostname#12], state info [ checkpoint = file:/tmp/spark-checkpoints/state, runId = e0c08a0a-5cf3-468b-888d-995173b23fbd, opId = 0, ver = 1, numPartitions = 200], Complete, 0, 2
HashAggregate(keys=[window#69-T3000ms, hostname#12], functions=[merge_sum(cast(bytes#47 as double))])
WholeStageCodegen (5)
StateStoreRestore [window#69-T3000ms, hostname#12], state info [ checkpoint = file:/tmp/spark-checkpoints/state, runId = e0c08a0a-5cf3-468b-888d-995173b23fbd, opId = 0, ver = 1, numPartitions = 200], 2
HashAggregate(keys=[window#69-T3000ms, hostname#12], functions=[merge_sum(cast(bytes#47 as double))])
WholeStageCodegen (4)
Exchange hashpartitioning(window#69-T3000ms, hostname#12, 200), true, [id=#202]
HashAggregate(keys=[window#69-T3000ms, hostname#12], functions=[partial_sum(cast(bytes#47 as double))])
Project [window#69-T3000ms, hostname#12, bytes#47]
Filter (((isnotnull(timestamp#1-T3000ms) AND isnotnull(window#69-T3000ms)) AND (timestamp#1-T3000ms >= window#69-T3000ms.start)) AND (timestamp#1-T3000ms < window#69-T3000ms.end))
Expand [ArrayBuffer(named_struct(start, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 0) - 2) * 30000000) + 0), LongType, TimestampType), end, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 0) - 2) * 30000000) + 60000000), LongType, TimestampType)), timestamp#1-T3000ms, hostname#12, bytes#47), ArrayBuffer(named_struct(start, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 1) - 2) * 30000000) + 0), LongType, TimestampType), end, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 1) - 2) * 30000000) + 60000000), LongType, TimestampType)), timestamp#1-T3000ms, hostname#12, bytes#47)], [window#69-T3000ms, timestamp#1-T3000ms, hostname#12, bytes#47]
Project [timestamp#1-T3000ms, split(logs#5, ,, -1)[1] AS hostname#12, split(logs#5, ,, -1)[6] AS bytes#47]
WholeStageCodegen (3)
EventTimeWatermark timestamp#1: timestamp, 3 seconds
Project [logs#5, timestamp#1]
WholeStageCodegen (2)
Generate explode(split(value#0, , -1)), [timestamp#1], false, [logs#5]
Project [value#0, timestamp#1]
WholeStageCodegen (1)
MicroBatchScan[value#0, timestamp#1] class org.apache.spark.sql.execution.streaming.sources.TextSocketTable$$anon$1
== Parsed Logical Plan ==
WriteToDataSourceV2 org.apache.spark.sql.execution.streaming.sources.MicroBatchWrite@133afe90
+- Sort [total_bytes#68 DESC NULLS LAST], true
+- Aggregate [window#69-T3000ms, hostname#12], [window#69-T3000ms AS window#57-T3000ms, hostname#12, sum(cast(bytes#47 as double)) AS total_bytes#68]
+- Filter ((timestamp#1-T3000ms >= window#69-T3000ms.start) AND (timestamp#1-T3000ms < window#69-T3000ms.end))
+- Expand [ArrayBuffer(named_struct(start, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) + cast(1 as bigint)) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) END + cast(0 as bigint)) - cast(2 as bigint)) * 30000000) + 0), LongType, TimestampType), end, precisetimestampconversion((((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) + cast(1 as bigint)) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) END + cast(0 as bigint)) - cast(2 as bigint)) * 30000000) + 0) + 60000000), LongType, TimestampType)), logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, url#30, responsecode#38, bytes#47), ArrayBuffer(named_struct(start, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) + cast(1 as bigint)) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) END + cast(1 as bigint)) - cast(2 as bigint)) * 30000000) + 0), LongType, TimestampType), end, precisetimestampconversion((((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) + cast(1 as bigint)) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) END + cast(1 as bigint)) - cast(2 as bigint)) * 30000000) + 0) + 60000000), LongType, TimestampType)), logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, url#30, responsecode#38, bytes#47)], [window#69-T3000ms, logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, url#30, responsecode#38, bytes#47]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, url#30, responsecode#38, split(logs#5, ,, -1)[6] AS bytes#47]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, url#30, split(logs#5, ,, -1)[5] AS responsecode#38]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, split(logs#5, ,, -1)[4] AS url#30]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, split(logs#5, ,, -1)[3] AS method#23]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, hostname#12, split(logs#5, ,, -1)[2] AS time#17]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, split(logs#5, ,, -1)[1] AS hostname#12]
+- Project [logs#5, timestamp#1-T3000ms, split(logs#5, ,, -1)[0] AS idx#8]
+- EventTimeWatermark timestamp#1: timestamp, 3 seconds
+- Project [logs#5, timestamp#1]
+- Generate explode(split(value#0, , -1)), false, [logs#5]
+- StreamingDataSourceV2Relation [value#0, timestamp#1], org.apache.spark.sql.execution.streaming.sources.TextSocketTable$$anon$1@64e6357e, TextSocketV2[host: stream-emulator.data-science-tools.svc.cluster.local, port: 5551], -1, 18
== Analyzed Logical Plan ==
WriteToDataSourceV2 org.apache.spark.sql.execution.streaming.sources.MicroBatchWrite@133afe90
+- Sort [total_bytes#68 DESC NULLS LAST], true
+- Aggregate [window#69-T3000ms, hostname#12], [window#69-T3000ms AS window#57-T3000ms, hostname#12, sum(cast(bytes#47 as double)) AS total_bytes#68]
+- Filter ((timestamp#1-T3000ms >= window#69-T3000ms.start) AND (timestamp#1-T3000ms < window#69-T3000ms.end))
+- Expand [ArrayBuffer(named_struct(start, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) + cast(1 as bigint)) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) END + cast(0 as bigint)) - cast(2 as bigint)) * 30000000) + 0), LongType, TimestampType), end, precisetimestampconversion((((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) + cast(1 as bigint)) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) END + cast(0 as bigint)) - cast(2 as bigint)) * 30000000) + 0) + 60000000), LongType, TimestampType)), logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, url#30, responsecode#38, bytes#47), ArrayBuffer(named_struct(start, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) + cast(1 as bigint)) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) END + cast(1 as bigint)) - cast(2 as bigint)) * 30000000) + 0), LongType, TimestampType), end, precisetimestampconversion((((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) + cast(1 as bigint)) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / cast(30000000 as double))) END + cast(1 as bigint)) - cast(2 as bigint)) * 30000000) + 0) + 60000000), LongType, TimestampType)), logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, url#30, responsecode#38, bytes#47)], [window#69-T3000ms, logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, url#30, responsecode#38, bytes#47]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, url#30, responsecode#38, split(logs#5, ,, -1)[6] AS bytes#47]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, url#30, split(logs#5, ,, -1)[5] AS responsecode#38]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, method#23, split(logs#5, ,, -1)[4] AS url#30]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, hostname#12, time#17, split(logs#5, ,, -1)[3] AS method#23]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, hostname#12, split(logs#5, ,, -1)[2] AS time#17]
+- Project [logs#5, timestamp#1-T3000ms, idx#8, split(logs#5, ,, -1)[1] AS hostname#12]
+- Project [logs#5, timestamp#1-T3000ms, split(logs#5, ,, -1)[0] AS idx#8]
+- EventTimeWatermark timestamp#1: timestamp, 3 seconds
+- Project [logs#5, timestamp#1]
+- Generate explode(split(value#0, , -1)), false, [logs#5]
+- StreamingDataSourceV2Relation [value#0, timestamp#1], org.apache.spark.sql.execution.streaming.sources.TextSocketTable$$anon$1@64e6357e, TextSocketV2[host: stream-emulator.data-science-tools.svc.cluster.local, port: 5551], -1, 18
== Optimized Logical Plan ==
WriteToDataSourceV2 org.apache.spark.sql.execution.streaming.sources.MicroBatchWrite@133afe90
+- Sort [total_bytes#68 DESC NULLS LAST], true
+- Aggregate [window#69-T3000ms, hostname#12], [window#69-T3000ms AS window#57-T3000ms, hostname#12, sum(cast(bytes#47 as double)) AS total_bytes#68]
+- Project [window#69-T3000ms, hostname#12, bytes#47]
+- Filter (((isnotnull(timestamp#1-T3000ms) AND isnotnull(window#69-T3000ms)) AND (timestamp#1-T3000ms >= window#69-T3000ms.start)) AND (timestamp#1-T3000ms < window#69-T3000ms.end))
+- Expand [ArrayBuffer(named_struct(start, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 0) - 2) * 30000000) + 0), LongType, TimestampType), end, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 0) - 2) * 30000000) + 60000000), LongType, TimestampType)), timestamp#1-T3000ms, hostname#12, bytes#47), ArrayBuffer(named_struct(start, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 1) - 2) * 30000000) + 0), LongType, TimestampType), end, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 1) - 2) * 30000000) + 60000000), LongType, TimestampType)), timestamp#1-T3000ms, hostname#12, bytes#47)], [window#69-T3000ms, timestamp#1-T3000ms, hostname#12, bytes#47]
+- Project [timestamp#1-T3000ms, split(logs#5, ,, -1)[1] AS hostname#12, split(logs#5, ,, -1)[6] AS bytes#47]
+- EventTimeWatermark timestamp#1: timestamp, 3 seconds
+- Project [logs#5, timestamp#1]
+- Generate explode(split(value#0, , -1)), [0], false, [logs#5]
+- StreamingDataSourceV2Relation [value#0, timestamp#1], org.apache.spark.sql.execution.streaming.sources.TextSocketTable$$anon$1@64e6357e, TextSocketV2[host: stream-emulator.data-science-tools.svc.cluster.local, port: 5551], -1, 18
== Physical Plan ==
WriteToDataSourceV2 org.apache.spark.sql.execution.streaming.sources.MicroBatchWrite@133afe90
+- *(7) Sort [total_bytes#68 DESC NULLS LAST], true, 0
+- Exchange rangepartitioning(total_bytes#68 DESC NULLS LAST, 200), true, [id=#214]
+- *(6) HashAggregate(keys=[window#69-T3000ms, hostname#12], functions=[sum(cast(bytes#47 as double))], output=[window#57-T3000ms, hostname#12, total_bytes#68])
+- StateStoreSave [window#69-T3000ms, hostname#12], state info [ checkpoint = file:/tmp/spark-checkpoints/state, runId = e0c08a0a-5cf3-468b-888d-995173b23fbd, opId = 0, ver = 1, numPartitions = 200], Complete, 0, 2
+- *(5) HashAggregate(keys=[window#69-T3000ms, hostname#12], functions=[merge_sum(cast(bytes#47 as double))], output=[window#69-T3000ms, hostname#12, sum#77])
+- StateStoreRestore [window#69-T3000ms, hostname#12], state info [ checkpoint = file:/tmp/spark-checkpoints/state, runId = e0c08a0a-5cf3-468b-888d-995173b23fbd, opId = 0, ver = 1, numPartitions = 200], 2
+- *(4) HashAggregate(keys=[window#69-T3000ms, hostname#12], functions=[merge_sum(cast(bytes#47 as double))], output=[window#69-T3000ms, hostname#12, sum#77])
+- Exchange hashpartitioning(window#69-T3000ms, hostname#12, 200), true, [id=#202]
+- *(3) HashAggregate(keys=[window#69-T3000ms, hostname#12], functions=[partial_sum(cast(bytes#47 as double))], output=[window#69-T3000ms, hostname#12, sum#77])
+- *(3) Project [window#69-T3000ms, hostname#12, bytes#47]
+- *(3) Filter (((isnotnull(timestamp#1-T3000ms) AND isnotnull(window#69-T3000ms)) AND (timestamp#1-T3000ms >= window#69-T3000ms.start)) AND (timestamp#1-T3000ms < window#69-T3000ms.end))
+- *(3) Expand [ArrayBuffer(named_struct(start, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 0) - 2) * 30000000) + 0), LongType, TimestampType), end, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 0) - 2) * 30000000) + 60000000), LongType, TimestampType)), timestamp#1-T3000ms, hostname#12, bytes#47), ArrayBuffer(named_struct(start, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 1) - 2) * 30000000) + 0), LongType, TimestampType), end, precisetimestampconversion(((((CASE WHEN (cast(CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) as double) = (cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) THEN (CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) + 1) ELSE CEIL((cast((precisetimestampconversion(timestamp#1-T3000ms, TimestampType, LongType) - 0) as double) / 3.0E7)) END + 1) - 2) * 30000000) + 60000000), LongType, TimestampType)), timestamp#1-T3000ms, hostname#12, bytes#47)], [window#69-T3000ms, timestamp#1-T3000ms, hostname#12, bytes#47]
+- *(3) Project [timestamp#1-T3000ms, split(logs#5, ,, -1)[1] AS hostname#12, split(logs#5, ,, -1)[6] AS bytes#47]
+- EventTimeWatermark timestamp#1: timestamp, 3 seconds
+- *(2) Project [logs#5, timestamp#1]
+- Generate explode(split(value#0, , -1)), [timestamp#1], false, [logs#5]
+- *(1) Project [value#0, timestamp#1]
+- MicroBatchScan[value#0, timestamp#1] class org.apache.spark.sql.execution.streaming.sources.TextSocketTable$$anon$1