|
0
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.Dataset.javaRDD(Dataset.scala:3270)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.CachePortImpl.<init>(CachePortImpl.java:66)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.OrderQuantityKpiProviderImpl.create(OrderQuantityKpiProviderImpl.java:38)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.OrderQuantityKpiProviderImpl.create(OrderQuantityKpiProviderImpl.java:11)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl$LifecycleControlledObjectHolder.<init>(CalculationObjectProviderImpl.java:23)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.lambda$lookupOrCreate$0(CalculationObjectProviderImpl.java:83)
java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(Unknown Source)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.lookupOrCreate(CalculationObjectProviderImpl.java:83)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.get(CalculationObjectProviderImpl.java:57)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.prepareStreams(KpiCalculationStreamingActivityImpl.java:47)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
|
2026/04/20 06:11:22
|
0.6 s
|
|
|
|
1
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.classic.DataFrameWriter.save(DataFrameWriter.scala:118)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.writeToStorage(FileStorageAdapterImpl.java:129)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.prepareKpiCurveUpdateStream(FileStorageAdapterImpl.java:118)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.prepareStreams(KpiCalculationStreamingActivityImpl.java:47)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:23
|
11 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
2
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.classic.DataFrameWriter.save(DataFrameWriter.scala:118)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.writeToStorage(FileStorageAdapterImpl.java:129)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.prepareKpiCurveUpdateStream(FileStorageAdapterImpl.java:118)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.prepareStreams(KpiCalculationStreamingActivityImpl.java:47)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:24
|
8 s
|
[0]
|
|
3
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:1090)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:1089)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:1086)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:1086)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.commands.WriteIntoDelta.writeAndReturnCommitData(WriteIntoDelta.scala:336)
org.apache.spark.sql.delta.commands.WriteIntoDelta.$anonfun$run$1(WriteIntoDelta.scala:109)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:247)
|
2026/04/20 06:11:32
|
0.6 s
|
[1]
|
|
4
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.OptimisticTransaction.recordOperation(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.OptimisticTransaction.recordDeltaOperation(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.OptimisticTransactionImpl.$anonfun$doCommitRetryIteratively$1(OptimisticTransaction.scala:2328)
org.apache.spark.sql.delta.OptimisticTransactionImpl.lockCommitIfEnabled(OptimisticTransaction.scala:2299)
org.apache.spark.sql.delta.OptimisticTransactionImpl.doCommitRetryIteratively(OptimisticTransaction.scala:2315)
org.apache.spark.sql.delta.OptimisticTransactionImpl.doCommitRetryIteratively$(OptimisticTransaction.scala:2311)
org.apache.spark.sql.delta.OptimisticTransaction.doCommitRetryIteratively(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.OptimisticTransactionImpl.$anonfun$commitImpl$1(OptimisticTransaction.scala:1515)
org.apache.spark.sql.delta.metering.DeltaLogging.recordFrameProfile(DeltaLogging.scala:171)
org.apache.spark.sql.delta.metering.DeltaLogging.recordFrameProfile$(DeltaLogging.scala:169)
org.apache.spark.sql.delta.OptimisticTransaction.recordFrameProfile(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
|
2026/04/20 06:11:33
|
0.2 s
|
[2]
|
|
|
5
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = PrepareStreams activityId = d00c14b1-1076-3b01-866a-3f58e692f0a1 workflowType = CalculateOrderProposalsWorkflow workflowId = 6331284e-72eb-32d1-9dda-4eb130ce3185 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.classic.DataFrameWriter.save(DataFrameWriter.scala:118)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.storePerformanceMetrics(FileStorageAdapterImpl.java:189)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.prepareStreams(KpiCalculationStreamingActivityImpl.java:50)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:34
|
5 s
|
[3][4][5][6][7]
|
|
|
6
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:40
|
0.5 s
|
|
|
|
7
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:41
|
7 s
|
[8]
|
|
|
8
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:48
|
82 ms
|
|
|
|
9
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:48
|
26 ms
|
|
|
|
10
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:49
|
16 ms
|
|
|
|
11
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:49
|
4 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
12
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:49
|
9 ms
|
|
|
13
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:50
|
0.5 s
|
[11]
|
|
14
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:11:50
|
3 s
|
[12][13][14][15][16]
|
|
|
16
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/04/20 06:14:25
|
2 s
|
[18]
|
|
|
17
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/04/20 06:14:27
|
0.2 s
|
[19]
|
|
|
18
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/04/20 06:14:28
|
0.2 s
|
[20]
|
|
|
19
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/04/20 06:14:28
|
0.4 s
|
[22]
|
|
|
20
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.util.threads.DeltaThreadPool.$anonfun$submit$1(DeltaThreadPool.scala:39)
java.base/java.util.concurrent.FutureTask.run(Unknown Source)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.$anonfun$run$1(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:77)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured$(SparkThreadLocalForwardingThreadPoolExecutor.scala:60)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:116)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.run(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:14:28
|
0.3 s
|
[21]
|
|
|
21
|
Delta: Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query: Compute snapshot for version: 5
+details
Delta: Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query: Compute snapshot for version: 5 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:14:30
|
0.9 s
|
[24]
|
|
|
22
|
Delta: Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query: Compute snapshot for version: 5
+details
Delta: Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query: Compute snapshot for version: 5 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:14:31
|
2 s
|
[25][26]
|
|
|
23
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:14:33
|
0.2 s
|
[27]
|
|
|
24
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:14:33
|
0.3 s
|
[28]
|
|
|
25
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:14:34
|
60 ms
|
|
|
|
26
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:14:34
|
2 s
|
[29]
|
|
|
27
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.classic.Dataset.isEmpty(Dataset.scala:559)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.InsightHandlerDataAccessImpl.mapExternalProductPlantIds(InsightHandlerDataAccessImpl.java:62)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.insights.InsightHandlerImpl.createUnknownDemandInsight(InsightHandlerImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.attachRawDemand(DemandSnapshotEnhancementImpl.java:65)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.withDemand(DemandSnapshotEnhancementImpl.java:35)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:56)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
|
2026/04/20 06:14:35
|
17 s
|
[30][31][32][33][34][35][36][37][38][39]
|
|
|
28
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.classic.Dataset.count(Dataset.scala:1499)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.InsightHandlerDataAccessImpl.mapExternalProductPlantIds(InsightHandlerDataAccessImpl.java:81)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.insights.InsightHandlerImpl.createUnknownDemandInsight(InsightHandlerImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.attachRawDemand(DemandSnapshotEnhancementImpl.java:65)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.withDemand(DemandSnapshotEnhancementImpl.java:35)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:56)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
|
2026/04/20 06:14:53
|
0.4 s
|
[40][41]
|
|
|
29
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.classic.Dataset.count(Dataset.scala:1499)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.InsightHandlerDataAccessImpl.mapExternalProductPlantIds(InsightHandlerDataAccessImpl.java:82)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.insights.InsightHandlerImpl.createUnknownDemandInsight(InsightHandlerImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.attachRawDemand(DemandSnapshotEnhancementImpl.java:65)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.withDemand(DemandSnapshotEnhancementImpl.java:35)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:56)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
|
2026/04/20 06:14:53
|
14 s
|
[43][45][46][47][48]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
30
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:14:54
|
21 ms
|
|
|
31
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:14:54
|
0.9 s
|
[42]
|
|
32
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PreparedDeltaFileIndex.matchingFiles(PrepareDeltaScan.scala:389)
org.apache.spark.sql.delta.files.TahoeFileIndex.listAddFiles(TahoeFileIndex.scala:111)
org.apache.spark.sql.delta.files.TahoeFileIndex.listFiles(TahoeFileIndex.scala:103)
org.apache.spark.sql.execution.FileSourceScanLike.selectedPartitions(DataSourceScanExec.scala:297)
org.apache.spark.sql.execution.FileSourceScanLike.selectedPartitions$(DataSourceScanExec.scala:288)
org.apache.spark.sql.execution.FileSourceScanExec.selectedPartitions$lzycompute(DataSourceScanExec.scala:607)
org.apache.spark.sql.execution.FileSourceScanExec.selectedPartitions(DataSourceScanExec.scala:607)
org.apache.spark.sql.execution.FileSourceScanLike.dynamicallySelectedPartitions(DataSourceScanExec.scala:330)
org.apache.spark.sql.execution.FileSourceScanLike.dynamicallySelectedPartitions$(DataSourceScanExec.scala:309)
|
2026/04/20 06:14:55
|
0.4 s
|
[44]
|
|
|
33
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/04/20 06:15:07
|
0.2 s
|
[49]
|
|
|
34
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/04/20 06:15:08
|
0.2 s
|
[50]
|
|
|
35
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:09
|
0.4 s
|
[51]
|
|
|
36
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:09
|
0.2 s
|
[52]
|
|
|
37
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:09
|
0.1 s
|
[53]
|
|
|
38
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:10
|
0.5 s
|
[54]
|
|
|
39
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:10
|
21 ms
|
|
|
|
40
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:10
|
1 s
|
[55]
|
|
|
41
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:12
|
0.3 s
|
[56]
|
|
|
42
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:12
|
20 ms
|
|
|
|
43
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:12
|
0.4 s
|
[57]
|
|
|
44
|
Delta: Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query: Compute snapshot for version: 4
+details
Delta: Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query: Compute snapshot for version: 4 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:13
|
0.6 s
|
[58]
|
|
|
45
|
Delta: Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query: Compute snapshot for version: 4
+details
Delta: Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query: Compute snapshot for version: 4 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:13
|
0.8 s
|
[59][60]
|
|
|
46
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:14
|
99 ms
|
[61]
|
|
|
47
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:14
|
0.2 s
|
[62]
|
|
|
48
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:15
|
17 ms
|
|
|
|
49
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:15
|
0.3 s
|
[63]
|
|
|
50
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.Dataset.first(Dataset.scala:2687)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ConfigurationSnapshotEnhancementImpl.fetchEnableLostSalesStockPeriod(ConfigurationSnapshotEnhancementImpl.java:41)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ConfigurationSnapshotEnhancementImpl.withConfiguration(ConfigurationSnapshotEnhancementImpl.java:33)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:58)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:15:15
|
0.6 s
|
[65][66]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
51
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:15
|
0.2 s
|
[64]
|
|
|
52
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.classic.Dataset.isEmpty(Dataset.scala:559)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.muListingCheck(ListingSnapshotEnhancementImpl.java:53)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.withListing(ListingSnapshotEnhancementImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:59)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:15:16
|
0.5 s
|
[68][69][70][71]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
53
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:16
|
0.1 s
|
[67]
|
|
|
54
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.Dataset.first(Dataset.scala:2687)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.muListingCheck(ListingSnapshotEnhancementImpl.java:54)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.withListing(ListingSnapshotEnhancementImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:59)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:15:16
|
0.5 s
|
[73][74]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
55
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:16
|
0.1 s
|
[72]
|
|
|
56
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.classic.DataFrameWriter.save(DataFrameWriter.scala:118)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.storeDataSnapshot(FileStorageAdapterImpl.java:262)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.DataSnapshotGenerationDataAccessImpl.storeDataSnapshot(DataSnapshotGenerationDataAccessImpl.java:39)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:62)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:15:17
|
1.8 min
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
57
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.classic.DataFrameWriter.save(DataFrameWriter.scala:118)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.storeDataSnapshot(FileStorageAdapterImpl.java:262)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.DataSnapshotGenerationDataAccessImpl.storeDataSnapshot(DataSnapshotGenerationDataAccessImpl.java:39)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:62)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:15:17
|
1.8 min
|
[82][83][84][85][86][87][88][89][90][91][92][93][94][95][96][97][98][99][100][101][102][103][104][105][106][107][108][109][110][111][112][113][114][115][116][117][118][119][120][121][122][123][124][125][126][127][128][129][130][131][132][133][134][135][136][137][138][139][140][141][142][143][144][145][146][147][148][149][150][151][152][153][154][155][156][157][158][159]
|
|
58
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:18
|
0.2 s
|
[75]
|
|
59
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:18
|
0.2 s
|
[76]
|
|
60
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:18
|
0.2 s
|
[77]
|
|
61
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:19
|
0.1 s
|
[78]
|
|
62
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:19
|
0.1 s
|
[79]
|
|
63
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:19
|
86 ms
|
[80]
|
|
64
|
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/04/20 06:15:19
|
0.1 s
|
[81]
|
|
65
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:1090)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:1089)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:1086)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:1086)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.commands.WriteIntoDelta.writeAndReturnCommitData(WriteIntoDelta.scala:336)
org.apache.spark.sql.delta.commands.WriteIntoDelta.$anonfun$run$1(WriteIntoDelta.scala:109)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:247)
|
2026/04/20 06:17:02
|
51 ms
|
[160]
|
|
66
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.OptimisticTransaction.recordOperation(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.OptimisticTransaction.recordDeltaOperation(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.OptimisticTransactionImpl.$anonfun$doCommitRetryIteratively$1(OptimisticTransaction.scala:2328)
org.apache.spark.sql.delta.OptimisticTransactionImpl.lockCommitIfEnabled(OptimisticTransaction.scala:2299)
org.apache.spark.sql.delta.OptimisticTransactionImpl.doCommitRetryIteratively(OptimisticTransaction.scala:2315)
org.apache.spark.sql.delta.OptimisticTransactionImpl.doCommitRetryIteratively$(OptimisticTransaction.scala:2311)
org.apache.spark.sql.delta.OptimisticTransaction.doCommitRetryIteratively(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.OptimisticTransactionImpl.$anonfun$commitImpl$1(OptimisticTransaction.scala:1515)
org.apache.spark.sql.delta.metering.DeltaLogging.recordFrameProfile(DeltaLogging.scala:171)
org.apache.spark.sql.delta.metering.DeltaLogging.recordFrameProfile$(DeltaLogging.scala:169)
org.apache.spark.sql.delta.OptimisticTransaction.recordFrameProfile(OptimisticTransaction.scala:169)
org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
|
2026/04/20 06:17:03
|
62 ms
|
[161]
|
|
|
67
|
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD
+details
replenishmentRunId = 10000000004 tenantId = 8793759653749352888 activityType = BufferDataSnapShot activityId = 36e0914c-7e4e-3a52-bc93-e306e07299ba workflowType = KpiPrepareDataSnapshotWorkflow workflowId = b49fb3c6-88b3-3c61-97e8-cbe3f98f9409 attempt = 1 cornerstoneTenantId = 8445 marketUnit = 10000_OP_PERF scenario = STANDARD org.apache.spark.sql.classic.DataFrameWriter.save(DataFrameWriter.scala:118)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.storePerformanceMetrics(FileStorageAdapterImpl.java:189)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.notifyShardFinished(CreateDataSnapshotForShardActivityImpl.java:58)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.DataSnapshotGenerationDataAccessImpl.storeDataSnapshot(DataSnapshotGenerationDataAccessImpl.java:40)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:62)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/04/20 06:17:04
|
3 s
|
[162][163][164][165][166]
|
|
|
72
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:17:15
|
8.4 min
|
[187][190][192]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
73
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:17:15
|
24 ms
|
|
|
74
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:17:15
|
8 s
|
[177]
|
|
75
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:17:23
|
0.7 s
|
[180]
|
|
76
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:17:25
|
10 s
|
[182][184][185]
|
|
77
|
Delta: stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1: Filtering files for query
+details
Delta: stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1: Filtering files for query org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:17:25
|
0.2 s
|
|
|
78
|
Delta: stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1: Filtering files for query
+details
Delta: stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1: Filtering files for query org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:17:25
|
43 ms
|
|
|
79
|
Delta: stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1: Filtering files for query
+details
Delta: stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1: Filtering files for query org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:17:25
|
2 s
|
[181]
|
|
80
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:17:27
|
0.6 s
|
[183]
|
|
81
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:18:47
|
0.2 s
|
|
|
82
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:20:40
|
19 ms
|
|
|
83
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:20:40
|
1.0 min
|
[191]
|
|
84
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:22:42
|
1.1 min
|
[193][194]
|
|
85
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:23:45
|
60 s
|
|
|
86
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:23:45
|
58 s
|
[195]
|
|
87
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 1 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:24:45
|
54 s
|
[196][197][198][199][200]
|
|
|
88
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:42:30
|
1.1 min
|
[208][209][211]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
89
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:42:30
|
8 ms
|
|
|
90
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:42:30
|
2 s
|
[201]
|
|
91
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:42:32
|
0.8 s
|
[202]
|
|
92
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:42:33
|
5 s
|
[204][206][207]
|
|
93
|
Delta: stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2: Filtering files for query
+details
Delta: stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2: Filtering files for query org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:42:33
|
0.2 s
|
[203]
|
|
94
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:42:34
|
0.4 s
|
[205]
|
|
95
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:42:41
|
65 ms
|
|
|
96
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:42:43
|
18 ms
|
|
|
97
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:42:43
|
0.2 s
|
[210]
|
|
98
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:43:31
|
2 s
|
[212]
|
|
99
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:43:33
|
2 s
|
|
|
100
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:43:33
|
1 s
|
[213]
|
|
101
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 2 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 06:43:35
|
3 s
|
[214][215][216][217][218]
|
|
|
102
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:07:50
|
1.1 min
|
[227][228][230]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
103
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:07:50
|
10 ms
|
|
|
104
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:07:50
|
1 s
|
[220]
|
|
105
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:07:51
|
0.3 s
|
[221]
|
|
106
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:07:52
|
4 s
|
[223][224][226]
|
|
107
|
Delta: stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3: Filtering files for query
+details
Delta: stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3: Filtering files for query org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:07:52
|
0.3 s
|
[222]
|
|
108
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:07:53
|
0.6 s
|
[225]
|
|
109
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:07:58
|
68 ms
|
|
|
110
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:08:01
|
12 ms
|
|
|
111
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:08:01
|
0.2 s
|
[229]
|
|
112
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:08:46
|
3 s
|
[231]
|
|
113
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:08:49
|
2 s
|
|
|
114
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:08:50
|
0.7 s
|
[232]
|
|
115
|
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3
+details
stream-10000000004-1
id = 0ca74851-d945-4eb8-90f7-ea42a7110a4d
runId = c871ced2-19a8-473b-bfe1-6d8b64ffe1bc
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/04/20 07:08:51
|
2 s
|
[233][234][235][236][237]
|
|