|
20681
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 09:14:33
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20683
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 09:14:33
|
0.1 s
|
[45864]
|
|
20684
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 09:14:33
|
0.1 s
|
[45865]
|
|
20685
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 09:14:34
|
0.1 s
|
[45866]
|
|
|
20692
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 09:26:47
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20693
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 09:26:47
|
0.1 s
|
[45883]
|
|
20694
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 09:26:47
|
0.1 s
|
[45884]
|
|
20695
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 09:26:48
|
0.1 s
|
[45885]
|
|
20696
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 09:26:48
|
0.1 s
|
[45886]
|
|
|
20703
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 09:40:53
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20704
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 09:40:53
|
0.1 s
|
[45903]
|
|
20705
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 09:40:54
|
0.1 s
|
[45904]
|
|
20706
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 09:40:54
|
0.1 s
|
[45905]
|
|
20707
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 09:40:54
|
0.1 s
|
[45906]
|
|
|
20714
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 10:00:01
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20715
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 10:00:01
|
0.1 s
|
[45923]
|
|
20716
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:00:02
|
0.1 s
|
[45924]
|
|
20717
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:00:02
|
0.1 s
|
[45925]
|
|
20718
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:00:02
|
0.1 s
|
[45926]
|
|
|
20725
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 10:06:16
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20726
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 10:06:17
|
0.1 s
|
[45943]
|
|
20727
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:06:17
|
0.1 s
|
[45944]
|
|
20728
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:06:17
|
98 ms
|
[45945]
|
|
20729
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:06:17
|
99 ms
|
[45946]
|
|
|
20736
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 10:11:27
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20737
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 10:11:27
|
0.1 s
|
[45963]
|
|
20738
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:11:27
|
0.1 s
|
[45964]
|
|
20739
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:11:28
|
91 ms
|
[45965]
|
|
20740
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:11:28
|
97 ms
|
[45966]
|
|
|
20747
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 10:23:41
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20748
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 10:23:41
|
0.1 s
|
[45983]
|
|
20749
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:23:42
|
0.1 s
|
[45984]
|
|
20750
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:23:42
|
0.1 s
|
[45985]
|
|
20751
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:23:42
|
0.1 s
|
[45986]
|
|
|
20758
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 10:37:55
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20759
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 10:37:55
|
0.1 s
|
[46003]
|
|
20760
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:37:55
|
0.1 s
|
[46004]
|
|
20761
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:37:56
|
0.1 s
|
[46005]
|
|
20762
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:37:56
|
0.1 s
|
[46006]
|
|
|
20769
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 10:45:00
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20770
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 10:45:00
|
0.1 s
|
[46023]
|
|
20771
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:45:01
|
0.1 s
|
[46024]
|
|
20772
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:45:01
|
0.1 s
|
[46025]
|
|
20773
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:45:01
|
0.1 s
|
[46026]
|
|
|
20780
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 10:51:15
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20781
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 10:51:15
|
0.1 s
|
[46043]
|
|
20782
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:51:16
|
98 ms
|
[46044]
|
|
20783
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:51:16
|
95 ms
|
[46045]
|
|
20784
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:51:16
|
0.1 s
|
[46046]
|
|
|
20791
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 10:59:30
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20792
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 10:59:30
|
0.1 s
|
[46063]
|
|
20793
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:59:30
|
0.1 s
|
[46064]
|
|
20794
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:59:30
|
0.1 s
|
[46065]
|
|
20795
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 10:59:31
|
0.1 s
|
[46066]
|
|
|
20802
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 11:11:44
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20803
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 11:11:44
|
0.1 s
|
[46083]
|
|
20804
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:11:44
|
0.1 s
|
[46084]
|
|
20805
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:11:44
|
0.1 s
|
[46085]
|
|
20806
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:11:45
|
0.1 s
|
[46086]
|
|
|
20813
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 11:22:40
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20814
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 11:22:40
|
0.1 s
|
[46103]
|
|
20815
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:22:40
|
0.1 s
|
[46104]
|
|
20816
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:22:41
|
0.1 s
|
[46105]
|
|
20817
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:22:41
|
0.1 s
|
[46106]
|
|
|
20824
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 11:30:01
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20825
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 11:30:01
|
0.1 s
|
[46123]
|
|
20826
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:30:01
|
98 ms
|
[46124]
|
|
20827
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:30:02
|
0.1 s
|
[46125]
|
|
20828
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:30:02
|
0.1 s
|
[46126]
|
|
|
20835
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 11:36:15
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20836
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 11:36:15
|
0.1 s
|
[46143]
|
|
20837
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:36:15
|
0.1 s
|
[46144]
|
|
20838
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:36:15
|
95 ms
|
[46145]
|
|
20839
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:36:16
|
95 ms
|
[46146]
|
|
|
20846
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 11:42:06
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20847
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 11:42:07
|
0.1 s
|
[46163]
|
|
20848
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:42:07
|
93 ms
|
[46164]
|
|
20849
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:42:07
|
0.1 s
|
[46165]
|
|
20850
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:42:07
|
97 ms
|
[46166]
|
|
|
20857
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 11:54:21
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20858
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 11:54:21
|
0.1 s
|
[46183]
|
|
20859
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:54:21
|
0.1 s
|
[46184]
|
|
20860
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:54:21
|
0.1 s
|
[46185]
|
|
20861
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 11:54:22
|
0.1 s
|
[46186]
|
|
|
20868
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 12:08:35
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20869
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 12:08:35
|
0.1 s
|
[46203]
|
|
20870
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:08:35
|
0.1 s
|
[46204]
|
|
20871
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:08:35
|
0.1 s
|
[46205]
|
|
20872
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:08:36
|
0.1 s
|
[46206]
|
|
|
20879
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 12:15:00
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20880
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 12:15:01
|
0.1 s
|
[46223]
|
|
20881
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:15:01
|
0.1 s
|
[46224]
|
|
20882
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:15:01
|
96 ms
|
[46225]
|
|
20883
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:15:02
|
0.1 s
|
[46226]
|
|
|
20890
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 12:21:14
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20891
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 12:21:14
|
0.1 s
|
[46243]
|
|
20892
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:21:15
|
0.1 s
|
[46244]
|
|
20893
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:21:15
|
0.1 s
|
[46245]
|
|
20894
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:21:15
|
0.1 s
|
[46246]
|
|
|
20901
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 12:29:29
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20902
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 12:29:29
|
0.1 s
|
[46263]
|
|
20903
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:29:29
|
0.1 s
|
[46264]
|
|
20904
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:29:30
|
0.1 s
|
[46265]
|
|
20905
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:29:30
|
0.1 s
|
[46266]
|
|
|
20912
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 12:41:43
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20913
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 12:41:43
|
0.1 s
|
[46283]
|
|
20914
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:41:43
|
0.2 s
|
[46284]
|
|
20915
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:41:43
|
0.1 s
|
[46285]
|
|
20916
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:41:44
|
0.1 s
|
[46286]
|
|
|
20923
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 12:52:47
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20924
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 12:52:47
|
0.1 s
|
[46303]
|
|
20925
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:52:47
|
0.1 s
|
[46304]
|
|
20926
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:52:48
|
0.1 s
|
[46305]
|
|
20927
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 12:52:48
|
0.1 s
|
[46306]
|
|
|
20934
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 13:00:01
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20935
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 13:00:01
|
0.1 s
|
[46323]
|
|
20936
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:00:01
|
0.1 s
|
[46324]
|
|
20937
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:00:01
|
0.1 s
|
[46325]
|
|
20938
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:00:02
|
0.1 s
|
[46326]
|
|
|
20945
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 13:06:14
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20946
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 13:06:14
|
0.1 s
|
[46343]
|
|
20947
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:06:15
|
0.1 s
|
[46344]
|
|
20948
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:06:15
|
93 ms
|
[46345]
|
|
20949
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:06:15
|
95 ms
|
[46346]
|
|
|
20956
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 13:14:28
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20957
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 13:14:28
|
0.1 s
|
[46363]
|
|
20958
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:14:28
|
0.1 s
|
[46364]
|
|
20959
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:14:29
|
0.1 s
|
[46365]
|
|
20960
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:14:29
|
0.1 s
|
[46366]
|
|
|
20967
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 13:24:40
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20968
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 13:24:40
|
0.1 s
|
[46383]
|
|
20969
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:24:40
|
0.1 s
|
[46384]
|
|
20970
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:24:41
|
0.1 s
|
[46385]
|
|
20971
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:24:41
|
0.1 s
|
[46386]
|
|
|
20978
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 13:38:54
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20979
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 13:38:54
|
0.1 s
|
[46403]
|
|
20980
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:38:54
|
0.1 s
|
[46404]
|
|
20981
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:38:55
|
0.1 s
|
[46405]
|
|
20982
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:38:55
|
0.1 s
|
[46406]
|
|
|
20989
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 13:45:00
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
20990
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 13:45:01
|
0.1 s
|
[46423]
|
|
20991
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:45:01
|
0.1 s
|
[46424]
|
|
20992
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:45:01
|
97 ms
|
[46425]
|
|
20993
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:45:01
|
0.1 s
|
[46426]
|
|
|
21000
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 13:51:14
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21001
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 13:51:15
|
0.1 s
|
[46443]
|
|
21002
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:51:15
|
0.1 s
|
[46444]
|
|
21003
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:51:15
|
0.1 s
|
[46445]
|
|
21004
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:51:15
|
0.1 s
|
[46446]
|
|
|
21011
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 13:59:28
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21012
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 13:59:28
|
0.1 s
|
[46463]
|
|
21013
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:59:28
|
0.1 s
|
[46464]
|
|
21014
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:59:29
|
0.1 s
|
[46465]
|
|
21015
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 13:59:29
|
0.1 s
|
[46466]
|
|
|
21022
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 14:11:42
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21023
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 14:11:42
|
0.2 s
|
[46483]
|
|
21024
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:11:42
|
0.1 s
|
[46484]
|
|
21025
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:11:43
|
0.1 s
|
[46485]
|
|
21026
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:11:43
|
0.1 s
|
[46486]
|
|
|
21033
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 14:25:56
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21034
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 14:25:56
|
0.1 s
|
[46503]
|
|
21035
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:25:57
|
0.1 s
|
[46504]
|
|
21036
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:25:57
|
0.1 s
|
[46505]
|
|
21037
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:25:57
|
0.1 s
|
[46506]
|
|
|
21044
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 14:45:00
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21045
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 14:45:01
|
0.1 s
|
[46523]
|
|
21046
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:45:01
|
0.1 s
|
[46524]
|
|
21047
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:45:01
|
0.1 s
|
[46525]
|
|
21048
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:45:01
|
0.1 s
|
[46526]
|
|
|
21055
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 14:51:16
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21056
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 14:51:16
|
0.1 s
|
[46543]
|
|
21057
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:51:16
|
0.1 s
|
[46544]
|
|
21058
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:51:16
|
95 ms
|
[46545]
|
|
21059
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:51:17
|
92 ms
|
[46546]
|
|
|
21066
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 14:59:29
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21067
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 14:59:29
|
0.1 s
|
[46563]
|
|
21068
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:59:30
|
0.1 s
|
[46564]
|
|
21069
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:59:30
|
0.1 s
|
[46565]
|
|
21070
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 14:59:30
|
0.1 s
|
[46566]
|
|
|
21077
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 15:11:43
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21078
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 15:11:43
|
0.1 s
|
[46583]
|
|
21079
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:11:43
|
0.1 s
|
[46584]
|
|
21080
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:11:43
|
0.1 s
|
[46585]
|
|
21081
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:11:44
|
0.1 s
|
[46586]
|
|
|
21088
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 15:25:57
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21089
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 15:25:57
|
0.1 s
|
[46603]
|
|
21090
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:25:57
|
0.1 s
|
[46604]
|
|
21091
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:25:57
|
0.1 s
|
[46605]
|
|
21092
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:25:58
|
0.1 s
|
[46606]
|
|
|
21099
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 15:30:01
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21100
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 15:30:01
|
0.1 s
|
[46623]
|
|
21101
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:30:01
|
0.1 s
|
[46624]
|
|
21102
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:30:01
|
0.1 s
|
[46625]
|
|
21103
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:30:02
|
0.1 s
|
[46626]
|
|
|
21110
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 15:36:17
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21111
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 15:36:17
|
0.1 s
|
[46643]
|
|
21112
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:36:18
|
0.1 s
|
[46644]
|
|
21113
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:36:18
|
0.1 s
|
[46645]
|
|
21114
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:36:18
|
99 ms
|
[46646]
|
|
|
21121
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 15:44:31
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21122
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 15:44:31
|
0.1 s
|
[46663]
|
|
21123
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:44:31
|
0.1 s
|
[46664]
|
|
21124
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:44:31
|
0.1 s
|
[46665]
|
|
21125
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:44:32
|
0.1 s
|
[46666]
|
|
|
21132
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 15:56:45
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21133
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 15:56:45
|
0.1 s
|
[46683]
|
|
21134
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:56:45
|
0.1 s
|
[46684]
|
|
21135
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:56:46
|
0.1 s
|
[46685]
|
|
21136
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 15:56:46
|
0.1 s
|
[46686]
|
|
|
21143
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 16:10:59
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21144
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 16:10:59
|
0.1 s
|
[46703]
|
|
21145
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:10:59
|
0.1 s
|
[46704]
|
|
21146
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:11:00
|
0.1 s
|
[46705]
|
|
21147
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:11:00
|
0.1 s
|
[46706]
|
|
|
21154
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 16:15:01
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21155
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 16:15:01
|
97 ms
|
[46723]
|
|
21156
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:15:01
|
0.1 s
|
[46724]
|
|
21157
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:15:01
|
90 ms
|
[46725]
|
|
21158
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:15:02
|
88 ms
|
[46726]
|
|
|
21165
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 16:18:18
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21166
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 16:18:18
|
0.1 s
|
[46743]
|
|
21167
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:18:18
|
98 ms
|
[46744]
|
|
21168
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:18:18
|
91 ms
|
[46745]
|
|
21169
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:18:18
|
94 ms
|
[46746]
|
|
|
21176
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 16:26:32
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21177
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 16:26:32
|
0.1 s
|
[46763]
|
|
21178
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:26:32
|
0.1 s
|
[46764]
|
|
21179
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:26:32
|
0.1 s
|
[46765]
|
|
21180
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:26:33
|
0.1 s
|
[46766]
|
|
|
21187
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 16:38:46
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21188
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 16:38:46
|
0.1 s
|
[46783]
|
|
21189
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:38:46
|
0.1 s
|
[46784]
|
|
21190
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:38:46
|
0.1 s
|
[46785]
|
|
21191
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:38:47
|
0.1 s
|
[46786]
|
|
|
21198
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 16:52:59
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21199
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 16:52:59
|
0.2 s
|
[46803]
|
|
21200
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:53:00
|
0.1 s
|
[46804]
|
|
21201
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:53:00
|
0.1 s
|
[46805]
|
|
21202
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 16:53:00
|
0.1 s
|
[46806]
|
|
|
21209
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 17:00:01
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21210
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 17:00:01
|
0.1 s
|
[46823]
|
|
21211
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:00:02
|
98 ms
|
[46824]
|
|
21212
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:00:02
|
0.1 s
|
[46825]
|
|
21213
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:00:02
|
95 ms
|
[46826]
|
|
|
21220
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 17:06:15
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21221
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 17:06:15
|
0.1 s
|
[46843]
|
|
21222
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:06:16
|
92 ms
|
[46844]
|
|
21223
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:06:16
|
97 ms
|
[46845]
|
|
21224
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:06:16
|
93 ms
|
[46846]
|
|
|
21231
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 17:14:29
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21232
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 17:14:29
|
0.2 s
|
[46863]
|
|
21233
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:14:29
|
0.1 s
|
[46864]
|
|
21234
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:14:29
|
0.1 s
|
[46865]
|
|
21235
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:14:30
|
0.1 s
|
[46866]
|
|
|
21242
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 17:26:43
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21243
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 17:26:43
|
0.1 s
|
[46883]
|
|
21244
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:26:43
|
0.1 s
|
[46884]
|
|
21245
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:26:43
|
0.1 s
|
[46885]
|
|
21246
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:26:44
|
0.1 s
|
[46886]
|
|
|
21253
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 17:40:57
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21254
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 17:40:57
|
0.1 s
|
[46903]
|
|
21255
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:40:57
|
0.1 s
|
[46904]
|
|
21256
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:40:57
|
0.1 s
|
[46905]
|
|
21257
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 17:40:58
|
0.1 s
|
[46906]
|
|
|
21264
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 18:00:01
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21265
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 18:00:01
|
0.1 s
|
[46923]
|
|
21266
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:00:01
|
0.1 s
|
[46924]
|
|
21267
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:00:02
|
0.1 s
|
[46925]
|
|
21268
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:00:02
|
0.1 s
|
[46926]
|
|
|
21275
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 18:02:59
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21276
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 18:02:59
|
95 ms
|
[46943]
|
|
21277
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:02:59
|
92 ms
|
[46944]
|
|
21278
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:02:59
|
94 ms
|
[46945]
|
|
21279
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:03:00
|
93 ms
|
[46946]
|
|
|
21286
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 18:08:04
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21287
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 18:08:04
|
0.1 s
|
[46963]
|
|
21288
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:08:04
|
94 ms
|
[46964]
|
|
21289
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:08:04
|
96 ms
|
[46965]
|
|
21290
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:08:05
|
0.1 s
|
[46966]
|
|
|
21297
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 18:20:18
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21298
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 18:20:18
|
0.1 s
|
[46983]
|
|
21299
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:20:18
|
0.1 s
|
[46984]
|
|
21300
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:20:18
|
0.1 s
|
[46985]
|
|
21301
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:20:19
|
0.1 s
|
[46986]
|
|
|
21308
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 18:34:31
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21309
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 18:34:31
|
0.1 s
|
[47003]
|
|
21310
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:34:31
|
0.1 s
|
[47004]
|
|
21311
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:34:32
|
0.1 s
|
[47005]
|
|
21312
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:34:32
|
0.1 s
|
[47006]
|
|
|
21319
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 18:45:00
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21320
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 18:45:01
|
0.1 s
|
[47023]
|
|
21321
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:45:01
|
0.1 s
|
[47024]
|
|
21322
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:45:01
|
0.1 s
|
[47025]
|
|
21323
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:45:02
|
0.1 s
|
[47026]
|
|
|
21330
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 18:51:15
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21331
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 18:51:15
|
0.1 s
|
[47043]
|
|
21332
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:51:16
|
96 ms
|
[47044]
|
|
21333
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:51:16
|
94 ms
|
[47045]
|
|
21334
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:51:16
|
94 ms
|
[47046]
|
|
|
21341
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 18:59:28
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21342
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 18:59:28
|
0.2 s
|
[47063]
|
|
21343
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:59:29
|
0.1 s
|
[47064]
|
|
21344
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:59:29
|
0.1 s
|
[47065]
|
|
21345
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 18:59:29
|
0.1 s
|
[47066]
|
|
|
21352
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 19:11:42
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21353
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 19:11:42
|
0.1 s
|
[47083]
|
|
21354
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:11:42
|
0.1 s
|
[47084]
|
|
21355
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:11:43
|
0.1 s
|
[47085]
|
|
21356
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:11:43
|
0.1 s
|
[47086]
|
|
|
21363
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 19:25:56
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21364
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 19:25:56
|
0.1 s
|
[47103]
|
|
21365
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:25:57
|
0.1 s
|
[47104]
|
|
21366
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:25:57
|
0.1 s
|
[47105]
|
|
21367
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:25:57
|
0.1 s
|
[47106]
|
|
|
21374
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 19:45:00
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21375
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 19:45:01
|
0.1 s
|
[47123]
|
|
21376
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:45:01
|
0.2 s
|
[47124]
|
|
21377
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:45:01
|
0.1 s
|
[47125]
|
|
21378
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:45:02
|
0.1 s
|
[47126]
|
|
|
21385
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 19:51:16
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21386
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 19:51:16
|
0.1 s
|
[47143]
|
|
21387
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:51:16
|
0.1 s
|
[47144]
|
|
21388
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:51:16
|
0.1 s
|
[47145]
|
|
21389
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:51:17
|
97 ms
|
[47146]
|
|
|
21396
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 19:59:30
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21397
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 19:59:30
|
0.1 s
|
[47163]
|
|
21398
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:59:31
|
0.1 s
|
[47164]
|
|
21399
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:59:31
|
0.1 s
|
[47165]
|
|
21400
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 19:59:31
|
0.1 s
|
[47166]
|
|
|
21407
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 20:10:43
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21408
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 20:10:43
|
0.1 s
|
[47183]
|
|
21409
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:10:43
|
0.1 s
|
[47184]
|
|
21410
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:10:44
|
0.1 s
|
[47185]
|
|
21411
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:10:44
|
0.1 s
|
[47186]
|
|
|
21418
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 20:21:42
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21419
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 20:21:42
|
0.1 s
|
[47203]
|
|
21420
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:21:42
|
0.1 s
|
[47204]
|
|
21421
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:21:43
|
0.1 s
|
[47205]
|
|
21422
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:21:43
|
0.1 s
|
[47206]
|
|
|
21429
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 20:30:01
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21430
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 20:30:01
|
0.1 s
|
[47223]
|
|
21431
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:30:01
|
0.1 s
|
[47224]
|
|
21432
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:30:01
|
0.1 s
|
[47225]
|
|
21433
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:30:02
|
0.1 s
|
[47226]
|
|
|
21440
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 20:36:17
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21441
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 20:36:17
|
0.1 s
|
[47243]
|
|
21442
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:36:18
|
96 ms
|
[47244]
|
|
21443
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:36:18
|
0.1 s
|
[47245]
|
|
21444
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:36:18
|
97 ms
|
[47246]
|
|
|
21451
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 20:44:31
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21452
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 20:44:31
|
0.1 s
|
[47263]
|
|
21453
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:44:31
|
0.1 s
|
[47264]
|
|
21454
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:44:32
|
0.2 s
|
[47265]
|
|
21455
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:44:32
|
0.1 s
|
[47266]
|
|
|
21462
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 20:54:07
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21463
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 20:54:07
|
0.1 s
|
[47283]
|
|
21464
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:54:08
|
0.1 s
|
[47284]
|
|
21465
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:54:08
|
0.1 s
|
[47285]
|
|
21466
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 20:54:08
|
0.1 s
|
[47286]
|
|
|
21473
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 21:08:22
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21474
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 21:08:22
|
0.1 s
|
[47303]
|
|
21475
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:08:22
|
0.1 s
|
[47304]
|
|
21476
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:08:23
|
0.1 s
|
[47305]
|
|
21477
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:08:23
|
0.1 s
|
[47306]
|
|
|
21484
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 21:15:00
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21485
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 21:15:01
|
0.1 s
|
[47323]
|
|
21486
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:15:01
|
0.1 s
|
[47324]
|
|
21487
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:15:01
|
97 ms
|
[47325]
|
|
21488
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:15:01
|
94 ms
|
[47326]
|
|
|
21495
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 21:20:01
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21496
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 21:20:01
|
0.1 s
|
[47343]
|
|
21497
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:20:01
|
91 ms
|
[47344]
|
|
21498
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:20:02
|
92 ms
|
[47345]
|
|
21499
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:20:02
|
94 ms
|
[47346]
|
|
|
21506
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 21:28:15
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21507
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 21:28:15
|
0.1 s
|
[47363]
|
|
21508
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:28:16
|
0.1 s
|
[47364]
|
|
21509
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:28:16
|
0.1 s
|
[47365]
|
|
21510
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:28:16
|
0.1 s
|
[47366]
|
|
|
21517
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 21:40:29
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21518
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 21:40:29
|
0.1 s
|
[47383]
|
|
21519
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:40:29
|
0.1 s
|
[47384]
|
|
21520
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:40:29
|
0.1 s
|
[47385]
|
|
21521
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:40:30
|
0.1 s
|
[47386]
|
|
|
21528
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 21:54:42
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21529
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 21:54:42
|
0.1 s
|
[47403]
|
|
21530
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:54:43
|
0.1 s
|
[47404]
|
|
21531
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:54:43
|
0.1 s
|
[47405]
|
|
21532
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 21:54:43
|
0.1 s
|
[47406]
|
|
|
21539
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 22:00:01
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21540
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 22:00:01
|
0.1 s
|
[47423]
|
|
21541
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:00:01
|
0.1 s
|
[47424]
|
|
21542
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:00:02
|
94 ms
|
[47425]
|
|
21543
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:00:02
|
88 ms
|
[47426]
|
|
|
21550
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 22:06:15
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21551
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 22:06:15
|
0.1 s
|
[47443]
|
|
21552
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:06:15
|
0.1 s
|
[47444]
|
|
21553
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:06:16
|
0.1 s
|
[47445]
|
|
21554
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:06:16
|
96 ms
|
[47446]
|
|
|
21561
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 22:14:29
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21562
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 22:14:29
|
0.1 s
|
[47463]
|
|
21563
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:14:30
|
0.1 s
|
[47464]
|
|
21564
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:14:30
|
0.1 s
|
[47465]
|
|
21565
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:14:30
|
0.1 s
|
[47466]
|
|
|
21572
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 22:26:48
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21573
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 22:26:48
|
0.1 s
|
[47483]
|
|
21574
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:26:49
|
0.1 s
|
[47484]
|
|
21575
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:26:49
|
0.1 s
|
[47485]
|
|
21576
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:26:49
|
0.1 s
|
[47486]
|
|
|
21583
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 22:41:02
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21584
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 22:41:03
|
0.1 s
|
[47503]
|
|
21585
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:41:03
|
0.1 s
|
[47504]
|
|
21586
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:41:03
|
0.1 s
|
[47505]
|
|
21587
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 22:41:03
|
0.1 s
|
[47506]
|
|
|
21594
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 23:00:01
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21595
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 23:00:01
|
0.1 s
|
[47523]
|
|
21596
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:00:01
|
0.1 s
|
[47524]
|
|
21597
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:00:01
|
0.1 s
|
[47525]
|
|
21598
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:00:02
|
0.1 s
|
[47526]
|
|
|
21605
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 23:06:15
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21606
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 23:06:15
|
0.1 s
|
[47543]
|
|
21607
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:06:16
|
94 ms
|
[47544]
|
|
21608
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:06:16
|
91 ms
|
[47545]
|
|
21609
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:06:16
|
0.1 s
|
[47546]
|
|
|
21616
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 23:14:29
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21617
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 23:14:30
|
0.1 s
|
[47563]
|
|
21618
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:14:30
|
0.1 s
|
[47564]
|
|
21619
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:14:30
|
0.1 s
|
[47565]
|
|
21620
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:14:30
|
0.1 s
|
[47566]
|
|
|
21627
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 23:26:43
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21628
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 23:26:44
|
0.1 s
|
[47583]
|
|
21629
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:26:44
|
0.1 s
|
[47584]
|
|
21630
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:26:44
|
0.1 s
|
[47585]
|
|
21631
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:26:45
|
0.1 s
|
[47586]
|
|
|
21638
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/01 23:40:59
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21639
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/01 23:40:59
|
0.1 s
|
[47603]
|
|
21640
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:40:59
|
0.1 s
|
[47604]
|
|
21641
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:41:00
|
0.1 s
|
[47605]
|
|
21642
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/01 23:41:00
|
0.1 s
|
[47606]
|
|
|
21649
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/02 00:00:12
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21650
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/02 00:00:12
|
0.1 s
|
[47623]
|
|
21651
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/02 00:00:12
|
0.1 s
|
[47624]
|
|
21652
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/02 00:00:12
|
0.1 s
|
[47625]
|
|
21653
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/02 00:00:13
|
0.1 s
|
[47626]
|
|
|
21660
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/02 00:03:15
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21661
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/02 00:03:15
|
97 ms
|
[47643]
|
|
21662
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/02 00:03:15
|
92 ms
|
[47644]
|
|
21663
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/02 00:03:16
|
93 ms
|
[47645]
|
|
21664
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/02 00:03:16
|
95 ms
|
[47646]
|
|
|
21671
|
toDataset$ at DeltaTable.scala:44
+details
toDataset$ at DeltaTable.scala:44 org.apache.spark.sql.delta.util.AnalysisHelper.toDataset$(AnalysisHelper.scala:81)
io.delta.tables.DeltaTable.toDataset(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$2(DeltaTableOperations.scala:47)
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
org.apache.spark.sql.delta.DeltaTableUtils$.withActiveSession(DeltaTable.scala:477)
io.delta.tables.execution.DeltaTableOperations.$anonfun$executeDelete$1(DeltaTableOperations.scala:43)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError(AnalysisHelper.scala:99)
org.apache.spark.sql.delta.util.AnalysisHelper.improveUnsupportedOpError$(AnalysisHelper.scala:85)
io.delta.tables.DeltaTable.improveUnsupportedOpError(DeltaTable.scala:44)
io.delta.tables.execution.DeltaTableOperations.executeDelete(DeltaTableOperations.scala:43)
io.delta.tables.execution.DeltaTableOperations.executeDelete$(DeltaTableOperations.scala:42)
io.delta.tables.DeltaTable.executeDelete(DeltaTable.scala:44)
io.delta.tables.DeltaTable.delete(DeltaTable.scala:185)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.wrkldDuplDeltaRemove(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:143)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.cleanUp(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:110)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:70)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
|
2026/05/02 00:08:37
|
1 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
21672
|
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
+details
$anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:761)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:760)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:141)
org.apache.spark.sql.delta.commands.DeleteCommand.performDelete(DeleteCommand.scala:241)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2(DeleteCommand.scala:128)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$2$adapted(DeleteCommand.scala:121)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:240)
org.apache.spark.sql.delta.commands.DeleteCommand.$anonfun$run$1(DeleteCommand.scala:121)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
|
2026/05/02 00:08:37
|
0.1 s
|
[47663]
|
|
21673
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/02 00:08:37
|
0.1 s
|
[47664]
|
|
21674
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/02 00:08:38
|
98 ms
|
[47665]
|
|
21675
|
Delta: Filtering files for query
+details
Delta: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:133)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:132)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:122)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:112)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:922)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:867)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:69)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:130)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:53)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:32)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:27)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:115)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:110)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:282)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:148)
|
2026/05/02 00:08:38
|
0.1 s
|
[47666]
|
|
|
20690
|
save at ApplicationUtils.java:269
+details
save at ApplicationUtils.java:269 org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.utils.ApplicationUtils.publishToBtpKafka(ApplicationUtils.java:269)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.publishSituations(SituationsUtil.java:129)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.lambda$processAndPublishSituations$2(SituationsUtil.java:98)
java.base/java.util.ArrayList.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.processAndPublishSituations(SituationsUtil.java:89)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:73)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/05/01 09:14:45
|
0.2 s
|
[45880]
|
|
|
20701
|
save at ApplicationUtils.java:269
+details
save at ApplicationUtils.java:269 org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.utils.ApplicationUtils.publishToBtpKafka(ApplicationUtils.java:269)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.publishSituations(SituationsUtil.java:129)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.lambda$processAndPublishSituations$2(SituationsUtil.java:98)
java.base/java.util.ArrayList.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.processAndPublishSituations(SituationsUtil.java:89)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:73)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/05/01 09:26:59
|
0.2 s
|
[45900]
|
|
|
20712
|
save at ApplicationUtils.java:269
+details
save at ApplicationUtils.java:269 org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.utils.ApplicationUtils.publishToBtpKafka(ApplicationUtils.java:269)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.publishSituations(SituationsUtil.java:129)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.lambda$processAndPublishSituations$2(SituationsUtil.java:98)
java.base/java.util.ArrayList.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.processAndPublishSituations(SituationsUtil.java:89)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:73)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/05/01 09:41:05
|
0.2 s
|
[45920]
|
|
|
20723
|
save at ApplicationUtils.java:269
+details
save at ApplicationUtils.java:269 org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.utils.ApplicationUtils.publishToBtpKafka(ApplicationUtils.java:269)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.publishSituations(SituationsUtil.java:129)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.lambda$processAndPublishSituations$2(SituationsUtil.java:98)
java.base/java.util.ArrayList.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.processAndPublishSituations(SituationsUtil.java:89)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:73)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/05/01 10:00:14
|
0.2 s
|
[45940]
|
|
|
20734
|
save at ApplicationUtils.java:269
+details
save at ApplicationUtils.java:269 org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.utils.ApplicationUtils.publishToBtpKafka(ApplicationUtils.java:269)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.publishSituations(SituationsUtil.java:129)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.lambda$processAndPublishSituations$2(SituationsUtil.java:98)
java.base/java.util.ArrayList.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.processAndPublishSituations(SituationsUtil.java:89)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:73)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/05/01 10:06:29
|
86 ms
|
[45960]
|
|
|
20745
|
save at ApplicationUtils.java:269
+details
save at ApplicationUtils.java:269 org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.utils.ApplicationUtils.publishToBtpKafka(ApplicationUtils.java:269)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.publishSituations(SituationsUtil.java:129)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.lambda$processAndPublishSituations$2(SituationsUtil.java:98)
java.base/java.util.ArrayList.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.processAndPublishSituations(SituationsUtil.java:89)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:73)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/05/01 10:11:39
|
0.3 s
|
[45980]
|
|
|
20756
|
save at ApplicationUtils.java:269
+details
save at ApplicationUtils.java:269 org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.utils.ApplicationUtils.publishToBtpKafka(ApplicationUtils.java:269)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.publishSituations(SituationsUtil.java:129)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.lambda$processAndPublishSituations$2(SituationsUtil.java:98)
java.base/java.util.ArrayList.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.processAndPublishSituations(SituationsUtil.java:89)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:73)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/05/01 10:23:53
|
0.1 s
|
[46000]
|
|
|
20767
|
save at ApplicationUtils.java:269
+details
save at ApplicationUtils.java:269 org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.utils.ApplicationUtils.publishToBtpKafka(ApplicationUtils.java:269)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.publishSituations(SituationsUtil.java:129)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.lambda$processAndPublishSituations$2(SituationsUtil.java:98)
java.base/java.util.ArrayList.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.processAndPublishSituations(SituationsUtil.java:89)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:73)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/05/01 10:38:07
|
0.2 s
|
[46020]
|
|
|
20778
|
save at ApplicationUtils.java:269
+details
save at ApplicationUtils.java:269 org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.utils.ApplicationUtils.publishToBtpKafka(ApplicationUtils.java:269)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.publishSituations(SituationsUtil.java:129)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.lambda$processAndPublishSituations$2(SituationsUtil.java:98)
java.base/java.util.ArrayList.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.extractionutils.SituationsUtil.processAndPublishSituations(SituationsUtil.java:89)
com.sap.s4hana.eureka.business.crpwrkloaddetermination.temporal.activity.NotifyWorkloadDetServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyWorkloadDetServiceReplRunFinishedActivityImpl.java:73)
jdk.internal.reflect.GeneratedMethodAccessor504.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/05/01 10:45:13
|
0.3 s
|
[46040]
|
|