Error

None

Details

    Traceback (most recent call last):
  File "/opt/acc-bpt/release_2026_01/acc_bpt/shared/BptV2.py", line 284, in build
    rawdata = self.bptStatic.fetch(vars, times)
              ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/acc_bpt/ps/statistics/ps_statistics_integrated_intensity_per_destination.py", line 50, in fetch
    return super().fetch(var_defs, time_ranges, cache)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/acc_bpt/ps/statistics/integrated_intensity_base.py", line 86, in fetch
    df = self.nxcals.fetch_vars_list(*time_ranges, vars_list=stats_vars, system='CMW') \
         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/acc_bpt/shared/utils_nxcals.py", line 385, in fetch_vars_list
    return self._fetch_vars_of_same_type(ts1, ts2, vars_like=None, vars_list=vars_list, acc_name=acc_name,
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/acc_bpt/shared/utils_nxcals.py", line 436, in _fetch_vars_of_same_type
    sdfs = self._get_vars_raw(nxcals_query, vars_like, vars_list, system)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/acc_bpt/shared/utils_nxcals.py", line 452, in _get_vars_raw
    return nxcals_query._get_varslist_raw(vars_list, system)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/acc_bpt/shared/utils_nxcals_queries.py", line 185, in _get_varslist_raw
    .buildDataset() \
     ^^^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/venv/lib/python3.11/site-packages/nxcals/api/extraction/data/common.py", line 111, in buildDataset
    return self.build()
           ^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/venv/lib/python3.11/site-packages/nxcals/api/extraction/data/common.py", line 114, in build
    return self._builder._build()
           ^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/venv/lib/python3.11/site-packages/nxcals/api/extraction/data/builders.py", line 33, in _build
    df = self._get_java_builder().build()
         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/venv/lib/python3.11/site-packages/py4j/java_gateway.py", line 1322, in __call__
    return_value = get_return_value(
                   ^^^^^^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/venv/lib/python3.11/site-packages/pyspark/errors/exceptions/captured.py", line 179, in deco
    return f(*a, **kw)
           ^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/venv/lib/python3.11/site-packages/py4j/protocol.py", line 326, in get_return_value
    raise Py4JJavaError(
py4j.protocol.Py4JJavaError: An error occurred while calling o16961.build.
: java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext.
This stopped SparkContext was created at:

org.apache.spark.api.java.JavaSparkContext.(JavaSparkContext.scala:58)
java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:374)
py4j.Gateway.invoke(Gateway.java:238)
py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)
py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)
py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)
py4j.ClientServerConnection.run(ClientServerConnection.java:106)
java.base/java.lang.Thread.run(Thread.java:829)

The currently active SparkContext was created at:

(No active SparkContext.)
         
	at org.apache.spark.SparkContext.assertNotStopped(SparkContext.scala:122)
	at org.apache.spark.SparkContext.$anonfun$parallelize$1(SparkContext.scala:937)
	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
	at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
	at org.apache.spark.SparkContext.withScope(SparkContext.scala:919)
	at org.apache.spark.SparkContext.parallelize(SparkContext.scala:936)
	at org.apache.spark.util.HadoopFSUtils$.parallelListLeafFilesInternal(HadoopFSUtils.scala:121)
	at org.apache.spark.util.HadoopFSUtils$.parallelListLeafFiles(HadoopFSUtils.scala:69)
	at org.apache.spark.sql.execution.datasources.InMemoryFileIndex$.bulkListLeafFiles(InMemoryFileIndex.scala:162)
	at org.apache.spark.sql.execution.datasources.InMemoryFileIndex.listLeafFiles(InMemoryFileIndex.scala:133)
	at org.apache.spark.sql.execution.datasources.InMemoryFileIndex.refresh0(InMemoryFileIndex.scala:96)
	at org.apache.spark.sql.execution.datasources.InMemoryFileIndex.(InMemoryFileIndex.scala:68)
	at org.apache.spark.sql.execution.datasources.DataSource.createInMemoryFileIndex(DataSource.scala:539)
	at org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:405)
	at org.apache.spark.sql.DataFrameReader.loadV1Source(DataFrameReader.scala:229)
	at org.apache.spark.sql.DataFrameReader.$anonfun$load$2(DataFrameReader.scala:211)
	at scala.Option.getOrElse(Option.scala:189)
	at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:211)
	at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:197)
	at cern.nxcals.api.extraction.data.spark.HdfsDatasetCreator.lambda$apply$0(HdfsDatasetCreator.java:58)
	at java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:195)
	at java.base/java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1655)
	at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
	at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
	at java.base/java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:913)
	at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
	at java.base/java.util.stream.ReferencePipeline.reduce(ReferencePipeline.java:558)
	at cern.nxcals.api.extraction.data.spark.HdfsDatasetCreator.apply(HdfsDatasetCreator.java:61)
	at cern.nxcals.api.extraction.data.spark.HdfsDatasetCreator.apply(HdfsDatasetCreator.java:31)
	at cern.nxcals.api.extraction.data.spark.SparkExtractionTaskProcessor.execute(SparkExtractionTaskProcessor.java:27)
	at cern.nxcals.api.extraction.data.spark.SparkExtractionTaskProcessor.execute(SparkExtractionTaskProcessor.java:13)
	at cern.nxcals.common.domain.HdfsExtractionTask.processWith(HdfsExtractionTask.java:20)
	at cern.nxcals.api.extraction.data.spark.SparkExtractionTaskExecutor.toOptionalDataset(SparkExtractionTaskExecutor.java:85)
	at java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:195)
	at java.base/java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1655)
	at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
	at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
	at java.base/java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:913)
	at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
	at java.base/java.util.stream.ReferencePipeline.reduce(ReferencePipeline.java:558)
	at cern.nxcals.api.extraction.data.spark.SparkExtractionTaskExecutor.execute(SparkExtractionTaskExecutor.java:56)
	at cern.nxcals.api.extraction.data.builders.SparkDatasetProducer.apply(SparkDatasetProducer.java:34)
	at cern.nxcals.api.extraction.data.builders.SparkDatasetProducer.apply(SparkDatasetProducer.java:17)
	at cern.nxcals.api.extraction.data.builders.fluent.QueryData.build(QueryData.java:167)
	at cern.nxcals.api.extraction.data.builders.fluent.VariableStageLoop.build(VariableStageLoop.java:18)
	at jdk.internal.reflect.GeneratedMethodAccessor82.invoke(Unknown Source)
	at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.base/java.lang.reflect.Method.invoke(Method.java:566)
	at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
	at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:374)
	at py4j.Gateway.invoke(Gateway.java:282)
	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
	at py4j.commands.CallCommand.execute(CallCommand.java:79)
	at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)
	at py4j.ClientServerConnection.run(ClientServerConnection.java:106)
	at java.base/java.lang.Thread.run(Thread.java:829)


During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/opt/acc-bpt/release_2026_01/jinja/build_static_plots.py", line 85, in _create_static_html_content
    output_html_file_path = driver.build(abs_plot_dest_file_html)
                            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/acc-bpt/release_2026_01/acc_bpt/shared/BptV2.py", line 304, in build
    raise RuntimeError(cause)
RuntimeError: None

    

Generated 2026-04-30 21:09:50.254449+02:00