Bioconductor Single Package Builder - Build History

Snapshot Date: 
URL:  https://git.bioconductor.org/packages/BiocHail
Last Changed Rev:  / Revision: 
Last Changed Date: 

Hostname OS /Arch BUILD CHECK BUILD BIN POST-PROCESSING
merida1 macOS 10.14.6 Mojave/x86_64   ERROR     skipped     skipped     skipped  
nebbiolo1 Linux (Ubuntu 22.04.1 LTS)/x86_64   OK     ERROR     skipped     OK  

merida1 Summary

[top]

Package: BiocHail
Version: 0.99.15
RVersion: 4.3
BiocVersion: 3.17
BuildCommand: /Library/Frameworks/R.framework/Versions/Current/Resources/bin/R CMD build --keep-empty-dirs --no-resave-data BiocHail
BuildTime: 5 minutes 3.68 seconds
CheckCommand:
CheckTime:
BuildBinCommand:
BuildBinTime:
PackageFileSize: -1.00 KiB
BuildID:: BiocHail_20230321131518
PreProcessing: Starting Git clone. Installing dependencies. Checking Git Clone. Installing package: BiocHail. Starting Build package.
PostProcessing: Finished Git clone. Package type: Software. Installing dependency status: 0. Checking git clone status: 0. Installing package status: 0. Build Package status: 1.

merida1 BUILD SRC output

[top]

===============================

 R CMD BUILD

===============================

* checking for file ‘BiocHail/DESCRIPTION’ ... OK
* preparing ‘BiocHail’:
* checking DESCRIPTION meta-information ... OK
* installing the package to build vignettes
* creating vignettes ... ERROR
--- re-building ‘gwas_tut.Rmd’ using rmarkdown
2023-03-21 13:16:17.363 WARN  NativeCodeLoader:60 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
Initializing Hail with default parameters...
2023-03-21 13:16:19.091 WARN  SparkContext:69 - Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:
org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
is.hail.backend.spark.SparkBackend$.configureAndCreateSparkContext(SparkBackend.scala:148)
is.hail.backend.spark.SparkBackend$.apply(SparkBackend.scala:230)
is.hail.backend.spark.SparkBackend.apply(SparkBackend.scala)
java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
java.base/java.lang.reflect.Method.invoke(Method.java:577)
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
py4j.Gateway.invoke(Gateway.java:282)
py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
py4j.commands.CallCommand.execute(CallCommand.java:79)
py4j.GatewayConnection.run(GatewayConnection.java:238)
java.base/java.lang.Thread.run(Thread.java:833)
Quitting from lines 57-61 (gwas_tut.Rmd) 
Error: processing vignette 'gwas_tut.Rmd' failed with diagnostics:
py4j.protocol.Py4JJavaError: An error occurred while calling z:is.hail.backend.spark.SparkBackend.apply.
<... omitted ...>ke(ReflectionEngine.java:357)
	at py4j.Gateway.invoke(Gateway.java:282)
	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
	at py4j.commands.CallCommand.execute(CallCommand.java:79)
	at py4j.GatewayConnection.run(GatewayConnection.java:238)
	at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: java.lang.ExceptionInInitializerError: Exception java.lang.IllegalAccessError: class org.apache.spark.storage.StorageUtils$ (in unnamed module @0x6671dc81) cannot access class sun.nio.ch.DirectBuffer (in module java.base) because module java.base does not export sun.nio.ch to unnamed module @0x6671dc81 [in thread "Thread-3"]
	at org.apache.spark.storage.StorageUtils$.<init>(StorageUtils.scala:213)
	at org.apache.spark.storage.StorageUtils$.<clinit>(StorageUtils.scala)
	... 19 more

See `reticulate::py_last_error()` for details
--- failed re-building ‘gwas_tut.Rmd’

--- re-building ‘large_t2t.Rmd’ using rmarkdown
2023-03-21 13:16:19.201 WARN  SparkContext:69 - Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:
org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
is.hail.backend.spark.SparkBackend$.configureAndCreateSparkContext(SparkBackend.scala:148)
is.hail.backend.spark.SparkBackend$.apply(SparkBackend.scala:230)
is.hail.backend.spark.SparkBackend.apply(SparkBackend.scala)
java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
java.base/java.lang.reflect.Method.invoke(Method.java:577)
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
py4j.Gateway.invoke(Gateway.java:282)
py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
py4j.commands.CallCommand.execute(CallCommand.java:79)
py4j.GatewayConnection.run(GatewayConnection.java:238)
java.base/java.lang.Thread.run(Thread.java:833)
Initializing Hail with default parameters...
2023-03-21 13:16:19.248 WARN  SparkContext:69 - Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:
org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
is.hail.backend.spark.SparkBackend$.configureAndCreateSparkContext(SparkBackend.scala:148)
is.hail.backend.spark.SparkBackend$.apply(SparkBackend.scala:230)
is.hail.backend.spark.SparkBackend.apply(SparkBackend.scala)
java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
java.base/java.lang.reflect.Method.invoke(Method.java:577)
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
py4j.Gateway.invoke(Gateway.java:282)
py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
py4j.commands.CallCommand.execute(CallCommand.java:79)
py4j.GatewayConnection.run(GatewayConnection.java:238)
java.base/java.lang.Thread.run(Thread.java:833)
Quitting from lines 65-75 (large_t2t.Rmd) 
Error: processing vignette 'large_t2t.Rmd' failed with diagnostics:
py4j.protocol.Py4JJavaError: An error occurred while calling z:is.hail.backend.spark.SparkBackend.apply.
<... omitted ...>ke(ReflectionEngine.java:357)
	at py4j.Gateway.invoke(Gateway.java:282)
	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
	at py4j.commands.CallCommand.execute(CallCommand.java:79)
	at py4j.GatewayConnection.run(GatewayConnection.java:238)
	at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: java.lang.ExceptionInInitializerError: Exception java.lang.IllegalAccessError: class org.apache.spark.storage.StorageUtils$ (in unnamed module @0x6671dc81) cannot access class sun.nio.ch.DirectBuffer (in module java.base) because module java.base does not export sun.nio.ch to unnamed module @0x6671dc81 [in thread "Thread-3"]
	at org.apache.spark.storage.StorageUtils$.<init>(StorageUtils.scala:213)
	at org.apache.spark.storage.StorageUtils$.<clinit>(StorageUtils.scala)
	... 19 more

See `reticulate::py_last_error()` for details
--- failed re-building ‘large_t2t.Rmd’

--- re-building ‘ukbb.Rmd’ using rmarkdown
2023-03-21 13:16:19.359 WARN  SparkContext:69 - Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:
org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
is.hail.backend.spark.SparkBackend$.configureAndCreateSparkContext(SparkBackend.scala:148)
is.hail.backend.spark.SparkBackend$.apply(SparkBackend.scala:230)
is.hail.backend.spark.SparkBackend.apply(SparkBackend.scala)
java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
java.base/java.lang.reflect.Method.invoke(Method.java:577)
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
py4j.Gateway.invoke(Gateway.java:282)
py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
py4j.commands.CallCommand.execute(CallCommand.java:79)
py4j.GatewayConnection.run(GatewayConnection.java:238)
java.base/java.lang.Thread.run(Thread.java:833)
Initializing Hail with default parameters...
2023-03-21 13:20:48.491 WARN  SparkContext:69 - Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:
org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
is.hail.backend.spark.SparkBackend$.configureAndCreateSparkContext(SparkBackend.scala:148)
is.hail.backend.spark.SparkBackend$.apply(SparkBackend.scala:230)
is.hail.backend.spark.SparkBackend.apply(SparkBackend.scala)
java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
java.base/java.lang.reflect.Method.invoke(Method.java:577)
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
py4j.Gateway.invoke(Gateway.java:282)
py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
py4j.commands.CallCommand.execute(CallCommand.java:79)
py4j.GatewayConnection.run(GatewayConnection.java:238)
java.base/java.lang.Thread.run(Thread.java:833)
Quitting from lines 38-42 (ukbb.Rmd) 
Error: processing vignette 'ukbb.Rmd' failed with diagnostics:
py4j.protocol.Py4JJavaError: An error occurred while calling z:is.hail.backend.spark.SparkBackend.apply.
<... omitted ...>ke(ReflectionEngine.java:357)
	at py4j.Gateway.invoke(Gateway.java:282)
	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
	at py4j.commands.CallCommand.execute(CallCommand.java:79)
	at py4j.GatewayConnection.run(GatewayConnection.java:238)
	at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: java.lang.ExceptionInInitializerError: Exception java.lang.IllegalAccessError: class org.apache.spark.storage.StorageUtils$ (in unnamed module @0x6671dc81) cannot access class sun.nio.ch.DirectBuffer (in module java.base) because module java.base does not export sun.nio.ch to unnamed module @0x6671dc81 [in thread "Thread-3"]
	at org.apache.spark.storage.StorageUtils$.<init>(StorageUtils.scala:213)
	at org.apache.spark.storage.StorageUtils$.<clinit>(StorageUtils.scala)
	... 19 more

See `reticulate::py_last_error()` for details
--- failed re-building ‘ukbb.Rmd’

SUMMARY: processing the following files failed:
  ‘gwas_tut.Rmd’ ‘large_t2t.Rmd’ ‘ukbb.Rmd’

Error: Vignette re-building failed.
Execution halted

merida1 CHECK output

[top]


                				

merida1 BUILD BIN output

[top]


                				

nebbiolo1 Summary

[top]

Package: BiocHail
Version: 0.99.15
RVersion: 4.3
BiocVersion: 3.17
BuildCommand: /home/biocbuild/bbs-3.17-bioc/R/bin/R CMD build --keep-empty-dirs --no-resave-data BiocHail
BuildTime: 2 minutes 53.10 seconds
CheckCommand: BiocCheckGitClone('BiocHail') && /home/biocbuild/bbs-3.17-bioc/R/bin/R CMD check --no-vignettes --timings BiocHail_0.99.15.tar.gz && BiocCheck('BiocHail_0.99.15.tar.gz', `new-package`=TRUE)
CheckTime: 0 minutes 5.47 seconds
BuildBinCommand:
BuildBinTime:
PackageFileSize: 1737.31 KiB
BuildID:: BiocHail_20230321131518
PreProcessing: Starting Git clone. Installing dependencies. Checking Git Clone. Installing package: BiocHail. Starting Build package. Starting Check package.
PostProcessing: Finished Git clone. Package type: Software. Installing dependency status: 0. Checking git clone status: 0. Installing package status: 0. Build Package status: 0. Checking Package status: 1.

nebbiolo1 BUILD SRC output

[top]

===============================

 R CMD BUILD

===============================

* checking for file ‘BiocHail/DESCRIPTION’ ... OK
* preparing ‘BiocHail’:
* checking DESCRIPTION meta-information ... OK
* installing the package to build vignettes
* creating vignettes ... OK
* checking for LF line-endings in source and make files and shell scripts
* checking for empty or unneeded directories
* looking to see if a ‘data/datalist’ file should be added
* building ‘BiocHail_0.99.15.tar.gz’


nebbiolo1 CHECK output

[top]

===============================

 BiocCheckGitClone('BiocHail')

===============================

─ BiocCheckVersion: 1.35.13
─ BiocVersion: 3.17
─ Package: BiocHail
─ PackageVersion: 0.99.15
─ sourceDir: /home/pkgbuild/packagebuilder/workers/jobs/2947/BiocHail_20230321131518/BiocHail
─ platform: unix
─ isTarBall: FALSE

* Checking valid files...
* Checking for stray BiocCheck output folders...
* Checking for inst/doc folders...
* Checking DESCRIPTION...
* Checking if DESCRIPTION is well formatted...
* Checking for valid maintainer...
* Checking CITATION...

─ BiocCheck results ──
0 ERRORS | 0 WARNINGS | 0 NOTES

For more details, run
    browseVignettes(package = 'BiocCheck')




===============================

 R CMD CHECK

===============================

* using log directory ‘/home/pkgbuild/packagebuilder/workers/jobs/2947/BiocHail_20230321131518/BiocHail.Rcheck’
* using R Under development (unstable) (2023-03-16 r83996)
* using platform: x86_64-pc-linux-gnu (64-bit)
* R was compiled by
    gcc (Ubuntu 11.3.0-1ubuntu1~22.04) 11.3.0
    GNU Fortran (Ubuntu 11.3.0-1ubuntu1~22.04) 11.3.0
* running under: Ubuntu 22.04.2 LTS
* using session charset: UTF-8
* using option ‘--no-vignettes’
* checking for file ‘BiocHail/DESCRIPTION’ ... OK
* this is package ‘BiocHail’ version ‘0.99.15’
* package encoding: UTF-8
* checking package namespace information ... OK
* checking package dependencies ... OK
* checking if this is a source package ... OK
* checking if there is a namespace ... OK
* checking for hidden files and directories ... NOTE
Found the following hidden files and directories:
  .BBSoptions
These were most likely included in error. See section ‘Package
structure’ in the ‘Writing R Extensions’ manual.
* checking for portable file names ... OK
* checking for sufficient/correct file permissions ... OK
* checking whether package ‘BiocHail’ can be installed ... [0s/0s] ERROR
Installation failed.
See ‘/home/pkgbuild/packagebuilder/workers/jobs/2947/BiocHail_20230321131518/BiocHail.Rcheck/00install.out’ for details.
* DONE

Status: 1 ERROR, 1 NOTE
See
  ‘/home/pkgbuild/packagebuilder/workers/jobs/2947/BiocHail_20230321131518/BiocHail.Rcheck/00check.log’
for details.





===============================

 BiocCheck('BiocHail_0.99.15.tar.gz')

===============================

─ BiocCheckVersion: 1.35.13
─ BiocVersion: 3.17
─ Package: BiocHail
─ PackageVersion: 0.99.15
─ sourceDir: /tmp/RtmpwgL7N2/filea6ea6404682f5/BiocHail
─ installDir: /tmp/RtmpwgL7N2/filea6ea626ff3f12
─ BiocCheckDir: /home/pkgbuild/packagebuilder/workers/jobs/2947/BiocHail_20230321131518/BiocHail.BiocCheck
─ platform: unix
─ isTarBall: TRUE

* Installing package...
* Checking for deprecated package usage...
* Checking for remote package usage...
* Checking for 'LazyData: true' usage...
* Checking version number...
* Checking for version number mismatch...
* Checking new package version number...
* Checking R version dependency...
    * NOTE: Update R version dependency from 4.2 to 4.3.0.
* Checking package size...
* Checking individual file sizes...
* Checking biocViews...
* Checking that biocViews are present...
* Checking package type based on biocViews...
    Software
* Checking for non-trivial biocViews...
* Checking that biocViews come from the same category...
* Checking biocViews validity...
* Checking for recommended biocViews...
* Checking build system compatibility...
* Checking for blank lines in DESCRIPTION...
* Checking if DESCRIPTION is well formatted...
* Checking for proper Description: field...
    * NOTE: The Description field in the DESCRIPTION is made up by less
      than 3 sentences. Please consider expanding this field, and
      structure it as a full paragraph
* Checking for whitespace in DESCRIPTION field names...
* Checking that Package field matches directory/tarball name...
* Checking for Version field...
* Checking for valid maintainer...
* Checking License: for restrictive use...
* Checking for pinned package versions...
* Checking DESCRIPTION/NAMESPACE consistency...
* Checking .Rbuildignore...
* Checking for stray BiocCheck output folders...
* Checking vignette directory...
    * NOTE: 'sessionInfo' not found in vignette(s)
* Checking whether vignette is built with 'R CMD build'...
* Checking package installation calls in R code...
* Checking for library/require of BiocHail...
* Checking coding practice...
    * NOTE: Avoid 1:...; use seq_len() or seq_along()
    * NOTE: Avoid using '=' for assignment and use '<-' instead
* Checking parsed R code in R directory, examples, vignettes...
* Checking function lengths...
* Checking man page documentation...
    * WARNING: Non-empty \value sections found in man pages.
    * ERROR: At least 80% of man pages documenting exported objects
      must have runnable examples.
    * NOTE: Usage of dontrun{} / donttest{} tags found in man page
      examples. 7% of man pages use at least one of these tags.
    * NOTE: Use donttest{} instead of dontrun{}.
* Checking package NEWS...
    * NOTE: Consider adding a NEWS file, so your package news will be
      included in Bioconductor release announcements.
* Checking unit tests...
* Checking skip_on_bioc() in tests...
* Checking formatting of DESCRIPTION, NAMESPACE, man pages, R source,
  and vignette source...
    * NOTE: Consider shorter lines; 124 lines (6%) are > 80 characters
      long.
    * NOTE: Consider 4 spaces instead of tabs; 1 lines (0%) contain
      tabs.
    * NOTE: Consider multiples of 4 spaces for line indents; 204 lines
      (10%) are not.
    See https://contributions.bioconductor.org/r-code.html
    See styler package: https://cran.r-project.org/package=styler as
      described in the BiocCheck vignette.
* Checking if package already exists in CRAN...
* Checking if new package already exists in Bioconductor...
* Checking for bioc-devel mailing list subscription...
    Maintainer is subscribed to bioc-devel.
* Checking for support site registration...
    Maintainer is registered at support site.
    Package name is in support site watched tags.

─ BiocCheck results ──
1 ERRORS | 1 WARNINGS | 11 NOTES

See the BiocHail.BiocCheck folder and run
    browseVignettes(package = 'BiocCheck')
for details.

nebbiolo1 BUILD BIN output

[top]