Bioconductor Single Package Builder - Build History

Snapshot Date: 
URL:  https://git.bioconductor.org/packages/BiocHail
Last Changed Rev:  / Revision: 
Last Changed Date: 

Hostname OS /Arch BUILD CHECK BUILD BIN POST-PROCESSING
merida1 macOS 10.14.6 Mojave/x86_64   ERROR     skipped     skipped     skipped  
nebbiolo1 Linux (Ubuntu 22.04.1 LTS)/x86_64   OK     ERROR     skipped     OK  

merida1 Summary

[top]

Package: BiocHail
Version: 0.99.15
RVersion: 4.3
BiocVersion: 3.17
BuildCommand: /Library/Frameworks/R.framework/Versions/Current/Resources/bin/R CMD build --keep-empty-dirs --no-resave-data BiocHail
BuildTime: 3 minutes 2.15 seconds
CheckCommand:
CheckTime:
BuildBinCommand:
BuildBinTime:
PackageFileSize: -1.00 KiB
BuildID:: BiocHail_20230321135853
PreProcessing: Starting Git clone. Installing dependencies. Checking Git Clone. Installing package: BiocHail. Starting Build package.
PostProcessing: Finished Git clone. Package type: Software. Installing dependency status: 0. Checking git clone status: 0. Installing package status: 0. Build Package status: 1.

merida1 BUILD SRC output

[top]

===============================

 R CMD BUILD

===============================

* checking for file ‘BiocHail/DESCRIPTION’ ... OK
* preparing ‘BiocHail’:
* checking DESCRIPTION meta-information ... OK
* installing the package to build vignettes
* creating vignettes ... ERROR
--- re-building ‘gwas_tut.Rmd’ using rmarkdown
2023-03-21 13:59:46.819 WARN  NativeCodeLoader:60 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
Initializing Hail with default parameters...
2023-03-21 13:59:48.374 WARN  SparkContext:69 - Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:
org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
is.hail.backend.spark.SparkBackend$.configureAndCreateSparkContext(SparkBackend.scala:148)
is.hail.backend.spark.SparkBackend$.apply(SparkBackend.scala:230)
is.hail.backend.spark.SparkBackend.apply(SparkBackend.scala)
java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
java.base/java.lang.reflect.Method.invoke(Method.java:577)
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
py4j.Gateway.invoke(Gateway.java:282)
py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
py4j.commands.CallCommand.execute(CallCommand.java:79)
py4j.GatewayConnection.run(GatewayConnection.java:238)
java.base/java.lang.Thread.run(Thread.java:833)
Quitting from lines 57-61 (gwas_tut.Rmd) 
Error: processing vignette 'gwas_tut.Rmd' failed with diagnostics:
py4j.protocol.Py4JJavaError: An error occurred while calling z:is.hail.backend.spark.SparkBackend.apply.
<... omitted ...>ke(ReflectionEngine.java:357)
	at py4j.Gateway.invoke(Gateway.java:282)
	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
	at py4j.commands.CallCommand.execute(CallCommand.java:79)
	at py4j.GatewayConnection.run(GatewayConnection.java:238)
	at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: java.lang.ExceptionInInitializerError: Exception java.lang.IllegalAccessError: class org.apache.spark.storage.StorageUtils$ (in unnamed module @0x6671dc81) cannot access class sun.nio.ch.DirectBuffer (in module java.base) because module java.base does not export sun.nio.ch to unnamed module @0x6671dc81 [in thread "Thread-3"]
	at org.apache.spark.storage.StorageUtils$.<init>(StorageUtils.scala:213)
	at org.apache.spark.storage.StorageUtils$.<clinit>(StorageUtils.scala)
	... 19 more

See `reticulate::py_last_error()` for details
--- failed re-building ‘gwas_tut.Rmd’

--- re-building ‘large_t2t.Rmd’ using rmarkdown
2023-03-21 13:59:48.495 WARN  SparkContext:69 - Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:
org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
is.hail.backend.spark.SparkBackend$.configureAndCreateSparkContext(SparkBackend.scala:148)
is.hail.backend.spark.SparkBackend$.apply(SparkBackend.scala:230)
is.hail.backend.spark.SparkBackend.apply(SparkBackend.scala)
java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
java.base/java.lang.reflect.Method.invoke(Method.java:577)
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
py4j.Gateway.invoke(Gateway.java:282)
py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
py4j.commands.CallCommand.execute(CallCommand.java:79)
py4j.GatewayConnection.run(GatewayConnection.java:238)
java.base/java.lang.Thread.run(Thread.java:833)
Initializing Hail with default parameters...
2023-03-21 13:59:48.544 WARN  SparkContext:69 - Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:
org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
is.hail.backend.spark.SparkBackend$.configureAndCreateSparkContext(SparkBackend.scala:148)
is.hail.backend.spark.SparkBackend$.apply(SparkBackend.scala:230)
is.hail.backend.spark.SparkBackend.apply(SparkBackend.scala)
java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
java.base/java.lang.reflect.Method.invoke(Method.java:577)
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
py4j.Gateway.invoke(Gateway.java:282)
py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
py4j.commands.CallCommand.execute(CallCommand.java:79)
py4j.GatewayConnection.run(GatewayConnection.java:238)
java.base/java.lang.Thread.run(Thread.java:833)
Quitting from lines 65-75 (large_t2t.Rmd) 
Error: processing vignette 'large_t2t.Rmd' failed with diagnostics:
py4j.protocol.Py4JJavaError: An error occurred while calling z:is.hail.backend.spark.SparkBackend.apply.
<... omitted ...>ke(ReflectionEngine.java:357)
	at py4j.Gateway.invoke(Gateway.java:282)
	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
	at py4j.commands.CallCommand.execute(CallCommand.java:79)
	at py4j.GatewayConnection.run(GatewayConnection.java:238)
	at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: java.lang.ExceptionInInitializerError: Exception java.lang.IllegalAccessError: class org.apache.spark.storage.StorageUtils$ (in unnamed module @0x6671dc81) cannot access class sun.nio.ch.DirectBuffer (in module java.base) because module java.base does not export sun.nio.ch to unnamed module @0x6671dc81 [in thread "Thread-3"]
	at org.apache.spark.storage.StorageUtils$.<init>(StorageUtils.scala:213)
	at org.apache.spark.storage.StorageUtils$.<clinit>(StorageUtils.scala)
	... 19 more

See `reticulate::py_last_error()` for details
--- failed re-building ‘large_t2t.Rmd’

--- re-building ‘ukbb.Rmd’ using rmarkdown
2023-03-21 13:59:48.666 WARN  SparkContext:69 - Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:
org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
is.hail.backend.spark.SparkBackend$.configureAndCreateSparkContext(SparkBackend.scala:148)
is.hail.backend.spark.SparkBackend$.apply(SparkBackend.scala:230)
is.hail.backend.spark.SparkBackend.apply(SparkBackend.scala)
java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
java.base/java.lang.reflect.Method.invoke(Method.java:577)
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
py4j.Gateway.invoke(Gateway.java:282)
py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
py4j.commands.CallCommand.execute(CallCommand.java:79)
py4j.GatewayConnection.run(GatewayConnection.java:238)
java.base/java.lang.Thread.run(Thread.java:833)
Initializing Hail with default parameters...
2023-03-21 14:02:18.767 WARN  SparkContext:69 - Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:
org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
is.hail.backend.spark.SparkBackend$.configureAndCreateSparkContext(SparkBackend.scala:148)
is.hail.backend.spark.SparkBackend$.apply(SparkBackend.scala:230)
is.hail.backend.spark.SparkBackend.apply(SparkBackend.scala)
java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:104)
java.base/java.lang.reflect.Method.invoke(Method.java:577)
py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
py4j.Gateway.invoke(Gateway.java:282)
py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
py4j.commands.CallCommand.execute(CallCommand.java:79)
py4j.GatewayConnection.run(GatewayConnection.java:238)
java.base/java.lang.Thread.run(Thread.java:833)
Quitting from lines 38-42 (ukbb.Rmd) 
Error: processing vignette 'ukbb.Rmd' failed with diagnostics:
py4j.protocol.Py4JJavaError: An error occurred while calling z:is.hail.backend.spark.SparkBackend.apply.
<... omitted ...>ke(ReflectionEngine.java:357)
	at py4j.Gateway.invoke(Gateway.java:282)
	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
	at py4j.commands.CallCommand.execute(CallCommand.java:79)
	at py4j.GatewayConnection.run(GatewayConnection.java:238)
	at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: java.lang.ExceptionInInitializerError: Exception java.lang.IllegalAccessError: class org.apache.spark.storage.StorageUtils$ (in unnamed module @0x6671dc81) cannot access class sun.nio.ch.DirectBuffer (in module java.base) because module java.base does not export sun.nio.ch to unnamed module @0x6671dc81 [in thread "Thread-3"]
	at org.apache.spark.storage.StorageUtils$.<init>(StorageUtils.scala:213)
	at org.apache.spark.storage.StorageUtils$.<clinit>(StorageUtils.scala)
	... 19 more

See `reticulate::py_last_error()` for details
--- failed re-building ‘ukbb.Rmd’

SUMMARY: processing the following files failed:
  ‘gwas_tut.Rmd’ ‘large_t2t.Rmd’ ‘ukbb.Rmd’

Error: Vignette re-building failed.
Execution halted

merida1 CHECK output

[top]


                				

merida1 BUILD BIN output

[top]


                				

nebbiolo1 Summary

[top]

Package: BiocHail
Version: 0.99.15
RVersion: 4.3
BiocVersion: 3.17
BuildCommand: /home/biocbuild/bbs-3.17-bioc/R/bin/R CMD build --keep-empty-dirs --no-resave-data BiocHail
BuildTime: 2 minutes 44.64 seconds
CheckCommand: BiocCheckGitClone('BiocHail') && /home/biocbuild/bbs-3.17-bioc/R/bin/R CMD check --no-vignettes --timings --library=/home/pkgbuild/packagebuilder/workers/jobs/2947/R-libs --install=check:/home/pkgbuild/packagebuilder/workers/jobs/2947/BiocHail_20230321135853/BiocHail.install-out.txt BiocHail_0.99.15.tar.gz && BiocCheck('BiocHail_0.99.15.tar.gz', `new-package`=TRUE)
CheckTime: 2 minutes 31.78 seconds
BuildBinCommand:
BuildBinTime:
PackageFileSize: 1737.31 KiB
BuildID:: BiocHail_20230321135853
PreProcessing: Starting Git clone. Installing dependencies. Checking Git Clone. Installing package: BiocHail. Starting Build package. Starting Check package.
PostProcessing: Finished Git clone. Package type: Software. Installing dependency status: 0. Checking git clone status: 0. Installing package status: 0. Build Package status: 0. Checking Package status: 1.

nebbiolo1 BUILD SRC output

[top]

===============================

 R CMD BUILD

===============================

* checking for file ‘BiocHail/DESCRIPTION’ ... OK
* preparing ‘BiocHail’:
* checking DESCRIPTION meta-information ... OK
* installing the package to build vignettes
* creating vignettes ... OK
* checking for LF line-endings in source and make files and shell scripts
* checking for empty or unneeded directories
* looking to see if a ‘data/datalist’ file should be added
* building ‘BiocHail_0.99.15.tar.gz’


nebbiolo1 CHECK output

[top]

===============================

 BiocCheckGitClone('BiocHail')

===============================

─ BiocCheckVersion: 1.35.13
─ BiocVersion: 3.17
─ Package: BiocHail
─ PackageVersion: 0.99.15
─ sourceDir: /home/pkgbuild/packagebuilder/workers/jobs/2947/BiocHail_20230321135853/BiocHail
─ platform: unix
─ isTarBall: FALSE

* Checking valid files...
* Checking for stray BiocCheck output folders...
* Checking for inst/doc folders...
* Checking DESCRIPTION...
* Checking if DESCRIPTION is well formatted...
* Checking for valid maintainer...
* Checking CITATION...

─ BiocCheck results ──
0 ERRORS | 0 WARNINGS | 0 NOTES

For more details, run
    browseVignettes(package = 'BiocCheck')




===============================

 R CMD CHECK

===============================

* using log directory ‘/home/pkgbuild/packagebuilder/workers/jobs/2947/BiocHail_20230321135853/BiocHail.Rcheck’
* using R Under development (unstable) (2023-03-16 r83996)
* using platform: x86_64-pc-linux-gnu (64-bit)
* R was compiled by
    gcc (Ubuntu 11.3.0-1ubuntu1~22.04) 11.3.0
    GNU Fortran (Ubuntu 11.3.0-1ubuntu1~22.04) 11.3.0
* running under: Ubuntu 22.04.2 LTS
* using session charset: UTF-8
* using option ‘--no-vignettes’
* checking for file ‘BiocHail/DESCRIPTION’ ... OK
* this is package ‘BiocHail’ version ‘0.99.15’
* package encoding: UTF-8
* checking package namespace information ... OK
* checking package dependencies ... OK
* checking if this is a source package ... OK
* checking if there is a namespace ... OK
* checking for hidden files and directories ... NOTE
Found the following hidden files and directories:
  .BBSoptions
These were most likely included in error. See section ‘Package
structure’ in the ‘Writing R Extensions’ manual.
* checking for portable file names ... OK
* checking for sufficient/correct file permissions ... OK
* checking whether package ‘BiocHail’ can be installed ... OK
* checking installed package size ... OK
* checking package directory ... OK
* checking for future file timestamps ... OK
* checking ‘build’ directory ... OK
* checking DESCRIPTION meta-information ... OK
* checking top-level files ... OK
* checking for left-over files ... OK
* checking index information ... OK
* checking package subdirectories ... OK
* checking R files for non-ASCII characters ... OK
* checking R files for syntax errors ... OK
* checking whether the package can be loaded ... [2s/2s] OK
* checking whether the package can be loaded with stated dependencies ... [2s/2s] OK
* checking whether the package can be unloaded cleanly ... [2s/2s] OK
* checking whether the namespace can be loaded with stated dependencies ... [2s/2s] OK
* checking whether the namespace can be unloaded cleanly ... [2s/2s] OK
* checking loading without being on the library search path ... [2s/2s] OK
* checking use of S3 registration ... OK
* checking dependencies in R code ... OK
* checking S3 generic/method consistency ... OK
* checking replacement functions ... OK
* checking foreign function calls ... OK
* checking R code for possible problems ... [7s/7s] OK
* checking Rd files ... [0s/0s] OK
* checking Rd metadata ... OK
* checking Rd line widths ... OK
* checking Rd cross-references ... OK
* checking for missing documentation entries ... OK
* checking for code/documentation mismatches ... OK
* checking Rd \usage sections ... NOTE
S3 methods shown with full name in documentation object 'as.data.frame.default':
  ‘as.data.frame.default’

The \usage entries for S3 methods should use the \method markup and not
their full name.
See chapter ‘Writing R documentation files’ in the ‘Writing R
Extensions’ manual.
* checking Rd contents ... OK
* checking for unstated dependencies in examples ... OK
* checking contents of ‘data’ directory ... OK
* checking data for non-ASCII characters ... [0s/0s] OK
* checking data for ASCII and uncompressed saves ... OK
* checking files in ‘vignettes’ ... OK
* checking examples ... [88s/85s] OK
Examples with CPU (user + system) or elapsed time > 5s
                                 user system elapsed
get_ukbb_sumstat_10kloci_mt    35.197 10.452  46.448
as.data.frame.hail.table.Table 10.408 23.877  17.195
* checking for unstated dependencies in ‘tests’ ... OK
* checking tests ...
  Running ‘test.R’ [40s/14s]
 [40s/14s] OK
* checking for unstated dependencies in vignettes ... OK
* checking package vignettes in ‘inst/doc’ ... OK
* checking running R code from vignettes ... SKIPPED
* checking re-building of vignette outputs ... SKIPPED
* checking PDF version of manual ... [3s/3s] OK
* DONE

Status: 2 NOTEs
See
  ‘/home/pkgbuild/packagebuilder/workers/jobs/2947/BiocHail_20230321135853/BiocHail.Rcheck/00check.log’
for details.






===============================

 BiocCheck('BiocHail_0.99.15.tar.gz')

===============================

─ BiocCheckVersion: 1.35.13
─ BiocVersion: 3.17
─ Package: BiocHail
─ PackageVersion: 0.99.15
─ sourceDir: /tmp/RtmpNVfEBi/filecad6f3020931e/BiocHail
─ installDir: /tmp/RtmpNVfEBi/filecad6f2ebd7c32
─ BiocCheckDir: /home/pkgbuild/packagebuilder/workers/jobs/2947/BiocHail_20230321135853/BiocHail.BiocCheck
─ platform: unix
─ isTarBall: TRUE

* Installing package...
* Checking for deprecated package usage...
* Checking for remote package usage...
* Checking for 'LazyData: true' usage...
* Checking version number...
* Checking for version number mismatch...
* Checking new package version number...
* Checking R version dependency...
    * NOTE: Update R version dependency from 4.2 to 4.3.0.
* Checking package size...
* Checking individual file sizes...
* Checking biocViews...
* Checking that biocViews are present...
* Checking package type based on biocViews...
    Software
* Checking for non-trivial biocViews...
* Checking that biocViews come from the same category...
* Checking biocViews validity...
* Checking for recommended biocViews...
* Checking build system compatibility...
* Checking for blank lines in DESCRIPTION...
* Checking if DESCRIPTION is well formatted...
* Checking for proper Description: field...
    * NOTE: The Description field in the DESCRIPTION is made up by less
      than 3 sentences. Please consider expanding this field, and
      structure it as a full paragraph
* Checking for whitespace in DESCRIPTION field names...
* Checking that Package field matches directory/tarball name...
* Checking for Version field...
* Checking for valid maintainer...
* Checking License: for restrictive use...
* Checking for pinned package versions...
* Checking DESCRIPTION/NAMESPACE consistency...
* Checking .Rbuildignore...
* Checking for stray BiocCheck output folders...
* Checking vignette directory...
    * NOTE: 'sessionInfo' not found in vignette(s)
* Checking whether vignette is built with 'R CMD build'...
* Checking package installation calls in R code...
* Checking for library/require of BiocHail...
* Checking coding practice...
    * NOTE: Avoid 1:...; use seq_len() or seq_along()
    * NOTE: Avoid using '=' for assignment and use '<-' instead
* Checking parsed R code in R directory, examples, vignettes...
* Checking function lengths...
* Checking man page documentation...
    * WARNING: Non-empty \value sections found in man pages.
    * ERROR: At least 80% of man pages documenting exported objects
      must have runnable examples.
    * NOTE: Usage of dontrun{} / donttest{} tags found in man page
      examples. 7% of man pages use at least one of these tags.
    * NOTE: Use donttest{} instead of dontrun{}.
* Checking package NEWS...
    * NOTE: Consider adding a NEWS file, so your package news will be
      included in Bioconductor release announcements.
* Checking unit tests...
* Checking skip_on_bioc() in tests...
* Checking formatting of DESCRIPTION, NAMESPACE, man pages, R source,
  and vignette source...
    * NOTE: Consider shorter lines; 124 lines (6%) are > 80 characters
      long.
    * NOTE: Consider 4 spaces instead of tabs; 1 lines (0%) contain
      tabs.
    * NOTE: Consider multiples of 4 spaces for line indents; 204 lines
      (10%) are not.
    See https://contributions.bioconductor.org/r-code.html
    See styler package: https://cran.r-project.org/package=styler as
      described in the BiocCheck vignette.
* Checking if package already exists in CRAN...
* Checking if new package already exists in Bioconductor...
* Checking for bioc-devel mailing list subscription...
    Maintainer is subscribed to bioc-devel.
* Checking for support site registration...
    Maintainer is registered at support site.
    Package name is in support site watched tags.

─ BiocCheck results ──
1 ERRORS | 1 WARNINGS | 11 NOTES

See the BiocHail.BiocCheck folder and run
    browseVignettes(package = 'BiocCheck')
for details.

nebbiolo1 BUILD BIN output

[top]