spark-tests


License

License

GroupId

GroupId

org.hammerlab
ArtifactId

ArtifactId

spark-tests_2.12
Last Version

Last Version

2.4.0
Release Date

Release Date

Type

Type

jar
Description

Description

spark-tests
spark-tests
Project URL

Project URL

https://github.com/hammerlab/spark-tests
Project Organization

Project Organization

org.hammerlab
Source Code Management

Source Code Management

https://github.com/hammerlab/spark-tests

Download spark-tests_2.12

How to add to project

<!-- https://jarcasting.com/artifacts/org.hammerlab/spark-tests_2.12/ -->
<dependency>
    <groupId>org.hammerlab</groupId>
    <artifactId>spark-tests_2.12</artifactId>
    <version>2.4.0</version>
</dependency>
// https://jarcasting.com/artifacts/org.hammerlab/spark-tests_2.12/
implementation 'org.hammerlab:spark-tests_2.12:2.4.0'
// https://jarcasting.com/artifacts/org.hammerlab/spark-tests_2.12/
implementation ("org.hammerlab:spark-tests_2.12:2.4.0")
'org.hammerlab:spark-tests_2.12:jar:2.4.0'
<dependency org="org.hammerlab" name="spark-tests_2.12" rev="2.4.0">
  <artifact name="spark-tests_2.12" type="jar" />
</dependency>
@Grapes(
@Grab(group='org.hammerlab', module='spark-tests_2.12', version='2.4.0')
)
libraryDependencies += "org.hammerlab" % "spark-tests_2.12" % "2.4.0"
[org.hammerlab/spark-tests_2.12 "2.4.0"]

Dependencies

compile (7)

Group / Artifact Type Version
org.scala-lang : scala-library jar 2.12.8
org.hammerlab : paths_2.12 jar 1.5.0
org.scalatest : scalatest_2.12 jar 3.0.4
org.clapper : grizzled-slf4j_2.12 jar 1.3.1
org.apache.spark : spark-core_2.12 jar 2.4.0
org.hammerlab : spark-util_2.12 jar 3.1.0
org.hammerlab.test : base_2.12 jar 1.1.0

provided (1)

Group / Artifact Type Version
com.google.guava : guava jar 14.0.1

test (1)

Group / Artifact Type Version
uk.org.lidalia : slf4j-test jar 1.1.0

Project Modules

There are no modules declared in this project.

spark-tests

Build Status Coverage Status Maven Central

Utilities for writing tests that use Apache Spark.

SparkSuite: a SparkContext for each test suite

Add configuration options in subclasses using sparkConf(…), cf. KryoSparkSuite:

sparkConf(
  // Register this class as its own KryoRegistrator
  "spark.kryo.registrator"  getClass.getCanonicalName,
  "spark.serializer"  "org.apache.spark.serializer.KryoSerializer",
  "spark.kryo.referenceTracking"  referenceTracking.toString,
  "spark.kryo.registrationRequired"  registrationRequired.toString
)

PerCaseSuite: SparkContext for each test case

KryoSparkSuite

SparkSuite implementation that provides hooks for kryo-registration:

register(
  classOf[Foo],
  "org.foo.Bar",
  classOf[Bar]  new BarSerializer
)

Also useful for subclassing once per-project and filling in that project's default Kryo registrar, then having concrete tests subclass that; see cf. hammerlab/guacamole and hammerlab/pageant for examples.

Miscellaneous RDD / Job / Stage utilities

  • rdd.Util: make an RDD with specific elements in specific partitions.
  • NumJobsUtil: verify the number of Spark jobs that have been run.
  • RDDSerialization: interface that allows for verifying that performing a serialization+deserialization round-trip on an RDD results in the same RDD.
org.hammerlab

Hammer Lab

We're a lab working to understand and improve the immune response to cancer

Versions

Version
2.4.0