t

com.holdenkarau.spark.testing

SharedSparkContext

trait SharedSparkContext extends BeforeAndAfterAll with SparkContextProvider

Shares a local SparkContext between all tests in a suite and closes it at the end. You can share between suites by enabling reuseContextIfPossible.

Self Type
SharedSparkContext with Suite
Linear Supertypes
SparkContextProvider, BeforeAndAfterAll, SuiteMixin, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. SharedSparkContext
  2. SparkContextProvider
  3. BeforeAndAfterAll
  4. SuiteMixin
  5. AnyRef
  6. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Abstract Value Members

  1. abstract def expectedTestCount(filter: Filter): Int
    Definition Classes
    SuiteMixin
  2. abstract def nestedSuites: IndexedSeq[Suite]
    Definition Classes
    SuiteMixin
  3. abstract def rerunner: Option[String]
    Definition Classes
    SuiteMixin
  4. abstract def runNestedSuites(args: Args): Status
    Attributes
    protected
    Definition Classes
    SuiteMixin
  5. abstract def runTest(testName: String, args: Args): Status
    Attributes
    protected
    Definition Classes
    SuiteMixin
  6. abstract def runTests(testName: Option[String], args: Args): Status
    Attributes
    protected
    Definition Classes
    SuiteMixin
  7. abstract def suiteId: String
    Definition Classes
    SuiteMixin
  8. abstract def suiteName: String
    Definition Classes
    SuiteMixin
  9. abstract def tags: Map[String, Set[String]]
    Definition Classes
    SuiteMixin
  10. abstract def testDataFor(testName: String, theConfigMap: ConfigMap): TestData
    Definition Classes
    SuiteMixin
  11. abstract def testNames: Set[String]
    Definition Classes
    SuiteMixin
  12. abstract val styleName: String
    Definition Classes
    SuiteMixin
    Annotations
    @deprecated
    Deprecated

    (Since version 3.1.0) The styleName lifecycle method has been deprecated and will be removed in a future version of ScalaTest with no replacement.

Concrete Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. def afterAll(): Unit
    Definition Classes
    SharedSparkContext → BeforeAndAfterAll
  5. def appID: String
    Definition Classes
    SparkContextProvider
  6. def appName: String
    Definition Classes
    SparkContextProvider
  7. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  8. def beforeAll(): Unit
    Definition Classes
    SharedSparkContext → BeforeAndAfterAll
  9. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native() @IntrinsicCandidate()
  10. def conf: SparkConf
    Definition Classes
    SparkContextProvider
  11. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  12. def equals(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  13. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native() @IntrinsicCandidate()
  14. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @native() @IntrinsicCandidate()
  15. val invokeBeforeAllAndAfterAllEvenIfNoTestsAreExpected: Boolean
    Definition Classes
    BeforeAndAfterAll
  16. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  17. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  18. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native() @IntrinsicCandidate()
  19. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native() @IntrinsicCandidate()
  20. implicit def reuseContextIfPossible: Boolean
    Attributes
    protected
  21. def run(testName: Option[String], args: Args): Status
    Definition Classes
    BeforeAndAfterAll → SuiteMixin
  22. def sc: SparkContext
  23. def setup(sc: SparkContext): Unit

    Setup work to be called when creating a new SparkContext.

    Setup work to be called when creating a new SparkContext. Default implementation currently sets a checkpoint directory.

    This _should_ be called by the context provider automatically.

    Definition Classes
    SparkContextProvider
  24. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  25. def toString(): String
    Definition Classes
    AnyRef → Any
  26. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  27. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  28. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Deprecated Value Members

  1. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] ) @Deprecated
    Deprecated

Inherited from SparkContextProvider

Inherited from BeforeAndAfterAll

Inherited from SuiteMixin

Inherited from AnyRef

Inherited from Any

Ungrouped