Class

org.apache.predictionio.controller

PDataSource

Related Doc: package controller

Permalink

abstract class PDataSource[TD, EI, Q, A] extends BaseDataSource[TD, EI, Q, A]

Base class of a parallel data source.

A parallel data source runs locally within a single machine, or in parallel on a cluster, to return data that is distributed across a cluster.

TD

Training data class.

EI

Evaluation Info class.

Q

Input query class.

A

Actual value class.

Linear Supertypes
BaseDataSource[TD, EI, Q, A], AbstractDoer, Serializable, Serializable, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. PDataSource
  2. BaseDataSource
  3. AbstractDoer
  4. Serializable
  5. Serializable
  6. AnyRef
  7. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new PDataSource()

    Permalink

Abstract Value Members

  1. abstract def readTraining(sc: SparkContext): TD

    Permalink

    Implement this method to only return training data from a data source

Concrete Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  5. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  6. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  7. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  8. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  9. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  10. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  11. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  12. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  13. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  14. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  15. def readEval(sc: SparkContext): Seq[(TD, EI, RDD[(Q, A)])]

    Permalink

    To provide evaluation feature for your engine, your must override this method to return data for evaluation from a data source.

    To provide evaluation feature for your engine, your must override this method to return data for evaluation from a data source. Returned data can optionally include a sequence of query and actual value pairs for evaluation purpose.

    The default implementation returns an empty sequence as a stub, so that an engine can be compiled without implementing evaluation.

  16. def readEvalBase(sc: SparkContext): Seq[(TD, EI, RDD[(Q, A)])]

    Permalink

    :: DeveloperApi :: Engine developer should not use this directly.

    :: DeveloperApi :: Engine developer should not use this directly. This is called by evaluation workflow to read training and validation data.

    sc

    Spark context

    returns

    Sets of training data, evaluation information, queries, and actual results

    Definition Classes
    PDataSourceBaseDataSource
  17. def readTrainingBase(sc: SparkContext): TD

    Permalink

    :: DeveloperApi :: Engine developer should not use this directly.

    :: DeveloperApi :: Engine developer should not use this directly. This is called by workflow to read training data.

    sc

    Spark context

    returns

    Training data

    Definition Classes
    PDataSourceBaseDataSource
  18. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  19. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  20. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  21. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  22. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Deprecated Value Members

  1. def read(sc: SparkContext): Seq[(TD, EI, RDD[(Q, A)])]

    Permalink
    Annotations
    @deprecated
    Deprecated

    (Since version 0.9.0) Use readEval() instead.

Inherited from BaseDataSource[TD, EI, Q, A]

Inherited from AbstractDoer

Inherited from Serializable

Inherited from Serializable

Inherited from AnyRef

Inherited from Any

Ungrouped