case class HdfsWriteStrategy(ingestPath: Path) extends AccumuloWriteStrategy with Product with Serializable

This strategy will perfom Accumulo bulk ingest. Bulk ingest requires that sorted records be written to the filesystem, preferbly HDFS, before Accumulo is able to ingest them. After the ingest is finished the nodes will likely go through a period of high load as they perform major compactions.

Note: Giving relative URLs will cause HDFS to use the fs.defaultFS property in core-site.xml. If not specified this will default to local ('file:/') system, this is undesriable.

ingestPath

Path where spark will write RDD records for ingest

Linear Supertypes
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. HdfsWriteStrategy
  2. Product
  3. Equals
  4. AccumuloWriteStrategy
  5. Serializable
  6. Serializable
  7. AnyRef
  8. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new HdfsWriteStrategy(ingestPath: Path)

    ingestPath

    Path where spark will write RDD records for ingest

Value Members

  1. val ingestPath: Path
  2. def write(kvPairs: RDD[(Key, Value)], instance: AccumuloInstance, table: String): Unit

    Requires that the RDD be pre-sorted

    Requires that the RDD be pre-sorted

    Definition Classes
    HdfsWriteStrategyAccumuloWriteStrategy