Object

geotrellis.spark.io.hadoop

HadoopGeoTiffRDD

Related Doc: package hadoop

Permalink

object HadoopGeoTiffRDD extends LazyLogging

Allows for reading of whole or windowed GeoTiff as RDD[(K, V)]s through Hadoop FileSystem API.

Linear Supertypes
LazyLogging, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. HadoopGeoTiffRDD
  2. LazyLogging
  3. AnyRef
  4. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Type Members

  1. case class Options(tiffExtensions: Seq[String] = ..., crs: Option[CRS] = None, timeTag: String = GEOTIFF_TIME_TAG_DEFAULT, timeFormat: String = GEOTIFF_TIME_FORMAT_DEFAULT, maxTileSize: Option[Int] = Some(DefaultMaxTileSize), numPartitions: Option[Int] = None, partitionBytes: Option[Long] = Some(DefaultPartitionBytes), chunkSize: Option[Int] = None) extends RasterReader.Options with Product with Serializable

    Permalink

    This case class contains the various parameters one can set when reading RDDs from Hadoop using Spark.

    This case class contains the various parameters one can set when reading RDDs from Hadoop using Spark.

    tiffExtensions

    Read all file with an extension contained in the given list.

    crs

    Override CRS of the input files. If None, the reader will use the file's original CRS.

    timeTag

    Name of tiff tag containing the timestamp for the tile.

    timeFormat

    Pattern for java.time.format.DateTimeFormatter to parse timeTag.

    maxTileSize

    Maximum allowed size of each tiles in output RDD. May result in a one input GeoTiff being split amongst multiple records if it exceeds this size. If no maximum tile size is specific, then each file is broken into 256x256 tiles. If None, then the whole file will be read in. This option is incompatible with numPartitions and anything set to that parameter will be ignored.

    numPartitions

    How many partitions Spark should create when it repartitions the data.

    partitionBytes

    Desired partition size in bytes, at least one item per partition will be assigned. If no size is specified, then partitions 128 Mb in size will be created by default. This option is incompatible with the numPartitions option. If both are set and maxTileSize isn't, then partitionBytes will be ignored in favor of numPartitions. However, if maxTileSize is set, then partitionBytes will be retained. If None and maxTileSize is defined, then the default partitionBytes' value will still be used. If maxTileSize is also None, then partitionBytes will remain None as well.

    chunkSize

    How many bytes should be read in at a time when reading a file. If None, then 65536 byte chunks will be read in at a time.

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. final val GEOTIFF_TIME_FORMAT_DEFAULT: String("yyyy:MM:dd HH:mm:ss")

    Permalink
  5. final val GEOTIFF_TIME_TAG_DEFAULT: String("TIFFTAG_DATETIME")

    Permalink
  6. object Options extends Serializable

    Permalink
  7. def apply[I, K, V](pathsToDimensions: RDD[(Path, (Int, Int))], uriToKey: (URI, I) ⇒ K, options: Options)(implicit rr: RasterReader[Options, (I, V)]): RDD[(K, V)]

    Permalink

    Creates a RDD[(K, V)] whose K and V depends on the type of the GeoTiff that is going to be read in.

    Creates a RDD[(K, V)] whose K and V depends on the type of the GeoTiff that is going to be read in.

    pathsToDimensions

    RDD keyed by GeoTiff path with (cols, rows) tuple as value.

    uriToKey

    A function to transform input key basing on the URI information.

    options

    An instance of Options that contains any user defined or default settings.

  8. def apply[K, V](path: Path, options: Options)(implicit sc: SparkContext, rr: RasterReader[Options, (K, V)]): RDD[(K, V)]

    Permalink

    Creates a RDD[(K, V)] whose K and V depends on the type of the GeoTiff that is going to be read in.

    Creates a RDD[(K, V)] whose K and V depends on the type of the GeoTiff that is going to be read in.

    path

    Hdfs GeoTiff path.

    options

    An instance of Options that contains any user defined or default settings.

  9. def apply[I, K, V](path: Path, uriToKey: (URI, I) ⇒ K, options: Options, geometry: Option[Geometry] = None)(implicit sc: SparkContext, rr: RasterReader[Options, (I, V)]): RDD[(K, V)]

    Permalink

    Creates a RDD[(K, V)] whose K and V depends on the type of the GeoTiff that is going to be read in.

    Creates a RDD[(K, V)] whose K and V depends on the type of the GeoTiff that is going to be read in.

    This function has two modes of operation: When options.maxTileSize is set windows will be read from GeoTiffs and their size and count will be balanced among partitions using partitionBytes option. Resulting partitions will be grouped in relation to GeoTiff segment layout.

    When maxTileSize is None the GeoTiffs will be read fully and balanced among partitions using either numPartitions or partitionBytes option.

    path

    HDFS GeoTiff path.

    uriToKey

    Function to transform input key basing on the URI information.

    options

    An instance of Options that contains any user defined or default settings.

    geometry

    An optional geometry to filter by. If this is provided, it is assumed that all GeoTiffs are in the same CRS, and that this geometry is in that CRS.

  10. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  11. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  12. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  13. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  14. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  15. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  16. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  17. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  18. lazy val logger: Logger

    Permalink
    Attributes
    protected
    Definition Classes
    LazyLogging
  19. def multiband[K](path: Path, options: Options)(implicit sc: SparkContext, rr: RasterReader[Options, (K, MultibandTile)]): RDD[(K, MultibandTile)]

    Permalink

    Creates RDDs with the [(K, V)] values where V is a MultibandTile.

    Creates RDDs with the [(K, V)] values where V is a MultibandTile. It assumes that the provided files are MultibandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    options

    An instance of Options that contains any user defined or default settings.

  20. def multiband[I, K](path: Path, uriToKey: (URI, I) ⇒ K, options: Options)(implicit sc: SparkContext, rr: RasterReader[Options, (I, MultibandTile)]): RDD[(K, MultibandTile)]

    Permalink

    Creates RDDs with the [(K, V)] values where V is a MultibandTile.

    Creates RDDs with the [(K, V)] values where V is a MultibandTile. It assumes that the provided files are MultibandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    uriToKey

    function to transform input key basing on the URI information.

    options

    An instance of Options that contains any user defined or default settings.

  21. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  22. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  23. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  24. def singleband[K](path: Path, options: Options)(implicit sc: SparkContext, rr: RasterReader[Options, (K, Tile)]): RDD[(K, Tile)]

    Permalink

    Creates RDDs with the [(K, V)] values where V is a Tile.

    Creates RDDs with the [(K, V)] values where V is a Tile. It assumes that the provided files are SinglebandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    options

    An instance of Options that contains any user defined or default settings.

  25. def singleband[I, K](path: Path, uriToKey: (URI, I) ⇒ K, options: Options)(implicit sc: SparkContext, rr: RasterReader[Options, (I, Tile)]): RDD[(K, Tile)]

    Permalink

    Creates RDDs with the [(K, V)] values where V is a Tile.

    Creates RDDs with the [(K, V)] values where V is a Tile. It assumes that the provided files are SinglebandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    uriToKey

    function to transform input key basing on the URI information.

    options

    An instance of Options that contains any user defined or default settings.

  26. def spatial(path: Path, uriToKey: (URI, ProjectedExtent) ⇒ ProjectedExtent, options: Options)(implicit sc: SparkContext): RDD[(ProjectedExtent, Tile)]

    Permalink

    Creates RDDs with the [(K, V)] values being ProjectedExtent and Tile, respectively.

    Creates RDDs with the [(K, V)] values being ProjectedExtent and Tile, respectively. It assumes that the provided files are SinglebandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    uriToKey

    function to transform input key basing on the URI information.

    options

    An instance of Options that contains any user defined or default settings.

  27. def spatial(path: Path, options: Options)(implicit sc: SparkContext): RDD[(ProjectedExtent, Tile)]

    Permalink

    Creates RDDs with the [(K, V)] values being ProjectedExtent and Tile, respectively.

    Creates RDDs with the [(K, V)] values being ProjectedExtent and Tile, respectively. It assumes that the provided files are SinglebandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    options

    An instance of Options that contains any user defined or default settings.

  28. def spatial(path: Path)(implicit sc: SparkContext): RDD[(ProjectedExtent, Tile)]

    Permalink

    Creates RDDs with the [(K, V)] values being ProjectedExtent and Tile, respectively.

    Creates RDDs with the [(K, V)] values being ProjectedExtent and Tile, respectively. It assumes that the provided files are SinglebandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

  29. def spatialMultiband(path: Path, uriToKey: (URI, ProjectedExtent) ⇒ ProjectedExtent, options: Options)(implicit sc: SparkContext): RDD[(ProjectedExtent, MultibandTile)]

    Permalink

    Creates RDDs with the [(K, V)] values being ProjectedExtent and MultibandTile, respectively.

    Creates RDDs with the [(K, V)] values being ProjectedExtent and MultibandTile, respectively. It assumes that the provided files are MultibandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    uriToKey

    function to transform input key basing on the URI information.

    options

    An instance of Options that contains any user defined or default settings.

  30. def spatialMultiband(path: Path, options: Options)(implicit sc: SparkContext): RDD[(ProjectedExtent, MultibandTile)]

    Permalink

    Creates RDDs with the [(K, V)] values being ProjectedExtent and MultibandTile, respectively.

    Creates RDDs with the [(K, V)] values being ProjectedExtent and MultibandTile, respectively. It assumes that the provided files are MultibandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    options

    An instance of Options that contains any user defined or default settings.

  31. def spatialMultiband(path: Path)(implicit sc: SparkContext): RDD[(ProjectedExtent, MultibandTile)]

    Permalink

    Creates RDDs with the [(K, V)] values being ProjectedExtent and MultibandTile, respectively.

    Creates RDDs with the [(K, V)] values being ProjectedExtent and MultibandTile, respectively. It assumes that the provided files are MultibandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

  32. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  33. def temporal(path: Path, uriToKey: (URI, TemporalProjectedExtent) ⇒ TemporalProjectedExtent, options: Options)(implicit sc: SparkContext): RDD[(TemporalProjectedExtent, Tile)]

    Permalink

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and Tile, respectively.

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and Tile, respectively. It assumes that the provided files are SinglebandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    uriToKey

    function to transform input key basing on the URI information.

    options

    An instance of Options that contains any user defined or default settings.

  34. def temporal(path: Path, options: Options)(implicit sc: SparkContext): RDD[(TemporalProjectedExtent, Tile)]

    Permalink

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and Tile, respectively.

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and Tile, respectively. It assumes that the provided files are SinglebandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    options

    An instance of Options that contains any user defined or default settings.

  35. def temporal(path: Path)(implicit sc: SparkContext): RDD[(TemporalProjectedExtent, Tile)]

    Permalink

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and Tile, respectively.

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and Tile, respectively. It assumes that the provided files are SinglebandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

  36. def temporalMultiband(path: Path, uriToKey: (URI, TemporalProjectedExtent) ⇒ TemporalProjectedExtent, options: Options)(implicit sc: SparkContext): RDD[(TemporalProjectedExtent, MultibandTile)]

    Permalink

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and MultibandTile, respectively.

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and MultibandTile, respectively. It assumes that the provided files are MultibandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    uriToKey

    function to transform input key basing on the URI information.

    options

    An instance of Options that contains any user defined or default settings.

  37. def temporalMultiband(path: Path, options: Options)(implicit sc: SparkContext): RDD[(TemporalProjectedExtent, MultibandTile)]

    Permalink

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and MultibandTile, respectively.

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and MultibandTile, respectively. It assumes that the provided files are MultibandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

    options

    An instance of Options that contains any user defined or default settings.

  38. def temporalMultiband(path: Path)(implicit sc: SparkContext): RDD[(TemporalProjectedExtent, MultibandTile)]

    Permalink

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and MultibandTile, respectively.

    Creates RDDs with the [(K, V)] values being TemporalProjectedExtent and MultibandTile, respectively. It assumes that the provided files are MultibandGeoTiffs.

    path

    Hadoop path to recursively search for GeoTiffs.

  39. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  40. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  41. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  42. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Inherited from LazyLogging

Inherited from AnyRef

Inherited from Any

Ungrouped