c

geotrellis.spark.store.hadoop

HadoopSparkLayerProvider

class HadoopSparkLayerProvider extends HadoopCollectionLayerProvider with LayerReaderProvider with LayerWriterProvider

Provides HadoopAttributeStore instance for URI with hdfs, hdfs+file, s3n, s3a, wasb and wasbs schemes. The uri represents Hadoop Path of catalog root. wasb and wasbs provide support for the Hadoop Azure connector. Additional configuration is required for this. This Provider intentinally does not handle the s3 scheme because the Hadoop implemintation is poor. That support is provided by HadoopAttributeStore

Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. HadoopSparkLayerProvider
  2. LayerWriterProvider
  3. LayerReaderProvider
  4. HadoopCollectionLayerProvider
  5. CollectionLayerReaderProvider
  6. ValueReaderProvider
  7. AttributeStoreProvider
  8. AnyRef
  9. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new HadoopSparkLayerProvider()

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def attributeStore(uri: URI): AttributeStore
  6. def canProcess(uri: URI): Boolean
  7. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()
  8. def collectionLayerReader(uri: URI, store: AttributeStore): HadoopCollectionLayerReader
  9. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  10. def equals(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  11. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  12. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  13. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  14. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  15. def layerReader(uri: URI, store: AttributeStore, sc: SparkContext): FilteringLayerReader[LayerId]
  16. def layerWriter(uri: URI, store: AttributeStore): LayerWriter[LayerId]
  17. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  18. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  19. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  20. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  21. def toString(): String
    Definition Classes
    AnyRef → Any
  22. def valueReader(uri: URI, store: AttributeStore): ValueReader[LayerId]
  23. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  24. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  25. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... ) @native()

Inherited from LayerWriterProvider

Inherited from LayerReaderProvider

Inherited from ValueReaderProvider

Inherited from AttributeStoreProvider

Inherited from AnyRef

Inherited from Any

Ungrouped