Packages

case class ParquetWrite(paths: Seq[String], formatName: String, supportsDataType: (DataType) => Boolean, info: LogicalWriteInfo) extends FileWrite with Logging with Product with Serializable

Linear Supertypes
Serializable, Product, Equals, Logging, FileWrite, Write, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. ParquetWrite
  2. Serializable
  3. Product
  4. Equals
  5. Logging
  6. FileWrite
  7. Write
  8. AnyRef
  9. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. Protected

Instance Constructors

  1. new ParquetWrite(paths: Seq[String], formatName: String, supportsDataType: (DataType) => Boolean, info: LogicalWriteInfo)

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##: Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.CloneNotSupportedException]) @native()
  6. def description(): String
    Definition Classes
    FileWrite → Write
  7. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  8. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.Throwable])
  9. val formatName: String
    Definition Classes
    ParquetWriteFileWrite
  10. final def getClass(): Class[_ <: AnyRef]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  11. val info: LogicalWriteInfo
    Definition Classes
    ParquetWriteFileWrite
  12. def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
    Attributes
    protected
    Definition Classes
    Logging
  13. def initializeLogIfNecessary(isInterpreter: Boolean): Unit
    Attributes
    protected
    Definition Classes
    Logging
  14. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  15. def isTraceEnabled(): Boolean
    Attributes
    protected
    Definition Classes
    Logging
  16. def log: Logger
    Attributes
    protected
    Definition Classes
    Logging
  17. def logDebug(msg: => String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  18. def logDebug(msg: => String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  19. def logError(msg: => String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  20. def logError(msg: => String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  21. def logInfo(msg: => String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  22. def logInfo(msg: => String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  23. def logName: String
    Attributes
    protected
    Definition Classes
    Logging
  24. def logTrace(msg: => String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  25. def logTrace(msg: => String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  26. def logWarning(msg: => String, throwable: Throwable): Unit
    Attributes
    protected
    Definition Classes
    Logging
  27. def logWarning(msg: => String): Unit
    Attributes
    protected
    Definition Classes
    Logging
  28. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  29. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  30. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  31. val paths: Seq[String]
    Definition Classes
    ParquetWriteFileWrite
  32. def prepareWrite(sqlConf: SQLConf, job: Job, options: Map[String, String], dataSchema: StructType): OutputWriterFactory

    Prepares a write job and returns an OutputWriterFactory.

    Prepares a write job and returns an OutputWriterFactory. Client side job preparation can be put here. For example, user defined output committer can be configured here by setting the output committer class in the conf of spark.sql.sources.outputCommitterClass.

    Definition Classes
    ParquetWriteFileWrite
  33. def productElementNames: Iterator[String]
    Definition Classes
    Product
  34. def supportedCustomMetrics(): Array[CustomMetric]
    Definition Classes
    Write
  35. val supportsDataType: (DataType) => Boolean
    Definition Classes
    ParquetWriteFileWrite
  36. final def synchronized[T0](arg0: => T0): T0
    Definition Classes
    AnyRef
  37. def toBatch(): BatchWrite
    Definition Classes
    FileWrite → Write
  38. def toStreaming(): StreamingWrite
    Definition Classes
    Write
  39. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])
  40. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])
  41. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException]) @native()

Inherited from Serializable

Inherited from Product

Inherited from Equals

Inherited from Logging

Inherited from FileWrite

Inherited from Write

Inherited from AnyRef

Inherited from Any

Ungrouped