Instance Constructors
-
new
H2OConf(jsc: JavaSparkContext)
-
new
H2OConf(sc: SparkContext)
Value Members
-
final
def
!=(arg0: Any): Boolean
-
final
def
##(): Int
-
final
def
==(arg0: Any): Boolean
-
final
def
asInstanceOf[T0]: T0
-
def
backendClusterMode: String
-
def
clientBasePort: Int
-
def
clientIcedDir: Option[String]
-
def
clientIp: Option[String]
-
def
clientNetworkMask: Option[String]
-
def
clientVerboseOutput: Boolean
-
def
clientWebPort: Int
-
-
def
cloudName: Option[String]
-
def
cloudTimeout: Int
-
def
contains(key: String): Boolean
-
def
defaultCloudSize: Int
-
def
disableGA: Boolean
-
def
drddMulFactor: Int
-
final
def
eq(arg0: AnyRef): Boolean
-
def
equals(arg0: Any): Boolean
-
def
finalize(): Unit
-
def
get(key: String, defaultValue: String): String
-
def
get(key: String): String
-
def
getAll: Array[(String, String)]
-
def
getBoolean(key: String, defaultValue: Boolean): Boolean
-
final
def
getClass(): Class[_]
-
def
getDouble(key: String, defaultValue: Double): Double
-
def
getInt(key: String, defaultValue: Int): Int
-
def
getLong(key: String, defaultValue: Long): Long
-
def
getOption(key: String): Option[String]
-
def
h2oClientLogDir: String
-
def
h2oClientLogLevel: String
-
def
h2oNodeLogDir: String
-
def
h2oNodeLogLevel: String
-
def
hashCode(): Int
-
def
hashLogin: Boolean
-
def
initializeLogIfNecessary(isInterpreter: Boolean): Unit
-
def
internalConfString: String
-
def
isClusterTopologyListenerEnabled: Boolean
-
def
isFailOnUnsupportedSparkParamEnabled: Boolean
-
def
isH2OReplEnabled: Boolean
-
final
def
isInstanceOf[T0]: Boolean
-
def
isSparkVersionCheckEnabled: Boolean
-
def
isTraceEnabled(): Boolean
-
def
jks: Option[String]
-
def
jksPass: Option[String]
-
def
kerberosLogin: Boolean
-
def
ldapLogin: Boolean
-
def
log: Logger
-
def
logDebug(msg: ⇒ String, throwable: Throwable): Unit
-
def
logDebug(msg: ⇒ String): Unit
-
def
logError(msg: ⇒ String, throwable: Throwable): Unit
-
def
logError(msg: ⇒ String): Unit
-
def
logInfo(msg: ⇒ String, throwable: Throwable): Unit
-
def
logInfo(msg: ⇒ String): Unit
-
def
logName: String
-
def
logTrace(msg: ⇒ String, throwable: Throwable): Unit
-
def
logTrace(msg: ⇒ String): Unit
-
def
logWarning(msg: ⇒ String, throwable: Throwable): Unit
-
def
logWarning(msg: ⇒ String): Unit
-
def
loginConf: Option[String]
-
final
def
ne(arg0: AnyRef): Boolean
-
def
nodeBasePort: Int
-
def
nodeIcedDir: Option[String]
-
def
nodeNetworkMask: Option[String]
-
final
def
notify(): Unit
-
final
def
notifyAll(): Unit
-
def
nthreads: Int
-
def
numH2OWorkers: Option[Int]
-
def
numRddRetries: Int
-
def
remove(key: String): H2OConf
-
def
runsInExternalClusterMode: Boolean
-
def
runsInInternalClusterMode: Boolean
-
val
sc: SparkContext
-
def
scalaIntDefaultNum: Int
-
def
set(key: String, value: String): H2OConf
-
def
setAll(settings: Traversable[(String, String)]): H2OConf
-
def
setClientIp(ip: String): H2OConf
-
def
setCloudName(cloudName: String): H2OConf
-
def
setExternalClusterMode(): H2OConf
-
def
setH2OClientLogLevel(level: String): H2OConf
-
def
setH2ONodeLogLevel(level: String): H2OConf
-
def
setInternalClusterMode(): H2OConf
-
def
setReplDisabled(): H2OConf
-
def
setReplEnabled(): H2OConf
-
val
sparkConf: SparkConf
-
def
subseqTries: Int
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
-
def
toString(): String
-
def
useFlatFile: Boolean
-
def
userName: Option[String]
-
final
def
wait(): Unit
-
final
def
wait(arg0: Long, arg1: Int): Unit
-
final
def
wait(arg0: Long): Unit
Inherited from Logging
Inherited from AnyRef
Inherited from Any
Configuration holder which is representing properties passed from user to Sparkling Water.