Automatically generate key store for H2O Flow SSL
This option can be set either to "internal" or "external" When set to "external" H2O Context is created by connecting to existing H2O cluster, otherwise it creates H2O cluster living in Spark - that means that each Spark executor will have one h2o instance running in it.
This option can be set either to "internal" or "external" When set to "external" H2O Context is created by connecting to existing H2O cluster, otherwise it creates H2O cluster living in Spark - that means that each Spark executor will have one h2o instance running in it. The internal is not recommended for big clusters and clusters where Spark executors are not stable.
Enable or disable web on h2o client node.
Enable or disable web on h2o client node. It is enabled by default. Disabling the web just on the client node just restricts everybody from accessing flow, the internal ports between client and rest of the cluster remain open
Extra properties passed to H2O client during startup.
Allows to override the base URL address of Flow UI, including the scheme, which is showed to the user.
Location of iced directory for the driver instance.
Ignore SPARK_PUBLIC_DNS setting on the H2O client.
Ignore SPARK_PUBLIC_DNS setting on the H2O client. The option still applies to the Spark application.
IP of H2O client node
Location of log directory for the driver instance.
H2O log level for client running in Spark driver
Subnet selector for H2O client - if the mask is specified then Spark network setup is not discussed.
Port on which H2O client publishes its API.
Port on which H2O client publishes its API. If already occupied, the next odd port is tried and so on
Print detailed messages to client stdout
Exact client port to access web UI.
Exact client port to access web UI.
The value -1
means automatic search for free port starting at spark.ext.h2o.port.base
.
Configuration property - name of H2O cloud
Configuration property - timeout for cloud up.
Enable/Disable listener which kills H2O when there is a change in underlying cluster's topology*
H2O's URL context path
Timeout in milliseconds specifying how often we check whether the client is still connected
Enable/Disable exit on unsupported Spark parameters.
Path to flow dir.
Extra http headers for Flow UI
Decide whether Scala cells are running synchronously or asynchronously
Number of max parallel Scala cell jobs.
Enable hash login.
Offset between the API(=web) port and the internal communication port; api_port + port_offset = h2o_port
Secure internal connections by automatically generated credentials
Path to Java KeyStore file.
Alias for certificate in keystore to secure Flow
Password for Java KeyStore file.
Enable Kerberos login.
Enable LDAP login.
Login configuration file.
If a scoring MOJO instance is not used within a Spark executor JVM for a given timeout in milliseconds, it's evicted from executor's cache.
Enable or disable web on h2o worker nodes.
Enable or disable web on h2o worker nodes. It is disabled by default for security reasons.
Extra properties passed to H2O nodes during startup.
Location of log directory for remote nodes.
H2O internal log level for launched remote nodes.
Subnet selector for H2O nodes running inside executors - if the mask is specified then Spark network setup is not discussed.
Configuration property - base port used for individual H2O nodes configuration.
Set how often in seconds stack traces are taken on each h2o node.
Set how often in seconds stack traces are taken on each h2o node. -1 represents that the stack traces are not collected.
Limit for number of threads used by H2O, default -1 means unlimited
Enable/Disable Sparkling-Water REPL *
Number of executors started at the start of h2o services, by default 1
Enable/Disable check for Spark version.
Path to Java KeyStore file used for the internal SSL communication.
Interval for updates of Spark UI in milliseconds
Override user name for cluster.