public final class PySparkJob extends GeneratedMessageV3 implements PySparkJobOrBuilder
A Cloud Dataproc job for running [Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN.Protobuf type
google.cloud.dataproc.v1.PySparkJob
Modifier and Type | Class and Description |
---|---|
static class |
PySparkJob.Builder
A Cloud Dataproc job for running
[Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html)
applications on YARN.
|
GeneratedMessageV3.BuilderParent, GeneratedMessageV3.ExtendableBuilder<MessageType extends GeneratedMessageV3.ExtendableMessage,BuilderType extends GeneratedMessageV3.ExtendableBuilder<MessageType,BuilderType>>, GeneratedMessageV3.ExtendableMessage<MessageType extends GeneratedMessageV3.ExtendableMessage>, GeneratedMessageV3.ExtendableMessageOrBuilder<MessageType extends GeneratedMessageV3.ExtendableMessage>, GeneratedMessageV3.FieldAccessorTable
Modifier and Type | Field and Description |
---|---|
static int |
ARCHIVE_URIS_FIELD_NUMBER |
static int |
ARGS_FIELD_NUMBER |
static int |
FILE_URIS_FIELD_NUMBER |
static int |
JAR_FILE_URIS_FIELD_NUMBER |
static int |
LOGGING_CONFIG_FIELD_NUMBER |
static int |
MAIN_PYTHON_FILE_URI_FIELD_NUMBER |
static int |
PROPERTIES_FIELD_NUMBER |
static int |
PYTHON_FILE_URIS_FIELD_NUMBER |
alwaysUseFieldBuilders, unknownFields
memoizedSize
memoizedHashCode
canUseUnsafe, computeStringSize, computeStringSizeNoTag, getAllFields, getDescriptorForType, getField, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, hasField, hasOneof, makeExtensionsImmutable, newBuilderForType, parseDelimitedWithIOException, parseDelimitedWithIOException, parseUnknownField, parseUnknownFieldProto3, parseWithIOException, parseWithIOException, parseWithIOException, parseWithIOException, serializeBooleanMapTo, serializeIntegerMapTo, serializeLongMapTo, serializeStringMapTo, writeReplace, writeString, writeStringNoTag
findInitializationErrors, getInitializationErrorString, hashBoolean, hashEnum, hashEnumList, hashFields, hashLong, toString
addAll, addAll, checkByteStringIsUtf8, toByteArray, toByteString, writeDelimitedTo, writeTo
clone, finalize, getClass, notify, notifyAll, wait, wait, wait
findInitializationErrors, getAllFields, getDescriptorForType, getField, getInitializationErrorString, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, hasField, hasOneof
toByteArray, toByteString, writeDelimitedTo, writeTo
public static final int MAIN_PYTHON_FILE_URI_FIELD_NUMBER
public static final int ARGS_FIELD_NUMBER
public static final int PYTHON_FILE_URIS_FIELD_NUMBER
public static final int JAR_FILE_URIS_FIELD_NUMBER
public static final int FILE_URIS_FIELD_NUMBER
public static final int ARCHIVE_URIS_FIELD_NUMBER
public static final int PROPERTIES_FIELD_NUMBER
public static final int LOGGING_CONFIG_FIELD_NUMBER
public final UnknownFieldSet getUnknownFields()
getUnknownFields
in interface MessageOrBuilder
getUnknownFields
in class GeneratedMessageV3
public static final Descriptors.Descriptor getDescriptor()
protected MapField internalGetMapField(int number)
internalGetMapField
in class GeneratedMessageV3
protected GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
internalGetFieldAccessorTable
in class GeneratedMessageV3
public java.lang.String getMainPythonFileUri()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1;
getMainPythonFileUri
in interface PySparkJobOrBuilder
public ByteString getMainPythonFileUriBytes()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1;
getMainPythonFileUriBytes
in interface PySparkJobOrBuilder
public ProtocolStringList getArgsList()
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2;
getArgsList
in interface PySparkJobOrBuilder
public int getArgsCount()
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2;
getArgsCount
in interface PySparkJobOrBuilder
public java.lang.String getArgs(int index)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2;
getArgs
in interface PySparkJobOrBuilder
public ByteString getArgsBytes(int index)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2;
getArgsBytes
in interface PySparkJobOrBuilder
public ProtocolStringList getPythonFileUrisList()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3;
getPythonFileUrisList
in interface PySparkJobOrBuilder
public int getPythonFileUrisCount()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3;
getPythonFileUrisCount
in interface PySparkJobOrBuilder
public java.lang.String getPythonFileUris(int index)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3;
getPythonFileUris
in interface PySparkJobOrBuilder
public ByteString getPythonFileUrisBytes(int index)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3;
getPythonFileUrisBytes
in interface PySparkJobOrBuilder
public ProtocolStringList getJarFileUrisList()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4;
getJarFileUrisList
in interface PySparkJobOrBuilder
public int getJarFileUrisCount()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4;
getJarFileUrisCount
in interface PySparkJobOrBuilder
public java.lang.String getJarFileUris(int index)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4;
getJarFileUris
in interface PySparkJobOrBuilder
public ByteString getJarFileUrisBytes(int index)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4;
getJarFileUrisBytes
in interface PySparkJobOrBuilder
public ProtocolStringList getFileUrisList()
Optional. HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
repeated string file_uris = 5;
getFileUrisList
in interface PySparkJobOrBuilder
public int getFileUrisCount()
Optional. HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
repeated string file_uris = 5;
getFileUrisCount
in interface PySparkJobOrBuilder
public java.lang.String getFileUris(int index)
Optional. HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
repeated string file_uris = 5;
getFileUris
in interface PySparkJobOrBuilder
public ByteString getFileUrisBytes(int index)
Optional. HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
repeated string file_uris = 5;
getFileUrisBytes
in interface PySparkJobOrBuilder
public ProtocolStringList getArchiveUrisList()
Optional. HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6;
getArchiveUrisList
in interface PySparkJobOrBuilder
public int getArchiveUrisCount()
Optional. HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6;
getArchiveUrisCount
in interface PySparkJobOrBuilder
public java.lang.String getArchiveUris(int index)
Optional. HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6;
getArchiveUris
in interface PySparkJobOrBuilder
public ByteString getArchiveUrisBytes(int index)
Optional. HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6;
getArchiveUrisBytes
in interface PySparkJobOrBuilder
public int getPropertiesCount()
PySparkJobOrBuilder
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;
getPropertiesCount
in interface PySparkJobOrBuilder
public boolean containsProperties(java.lang.String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;
containsProperties
in interface PySparkJobOrBuilder
@Deprecated public java.util.Map<java.lang.String,java.lang.String> getProperties()
getPropertiesMap()
instead.getProperties
in interface PySparkJobOrBuilder
public java.util.Map<java.lang.String,java.lang.String> getPropertiesMap()
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;
getPropertiesMap
in interface PySparkJobOrBuilder
public java.lang.String getPropertiesOrDefault(java.lang.String key, java.lang.String defaultValue)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;
getPropertiesOrDefault
in interface PySparkJobOrBuilder
public java.lang.String getPropertiesOrThrow(java.lang.String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;
getPropertiesOrThrow
in interface PySparkJobOrBuilder
public boolean hasLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8;
hasLoggingConfig
in interface PySparkJobOrBuilder
public LoggingConfig getLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8;
getLoggingConfig
in interface PySparkJobOrBuilder
public LoggingConfigOrBuilder getLoggingConfigOrBuilder()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8;
getLoggingConfigOrBuilder
in interface PySparkJobOrBuilder
public final boolean isInitialized()
isInitialized
in interface MessageLiteOrBuilder
isInitialized
in class GeneratedMessageV3
public void writeTo(CodedOutputStream output) throws java.io.IOException
writeTo
in interface MessageLite
writeTo
in class GeneratedMessageV3
java.io.IOException
public int getSerializedSize()
getSerializedSize
in interface MessageLite
getSerializedSize
in class GeneratedMessageV3
public boolean equals(java.lang.Object obj)
equals
in interface Message
equals
in class AbstractMessage
public int hashCode()
hashCode
in interface Message
hashCode
in class AbstractMessage
public static PySparkJob parseFrom(java.nio.ByteBuffer data) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(java.nio.ByteBuffer data, ExtensionRegistryLite extensionRegistry) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(ByteString data) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(ByteString data, ExtensionRegistryLite extensionRegistry) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(byte[] data) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(byte[] data, ExtensionRegistryLite extensionRegistry) throws InvalidProtocolBufferException
InvalidProtocolBufferException
public static PySparkJob parseFrom(java.io.InputStream input) throws java.io.IOException
java.io.IOException
public static PySparkJob parseFrom(java.io.InputStream input, ExtensionRegistryLite extensionRegistry) throws java.io.IOException
java.io.IOException
public static PySparkJob parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException
java.io.IOException
public static PySparkJob parseDelimitedFrom(java.io.InputStream input, ExtensionRegistryLite extensionRegistry) throws java.io.IOException
java.io.IOException
public static PySparkJob parseFrom(CodedInputStream input) throws java.io.IOException
java.io.IOException
public static PySparkJob parseFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry) throws java.io.IOException
java.io.IOException
public PySparkJob.Builder newBuilderForType()
newBuilderForType
in interface Message
newBuilderForType
in interface MessageLite
public static PySparkJob.Builder newBuilder()
public static PySparkJob.Builder newBuilder(PySparkJob prototype)
public PySparkJob.Builder toBuilder()
toBuilder
in interface Message
toBuilder
in interface MessageLite
protected PySparkJob.Builder newBuilderForType(GeneratedMessageV3.BuilderParent parent)
newBuilderForType
in class GeneratedMessageV3
public static PySparkJob getDefaultInstance()
public static Parser<PySparkJob> parser()
public Parser<PySparkJob> getParserForType()
getParserForType
in interface Message
getParserForType
in interface MessageLite
getParserForType
in class GeneratedMessageV3
public PySparkJob getDefaultInstanceForType()
getDefaultInstanceForType
in interface MessageLiteOrBuilder
getDefaultInstanceForType
in interface MessageOrBuilder