public final class PySparkJob extends GeneratedMessageV3 implements PySparkJobOrBuilder
A Cloud Dataproc job for running [Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html) applications on YARN.Protobuf type
google.cloud.dataproc.v1.PySparkJob| Modifier and Type | Class and Description |
|---|---|
static class |
PySparkJob.Builder
A Cloud Dataproc job for running
[Apache PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html)
applications on YARN.
|
GeneratedMessageV3.BuilderParent, GeneratedMessageV3.ExtendableBuilder<MessageType extends GeneratedMessageV3.ExtendableMessage,BuilderType extends GeneratedMessageV3.ExtendableBuilder<MessageType,BuilderType>>, GeneratedMessageV3.ExtendableMessage<MessageType extends GeneratedMessageV3.ExtendableMessage>, GeneratedMessageV3.ExtendableMessageOrBuilder<MessageType extends GeneratedMessageV3.ExtendableMessage>, GeneratedMessageV3.FieldAccessorTable| Modifier and Type | Field and Description |
|---|---|
static int |
ARCHIVE_URIS_FIELD_NUMBER |
static int |
ARGS_FIELD_NUMBER |
static int |
FILE_URIS_FIELD_NUMBER |
static int |
JAR_FILE_URIS_FIELD_NUMBER |
static int |
LOGGING_CONFIG_FIELD_NUMBER |
static int |
MAIN_PYTHON_FILE_URI_FIELD_NUMBER |
static int |
PROPERTIES_FIELD_NUMBER |
static int |
PYTHON_FILE_URIS_FIELD_NUMBER |
alwaysUseFieldBuilders, unknownFieldsmemoizedSizememoizedHashCodecanUseUnsafe, computeStringSize, computeStringSizeNoTag, getAllFields, getDescriptorForType, getField, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, hasField, hasOneof, makeExtensionsImmutable, newBuilderForType, parseDelimitedWithIOException, parseDelimitedWithIOException, parseUnknownField, parseUnknownFieldProto3, parseWithIOException, parseWithIOException, parseWithIOException, parseWithIOException, serializeBooleanMapTo, serializeIntegerMapTo, serializeLongMapTo, serializeStringMapTo, writeReplace, writeString, writeStringNoTagfindInitializationErrors, getInitializationErrorString, hashBoolean, hashEnum, hashEnumList, hashFields, hashLong, toStringaddAll, addAll, checkByteStringIsUtf8, toByteArray, toByteString, writeDelimitedTo, writeToclone, finalize, getClass, notify, notifyAll, wait, wait, waitfindInitializationErrors, getAllFields, getDescriptorForType, getField, getInitializationErrorString, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, hasField, hasOneoftoByteArray, toByteString, writeDelimitedTo, writeTopublic static final int MAIN_PYTHON_FILE_URI_FIELD_NUMBER
public static final int ARGS_FIELD_NUMBER
public static final int PYTHON_FILE_URIS_FIELD_NUMBER
public static final int JAR_FILE_URIS_FIELD_NUMBER
public static final int FILE_URIS_FIELD_NUMBER
public static final int ARCHIVE_URIS_FIELD_NUMBER
public static final int PROPERTIES_FIELD_NUMBER
public static final int LOGGING_CONFIG_FIELD_NUMBER
public final UnknownFieldSet getUnknownFields()
getUnknownFields in interface MessageOrBuildergetUnknownFields in class GeneratedMessageV3public static final Descriptors.Descriptor getDescriptor()
protected MapField internalGetMapField(int number)
internalGetMapField in class GeneratedMessageV3protected GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()
internalGetFieldAccessorTable in class GeneratedMessageV3public java.lang.String getMainPythonFileUri()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1;getMainPythonFileUri in interface PySparkJobOrBuilderpublic ByteString getMainPythonFileUriBytes()
Required. The HCFS URI of the main Python file to use as the driver. Must be a .py file.
string main_python_file_uri = 1;getMainPythonFileUriBytes in interface PySparkJobOrBuilderpublic ProtocolStringList getArgsList()
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2;getArgsList in interface PySparkJobOrBuilderpublic int getArgsCount()
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2;getArgsCount in interface PySparkJobOrBuilderpublic java.lang.String getArgs(int index)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2;getArgs in interface PySparkJobOrBuilderpublic ByteString getArgsBytes(int index)
Optional. The arguments to pass to the driver. Do not include arguments, such as `--conf`, that can be set as job properties, since a collision may occur that causes an incorrect job submission.
repeated string args = 2;getArgsBytes in interface PySparkJobOrBuilderpublic ProtocolStringList getPythonFileUrisList()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3;getPythonFileUrisList in interface PySparkJobOrBuilderpublic int getPythonFileUrisCount()
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3;getPythonFileUrisCount in interface PySparkJobOrBuilderpublic java.lang.String getPythonFileUris(int index)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3;getPythonFileUris in interface PySparkJobOrBuilderpublic ByteString getPythonFileUrisBytes(int index)
Optional. HCFS file URIs of Python files to pass to the PySpark framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3;getPythonFileUrisBytes in interface PySparkJobOrBuilderpublic ProtocolStringList getJarFileUrisList()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4;getJarFileUrisList in interface PySparkJobOrBuilderpublic int getJarFileUrisCount()
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4;getJarFileUrisCount in interface PySparkJobOrBuilderpublic java.lang.String getJarFileUris(int index)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4;getJarFileUris in interface PySparkJobOrBuilderpublic ByteString getJarFileUrisBytes(int index)
Optional. HCFS URIs of jar files to add to the CLASSPATHs of the Python driver and tasks.
repeated string jar_file_uris = 4;getJarFileUrisBytes in interface PySparkJobOrBuilderpublic ProtocolStringList getFileUrisList()
Optional. HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
repeated string file_uris = 5;getFileUrisList in interface PySparkJobOrBuilderpublic int getFileUrisCount()
Optional. HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
repeated string file_uris = 5;getFileUrisCount in interface PySparkJobOrBuilderpublic java.lang.String getFileUris(int index)
Optional. HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
repeated string file_uris = 5;getFileUris in interface PySparkJobOrBuilderpublic ByteString getFileUrisBytes(int index)
Optional. HCFS URIs of files to be copied to the working directory of Python drivers and distributed tasks. Useful for naively parallel tasks.
repeated string file_uris = 5;getFileUrisBytes in interface PySparkJobOrBuilderpublic ProtocolStringList getArchiveUrisList()
Optional. HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6;getArchiveUrisList in interface PySparkJobOrBuilderpublic int getArchiveUrisCount()
Optional. HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6;getArchiveUrisCount in interface PySparkJobOrBuilderpublic java.lang.String getArchiveUris(int index)
Optional. HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6;getArchiveUris in interface PySparkJobOrBuilderpublic ByteString getArchiveUrisBytes(int index)
Optional. HCFS URIs of archives to be extracted in the working directory of .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6;getArchiveUrisBytes in interface PySparkJobOrBuilderpublic int getPropertiesCount()
PySparkJobOrBuilderOptional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;getPropertiesCount in interface PySparkJobOrBuilderpublic boolean containsProperties(java.lang.String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;containsProperties in interface PySparkJobOrBuilder@Deprecated public java.util.Map<java.lang.String,java.lang.String> getProperties()
getPropertiesMap() instead.getProperties in interface PySparkJobOrBuilderpublic java.util.Map<java.lang.String,java.lang.String> getPropertiesMap()
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;getPropertiesMap in interface PySparkJobOrBuilderpublic java.lang.String getPropertiesOrDefault(java.lang.String key,
java.lang.String defaultValue)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;getPropertiesOrDefault in interface PySparkJobOrBuilderpublic java.lang.String getPropertiesOrThrow(java.lang.String key)
Optional. A mapping of property names to values, used to configure PySpark. Properties that conflict with values set by the Cloud Dataproc API may be overwritten. Can include properties set in /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;getPropertiesOrThrow in interface PySparkJobOrBuilderpublic boolean hasLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8;hasLoggingConfig in interface PySparkJobOrBuilderpublic LoggingConfig getLoggingConfig()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8;getLoggingConfig in interface PySparkJobOrBuilderpublic LoggingConfigOrBuilder getLoggingConfigOrBuilder()
Optional. The runtime log config for job execution.
.google.cloud.dataproc.v1.LoggingConfig logging_config = 8;getLoggingConfigOrBuilder in interface PySparkJobOrBuilderpublic final boolean isInitialized()
isInitialized in interface MessageLiteOrBuilderisInitialized in class GeneratedMessageV3public void writeTo(CodedOutputStream output) throws java.io.IOException
writeTo in interface MessageLitewriteTo in class GeneratedMessageV3java.io.IOExceptionpublic int getSerializedSize()
getSerializedSize in interface MessageLitegetSerializedSize in class GeneratedMessageV3public boolean equals(java.lang.Object obj)
equals in interface Messageequals in class AbstractMessagepublic int hashCode()
hashCode in interface MessagehashCode in class AbstractMessagepublic static PySparkJob parseFrom(java.nio.ByteBuffer data) throws InvalidProtocolBufferException
InvalidProtocolBufferExceptionpublic static PySparkJob parseFrom(java.nio.ByteBuffer data, ExtensionRegistryLite extensionRegistry) throws InvalidProtocolBufferException
InvalidProtocolBufferExceptionpublic static PySparkJob parseFrom(ByteString data) throws InvalidProtocolBufferException
InvalidProtocolBufferExceptionpublic static PySparkJob parseFrom(ByteString data, ExtensionRegistryLite extensionRegistry) throws InvalidProtocolBufferException
InvalidProtocolBufferExceptionpublic static PySparkJob parseFrom(byte[] data) throws InvalidProtocolBufferException
InvalidProtocolBufferExceptionpublic static PySparkJob parseFrom(byte[] data, ExtensionRegistryLite extensionRegistry) throws InvalidProtocolBufferException
InvalidProtocolBufferExceptionpublic static PySparkJob parseFrom(java.io.InputStream input) throws java.io.IOException
java.io.IOExceptionpublic static PySparkJob parseFrom(java.io.InputStream input, ExtensionRegistryLite extensionRegistry) throws java.io.IOException
java.io.IOExceptionpublic static PySparkJob parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException
java.io.IOExceptionpublic static PySparkJob parseDelimitedFrom(java.io.InputStream input, ExtensionRegistryLite extensionRegistry) throws java.io.IOException
java.io.IOExceptionpublic static PySparkJob parseFrom(CodedInputStream input) throws java.io.IOException
java.io.IOExceptionpublic static PySparkJob parseFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry) throws java.io.IOException
java.io.IOExceptionpublic PySparkJob.Builder newBuilderForType()
newBuilderForType in interface MessagenewBuilderForType in interface MessageLitepublic static PySparkJob.Builder newBuilder()
public static PySparkJob.Builder newBuilder(PySparkJob prototype)
public PySparkJob.Builder toBuilder()
toBuilder in interface MessagetoBuilder in interface MessageLiteprotected PySparkJob.Builder newBuilderForType(GeneratedMessageV3.BuilderParent parent)
newBuilderForType in class GeneratedMessageV3public static PySparkJob getDefaultInstance()
public static Parser<PySparkJob> parser()
public Parser<PySparkJob> getParserForType()
getParserForType in interface MessagegetParserForType in interface MessageLitegetParserForType in class GeneratedMessageV3public PySparkJob getDefaultInstanceForType()
getDefaultInstanceForType in interface MessageLiteOrBuildergetDefaultInstanceForType in interface MessageOrBuilder