|
@@ -1,1963 +0,0 @@
|
|
|
-/**
|
|
|
- * Licensed to the Apache Software Foundation (ASF) under one
|
|
|
- * or more contributor license agreements. See the NOTICE file
|
|
|
- * distributed with this work for additional information
|
|
|
- * regarding copyright ownership. The ASF licenses this file
|
|
|
- * to you under the Apache License, Version 2.0 (the
|
|
|
- * "License"); you may not use this file except in compliance
|
|
|
- * with the License. You may obtain a copy of the License at
|
|
|
- *
|
|
|
- * http://www.apache.org/licenses/LICENSE-2.0
|
|
|
- *
|
|
|
- * Unless required by applicable law or agreed to in writing, software
|
|
|
- * distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
- * See the License for the specific language governing permissions and
|
|
|
- * limitations under the License.
|
|
|
- */
|
|
|
-// Generated by the protocol buffer compiler. DO NOT EDIT!
|
|
|
-// source: hadoop_rpc.proto
|
|
|
-
|
|
|
-package org.apache.hadoop.ipc.protobuf;
|
|
|
-
|
|
|
-public final class HadoopRpcProtos {
|
|
|
- private HadoopRpcProtos() {}
|
|
|
- public static void registerAllExtensions(
|
|
|
- com.google.protobuf.ExtensionRegistry registry) {
|
|
|
- }
|
|
|
- public interface HadoopRpcRequestProtoOrBuilder
|
|
|
- extends com.google.protobuf.MessageOrBuilder {
|
|
|
-
|
|
|
- // required string methodName = 1;
|
|
|
- boolean hasMethodName();
|
|
|
- String getMethodName();
|
|
|
-
|
|
|
- // optional bytes request = 2;
|
|
|
- boolean hasRequest();
|
|
|
- com.google.protobuf.ByteString getRequest();
|
|
|
-
|
|
|
- // required string declaringClassProtocolName = 3;
|
|
|
- boolean hasDeclaringClassProtocolName();
|
|
|
- String getDeclaringClassProtocolName();
|
|
|
-
|
|
|
- // required uint64 clientProtocolVersion = 4;
|
|
|
- boolean hasClientProtocolVersion();
|
|
|
- long getClientProtocolVersion();
|
|
|
- }
|
|
|
- public static final class HadoopRpcRequestProto extends
|
|
|
- com.google.protobuf.GeneratedMessage
|
|
|
- implements HadoopRpcRequestProtoOrBuilder {
|
|
|
- // Use HadoopRpcRequestProto.newBuilder() to construct.
|
|
|
- private HadoopRpcRequestProto(Builder builder) {
|
|
|
- super(builder);
|
|
|
- }
|
|
|
- private HadoopRpcRequestProto(boolean noInit) {}
|
|
|
-
|
|
|
- private static final HadoopRpcRequestProto defaultInstance;
|
|
|
- public static HadoopRpcRequestProto getDefaultInstance() {
|
|
|
- return defaultInstance;
|
|
|
- }
|
|
|
-
|
|
|
- public HadoopRpcRequestProto getDefaultInstanceForType() {
|
|
|
- return defaultInstance;
|
|
|
- }
|
|
|
-
|
|
|
- public static final com.google.protobuf.Descriptors.Descriptor
|
|
|
- getDescriptor() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcRequestProto_descriptor;
|
|
|
- }
|
|
|
-
|
|
|
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
|
|
- internalGetFieldAccessorTable() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcRequestProto_fieldAccessorTable;
|
|
|
- }
|
|
|
-
|
|
|
- private int bitField0_;
|
|
|
- // required string methodName = 1;
|
|
|
- public static final int METHODNAME_FIELD_NUMBER = 1;
|
|
|
- private java.lang.Object methodName_;
|
|
|
- public boolean hasMethodName() {
|
|
|
- return ((bitField0_ & 0x00000001) == 0x00000001);
|
|
|
- }
|
|
|
- public String getMethodName() {
|
|
|
- java.lang.Object ref = methodName_;
|
|
|
- if (ref instanceof String) {
|
|
|
- return (String) ref;
|
|
|
- } else {
|
|
|
- com.google.protobuf.ByteString bs =
|
|
|
- (com.google.protobuf.ByteString) ref;
|
|
|
- String s = bs.toStringUtf8();
|
|
|
- if (com.google.protobuf.Internal.isValidUtf8(bs)) {
|
|
|
- methodName_ = s;
|
|
|
- }
|
|
|
- return s;
|
|
|
- }
|
|
|
- }
|
|
|
- private com.google.protobuf.ByteString getMethodNameBytes() {
|
|
|
- java.lang.Object ref = methodName_;
|
|
|
- if (ref instanceof String) {
|
|
|
- com.google.protobuf.ByteString b =
|
|
|
- com.google.protobuf.ByteString.copyFromUtf8((String) ref);
|
|
|
- methodName_ = b;
|
|
|
- return b;
|
|
|
- } else {
|
|
|
- return (com.google.protobuf.ByteString) ref;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- // optional bytes request = 2;
|
|
|
- public static final int REQUEST_FIELD_NUMBER = 2;
|
|
|
- private com.google.protobuf.ByteString request_;
|
|
|
- public boolean hasRequest() {
|
|
|
- return ((bitField0_ & 0x00000002) == 0x00000002);
|
|
|
- }
|
|
|
- public com.google.protobuf.ByteString getRequest() {
|
|
|
- return request_;
|
|
|
- }
|
|
|
-
|
|
|
- // required string declaringClassProtocolName = 3;
|
|
|
- public static final int DECLARINGCLASSPROTOCOLNAME_FIELD_NUMBER = 3;
|
|
|
- private java.lang.Object declaringClassProtocolName_;
|
|
|
- public boolean hasDeclaringClassProtocolName() {
|
|
|
- return ((bitField0_ & 0x00000004) == 0x00000004);
|
|
|
- }
|
|
|
- public String getDeclaringClassProtocolName() {
|
|
|
- java.lang.Object ref = declaringClassProtocolName_;
|
|
|
- if (ref instanceof String) {
|
|
|
- return (String) ref;
|
|
|
- } else {
|
|
|
- com.google.protobuf.ByteString bs =
|
|
|
- (com.google.protobuf.ByteString) ref;
|
|
|
- String s = bs.toStringUtf8();
|
|
|
- if (com.google.protobuf.Internal.isValidUtf8(bs)) {
|
|
|
- declaringClassProtocolName_ = s;
|
|
|
- }
|
|
|
- return s;
|
|
|
- }
|
|
|
- }
|
|
|
- private com.google.protobuf.ByteString getDeclaringClassProtocolNameBytes() {
|
|
|
- java.lang.Object ref = declaringClassProtocolName_;
|
|
|
- if (ref instanceof String) {
|
|
|
- com.google.protobuf.ByteString b =
|
|
|
- com.google.protobuf.ByteString.copyFromUtf8((String) ref);
|
|
|
- declaringClassProtocolName_ = b;
|
|
|
- return b;
|
|
|
- } else {
|
|
|
- return (com.google.protobuf.ByteString) ref;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- // required uint64 clientProtocolVersion = 4;
|
|
|
- public static final int CLIENTPROTOCOLVERSION_FIELD_NUMBER = 4;
|
|
|
- private long clientProtocolVersion_;
|
|
|
- public boolean hasClientProtocolVersion() {
|
|
|
- return ((bitField0_ & 0x00000008) == 0x00000008);
|
|
|
- }
|
|
|
- public long getClientProtocolVersion() {
|
|
|
- return clientProtocolVersion_;
|
|
|
- }
|
|
|
-
|
|
|
- private void initFields() {
|
|
|
- methodName_ = "";
|
|
|
- request_ = com.google.protobuf.ByteString.EMPTY;
|
|
|
- declaringClassProtocolName_ = "";
|
|
|
- clientProtocolVersion_ = 0L;
|
|
|
- }
|
|
|
- private byte memoizedIsInitialized = -1;
|
|
|
- public final boolean isInitialized() {
|
|
|
- byte isInitialized = memoizedIsInitialized;
|
|
|
- if (isInitialized != -1) return isInitialized == 1;
|
|
|
-
|
|
|
- if (!hasMethodName()) {
|
|
|
- memoizedIsInitialized = 0;
|
|
|
- return false;
|
|
|
- }
|
|
|
- if (!hasDeclaringClassProtocolName()) {
|
|
|
- memoizedIsInitialized = 0;
|
|
|
- return false;
|
|
|
- }
|
|
|
- if (!hasClientProtocolVersion()) {
|
|
|
- memoizedIsInitialized = 0;
|
|
|
- return false;
|
|
|
- }
|
|
|
- memoizedIsInitialized = 1;
|
|
|
- return true;
|
|
|
- }
|
|
|
-
|
|
|
- public void writeTo(com.google.protobuf.CodedOutputStream output)
|
|
|
- throws java.io.IOException {
|
|
|
- getSerializedSize();
|
|
|
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
|
|
- output.writeBytes(1, getMethodNameBytes());
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
|
|
|
- output.writeBytes(2, request_);
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
|
|
|
- output.writeBytes(3, getDeclaringClassProtocolNameBytes());
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
|
|
|
- output.writeUInt64(4, clientProtocolVersion_);
|
|
|
- }
|
|
|
- getUnknownFields().writeTo(output);
|
|
|
- }
|
|
|
-
|
|
|
- private int memoizedSerializedSize = -1;
|
|
|
- public int getSerializedSize() {
|
|
|
- int size = memoizedSerializedSize;
|
|
|
- if (size != -1) return size;
|
|
|
-
|
|
|
- size = 0;
|
|
|
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
|
|
- size += com.google.protobuf.CodedOutputStream
|
|
|
- .computeBytesSize(1, getMethodNameBytes());
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
|
|
|
- size += com.google.protobuf.CodedOutputStream
|
|
|
- .computeBytesSize(2, request_);
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
|
|
|
- size += com.google.protobuf.CodedOutputStream
|
|
|
- .computeBytesSize(3, getDeclaringClassProtocolNameBytes());
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000008) == 0x00000008)) {
|
|
|
- size += com.google.protobuf.CodedOutputStream
|
|
|
- .computeUInt64Size(4, clientProtocolVersion_);
|
|
|
- }
|
|
|
- size += getUnknownFields().getSerializedSize();
|
|
|
- memoizedSerializedSize = size;
|
|
|
- return size;
|
|
|
- }
|
|
|
-
|
|
|
- private static final long serialVersionUID = 0L;
|
|
|
- @java.lang.Override
|
|
|
- protected java.lang.Object writeReplace()
|
|
|
- throws java.io.ObjectStreamException {
|
|
|
- return super.writeReplace();
|
|
|
- }
|
|
|
-
|
|
|
- @java.lang.Override
|
|
|
- public boolean equals(final java.lang.Object obj) {
|
|
|
- if (obj == this) {
|
|
|
- return true;
|
|
|
- }
|
|
|
- if (!(obj instanceof org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto)) {
|
|
|
- return super.equals(obj);
|
|
|
- }
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto other = (org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto) obj;
|
|
|
-
|
|
|
- boolean result = true;
|
|
|
- result = result && (hasMethodName() == other.hasMethodName());
|
|
|
- if (hasMethodName()) {
|
|
|
- result = result && getMethodName()
|
|
|
- .equals(other.getMethodName());
|
|
|
- }
|
|
|
- result = result && (hasRequest() == other.hasRequest());
|
|
|
- if (hasRequest()) {
|
|
|
- result = result && getRequest()
|
|
|
- .equals(other.getRequest());
|
|
|
- }
|
|
|
- result = result && (hasDeclaringClassProtocolName() == other.hasDeclaringClassProtocolName());
|
|
|
- if (hasDeclaringClassProtocolName()) {
|
|
|
- result = result && getDeclaringClassProtocolName()
|
|
|
- .equals(other.getDeclaringClassProtocolName());
|
|
|
- }
|
|
|
- result = result && (hasClientProtocolVersion() == other.hasClientProtocolVersion());
|
|
|
- if (hasClientProtocolVersion()) {
|
|
|
- result = result && (getClientProtocolVersion()
|
|
|
- == other.getClientProtocolVersion());
|
|
|
- }
|
|
|
- result = result &&
|
|
|
- getUnknownFields().equals(other.getUnknownFields());
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- @java.lang.Override
|
|
|
- public int hashCode() {
|
|
|
- int hash = 41;
|
|
|
- hash = (19 * hash) + getDescriptorForType().hashCode();
|
|
|
- if (hasMethodName()) {
|
|
|
- hash = (37 * hash) + METHODNAME_FIELD_NUMBER;
|
|
|
- hash = (53 * hash) + getMethodName().hashCode();
|
|
|
- }
|
|
|
- if (hasRequest()) {
|
|
|
- hash = (37 * hash) + REQUEST_FIELD_NUMBER;
|
|
|
- hash = (53 * hash) + getRequest().hashCode();
|
|
|
- }
|
|
|
- if (hasDeclaringClassProtocolName()) {
|
|
|
- hash = (37 * hash) + DECLARINGCLASSPROTOCOLNAME_FIELD_NUMBER;
|
|
|
- hash = (53 * hash) + getDeclaringClassProtocolName().hashCode();
|
|
|
- }
|
|
|
- if (hasClientProtocolVersion()) {
|
|
|
- hash = (37 * hash) + CLIENTPROTOCOLVERSION_FIELD_NUMBER;
|
|
|
- hash = (53 * hash) + hashLong(getClientProtocolVersion());
|
|
|
- }
|
|
|
- hash = (29 * hash) + getUnknownFields().hashCode();
|
|
|
- return hash;
|
|
|
- }
|
|
|
-
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto parseFrom(
|
|
|
- com.google.protobuf.ByteString data)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto parseFrom(
|
|
|
- com.google.protobuf.ByteString data,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto parseFrom(byte[] data)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto parseFrom(
|
|
|
- byte[] data,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto parseFrom(java.io.InputStream input)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto parseFrom(
|
|
|
- java.io.InputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto parseDelimitedFrom(java.io.InputStream input)
|
|
|
- throws java.io.IOException {
|
|
|
- Builder builder = newBuilder();
|
|
|
- if (builder.mergeDelimitedFrom(input)) {
|
|
|
- return builder.buildParsed();
|
|
|
- } else {
|
|
|
- return null;
|
|
|
- }
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto parseDelimitedFrom(
|
|
|
- java.io.InputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- Builder builder = newBuilder();
|
|
|
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
|
|
|
- return builder.buildParsed();
|
|
|
- } else {
|
|
|
- return null;
|
|
|
- }
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto parseFrom(
|
|
|
- com.google.protobuf.CodedInputStream input)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto parseFrom(
|
|
|
- com.google.protobuf.CodedInputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
-
|
|
|
- public static Builder newBuilder() { return Builder.create(); }
|
|
|
- public Builder newBuilderForType() { return newBuilder(); }
|
|
|
- public static Builder newBuilder(org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto prototype) {
|
|
|
- return newBuilder().mergeFrom(prototype);
|
|
|
- }
|
|
|
- public Builder toBuilder() { return newBuilder(this); }
|
|
|
-
|
|
|
- @java.lang.Override
|
|
|
- protected Builder newBuilderForType(
|
|
|
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
|
|
|
- Builder builder = new Builder(parent);
|
|
|
- return builder;
|
|
|
- }
|
|
|
- public static final class Builder extends
|
|
|
- com.google.protobuf.GeneratedMessage.Builder<Builder>
|
|
|
- implements org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProtoOrBuilder {
|
|
|
- public static final com.google.protobuf.Descriptors.Descriptor
|
|
|
- getDescriptor() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcRequestProto_descriptor;
|
|
|
- }
|
|
|
-
|
|
|
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
|
|
- internalGetFieldAccessorTable() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcRequestProto_fieldAccessorTable;
|
|
|
- }
|
|
|
-
|
|
|
- // Construct using org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto.newBuilder()
|
|
|
- private Builder() {
|
|
|
- maybeForceBuilderInitialization();
|
|
|
- }
|
|
|
-
|
|
|
- private Builder(BuilderParent parent) {
|
|
|
- super(parent);
|
|
|
- maybeForceBuilderInitialization();
|
|
|
- }
|
|
|
- private void maybeForceBuilderInitialization() {
|
|
|
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
|
|
|
- }
|
|
|
- }
|
|
|
- private static Builder create() {
|
|
|
- return new Builder();
|
|
|
- }
|
|
|
-
|
|
|
- public Builder clear() {
|
|
|
- super.clear();
|
|
|
- methodName_ = "";
|
|
|
- bitField0_ = (bitField0_ & ~0x00000001);
|
|
|
- request_ = com.google.protobuf.ByteString.EMPTY;
|
|
|
- bitField0_ = (bitField0_ & ~0x00000002);
|
|
|
- declaringClassProtocolName_ = "";
|
|
|
- bitField0_ = (bitField0_ & ~0x00000004);
|
|
|
- clientProtocolVersion_ = 0L;
|
|
|
- bitField0_ = (bitField0_ & ~0x00000008);
|
|
|
- return this;
|
|
|
- }
|
|
|
-
|
|
|
- public Builder clone() {
|
|
|
- return create().mergeFrom(buildPartial());
|
|
|
- }
|
|
|
-
|
|
|
- public com.google.protobuf.Descriptors.Descriptor
|
|
|
- getDescriptorForType() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto.getDescriptor();
|
|
|
- }
|
|
|
-
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto getDefaultInstanceForType() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto.getDefaultInstance();
|
|
|
- }
|
|
|
-
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto build() {
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto result = buildPartial();
|
|
|
- if (!result.isInitialized()) {
|
|
|
- throw newUninitializedMessageException(result);
|
|
|
- }
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- private org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto buildParsed()
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto result = buildPartial();
|
|
|
- if (!result.isInitialized()) {
|
|
|
- throw newUninitializedMessageException(
|
|
|
- result).asInvalidProtocolBufferException();
|
|
|
- }
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto buildPartial() {
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto result = new org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto(this);
|
|
|
- int from_bitField0_ = bitField0_;
|
|
|
- int to_bitField0_ = 0;
|
|
|
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
|
|
|
- to_bitField0_ |= 0x00000001;
|
|
|
- }
|
|
|
- result.methodName_ = methodName_;
|
|
|
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
|
|
|
- to_bitField0_ |= 0x00000002;
|
|
|
- }
|
|
|
- result.request_ = request_;
|
|
|
- if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
|
|
|
- to_bitField0_ |= 0x00000004;
|
|
|
- }
|
|
|
- result.declaringClassProtocolName_ = declaringClassProtocolName_;
|
|
|
- if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
|
|
|
- to_bitField0_ |= 0x00000008;
|
|
|
- }
|
|
|
- result.clientProtocolVersion_ = clientProtocolVersion_;
|
|
|
- result.bitField0_ = to_bitField0_;
|
|
|
- onBuilt();
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- public Builder mergeFrom(com.google.protobuf.Message other) {
|
|
|
- if (other instanceof org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto) {
|
|
|
- return mergeFrom((org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto)other);
|
|
|
- } else {
|
|
|
- super.mergeFrom(other);
|
|
|
- return this;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- public Builder mergeFrom(org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto other) {
|
|
|
- if (other == org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto.getDefaultInstance()) return this;
|
|
|
- if (other.hasMethodName()) {
|
|
|
- setMethodName(other.getMethodName());
|
|
|
- }
|
|
|
- if (other.hasRequest()) {
|
|
|
- setRequest(other.getRequest());
|
|
|
- }
|
|
|
- if (other.hasDeclaringClassProtocolName()) {
|
|
|
- setDeclaringClassProtocolName(other.getDeclaringClassProtocolName());
|
|
|
- }
|
|
|
- if (other.hasClientProtocolVersion()) {
|
|
|
- setClientProtocolVersion(other.getClientProtocolVersion());
|
|
|
- }
|
|
|
- this.mergeUnknownFields(other.getUnknownFields());
|
|
|
- return this;
|
|
|
- }
|
|
|
-
|
|
|
- public final boolean isInitialized() {
|
|
|
- if (!hasMethodName()) {
|
|
|
-
|
|
|
- return false;
|
|
|
- }
|
|
|
- if (!hasDeclaringClassProtocolName()) {
|
|
|
-
|
|
|
- return false;
|
|
|
- }
|
|
|
- if (!hasClientProtocolVersion()) {
|
|
|
-
|
|
|
- return false;
|
|
|
- }
|
|
|
- return true;
|
|
|
- }
|
|
|
-
|
|
|
- public Builder mergeFrom(
|
|
|
- com.google.protobuf.CodedInputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
|
|
|
- com.google.protobuf.UnknownFieldSet.newBuilder(
|
|
|
- this.getUnknownFields());
|
|
|
- while (true) {
|
|
|
- int tag = input.readTag();
|
|
|
- switch (tag) {
|
|
|
- case 0:
|
|
|
- this.setUnknownFields(unknownFields.build());
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- default: {
|
|
|
- if (!parseUnknownField(input, unknownFields,
|
|
|
- extensionRegistry, tag)) {
|
|
|
- this.setUnknownFields(unknownFields.build());
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- break;
|
|
|
- }
|
|
|
- case 10: {
|
|
|
- bitField0_ |= 0x00000001;
|
|
|
- methodName_ = input.readBytes();
|
|
|
- break;
|
|
|
- }
|
|
|
- case 18: {
|
|
|
- bitField0_ |= 0x00000002;
|
|
|
- request_ = input.readBytes();
|
|
|
- break;
|
|
|
- }
|
|
|
- case 26: {
|
|
|
- bitField0_ |= 0x00000004;
|
|
|
- declaringClassProtocolName_ = input.readBytes();
|
|
|
- break;
|
|
|
- }
|
|
|
- case 32: {
|
|
|
- bitField0_ |= 0x00000008;
|
|
|
- clientProtocolVersion_ = input.readUInt64();
|
|
|
- break;
|
|
|
- }
|
|
|
- }
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- private int bitField0_;
|
|
|
-
|
|
|
- // required string methodName = 1;
|
|
|
- private java.lang.Object methodName_ = "";
|
|
|
- public boolean hasMethodName() {
|
|
|
- return ((bitField0_ & 0x00000001) == 0x00000001);
|
|
|
- }
|
|
|
- public String getMethodName() {
|
|
|
- java.lang.Object ref = methodName_;
|
|
|
- if (!(ref instanceof String)) {
|
|
|
- String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
|
|
|
- methodName_ = s;
|
|
|
- return s;
|
|
|
- } else {
|
|
|
- return (String) ref;
|
|
|
- }
|
|
|
- }
|
|
|
- public Builder setMethodName(String value) {
|
|
|
- if (value == null) {
|
|
|
- throw new NullPointerException();
|
|
|
- }
|
|
|
- bitField0_ |= 0x00000001;
|
|
|
- methodName_ = value;
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- public Builder clearMethodName() {
|
|
|
- bitField0_ = (bitField0_ & ~0x00000001);
|
|
|
- methodName_ = getDefaultInstance().getMethodName();
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- void setMethodName(com.google.protobuf.ByteString value) {
|
|
|
- bitField0_ |= 0x00000001;
|
|
|
- methodName_ = value;
|
|
|
- onChanged();
|
|
|
- }
|
|
|
-
|
|
|
- // optional bytes request = 2;
|
|
|
- private com.google.protobuf.ByteString request_ = com.google.protobuf.ByteString.EMPTY;
|
|
|
- public boolean hasRequest() {
|
|
|
- return ((bitField0_ & 0x00000002) == 0x00000002);
|
|
|
- }
|
|
|
- public com.google.protobuf.ByteString getRequest() {
|
|
|
- return request_;
|
|
|
- }
|
|
|
- public Builder setRequest(com.google.protobuf.ByteString value) {
|
|
|
- if (value == null) {
|
|
|
- throw new NullPointerException();
|
|
|
- }
|
|
|
- bitField0_ |= 0x00000002;
|
|
|
- request_ = value;
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- public Builder clearRequest() {
|
|
|
- bitField0_ = (bitField0_ & ~0x00000002);
|
|
|
- request_ = getDefaultInstance().getRequest();
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
-
|
|
|
- // required string declaringClassProtocolName = 3;
|
|
|
- private java.lang.Object declaringClassProtocolName_ = "";
|
|
|
- public boolean hasDeclaringClassProtocolName() {
|
|
|
- return ((bitField0_ & 0x00000004) == 0x00000004);
|
|
|
- }
|
|
|
- public String getDeclaringClassProtocolName() {
|
|
|
- java.lang.Object ref = declaringClassProtocolName_;
|
|
|
- if (!(ref instanceof String)) {
|
|
|
- String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
|
|
|
- declaringClassProtocolName_ = s;
|
|
|
- return s;
|
|
|
- } else {
|
|
|
- return (String) ref;
|
|
|
- }
|
|
|
- }
|
|
|
- public Builder setDeclaringClassProtocolName(String value) {
|
|
|
- if (value == null) {
|
|
|
- throw new NullPointerException();
|
|
|
- }
|
|
|
- bitField0_ |= 0x00000004;
|
|
|
- declaringClassProtocolName_ = value;
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- public Builder clearDeclaringClassProtocolName() {
|
|
|
- bitField0_ = (bitField0_ & ~0x00000004);
|
|
|
- declaringClassProtocolName_ = getDefaultInstance().getDeclaringClassProtocolName();
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- void setDeclaringClassProtocolName(com.google.protobuf.ByteString value) {
|
|
|
- bitField0_ |= 0x00000004;
|
|
|
- declaringClassProtocolName_ = value;
|
|
|
- onChanged();
|
|
|
- }
|
|
|
-
|
|
|
- // required uint64 clientProtocolVersion = 4;
|
|
|
- private long clientProtocolVersion_ ;
|
|
|
- public boolean hasClientProtocolVersion() {
|
|
|
- return ((bitField0_ & 0x00000008) == 0x00000008);
|
|
|
- }
|
|
|
- public long getClientProtocolVersion() {
|
|
|
- return clientProtocolVersion_;
|
|
|
- }
|
|
|
- public Builder setClientProtocolVersion(long value) {
|
|
|
- bitField0_ |= 0x00000008;
|
|
|
- clientProtocolVersion_ = value;
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- public Builder clearClientProtocolVersion() {
|
|
|
- bitField0_ = (bitField0_ & ~0x00000008);
|
|
|
- clientProtocolVersion_ = 0L;
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
-
|
|
|
- // @@protoc_insertion_point(builder_scope:HadoopRpcRequestProto)
|
|
|
- }
|
|
|
-
|
|
|
- static {
|
|
|
- defaultInstance = new HadoopRpcRequestProto(true);
|
|
|
- defaultInstance.initFields();
|
|
|
- }
|
|
|
-
|
|
|
- // @@protoc_insertion_point(class_scope:HadoopRpcRequestProto)
|
|
|
- }
|
|
|
-
|
|
|
- public interface HadoopRpcExceptionProtoOrBuilder
|
|
|
- extends com.google.protobuf.MessageOrBuilder {
|
|
|
-
|
|
|
- // optional string exceptionName = 1;
|
|
|
- boolean hasExceptionName();
|
|
|
- String getExceptionName();
|
|
|
-
|
|
|
- // optional string stackTrace = 2;
|
|
|
- boolean hasStackTrace();
|
|
|
- String getStackTrace();
|
|
|
- }
|
|
|
- public static final class HadoopRpcExceptionProto extends
|
|
|
- com.google.protobuf.GeneratedMessage
|
|
|
- implements HadoopRpcExceptionProtoOrBuilder {
|
|
|
- // Use HadoopRpcExceptionProto.newBuilder() to construct.
|
|
|
- private HadoopRpcExceptionProto(Builder builder) {
|
|
|
- super(builder);
|
|
|
- }
|
|
|
- private HadoopRpcExceptionProto(boolean noInit) {}
|
|
|
-
|
|
|
- private static final HadoopRpcExceptionProto defaultInstance;
|
|
|
- public static HadoopRpcExceptionProto getDefaultInstance() {
|
|
|
- return defaultInstance;
|
|
|
- }
|
|
|
-
|
|
|
- public HadoopRpcExceptionProto getDefaultInstanceForType() {
|
|
|
- return defaultInstance;
|
|
|
- }
|
|
|
-
|
|
|
- public static final com.google.protobuf.Descriptors.Descriptor
|
|
|
- getDescriptor() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcExceptionProto_descriptor;
|
|
|
- }
|
|
|
-
|
|
|
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
|
|
- internalGetFieldAccessorTable() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcExceptionProto_fieldAccessorTable;
|
|
|
- }
|
|
|
-
|
|
|
- private int bitField0_;
|
|
|
- // optional string exceptionName = 1;
|
|
|
- public static final int EXCEPTIONNAME_FIELD_NUMBER = 1;
|
|
|
- private java.lang.Object exceptionName_;
|
|
|
- public boolean hasExceptionName() {
|
|
|
- return ((bitField0_ & 0x00000001) == 0x00000001);
|
|
|
- }
|
|
|
- public String getExceptionName() {
|
|
|
- java.lang.Object ref = exceptionName_;
|
|
|
- if (ref instanceof String) {
|
|
|
- return (String) ref;
|
|
|
- } else {
|
|
|
- com.google.protobuf.ByteString bs =
|
|
|
- (com.google.protobuf.ByteString) ref;
|
|
|
- String s = bs.toStringUtf8();
|
|
|
- if (com.google.protobuf.Internal.isValidUtf8(bs)) {
|
|
|
- exceptionName_ = s;
|
|
|
- }
|
|
|
- return s;
|
|
|
- }
|
|
|
- }
|
|
|
- private com.google.protobuf.ByteString getExceptionNameBytes() {
|
|
|
- java.lang.Object ref = exceptionName_;
|
|
|
- if (ref instanceof String) {
|
|
|
- com.google.protobuf.ByteString b =
|
|
|
- com.google.protobuf.ByteString.copyFromUtf8((String) ref);
|
|
|
- exceptionName_ = b;
|
|
|
- return b;
|
|
|
- } else {
|
|
|
- return (com.google.protobuf.ByteString) ref;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- // optional string stackTrace = 2;
|
|
|
- public static final int STACKTRACE_FIELD_NUMBER = 2;
|
|
|
- private java.lang.Object stackTrace_;
|
|
|
- public boolean hasStackTrace() {
|
|
|
- return ((bitField0_ & 0x00000002) == 0x00000002);
|
|
|
- }
|
|
|
- public String getStackTrace() {
|
|
|
- java.lang.Object ref = stackTrace_;
|
|
|
- if (ref instanceof String) {
|
|
|
- return (String) ref;
|
|
|
- } else {
|
|
|
- com.google.protobuf.ByteString bs =
|
|
|
- (com.google.protobuf.ByteString) ref;
|
|
|
- String s = bs.toStringUtf8();
|
|
|
- if (com.google.protobuf.Internal.isValidUtf8(bs)) {
|
|
|
- stackTrace_ = s;
|
|
|
- }
|
|
|
- return s;
|
|
|
- }
|
|
|
- }
|
|
|
- private com.google.protobuf.ByteString getStackTraceBytes() {
|
|
|
- java.lang.Object ref = stackTrace_;
|
|
|
- if (ref instanceof String) {
|
|
|
- com.google.protobuf.ByteString b =
|
|
|
- com.google.protobuf.ByteString.copyFromUtf8((String) ref);
|
|
|
- stackTrace_ = b;
|
|
|
- return b;
|
|
|
- } else {
|
|
|
- return (com.google.protobuf.ByteString) ref;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- private void initFields() {
|
|
|
- exceptionName_ = "";
|
|
|
- stackTrace_ = "";
|
|
|
- }
|
|
|
- private byte memoizedIsInitialized = -1;
|
|
|
- public final boolean isInitialized() {
|
|
|
- byte isInitialized = memoizedIsInitialized;
|
|
|
- if (isInitialized != -1) return isInitialized == 1;
|
|
|
-
|
|
|
- memoizedIsInitialized = 1;
|
|
|
- return true;
|
|
|
- }
|
|
|
-
|
|
|
- public void writeTo(com.google.protobuf.CodedOutputStream output)
|
|
|
- throws java.io.IOException {
|
|
|
- getSerializedSize();
|
|
|
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
|
|
- output.writeBytes(1, getExceptionNameBytes());
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
|
|
|
- output.writeBytes(2, getStackTraceBytes());
|
|
|
- }
|
|
|
- getUnknownFields().writeTo(output);
|
|
|
- }
|
|
|
-
|
|
|
- private int memoizedSerializedSize = -1;
|
|
|
- public int getSerializedSize() {
|
|
|
- int size = memoizedSerializedSize;
|
|
|
- if (size != -1) return size;
|
|
|
-
|
|
|
- size = 0;
|
|
|
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
|
|
- size += com.google.protobuf.CodedOutputStream
|
|
|
- .computeBytesSize(1, getExceptionNameBytes());
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
|
|
|
- size += com.google.protobuf.CodedOutputStream
|
|
|
- .computeBytesSize(2, getStackTraceBytes());
|
|
|
- }
|
|
|
- size += getUnknownFields().getSerializedSize();
|
|
|
- memoizedSerializedSize = size;
|
|
|
- return size;
|
|
|
- }
|
|
|
-
|
|
|
- private static final long serialVersionUID = 0L;
|
|
|
- @java.lang.Override
|
|
|
- protected java.lang.Object writeReplace()
|
|
|
- throws java.io.ObjectStreamException {
|
|
|
- return super.writeReplace();
|
|
|
- }
|
|
|
-
|
|
|
- @java.lang.Override
|
|
|
- public boolean equals(final java.lang.Object obj) {
|
|
|
- if (obj == this) {
|
|
|
- return true;
|
|
|
- }
|
|
|
- if (!(obj instanceof org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto)) {
|
|
|
- return super.equals(obj);
|
|
|
- }
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto other = (org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto) obj;
|
|
|
-
|
|
|
- boolean result = true;
|
|
|
- result = result && (hasExceptionName() == other.hasExceptionName());
|
|
|
- if (hasExceptionName()) {
|
|
|
- result = result && getExceptionName()
|
|
|
- .equals(other.getExceptionName());
|
|
|
- }
|
|
|
- result = result && (hasStackTrace() == other.hasStackTrace());
|
|
|
- if (hasStackTrace()) {
|
|
|
- result = result && getStackTrace()
|
|
|
- .equals(other.getStackTrace());
|
|
|
- }
|
|
|
- result = result &&
|
|
|
- getUnknownFields().equals(other.getUnknownFields());
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- @java.lang.Override
|
|
|
- public int hashCode() {
|
|
|
- int hash = 41;
|
|
|
- hash = (19 * hash) + getDescriptorForType().hashCode();
|
|
|
- if (hasExceptionName()) {
|
|
|
- hash = (37 * hash) + EXCEPTIONNAME_FIELD_NUMBER;
|
|
|
- hash = (53 * hash) + getExceptionName().hashCode();
|
|
|
- }
|
|
|
- if (hasStackTrace()) {
|
|
|
- hash = (37 * hash) + STACKTRACE_FIELD_NUMBER;
|
|
|
- hash = (53 * hash) + getStackTrace().hashCode();
|
|
|
- }
|
|
|
- hash = (29 * hash) + getUnknownFields().hashCode();
|
|
|
- return hash;
|
|
|
- }
|
|
|
-
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto parseFrom(
|
|
|
- com.google.protobuf.ByteString data)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto parseFrom(
|
|
|
- com.google.protobuf.ByteString data,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto parseFrom(byte[] data)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto parseFrom(
|
|
|
- byte[] data,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto parseFrom(java.io.InputStream input)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto parseFrom(
|
|
|
- java.io.InputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto parseDelimitedFrom(java.io.InputStream input)
|
|
|
- throws java.io.IOException {
|
|
|
- Builder builder = newBuilder();
|
|
|
- if (builder.mergeDelimitedFrom(input)) {
|
|
|
- return builder.buildParsed();
|
|
|
- } else {
|
|
|
- return null;
|
|
|
- }
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto parseDelimitedFrom(
|
|
|
- java.io.InputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- Builder builder = newBuilder();
|
|
|
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
|
|
|
- return builder.buildParsed();
|
|
|
- } else {
|
|
|
- return null;
|
|
|
- }
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto parseFrom(
|
|
|
- com.google.protobuf.CodedInputStream input)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto parseFrom(
|
|
|
- com.google.protobuf.CodedInputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
-
|
|
|
- public static Builder newBuilder() { return Builder.create(); }
|
|
|
- public Builder newBuilderForType() { return newBuilder(); }
|
|
|
- public static Builder newBuilder(org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto prototype) {
|
|
|
- return newBuilder().mergeFrom(prototype);
|
|
|
- }
|
|
|
- public Builder toBuilder() { return newBuilder(this); }
|
|
|
-
|
|
|
- @java.lang.Override
|
|
|
- protected Builder newBuilderForType(
|
|
|
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
|
|
|
- Builder builder = new Builder(parent);
|
|
|
- return builder;
|
|
|
- }
|
|
|
- public static final class Builder extends
|
|
|
- com.google.protobuf.GeneratedMessage.Builder<Builder>
|
|
|
- implements org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProtoOrBuilder {
|
|
|
- public static final com.google.protobuf.Descriptors.Descriptor
|
|
|
- getDescriptor() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcExceptionProto_descriptor;
|
|
|
- }
|
|
|
-
|
|
|
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
|
|
- internalGetFieldAccessorTable() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcExceptionProto_fieldAccessorTable;
|
|
|
- }
|
|
|
-
|
|
|
- // Construct using org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.newBuilder()
|
|
|
- private Builder() {
|
|
|
- maybeForceBuilderInitialization();
|
|
|
- }
|
|
|
-
|
|
|
- private Builder(BuilderParent parent) {
|
|
|
- super(parent);
|
|
|
- maybeForceBuilderInitialization();
|
|
|
- }
|
|
|
- private void maybeForceBuilderInitialization() {
|
|
|
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
|
|
|
- }
|
|
|
- }
|
|
|
- private static Builder create() {
|
|
|
- return new Builder();
|
|
|
- }
|
|
|
-
|
|
|
- public Builder clear() {
|
|
|
- super.clear();
|
|
|
- exceptionName_ = "";
|
|
|
- bitField0_ = (bitField0_ & ~0x00000001);
|
|
|
- stackTrace_ = "";
|
|
|
- bitField0_ = (bitField0_ & ~0x00000002);
|
|
|
- return this;
|
|
|
- }
|
|
|
-
|
|
|
- public Builder clone() {
|
|
|
- return create().mergeFrom(buildPartial());
|
|
|
- }
|
|
|
-
|
|
|
- public com.google.protobuf.Descriptors.Descriptor
|
|
|
- getDescriptorForType() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.getDescriptor();
|
|
|
- }
|
|
|
-
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto getDefaultInstanceForType() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.getDefaultInstance();
|
|
|
- }
|
|
|
-
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto build() {
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto result = buildPartial();
|
|
|
- if (!result.isInitialized()) {
|
|
|
- throw newUninitializedMessageException(result);
|
|
|
- }
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- private org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto buildParsed()
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto result = buildPartial();
|
|
|
- if (!result.isInitialized()) {
|
|
|
- throw newUninitializedMessageException(
|
|
|
- result).asInvalidProtocolBufferException();
|
|
|
- }
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto buildPartial() {
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto result = new org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto(this);
|
|
|
- int from_bitField0_ = bitField0_;
|
|
|
- int to_bitField0_ = 0;
|
|
|
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
|
|
|
- to_bitField0_ |= 0x00000001;
|
|
|
- }
|
|
|
- result.exceptionName_ = exceptionName_;
|
|
|
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
|
|
|
- to_bitField0_ |= 0x00000002;
|
|
|
- }
|
|
|
- result.stackTrace_ = stackTrace_;
|
|
|
- result.bitField0_ = to_bitField0_;
|
|
|
- onBuilt();
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- public Builder mergeFrom(com.google.protobuf.Message other) {
|
|
|
- if (other instanceof org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto) {
|
|
|
- return mergeFrom((org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto)other);
|
|
|
- } else {
|
|
|
- super.mergeFrom(other);
|
|
|
- return this;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- public Builder mergeFrom(org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto other) {
|
|
|
- if (other == org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.getDefaultInstance()) return this;
|
|
|
- if (other.hasExceptionName()) {
|
|
|
- setExceptionName(other.getExceptionName());
|
|
|
- }
|
|
|
- if (other.hasStackTrace()) {
|
|
|
- setStackTrace(other.getStackTrace());
|
|
|
- }
|
|
|
- this.mergeUnknownFields(other.getUnknownFields());
|
|
|
- return this;
|
|
|
- }
|
|
|
-
|
|
|
- public final boolean isInitialized() {
|
|
|
- return true;
|
|
|
- }
|
|
|
-
|
|
|
- public Builder mergeFrom(
|
|
|
- com.google.protobuf.CodedInputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
|
|
|
- com.google.protobuf.UnknownFieldSet.newBuilder(
|
|
|
- this.getUnknownFields());
|
|
|
- while (true) {
|
|
|
- int tag = input.readTag();
|
|
|
- switch (tag) {
|
|
|
- case 0:
|
|
|
- this.setUnknownFields(unknownFields.build());
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- default: {
|
|
|
- if (!parseUnknownField(input, unknownFields,
|
|
|
- extensionRegistry, tag)) {
|
|
|
- this.setUnknownFields(unknownFields.build());
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- break;
|
|
|
- }
|
|
|
- case 10: {
|
|
|
- bitField0_ |= 0x00000001;
|
|
|
- exceptionName_ = input.readBytes();
|
|
|
- break;
|
|
|
- }
|
|
|
- case 18: {
|
|
|
- bitField0_ |= 0x00000002;
|
|
|
- stackTrace_ = input.readBytes();
|
|
|
- break;
|
|
|
- }
|
|
|
- }
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- private int bitField0_;
|
|
|
-
|
|
|
- // optional string exceptionName = 1;
|
|
|
- private java.lang.Object exceptionName_ = "";
|
|
|
- public boolean hasExceptionName() {
|
|
|
- return ((bitField0_ & 0x00000001) == 0x00000001);
|
|
|
- }
|
|
|
- public String getExceptionName() {
|
|
|
- java.lang.Object ref = exceptionName_;
|
|
|
- if (!(ref instanceof String)) {
|
|
|
- String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
|
|
|
- exceptionName_ = s;
|
|
|
- return s;
|
|
|
- } else {
|
|
|
- return (String) ref;
|
|
|
- }
|
|
|
- }
|
|
|
- public Builder setExceptionName(String value) {
|
|
|
- if (value == null) {
|
|
|
- throw new NullPointerException();
|
|
|
- }
|
|
|
- bitField0_ |= 0x00000001;
|
|
|
- exceptionName_ = value;
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- public Builder clearExceptionName() {
|
|
|
- bitField0_ = (bitField0_ & ~0x00000001);
|
|
|
- exceptionName_ = getDefaultInstance().getExceptionName();
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- void setExceptionName(com.google.protobuf.ByteString value) {
|
|
|
- bitField0_ |= 0x00000001;
|
|
|
- exceptionName_ = value;
|
|
|
- onChanged();
|
|
|
- }
|
|
|
-
|
|
|
- // optional string stackTrace = 2;
|
|
|
- private java.lang.Object stackTrace_ = "";
|
|
|
- public boolean hasStackTrace() {
|
|
|
- return ((bitField0_ & 0x00000002) == 0x00000002);
|
|
|
- }
|
|
|
- public String getStackTrace() {
|
|
|
- java.lang.Object ref = stackTrace_;
|
|
|
- if (!(ref instanceof String)) {
|
|
|
- String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
|
|
|
- stackTrace_ = s;
|
|
|
- return s;
|
|
|
- } else {
|
|
|
- return (String) ref;
|
|
|
- }
|
|
|
- }
|
|
|
- public Builder setStackTrace(String value) {
|
|
|
- if (value == null) {
|
|
|
- throw new NullPointerException();
|
|
|
- }
|
|
|
- bitField0_ |= 0x00000002;
|
|
|
- stackTrace_ = value;
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- public Builder clearStackTrace() {
|
|
|
- bitField0_ = (bitField0_ & ~0x00000002);
|
|
|
- stackTrace_ = getDefaultInstance().getStackTrace();
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- void setStackTrace(com.google.protobuf.ByteString value) {
|
|
|
- bitField0_ |= 0x00000002;
|
|
|
- stackTrace_ = value;
|
|
|
- onChanged();
|
|
|
- }
|
|
|
-
|
|
|
- // @@protoc_insertion_point(builder_scope:HadoopRpcExceptionProto)
|
|
|
- }
|
|
|
-
|
|
|
- static {
|
|
|
- defaultInstance = new HadoopRpcExceptionProto(true);
|
|
|
- defaultInstance.initFields();
|
|
|
- }
|
|
|
-
|
|
|
- // @@protoc_insertion_point(class_scope:HadoopRpcExceptionProto)
|
|
|
- }
|
|
|
-
|
|
|
- public interface HadoopRpcResponseProtoOrBuilder
|
|
|
- extends com.google.protobuf.MessageOrBuilder {
|
|
|
-
|
|
|
- // required .HadoopRpcResponseProto.ResponseStatus status = 1;
|
|
|
- boolean hasStatus();
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus getStatus();
|
|
|
-
|
|
|
- // optional bytes response = 2;
|
|
|
- boolean hasResponse();
|
|
|
- com.google.protobuf.ByteString getResponse();
|
|
|
-
|
|
|
- // optional .HadoopRpcExceptionProto exception = 3;
|
|
|
- boolean hasException();
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto getException();
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProtoOrBuilder getExceptionOrBuilder();
|
|
|
- }
|
|
|
- public static final class HadoopRpcResponseProto extends
|
|
|
- com.google.protobuf.GeneratedMessage
|
|
|
- implements HadoopRpcResponseProtoOrBuilder {
|
|
|
- // Use HadoopRpcResponseProto.newBuilder() to construct.
|
|
|
- private HadoopRpcResponseProto(Builder builder) {
|
|
|
- super(builder);
|
|
|
- }
|
|
|
- private HadoopRpcResponseProto(boolean noInit) {}
|
|
|
-
|
|
|
- private static final HadoopRpcResponseProto defaultInstance;
|
|
|
- public static HadoopRpcResponseProto getDefaultInstance() {
|
|
|
- return defaultInstance;
|
|
|
- }
|
|
|
-
|
|
|
- public HadoopRpcResponseProto getDefaultInstanceForType() {
|
|
|
- return defaultInstance;
|
|
|
- }
|
|
|
-
|
|
|
- public static final com.google.protobuf.Descriptors.Descriptor
|
|
|
- getDescriptor() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcResponseProto_descriptor;
|
|
|
- }
|
|
|
-
|
|
|
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
|
|
- internalGetFieldAccessorTable() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcResponseProto_fieldAccessorTable;
|
|
|
- }
|
|
|
-
|
|
|
- public enum ResponseStatus
|
|
|
- implements com.google.protobuf.ProtocolMessageEnum {
|
|
|
- SUCCESS(0, 1),
|
|
|
- ERRROR(1, 2),
|
|
|
- ;
|
|
|
-
|
|
|
- public static final int SUCCESS_VALUE = 1;
|
|
|
- public static final int ERRROR_VALUE = 2;
|
|
|
-
|
|
|
-
|
|
|
- public final int getNumber() { return value; }
|
|
|
-
|
|
|
- public static ResponseStatus valueOf(int value) {
|
|
|
- switch (value) {
|
|
|
- case 1: return SUCCESS;
|
|
|
- case 2: return ERRROR;
|
|
|
- default: return null;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- public static com.google.protobuf.Internal.EnumLiteMap<ResponseStatus>
|
|
|
- internalGetValueMap() {
|
|
|
- return internalValueMap;
|
|
|
- }
|
|
|
- private static com.google.protobuf.Internal.EnumLiteMap<ResponseStatus>
|
|
|
- internalValueMap =
|
|
|
- new com.google.protobuf.Internal.EnumLiteMap<ResponseStatus>() {
|
|
|
- public ResponseStatus findValueByNumber(int number) {
|
|
|
- return ResponseStatus.valueOf(number);
|
|
|
- }
|
|
|
- };
|
|
|
-
|
|
|
- public final com.google.protobuf.Descriptors.EnumValueDescriptor
|
|
|
- getValueDescriptor() {
|
|
|
- return getDescriptor().getValues().get(index);
|
|
|
- }
|
|
|
- public final com.google.protobuf.Descriptors.EnumDescriptor
|
|
|
- getDescriptorForType() {
|
|
|
- return getDescriptor();
|
|
|
- }
|
|
|
- public static final com.google.protobuf.Descriptors.EnumDescriptor
|
|
|
- getDescriptor() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.getDescriptor().getEnumTypes().get(0);
|
|
|
- }
|
|
|
-
|
|
|
- private static final ResponseStatus[] VALUES = {
|
|
|
- SUCCESS, ERRROR,
|
|
|
- };
|
|
|
-
|
|
|
- public static ResponseStatus valueOf(
|
|
|
- com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
|
|
|
- if (desc.getType() != getDescriptor()) {
|
|
|
- throw new java.lang.IllegalArgumentException(
|
|
|
- "EnumValueDescriptor is not for this type.");
|
|
|
- }
|
|
|
- return VALUES[desc.getIndex()];
|
|
|
- }
|
|
|
-
|
|
|
- private final int index;
|
|
|
- private final int value;
|
|
|
-
|
|
|
- private ResponseStatus(int index, int value) {
|
|
|
- this.index = index;
|
|
|
- this.value = value;
|
|
|
- }
|
|
|
-
|
|
|
- // @@protoc_insertion_point(enum_scope:HadoopRpcResponseProto.ResponseStatus)
|
|
|
- }
|
|
|
-
|
|
|
- private int bitField0_;
|
|
|
- // required .HadoopRpcResponseProto.ResponseStatus status = 1;
|
|
|
- public static final int STATUS_FIELD_NUMBER = 1;
|
|
|
- private org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus status_;
|
|
|
- public boolean hasStatus() {
|
|
|
- return ((bitField0_ & 0x00000001) == 0x00000001);
|
|
|
- }
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus getStatus() {
|
|
|
- return status_;
|
|
|
- }
|
|
|
-
|
|
|
- // optional bytes response = 2;
|
|
|
- public static final int RESPONSE_FIELD_NUMBER = 2;
|
|
|
- private com.google.protobuf.ByteString response_;
|
|
|
- public boolean hasResponse() {
|
|
|
- return ((bitField0_ & 0x00000002) == 0x00000002);
|
|
|
- }
|
|
|
- public com.google.protobuf.ByteString getResponse() {
|
|
|
- return response_;
|
|
|
- }
|
|
|
-
|
|
|
- // optional .HadoopRpcExceptionProto exception = 3;
|
|
|
- public static final int EXCEPTION_FIELD_NUMBER = 3;
|
|
|
- private org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto exception_;
|
|
|
- public boolean hasException() {
|
|
|
- return ((bitField0_ & 0x00000004) == 0x00000004);
|
|
|
- }
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto getException() {
|
|
|
- return exception_;
|
|
|
- }
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProtoOrBuilder getExceptionOrBuilder() {
|
|
|
- return exception_;
|
|
|
- }
|
|
|
-
|
|
|
- private void initFields() {
|
|
|
- status_ = org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus.SUCCESS;
|
|
|
- response_ = com.google.protobuf.ByteString.EMPTY;
|
|
|
- exception_ = org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.getDefaultInstance();
|
|
|
- }
|
|
|
- private byte memoizedIsInitialized = -1;
|
|
|
- public final boolean isInitialized() {
|
|
|
- byte isInitialized = memoizedIsInitialized;
|
|
|
- if (isInitialized != -1) return isInitialized == 1;
|
|
|
-
|
|
|
- if (!hasStatus()) {
|
|
|
- memoizedIsInitialized = 0;
|
|
|
- return false;
|
|
|
- }
|
|
|
- memoizedIsInitialized = 1;
|
|
|
- return true;
|
|
|
- }
|
|
|
-
|
|
|
- public void writeTo(com.google.protobuf.CodedOutputStream output)
|
|
|
- throws java.io.IOException {
|
|
|
- getSerializedSize();
|
|
|
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
|
|
- output.writeEnum(1, status_.getNumber());
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
|
|
|
- output.writeBytes(2, response_);
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
|
|
|
- output.writeMessage(3, exception_);
|
|
|
- }
|
|
|
- getUnknownFields().writeTo(output);
|
|
|
- }
|
|
|
-
|
|
|
- private int memoizedSerializedSize = -1;
|
|
|
- public int getSerializedSize() {
|
|
|
- int size = memoizedSerializedSize;
|
|
|
- if (size != -1) return size;
|
|
|
-
|
|
|
- size = 0;
|
|
|
- if (((bitField0_ & 0x00000001) == 0x00000001)) {
|
|
|
- size += com.google.protobuf.CodedOutputStream
|
|
|
- .computeEnumSize(1, status_.getNumber());
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000002) == 0x00000002)) {
|
|
|
- size += com.google.protobuf.CodedOutputStream
|
|
|
- .computeBytesSize(2, response_);
|
|
|
- }
|
|
|
- if (((bitField0_ & 0x00000004) == 0x00000004)) {
|
|
|
- size += com.google.protobuf.CodedOutputStream
|
|
|
- .computeMessageSize(3, exception_);
|
|
|
- }
|
|
|
- size += getUnknownFields().getSerializedSize();
|
|
|
- memoizedSerializedSize = size;
|
|
|
- return size;
|
|
|
- }
|
|
|
-
|
|
|
- private static final long serialVersionUID = 0L;
|
|
|
- @java.lang.Override
|
|
|
- protected java.lang.Object writeReplace()
|
|
|
- throws java.io.ObjectStreamException {
|
|
|
- return super.writeReplace();
|
|
|
- }
|
|
|
-
|
|
|
- @java.lang.Override
|
|
|
- public boolean equals(final java.lang.Object obj) {
|
|
|
- if (obj == this) {
|
|
|
- return true;
|
|
|
- }
|
|
|
- if (!(obj instanceof org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto)) {
|
|
|
- return super.equals(obj);
|
|
|
- }
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto other = (org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto) obj;
|
|
|
-
|
|
|
- boolean result = true;
|
|
|
- result = result && (hasStatus() == other.hasStatus());
|
|
|
- if (hasStatus()) {
|
|
|
- result = result &&
|
|
|
- (getStatus() == other.getStatus());
|
|
|
- }
|
|
|
- result = result && (hasResponse() == other.hasResponse());
|
|
|
- if (hasResponse()) {
|
|
|
- result = result && getResponse()
|
|
|
- .equals(other.getResponse());
|
|
|
- }
|
|
|
- result = result && (hasException() == other.hasException());
|
|
|
- if (hasException()) {
|
|
|
- result = result && getException()
|
|
|
- .equals(other.getException());
|
|
|
- }
|
|
|
- result = result &&
|
|
|
- getUnknownFields().equals(other.getUnknownFields());
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- @java.lang.Override
|
|
|
- public int hashCode() {
|
|
|
- int hash = 41;
|
|
|
- hash = (19 * hash) + getDescriptorForType().hashCode();
|
|
|
- if (hasStatus()) {
|
|
|
- hash = (37 * hash) + STATUS_FIELD_NUMBER;
|
|
|
- hash = (53 * hash) + hashEnum(getStatus());
|
|
|
- }
|
|
|
- if (hasResponse()) {
|
|
|
- hash = (37 * hash) + RESPONSE_FIELD_NUMBER;
|
|
|
- hash = (53 * hash) + getResponse().hashCode();
|
|
|
- }
|
|
|
- if (hasException()) {
|
|
|
- hash = (37 * hash) + EXCEPTION_FIELD_NUMBER;
|
|
|
- hash = (53 * hash) + getException().hashCode();
|
|
|
- }
|
|
|
- hash = (29 * hash) + getUnknownFields().hashCode();
|
|
|
- return hash;
|
|
|
- }
|
|
|
-
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto parseFrom(
|
|
|
- com.google.protobuf.ByteString data)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto parseFrom(
|
|
|
- com.google.protobuf.ByteString data,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto parseFrom(byte[] data)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto parseFrom(
|
|
|
- byte[] data,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- return newBuilder().mergeFrom(data, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto parseFrom(java.io.InputStream input)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto parseFrom(
|
|
|
- java.io.InputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto parseDelimitedFrom(java.io.InputStream input)
|
|
|
- throws java.io.IOException {
|
|
|
- Builder builder = newBuilder();
|
|
|
- if (builder.mergeDelimitedFrom(input)) {
|
|
|
- return builder.buildParsed();
|
|
|
- } else {
|
|
|
- return null;
|
|
|
- }
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto parseDelimitedFrom(
|
|
|
- java.io.InputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- Builder builder = newBuilder();
|
|
|
- if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
|
|
|
- return builder.buildParsed();
|
|
|
- } else {
|
|
|
- return null;
|
|
|
- }
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto parseFrom(
|
|
|
- com.google.protobuf.CodedInputStream input)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input).buildParsed();
|
|
|
- }
|
|
|
- public static org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto parseFrom(
|
|
|
- com.google.protobuf.CodedInputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- return newBuilder().mergeFrom(input, extensionRegistry)
|
|
|
- .buildParsed();
|
|
|
- }
|
|
|
-
|
|
|
- public static Builder newBuilder() { return Builder.create(); }
|
|
|
- public Builder newBuilderForType() { return newBuilder(); }
|
|
|
- public static Builder newBuilder(org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto prototype) {
|
|
|
- return newBuilder().mergeFrom(prototype);
|
|
|
- }
|
|
|
- public Builder toBuilder() { return newBuilder(this); }
|
|
|
-
|
|
|
- @java.lang.Override
|
|
|
- protected Builder newBuilderForType(
|
|
|
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
|
|
|
- Builder builder = new Builder(parent);
|
|
|
- return builder;
|
|
|
- }
|
|
|
- public static final class Builder extends
|
|
|
- com.google.protobuf.GeneratedMessage.Builder<Builder>
|
|
|
- implements org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProtoOrBuilder {
|
|
|
- public static final com.google.protobuf.Descriptors.Descriptor
|
|
|
- getDescriptor() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcResponseProto_descriptor;
|
|
|
- }
|
|
|
-
|
|
|
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
|
|
- internalGetFieldAccessorTable() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.internal_static_HadoopRpcResponseProto_fieldAccessorTable;
|
|
|
- }
|
|
|
-
|
|
|
- // Construct using org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.newBuilder()
|
|
|
- private Builder() {
|
|
|
- maybeForceBuilderInitialization();
|
|
|
- }
|
|
|
-
|
|
|
- private Builder(BuilderParent parent) {
|
|
|
- super(parent);
|
|
|
- maybeForceBuilderInitialization();
|
|
|
- }
|
|
|
- private void maybeForceBuilderInitialization() {
|
|
|
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
|
|
|
- getExceptionFieldBuilder();
|
|
|
- }
|
|
|
- }
|
|
|
- private static Builder create() {
|
|
|
- return new Builder();
|
|
|
- }
|
|
|
-
|
|
|
- public Builder clear() {
|
|
|
- super.clear();
|
|
|
- status_ = org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus.SUCCESS;
|
|
|
- bitField0_ = (bitField0_ & ~0x00000001);
|
|
|
- response_ = com.google.protobuf.ByteString.EMPTY;
|
|
|
- bitField0_ = (bitField0_ & ~0x00000002);
|
|
|
- if (exceptionBuilder_ == null) {
|
|
|
- exception_ = org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.getDefaultInstance();
|
|
|
- } else {
|
|
|
- exceptionBuilder_.clear();
|
|
|
- }
|
|
|
- bitField0_ = (bitField0_ & ~0x00000004);
|
|
|
- return this;
|
|
|
- }
|
|
|
-
|
|
|
- public Builder clone() {
|
|
|
- return create().mergeFrom(buildPartial());
|
|
|
- }
|
|
|
-
|
|
|
- public com.google.protobuf.Descriptors.Descriptor
|
|
|
- getDescriptorForType() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.getDescriptor();
|
|
|
- }
|
|
|
-
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto getDefaultInstanceForType() {
|
|
|
- return org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.getDefaultInstance();
|
|
|
- }
|
|
|
-
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto build() {
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto result = buildPartial();
|
|
|
- if (!result.isInitialized()) {
|
|
|
- throw newUninitializedMessageException(result);
|
|
|
- }
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- private org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto buildParsed()
|
|
|
- throws com.google.protobuf.InvalidProtocolBufferException {
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto result = buildPartial();
|
|
|
- if (!result.isInitialized()) {
|
|
|
- throw newUninitializedMessageException(
|
|
|
- result).asInvalidProtocolBufferException();
|
|
|
- }
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto buildPartial() {
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto result = new org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto(this);
|
|
|
- int from_bitField0_ = bitField0_;
|
|
|
- int to_bitField0_ = 0;
|
|
|
- if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
|
|
|
- to_bitField0_ |= 0x00000001;
|
|
|
- }
|
|
|
- result.status_ = status_;
|
|
|
- if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
|
|
|
- to_bitField0_ |= 0x00000002;
|
|
|
- }
|
|
|
- result.response_ = response_;
|
|
|
- if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
|
|
|
- to_bitField0_ |= 0x00000004;
|
|
|
- }
|
|
|
- if (exceptionBuilder_ == null) {
|
|
|
- result.exception_ = exception_;
|
|
|
- } else {
|
|
|
- result.exception_ = exceptionBuilder_.build();
|
|
|
- }
|
|
|
- result.bitField0_ = to_bitField0_;
|
|
|
- onBuilt();
|
|
|
- return result;
|
|
|
- }
|
|
|
-
|
|
|
- public Builder mergeFrom(com.google.protobuf.Message other) {
|
|
|
- if (other instanceof org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto) {
|
|
|
- return mergeFrom((org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto)other);
|
|
|
- } else {
|
|
|
- super.mergeFrom(other);
|
|
|
- return this;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- public Builder mergeFrom(org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto other) {
|
|
|
- if (other == org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.getDefaultInstance()) return this;
|
|
|
- if (other.hasStatus()) {
|
|
|
- setStatus(other.getStatus());
|
|
|
- }
|
|
|
- if (other.hasResponse()) {
|
|
|
- setResponse(other.getResponse());
|
|
|
- }
|
|
|
- if (other.hasException()) {
|
|
|
- mergeException(other.getException());
|
|
|
- }
|
|
|
- this.mergeUnknownFields(other.getUnknownFields());
|
|
|
- return this;
|
|
|
- }
|
|
|
-
|
|
|
- public final boolean isInitialized() {
|
|
|
- if (!hasStatus()) {
|
|
|
-
|
|
|
- return false;
|
|
|
- }
|
|
|
- return true;
|
|
|
- }
|
|
|
-
|
|
|
- public Builder mergeFrom(
|
|
|
- com.google.protobuf.CodedInputStream input,
|
|
|
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
|
|
|
- throws java.io.IOException {
|
|
|
- com.google.protobuf.UnknownFieldSet.Builder unknownFields =
|
|
|
- com.google.protobuf.UnknownFieldSet.newBuilder(
|
|
|
- this.getUnknownFields());
|
|
|
- while (true) {
|
|
|
- int tag = input.readTag();
|
|
|
- switch (tag) {
|
|
|
- case 0:
|
|
|
- this.setUnknownFields(unknownFields.build());
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- default: {
|
|
|
- if (!parseUnknownField(input, unknownFields,
|
|
|
- extensionRegistry, tag)) {
|
|
|
- this.setUnknownFields(unknownFields.build());
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- break;
|
|
|
- }
|
|
|
- case 8: {
|
|
|
- int rawValue = input.readEnum();
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus value = org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus.valueOf(rawValue);
|
|
|
- if (value == null) {
|
|
|
- unknownFields.mergeVarintField(1, rawValue);
|
|
|
- } else {
|
|
|
- bitField0_ |= 0x00000001;
|
|
|
- status_ = value;
|
|
|
- }
|
|
|
- break;
|
|
|
- }
|
|
|
- case 18: {
|
|
|
- bitField0_ |= 0x00000002;
|
|
|
- response_ = input.readBytes();
|
|
|
- break;
|
|
|
- }
|
|
|
- case 26: {
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.Builder subBuilder = org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.newBuilder();
|
|
|
- if (hasException()) {
|
|
|
- subBuilder.mergeFrom(getException());
|
|
|
- }
|
|
|
- input.readMessage(subBuilder, extensionRegistry);
|
|
|
- setException(subBuilder.buildPartial());
|
|
|
- break;
|
|
|
- }
|
|
|
- }
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- private int bitField0_;
|
|
|
-
|
|
|
- // required .HadoopRpcResponseProto.ResponseStatus status = 1;
|
|
|
- private org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus status_ = org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus.SUCCESS;
|
|
|
- public boolean hasStatus() {
|
|
|
- return ((bitField0_ & 0x00000001) == 0x00000001);
|
|
|
- }
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus getStatus() {
|
|
|
- return status_;
|
|
|
- }
|
|
|
- public Builder setStatus(org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus value) {
|
|
|
- if (value == null) {
|
|
|
- throw new NullPointerException();
|
|
|
- }
|
|
|
- bitField0_ |= 0x00000001;
|
|
|
- status_ = value;
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- public Builder clearStatus() {
|
|
|
- bitField0_ = (bitField0_ & ~0x00000001);
|
|
|
- status_ = org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.ResponseStatus.SUCCESS;
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
-
|
|
|
- // optional bytes response = 2;
|
|
|
- private com.google.protobuf.ByteString response_ = com.google.protobuf.ByteString.EMPTY;
|
|
|
- public boolean hasResponse() {
|
|
|
- return ((bitField0_ & 0x00000002) == 0x00000002);
|
|
|
- }
|
|
|
- public com.google.protobuf.ByteString getResponse() {
|
|
|
- return response_;
|
|
|
- }
|
|
|
- public Builder setResponse(com.google.protobuf.ByteString value) {
|
|
|
- if (value == null) {
|
|
|
- throw new NullPointerException();
|
|
|
- }
|
|
|
- bitField0_ |= 0x00000002;
|
|
|
- response_ = value;
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
- public Builder clearResponse() {
|
|
|
- bitField0_ = (bitField0_ & ~0x00000002);
|
|
|
- response_ = getDefaultInstance().getResponse();
|
|
|
- onChanged();
|
|
|
- return this;
|
|
|
- }
|
|
|
-
|
|
|
- // optional .HadoopRpcExceptionProto exception = 3;
|
|
|
- private org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto exception_ = org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.getDefaultInstance();
|
|
|
- private com.google.protobuf.SingleFieldBuilder<
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto, org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.Builder, org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProtoOrBuilder> exceptionBuilder_;
|
|
|
- public boolean hasException() {
|
|
|
- return ((bitField0_ & 0x00000004) == 0x00000004);
|
|
|
- }
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto getException() {
|
|
|
- if (exceptionBuilder_ == null) {
|
|
|
- return exception_;
|
|
|
- } else {
|
|
|
- return exceptionBuilder_.getMessage();
|
|
|
- }
|
|
|
- }
|
|
|
- public Builder setException(org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto value) {
|
|
|
- if (exceptionBuilder_ == null) {
|
|
|
- if (value == null) {
|
|
|
- throw new NullPointerException();
|
|
|
- }
|
|
|
- exception_ = value;
|
|
|
- onChanged();
|
|
|
- } else {
|
|
|
- exceptionBuilder_.setMessage(value);
|
|
|
- }
|
|
|
- bitField0_ |= 0x00000004;
|
|
|
- return this;
|
|
|
- }
|
|
|
- public Builder setException(
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.Builder builderForValue) {
|
|
|
- if (exceptionBuilder_ == null) {
|
|
|
- exception_ = builderForValue.build();
|
|
|
- onChanged();
|
|
|
- } else {
|
|
|
- exceptionBuilder_.setMessage(builderForValue.build());
|
|
|
- }
|
|
|
- bitField0_ |= 0x00000004;
|
|
|
- return this;
|
|
|
- }
|
|
|
- public Builder mergeException(org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto value) {
|
|
|
- if (exceptionBuilder_ == null) {
|
|
|
- if (((bitField0_ & 0x00000004) == 0x00000004) &&
|
|
|
- exception_ != org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.getDefaultInstance()) {
|
|
|
- exception_ =
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.newBuilder(exception_).mergeFrom(value).buildPartial();
|
|
|
- } else {
|
|
|
- exception_ = value;
|
|
|
- }
|
|
|
- onChanged();
|
|
|
- } else {
|
|
|
- exceptionBuilder_.mergeFrom(value);
|
|
|
- }
|
|
|
- bitField0_ |= 0x00000004;
|
|
|
- return this;
|
|
|
- }
|
|
|
- public Builder clearException() {
|
|
|
- if (exceptionBuilder_ == null) {
|
|
|
- exception_ = org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.getDefaultInstance();
|
|
|
- onChanged();
|
|
|
- } else {
|
|
|
- exceptionBuilder_.clear();
|
|
|
- }
|
|
|
- bitField0_ = (bitField0_ & ~0x00000004);
|
|
|
- return this;
|
|
|
- }
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.Builder getExceptionBuilder() {
|
|
|
- bitField0_ |= 0x00000004;
|
|
|
- onChanged();
|
|
|
- return getExceptionFieldBuilder().getBuilder();
|
|
|
- }
|
|
|
- public org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProtoOrBuilder getExceptionOrBuilder() {
|
|
|
- if (exceptionBuilder_ != null) {
|
|
|
- return exceptionBuilder_.getMessageOrBuilder();
|
|
|
- } else {
|
|
|
- return exception_;
|
|
|
- }
|
|
|
- }
|
|
|
- private com.google.protobuf.SingleFieldBuilder<
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto, org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.Builder, org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProtoOrBuilder>
|
|
|
- getExceptionFieldBuilder() {
|
|
|
- if (exceptionBuilder_ == null) {
|
|
|
- exceptionBuilder_ = new com.google.protobuf.SingleFieldBuilder<
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto, org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.Builder, org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProtoOrBuilder>(
|
|
|
- exception_,
|
|
|
- getParentForChildren(),
|
|
|
- isClean());
|
|
|
- exception_ = null;
|
|
|
- }
|
|
|
- return exceptionBuilder_;
|
|
|
- }
|
|
|
-
|
|
|
- // @@protoc_insertion_point(builder_scope:HadoopRpcResponseProto)
|
|
|
- }
|
|
|
-
|
|
|
- static {
|
|
|
- defaultInstance = new HadoopRpcResponseProto(true);
|
|
|
- defaultInstance.initFields();
|
|
|
- }
|
|
|
-
|
|
|
- // @@protoc_insertion_point(class_scope:HadoopRpcResponseProto)
|
|
|
- }
|
|
|
-
|
|
|
- private static com.google.protobuf.Descriptors.Descriptor
|
|
|
- internal_static_HadoopRpcRequestProto_descriptor;
|
|
|
- private static
|
|
|
- com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
|
|
- internal_static_HadoopRpcRequestProto_fieldAccessorTable;
|
|
|
- private static com.google.protobuf.Descriptors.Descriptor
|
|
|
- internal_static_HadoopRpcExceptionProto_descriptor;
|
|
|
- private static
|
|
|
- com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
|
|
- internal_static_HadoopRpcExceptionProto_fieldAccessorTable;
|
|
|
- private static com.google.protobuf.Descriptors.Descriptor
|
|
|
- internal_static_HadoopRpcResponseProto_descriptor;
|
|
|
- private static
|
|
|
- com.google.protobuf.GeneratedMessage.FieldAccessorTable
|
|
|
- internal_static_HadoopRpcResponseProto_fieldAccessorTable;
|
|
|
-
|
|
|
- public static com.google.protobuf.Descriptors.FileDescriptor
|
|
|
- getDescriptor() {
|
|
|
- return descriptor;
|
|
|
- }
|
|
|
- private static com.google.protobuf.Descriptors.FileDescriptor
|
|
|
- descriptor;
|
|
|
- static {
|
|
|
- java.lang.String[] descriptorData = {
|
|
|
- "\n\020hadoop_rpc.proto\"\177\n\025HadoopRpcRequestPr" +
|
|
|
- "oto\022\022\n\nmethodName\030\001 \002(\t\022\017\n\007request\030\002 \001(\014" +
|
|
|
- "\022\"\n\032declaringClassProtocolName\030\003 \002(\t\022\035\n\025" +
|
|
|
- "clientProtocolVersion\030\004 \002(\004\"D\n\027HadoopRpc" +
|
|
|
- "ExceptionProto\022\025\n\rexceptionName\030\001 \001(\t\022\022\n" +
|
|
|
- "\nstackTrace\030\002 \001(\t\"\272\001\n\026HadoopRpcResponseP" +
|
|
|
- "roto\0226\n\006status\030\001 \002(\0162&.HadoopRpcResponse" +
|
|
|
- "Proto.ResponseStatus\022\020\n\010response\030\002 \001(\014\022+" +
|
|
|
- "\n\texception\030\003 \001(\0132\030.HadoopRpcExceptionPr" +
|
|
|
- "oto\")\n\016ResponseStatus\022\013\n\007SUCCESS\020\001\022\n\n\006ER",
|
|
|
- "RROR\020\002B4\n\036org.apache.hadoop.ipc.protobuf" +
|
|
|
- "B\017HadoopRpcProtos\240\001\001"
|
|
|
- };
|
|
|
- com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
|
|
|
- new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
|
|
|
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
|
|
|
- com.google.protobuf.Descriptors.FileDescriptor root) {
|
|
|
- descriptor = root;
|
|
|
- internal_static_HadoopRpcRequestProto_descriptor =
|
|
|
- getDescriptor().getMessageTypes().get(0);
|
|
|
- internal_static_HadoopRpcRequestProto_fieldAccessorTable = new
|
|
|
- com.google.protobuf.GeneratedMessage.FieldAccessorTable(
|
|
|
- internal_static_HadoopRpcRequestProto_descriptor,
|
|
|
- new java.lang.String[] { "MethodName", "Request", "DeclaringClassProtocolName", "ClientProtocolVersion", },
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto.class,
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcRequestProto.Builder.class);
|
|
|
- internal_static_HadoopRpcExceptionProto_descriptor =
|
|
|
- getDescriptor().getMessageTypes().get(1);
|
|
|
- internal_static_HadoopRpcExceptionProto_fieldAccessorTable = new
|
|
|
- com.google.protobuf.GeneratedMessage.FieldAccessorTable(
|
|
|
- internal_static_HadoopRpcExceptionProto_descriptor,
|
|
|
- new java.lang.String[] { "ExceptionName", "StackTrace", },
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.class,
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcExceptionProto.Builder.class);
|
|
|
- internal_static_HadoopRpcResponseProto_descriptor =
|
|
|
- getDescriptor().getMessageTypes().get(2);
|
|
|
- internal_static_HadoopRpcResponseProto_fieldAccessorTable = new
|
|
|
- com.google.protobuf.GeneratedMessage.FieldAccessorTable(
|
|
|
- internal_static_HadoopRpcResponseProto_descriptor,
|
|
|
- new java.lang.String[] { "Status", "Response", "Exception", },
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.class,
|
|
|
- org.apache.hadoop.ipc.protobuf.HadoopRpcProtos.HadoopRpcResponseProto.Builder.class);
|
|
|
- return null;
|
|
|
- }
|
|
|
- };
|
|
|
- com.google.protobuf.Descriptors.FileDescriptor
|
|
|
- .internalBuildGeneratedFileFrom(descriptorData,
|
|
|
- new com.google.protobuf.Descriptors.FileDescriptor[] {
|
|
|
- }, assigner);
|
|
|
- }
|
|
|
-
|
|
|
- // @@protoc_insertion_point(outer_class_scope)
|
|
|
-}
|