001// Generated by the protocol buffer compiler.  DO NOT EDIT!
002// source: QJournalProtocol.proto
003
004package org.apache.hadoop.hdfs.qjournal.protocol;
005
006public final class QJournalProtocolProtos {
007  private QJournalProtocolProtos() {}
008  public static void registerAllExtensions(
009      com.google.protobuf.ExtensionRegistry registry) {
010  }
011  public interface JournalIdProtoOrBuilder
012      extends com.google.protobuf.MessageOrBuilder {
013
014    // required string identifier = 1;
015    /**
016     * <code>required string identifier = 1;</code>
017     */
018    boolean hasIdentifier();
019    /**
020     * <code>required string identifier = 1;</code>
021     */
022    java.lang.String getIdentifier();
023    /**
024     * <code>required string identifier = 1;</code>
025     */
026    com.google.protobuf.ByteString
027        getIdentifierBytes();
028  }
029  /**
030   * Protobuf type {@code hadoop.hdfs.qjournal.JournalIdProto}
031   */
032  public static final class JournalIdProto extends
033      com.google.protobuf.GeneratedMessage
034      implements JournalIdProtoOrBuilder {
035    // Use JournalIdProto.newBuilder() to construct.
036    private JournalIdProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
037      super(builder);
038      this.unknownFields = builder.getUnknownFields();
039    }
040    private JournalIdProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
041
042    private static final JournalIdProto defaultInstance;
043    public static JournalIdProto getDefaultInstance() {
044      return defaultInstance;
045    }
046
047    public JournalIdProto getDefaultInstanceForType() {
048      return defaultInstance;
049    }
050
051    private final com.google.protobuf.UnknownFieldSet unknownFields;
052    @java.lang.Override
053    public final com.google.protobuf.UnknownFieldSet
054        getUnknownFields() {
055      return this.unknownFields;
056    }
057    private JournalIdProto(
058        com.google.protobuf.CodedInputStream input,
059        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
060        throws com.google.protobuf.InvalidProtocolBufferException {
061      initFields();
062      int mutable_bitField0_ = 0;
063      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
064          com.google.protobuf.UnknownFieldSet.newBuilder();
065      try {
066        boolean done = false;
067        while (!done) {
068          int tag = input.readTag();
069          switch (tag) {
070            case 0:
071              done = true;
072              break;
073            default: {
074              if (!parseUnknownField(input, unknownFields,
075                                     extensionRegistry, tag)) {
076                done = true;
077              }
078              break;
079            }
080            case 10: {
081              bitField0_ |= 0x00000001;
082              identifier_ = input.readBytes();
083              break;
084            }
085          }
086        }
087      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
088        throw e.setUnfinishedMessage(this);
089      } catch (java.io.IOException e) {
090        throw new com.google.protobuf.InvalidProtocolBufferException(
091            e.getMessage()).setUnfinishedMessage(this);
092      } finally {
093        this.unknownFields = unknownFields.build();
094        makeExtensionsImmutable();
095      }
096    }
097    public static final com.google.protobuf.Descriptors.Descriptor
098        getDescriptor() {
099      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalIdProto_descriptor;
100    }
101
102    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
103        internalGetFieldAccessorTable() {
104      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalIdProto_fieldAccessorTable
105          .ensureFieldAccessorsInitialized(
106              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder.class);
107    }
108
109    public static com.google.protobuf.Parser<JournalIdProto> PARSER =
110        new com.google.protobuf.AbstractParser<JournalIdProto>() {
111      public JournalIdProto parsePartialFrom(
112          com.google.protobuf.CodedInputStream input,
113          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
114          throws com.google.protobuf.InvalidProtocolBufferException {
115        return new JournalIdProto(input, extensionRegistry);
116      }
117    };
118
119    @java.lang.Override
120    public com.google.protobuf.Parser<JournalIdProto> getParserForType() {
121      return PARSER;
122    }
123
124    private int bitField0_;
125    // required string identifier = 1;
126    public static final int IDENTIFIER_FIELD_NUMBER = 1;
127    private java.lang.Object identifier_;
128    /**
129     * <code>required string identifier = 1;</code>
130     */
131    public boolean hasIdentifier() {
132      return ((bitField0_ & 0x00000001) == 0x00000001);
133    }
134    /**
135     * <code>required string identifier = 1;</code>
136     */
137    public java.lang.String getIdentifier() {
138      java.lang.Object ref = identifier_;
139      if (ref instanceof java.lang.String) {
140        return (java.lang.String) ref;
141      } else {
142        com.google.protobuf.ByteString bs = 
143            (com.google.protobuf.ByteString) ref;
144        java.lang.String s = bs.toStringUtf8();
145        if (bs.isValidUtf8()) {
146          identifier_ = s;
147        }
148        return s;
149      }
150    }
151    /**
152     * <code>required string identifier = 1;</code>
153     */
154    public com.google.protobuf.ByteString
155        getIdentifierBytes() {
156      java.lang.Object ref = identifier_;
157      if (ref instanceof java.lang.String) {
158        com.google.protobuf.ByteString b = 
159            com.google.protobuf.ByteString.copyFromUtf8(
160                (java.lang.String) ref);
161        identifier_ = b;
162        return b;
163      } else {
164        return (com.google.protobuf.ByteString) ref;
165      }
166    }
167
168    private void initFields() {
169      identifier_ = "";
170    }
171    private byte memoizedIsInitialized = -1;
172    public final boolean isInitialized() {
173      byte isInitialized = memoizedIsInitialized;
174      if (isInitialized != -1) return isInitialized == 1;
175
176      if (!hasIdentifier()) {
177        memoizedIsInitialized = 0;
178        return false;
179      }
180      memoizedIsInitialized = 1;
181      return true;
182    }
183
184    public void writeTo(com.google.protobuf.CodedOutputStream output)
185                        throws java.io.IOException {
186      getSerializedSize();
187      if (((bitField0_ & 0x00000001) == 0x00000001)) {
188        output.writeBytes(1, getIdentifierBytes());
189      }
190      getUnknownFields().writeTo(output);
191    }
192
193    private int memoizedSerializedSize = -1;
194    public int getSerializedSize() {
195      int size = memoizedSerializedSize;
196      if (size != -1) return size;
197
198      size = 0;
199      if (((bitField0_ & 0x00000001) == 0x00000001)) {
200        size += com.google.protobuf.CodedOutputStream
201          .computeBytesSize(1, getIdentifierBytes());
202      }
203      size += getUnknownFields().getSerializedSize();
204      memoizedSerializedSize = size;
205      return size;
206    }
207
208    private static final long serialVersionUID = 0L;
209    @java.lang.Override
210    protected java.lang.Object writeReplace()
211        throws java.io.ObjectStreamException {
212      return super.writeReplace();
213    }
214
215    @java.lang.Override
216    public boolean equals(final java.lang.Object obj) {
217      if (obj == this) {
218       return true;
219      }
220      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto)) {
221        return super.equals(obj);
222      }
223      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto) obj;
224
225      boolean result = true;
226      result = result && (hasIdentifier() == other.hasIdentifier());
227      if (hasIdentifier()) {
228        result = result && getIdentifier()
229            .equals(other.getIdentifier());
230      }
231      result = result &&
232          getUnknownFields().equals(other.getUnknownFields());
233      return result;
234    }
235
236    private int memoizedHashCode = 0;
237    @java.lang.Override
238    public int hashCode() {
239      if (memoizedHashCode != 0) {
240        return memoizedHashCode;
241      }
242      int hash = 41;
243      hash = (19 * hash) + getDescriptorForType().hashCode();
244      if (hasIdentifier()) {
245        hash = (37 * hash) + IDENTIFIER_FIELD_NUMBER;
246        hash = (53 * hash) + getIdentifier().hashCode();
247      }
248      hash = (29 * hash) + getUnknownFields().hashCode();
249      memoizedHashCode = hash;
250      return hash;
251    }
252
253    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
254        com.google.protobuf.ByteString data)
255        throws com.google.protobuf.InvalidProtocolBufferException {
256      return PARSER.parseFrom(data);
257    }
258    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
259        com.google.protobuf.ByteString data,
260        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
261        throws com.google.protobuf.InvalidProtocolBufferException {
262      return PARSER.parseFrom(data, extensionRegistry);
263    }
264    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(byte[] data)
265        throws com.google.protobuf.InvalidProtocolBufferException {
266      return PARSER.parseFrom(data);
267    }
268    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
269        byte[] data,
270        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
271        throws com.google.protobuf.InvalidProtocolBufferException {
272      return PARSER.parseFrom(data, extensionRegistry);
273    }
274    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(java.io.InputStream input)
275        throws java.io.IOException {
276      return PARSER.parseFrom(input);
277    }
278    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
279        java.io.InputStream input,
280        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
281        throws java.io.IOException {
282      return PARSER.parseFrom(input, extensionRegistry);
283    }
284    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseDelimitedFrom(java.io.InputStream input)
285        throws java.io.IOException {
286      return PARSER.parseDelimitedFrom(input);
287    }
288    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseDelimitedFrom(
289        java.io.InputStream input,
290        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
291        throws java.io.IOException {
292      return PARSER.parseDelimitedFrom(input, extensionRegistry);
293    }
294    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
295        com.google.protobuf.CodedInputStream input)
296        throws java.io.IOException {
297      return PARSER.parseFrom(input);
298    }
299    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
300        com.google.protobuf.CodedInputStream input,
301        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
302        throws java.io.IOException {
303      return PARSER.parseFrom(input, extensionRegistry);
304    }
305
306    public static Builder newBuilder() { return Builder.create(); }
307    public Builder newBuilderForType() { return newBuilder(); }
308    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto prototype) {
309      return newBuilder().mergeFrom(prototype);
310    }
311    public Builder toBuilder() { return newBuilder(this); }
312
313    @java.lang.Override
314    protected Builder newBuilderForType(
315        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
316      Builder builder = new Builder(parent);
317      return builder;
318    }
319    /**
320     * Protobuf type {@code hadoop.hdfs.qjournal.JournalIdProto}
321     */
322    public static final class Builder extends
323        com.google.protobuf.GeneratedMessage.Builder<Builder>
324       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder {
325      public static final com.google.protobuf.Descriptors.Descriptor
326          getDescriptor() {
327        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalIdProto_descriptor;
328      }
329
330      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
331          internalGetFieldAccessorTable() {
332        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalIdProto_fieldAccessorTable
333            .ensureFieldAccessorsInitialized(
334                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder.class);
335      }
336
337      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder()
338      private Builder() {
339        maybeForceBuilderInitialization();
340      }
341
342      private Builder(
343          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
344        super(parent);
345        maybeForceBuilderInitialization();
346      }
347      private void maybeForceBuilderInitialization() {
348        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
349        }
350      }
351      private static Builder create() {
352        return new Builder();
353      }
354
355      public Builder clear() {
356        super.clear();
357        identifier_ = "";
358        bitField0_ = (bitField0_ & ~0x00000001);
359        return this;
360      }
361
362      public Builder clone() {
363        return create().mergeFrom(buildPartial());
364      }
365
366      public com.google.protobuf.Descriptors.Descriptor
367          getDescriptorForType() {
368        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalIdProto_descriptor;
369      }
370
371      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getDefaultInstanceForType() {
372        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
373      }
374
375      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto build() {
376        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto result = buildPartial();
377        if (!result.isInitialized()) {
378          throw newUninitializedMessageException(result);
379        }
380        return result;
381      }
382
383      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto buildPartial() {
384        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto(this);
385        int from_bitField0_ = bitField0_;
386        int to_bitField0_ = 0;
387        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
388          to_bitField0_ |= 0x00000001;
389        }
390        result.identifier_ = identifier_;
391        result.bitField0_ = to_bitField0_;
392        onBuilt();
393        return result;
394      }
395
396      public Builder mergeFrom(com.google.protobuf.Message other) {
397        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto) {
398          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto)other);
399        } else {
400          super.mergeFrom(other);
401          return this;
402        }
403      }
404
405      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto other) {
406        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) return this;
407        if (other.hasIdentifier()) {
408          bitField0_ |= 0x00000001;
409          identifier_ = other.identifier_;
410          onChanged();
411        }
412        this.mergeUnknownFields(other.getUnknownFields());
413        return this;
414      }
415
416      public final boolean isInitialized() {
417        if (!hasIdentifier()) {
418          
419          return false;
420        }
421        return true;
422      }
423
424      public Builder mergeFrom(
425          com.google.protobuf.CodedInputStream input,
426          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
427          throws java.io.IOException {
428        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parsedMessage = null;
429        try {
430          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
431        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
432          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto) e.getUnfinishedMessage();
433          throw e;
434        } finally {
435          if (parsedMessage != null) {
436            mergeFrom(parsedMessage);
437          }
438        }
439        return this;
440      }
441      private int bitField0_;
442
443      // required string identifier = 1;
444      private java.lang.Object identifier_ = "";
445      /**
446       * <code>required string identifier = 1;</code>
447       */
448      public boolean hasIdentifier() {
449        return ((bitField0_ & 0x00000001) == 0x00000001);
450      }
451      /**
452       * <code>required string identifier = 1;</code>
453       */
454      public java.lang.String getIdentifier() {
455        java.lang.Object ref = identifier_;
456        if (!(ref instanceof java.lang.String)) {
457          java.lang.String s = ((com.google.protobuf.ByteString) ref)
458              .toStringUtf8();
459          identifier_ = s;
460          return s;
461        } else {
462          return (java.lang.String) ref;
463        }
464      }
465      /**
466       * <code>required string identifier = 1;</code>
467       */
468      public com.google.protobuf.ByteString
469          getIdentifierBytes() {
470        java.lang.Object ref = identifier_;
471        if (ref instanceof String) {
472          com.google.protobuf.ByteString b = 
473              com.google.protobuf.ByteString.copyFromUtf8(
474                  (java.lang.String) ref);
475          identifier_ = b;
476          return b;
477        } else {
478          return (com.google.protobuf.ByteString) ref;
479        }
480      }
481      /**
482       * <code>required string identifier = 1;</code>
483       */
484      public Builder setIdentifier(
485          java.lang.String value) {
486        if (value == null) {
487    throw new NullPointerException();
488  }
489  bitField0_ |= 0x00000001;
490        identifier_ = value;
491        onChanged();
492        return this;
493      }
494      /**
495       * <code>required string identifier = 1;</code>
496       */
497      public Builder clearIdentifier() {
498        bitField0_ = (bitField0_ & ~0x00000001);
499        identifier_ = getDefaultInstance().getIdentifier();
500        onChanged();
501        return this;
502      }
503      /**
504       * <code>required string identifier = 1;</code>
505       */
506      public Builder setIdentifierBytes(
507          com.google.protobuf.ByteString value) {
508        if (value == null) {
509    throw new NullPointerException();
510  }
511  bitField0_ |= 0x00000001;
512        identifier_ = value;
513        onChanged();
514        return this;
515      }
516
517      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.JournalIdProto)
518    }
519
520    static {
521      defaultInstance = new JournalIdProto(true);
522      defaultInstance.initFields();
523    }
524
525    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.JournalIdProto)
526  }
527
528  public interface RequestInfoProtoOrBuilder
529      extends com.google.protobuf.MessageOrBuilder {
530
531    // required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;
532    /**
533     * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
534     */
535    boolean hasJournalId();
536    /**
537     * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
538     */
539    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId();
540    /**
541     * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
542     */
543    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder();
544
545    // required uint64 epoch = 2;
546    /**
547     * <code>required uint64 epoch = 2;</code>
548     */
549    boolean hasEpoch();
550    /**
551     * <code>required uint64 epoch = 2;</code>
552     */
553    long getEpoch();
554
555    // required uint64 ipcSerialNumber = 3;
556    /**
557     * <code>required uint64 ipcSerialNumber = 3;</code>
558     */
559    boolean hasIpcSerialNumber();
560    /**
561     * <code>required uint64 ipcSerialNumber = 3;</code>
562     */
563    long getIpcSerialNumber();
564
565    // optional uint64 committedTxId = 4;
566    /**
567     * <code>optional uint64 committedTxId = 4;</code>
568     *
569     * <pre>
570     * Whenever a writer makes a request, it informs
571     * the node of the latest committed txid. This may
572     * be higher than the transaction data included in the
573     * request itself, eg in the case that the node has
574     * fallen behind.
575     * </pre>
576     */
577    boolean hasCommittedTxId();
578    /**
579     * <code>optional uint64 committedTxId = 4;</code>
580     *
581     * <pre>
582     * Whenever a writer makes a request, it informs
583     * the node of the latest committed txid. This may
584     * be higher than the transaction data included in the
585     * request itself, eg in the case that the node has
586     * fallen behind.
587     * </pre>
588     */
589    long getCommittedTxId();
590  }
591  /**
592   * Protobuf type {@code hadoop.hdfs.qjournal.RequestInfoProto}
593   */
594  public static final class RequestInfoProto extends
595      com.google.protobuf.GeneratedMessage
596      implements RequestInfoProtoOrBuilder {
597    // Use RequestInfoProto.newBuilder() to construct.
598    private RequestInfoProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
599      super(builder);
600      this.unknownFields = builder.getUnknownFields();
601    }
602    private RequestInfoProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
603
604    private static final RequestInfoProto defaultInstance;
605    public static RequestInfoProto getDefaultInstance() {
606      return defaultInstance;
607    }
608
609    public RequestInfoProto getDefaultInstanceForType() {
610      return defaultInstance;
611    }
612
613    private final com.google.protobuf.UnknownFieldSet unknownFields;
614    @java.lang.Override
615    public final com.google.protobuf.UnknownFieldSet
616        getUnknownFields() {
617      return this.unknownFields;
618    }
619    private RequestInfoProto(
620        com.google.protobuf.CodedInputStream input,
621        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
622        throws com.google.protobuf.InvalidProtocolBufferException {
623      initFields();
624      int mutable_bitField0_ = 0;
625      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
626          com.google.protobuf.UnknownFieldSet.newBuilder();
627      try {
628        boolean done = false;
629        while (!done) {
630          int tag = input.readTag();
631          switch (tag) {
632            case 0:
633              done = true;
634              break;
635            default: {
636              if (!parseUnknownField(input, unknownFields,
637                                     extensionRegistry, tag)) {
638                done = true;
639              }
640              break;
641            }
642            case 10: {
643              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
644              if (((bitField0_ & 0x00000001) == 0x00000001)) {
645                subBuilder = journalId_.toBuilder();
646              }
647              journalId_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
648              if (subBuilder != null) {
649                subBuilder.mergeFrom(journalId_);
650                journalId_ = subBuilder.buildPartial();
651              }
652              bitField0_ |= 0x00000001;
653              break;
654            }
655            case 16: {
656              bitField0_ |= 0x00000002;
657              epoch_ = input.readUInt64();
658              break;
659            }
660            case 24: {
661              bitField0_ |= 0x00000004;
662              ipcSerialNumber_ = input.readUInt64();
663              break;
664            }
665            case 32: {
666              bitField0_ |= 0x00000008;
667              committedTxId_ = input.readUInt64();
668              break;
669            }
670          }
671        }
672      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
673        throw e.setUnfinishedMessage(this);
674      } catch (java.io.IOException e) {
675        throw new com.google.protobuf.InvalidProtocolBufferException(
676            e.getMessage()).setUnfinishedMessage(this);
677      } finally {
678        this.unknownFields = unknownFields.build();
679        makeExtensionsImmutable();
680      }
681    }
682    public static final com.google.protobuf.Descriptors.Descriptor
683        getDescriptor() {
684      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_RequestInfoProto_descriptor;
685    }
686
687    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
688        internalGetFieldAccessorTable() {
689      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_RequestInfoProto_fieldAccessorTable
690          .ensureFieldAccessorsInitialized(
691              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder.class);
692    }
693
694    public static com.google.protobuf.Parser<RequestInfoProto> PARSER =
695        new com.google.protobuf.AbstractParser<RequestInfoProto>() {
696      public RequestInfoProto parsePartialFrom(
697          com.google.protobuf.CodedInputStream input,
698          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
699          throws com.google.protobuf.InvalidProtocolBufferException {
700        return new RequestInfoProto(input, extensionRegistry);
701      }
702    };
703
704    @java.lang.Override
705    public com.google.protobuf.Parser<RequestInfoProto> getParserForType() {
706      return PARSER;
707    }
708
709    private int bitField0_;
710    // required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;
711    public static final int JOURNALID_FIELD_NUMBER = 1;
712    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto journalId_;
713    /**
714     * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
715     */
716    public boolean hasJournalId() {
717      return ((bitField0_ & 0x00000001) == 0x00000001);
718    }
719    /**
720     * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
721     */
722    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId() {
723      return journalId_;
724    }
725    /**
726     * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
727     */
728    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder() {
729      return journalId_;
730    }
731
732    // required uint64 epoch = 2;
733    public static final int EPOCH_FIELD_NUMBER = 2;
734    private long epoch_;
735    /**
736     * <code>required uint64 epoch = 2;</code>
737     */
738    public boolean hasEpoch() {
739      return ((bitField0_ & 0x00000002) == 0x00000002);
740    }
741    /**
742     * <code>required uint64 epoch = 2;</code>
743     */
744    public long getEpoch() {
745      return epoch_;
746    }
747
748    // required uint64 ipcSerialNumber = 3;
749    public static final int IPCSERIALNUMBER_FIELD_NUMBER = 3;
750    private long ipcSerialNumber_;
751    /**
752     * <code>required uint64 ipcSerialNumber = 3;</code>
753     */
754    public boolean hasIpcSerialNumber() {
755      return ((bitField0_ & 0x00000004) == 0x00000004);
756    }
757    /**
758     * <code>required uint64 ipcSerialNumber = 3;</code>
759     */
760    public long getIpcSerialNumber() {
761      return ipcSerialNumber_;
762    }
763
764    // optional uint64 committedTxId = 4;
765    public static final int COMMITTEDTXID_FIELD_NUMBER = 4;
766    private long committedTxId_;
767    /**
768     * <code>optional uint64 committedTxId = 4;</code>
769     *
770     * <pre>
771     * Whenever a writer makes a request, it informs
772     * the node of the latest committed txid. This may
773     * be higher than the transaction data included in the
774     * request itself, eg in the case that the node has
775     * fallen behind.
776     * </pre>
777     */
778    public boolean hasCommittedTxId() {
779      return ((bitField0_ & 0x00000008) == 0x00000008);
780    }
781    /**
782     * <code>optional uint64 committedTxId = 4;</code>
783     *
784     * <pre>
785     * Whenever a writer makes a request, it informs
786     * the node of the latest committed txid. This may
787     * be higher than the transaction data included in the
788     * request itself, eg in the case that the node has
789     * fallen behind.
790     * </pre>
791     */
792    public long getCommittedTxId() {
793      return committedTxId_;
794    }
795
796    private void initFields() {
797      journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
798      epoch_ = 0L;
799      ipcSerialNumber_ = 0L;
800      committedTxId_ = 0L;
801    }
802    private byte memoizedIsInitialized = -1;
803    public final boolean isInitialized() {
804      byte isInitialized = memoizedIsInitialized;
805      if (isInitialized != -1) return isInitialized == 1;
806
807      if (!hasJournalId()) {
808        memoizedIsInitialized = 0;
809        return false;
810      }
811      if (!hasEpoch()) {
812        memoizedIsInitialized = 0;
813        return false;
814      }
815      if (!hasIpcSerialNumber()) {
816        memoizedIsInitialized = 0;
817        return false;
818      }
819      if (!getJournalId().isInitialized()) {
820        memoizedIsInitialized = 0;
821        return false;
822      }
823      memoizedIsInitialized = 1;
824      return true;
825    }
826
827    public void writeTo(com.google.protobuf.CodedOutputStream output)
828                        throws java.io.IOException {
829      getSerializedSize();
830      if (((bitField0_ & 0x00000001) == 0x00000001)) {
831        output.writeMessage(1, journalId_);
832      }
833      if (((bitField0_ & 0x00000002) == 0x00000002)) {
834        output.writeUInt64(2, epoch_);
835      }
836      if (((bitField0_ & 0x00000004) == 0x00000004)) {
837        output.writeUInt64(3, ipcSerialNumber_);
838      }
839      if (((bitField0_ & 0x00000008) == 0x00000008)) {
840        output.writeUInt64(4, committedTxId_);
841      }
842      getUnknownFields().writeTo(output);
843    }
844
845    private int memoizedSerializedSize = -1;
846    public int getSerializedSize() {
847      int size = memoizedSerializedSize;
848      if (size != -1) return size;
849
850      size = 0;
851      if (((bitField0_ & 0x00000001) == 0x00000001)) {
852        size += com.google.protobuf.CodedOutputStream
853          .computeMessageSize(1, journalId_);
854      }
855      if (((bitField0_ & 0x00000002) == 0x00000002)) {
856        size += com.google.protobuf.CodedOutputStream
857          .computeUInt64Size(2, epoch_);
858      }
859      if (((bitField0_ & 0x00000004) == 0x00000004)) {
860        size += com.google.protobuf.CodedOutputStream
861          .computeUInt64Size(3, ipcSerialNumber_);
862      }
863      if (((bitField0_ & 0x00000008) == 0x00000008)) {
864        size += com.google.protobuf.CodedOutputStream
865          .computeUInt64Size(4, committedTxId_);
866      }
867      size += getUnknownFields().getSerializedSize();
868      memoizedSerializedSize = size;
869      return size;
870    }
871
872    private static final long serialVersionUID = 0L;
873    @java.lang.Override
874    protected java.lang.Object writeReplace()
875        throws java.io.ObjectStreamException {
876      return super.writeReplace();
877    }
878
879    @java.lang.Override
880    public boolean equals(final java.lang.Object obj) {
881      if (obj == this) {
882       return true;
883      }
884      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto)) {
885        return super.equals(obj);
886      }
887      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto) obj;
888
889      boolean result = true;
890      result = result && (hasJournalId() == other.hasJournalId());
891      if (hasJournalId()) {
892        result = result && getJournalId()
893            .equals(other.getJournalId());
894      }
895      result = result && (hasEpoch() == other.hasEpoch());
896      if (hasEpoch()) {
897        result = result && (getEpoch()
898            == other.getEpoch());
899      }
900      result = result && (hasIpcSerialNumber() == other.hasIpcSerialNumber());
901      if (hasIpcSerialNumber()) {
902        result = result && (getIpcSerialNumber()
903            == other.getIpcSerialNumber());
904      }
905      result = result && (hasCommittedTxId() == other.hasCommittedTxId());
906      if (hasCommittedTxId()) {
907        result = result && (getCommittedTxId()
908            == other.getCommittedTxId());
909      }
910      result = result &&
911          getUnknownFields().equals(other.getUnknownFields());
912      return result;
913    }
914
915    private int memoizedHashCode = 0;
916    @java.lang.Override
917    public int hashCode() {
918      if (memoizedHashCode != 0) {
919        return memoizedHashCode;
920      }
921      int hash = 41;
922      hash = (19 * hash) + getDescriptorForType().hashCode();
923      if (hasJournalId()) {
924        hash = (37 * hash) + JOURNALID_FIELD_NUMBER;
925        hash = (53 * hash) + getJournalId().hashCode();
926      }
927      if (hasEpoch()) {
928        hash = (37 * hash) + EPOCH_FIELD_NUMBER;
929        hash = (53 * hash) + hashLong(getEpoch());
930      }
931      if (hasIpcSerialNumber()) {
932        hash = (37 * hash) + IPCSERIALNUMBER_FIELD_NUMBER;
933        hash = (53 * hash) + hashLong(getIpcSerialNumber());
934      }
935      if (hasCommittedTxId()) {
936        hash = (37 * hash) + COMMITTEDTXID_FIELD_NUMBER;
937        hash = (53 * hash) + hashLong(getCommittedTxId());
938      }
939      hash = (29 * hash) + getUnknownFields().hashCode();
940      memoizedHashCode = hash;
941      return hash;
942    }
943
944    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
945        com.google.protobuf.ByteString data)
946        throws com.google.protobuf.InvalidProtocolBufferException {
947      return PARSER.parseFrom(data);
948    }
949    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
950        com.google.protobuf.ByteString data,
951        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
952        throws com.google.protobuf.InvalidProtocolBufferException {
953      return PARSER.parseFrom(data, extensionRegistry);
954    }
955    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(byte[] data)
956        throws com.google.protobuf.InvalidProtocolBufferException {
957      return PARSER.parseFrom(data);
958    }
959    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
960        byte[] data,
961        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
962        throws com.google.protobuf.InvalidProtocolBufferException {
963      return PARSER.parseFrom(data, extensionRegistry);
964    }
965    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(java.io.InputStream input)
966        throws java.io.IOException {
967      return PARSER.parseFrom(input);
968    }
969    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
970        java.io.InputStream input,
971        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
972        throws java.io.IOException {
973      return PARSER.parseFrom(input, extensionRegistry);
974    }
975    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseDelimitedFrom(java.io.InputStream input)
976        throws java.io.IOException {
977      return PARSER.parseDelimitedFrom(input);
978    }
979    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseDelimitedFrom(
980        java.io.InputStream input,
981        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
982        throws java.io.IOException {
983      return PARSER.parseDelimitedFrom(input, extensionRegistry);
984    }
985    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
986        com.google.protobuf.CodedInputStream input)
987        throws java.io.IOException {
988      return PARSER.parseFrom(input);
989    }
990    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
991        com.google.protobuf.CodedInputStream input,
992        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
993        throws java.io.IOException {
994      return PARSER.parseFrom(input, extensionRegistry);
995    }
996
997    public static Builder newBuilder() { return Builder.create(); }
998    public Builder newBuilderForType() { return newBuilder(); }
999    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto prototype) {
1000      return newBuilder().mergeFrom(prototype);
1001    }
1002    public Builder toBuilder() { return newBuilder(this); }
1003
1004    @java.lang.Override
1005    protected Builder newBuilderForType(
1006        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1007      Builder builder = new Builder(parent);
1008      return builder;
1009    }
1010    /**
1011     * Protobuf type {@code hadoop.hdfs.qjournal.RequestInfoProto}
1012     */
1013    public static final class Builder extends
1014        com.google.protobuf.GeneratedMessage.Builder<Builder>
1015       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder {
1016      public static final com.google.protobuf.Descriptors.Descriptor
1017          getDescriptor() {
1018        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_RequestInfoProto_descriptor;
1019      }
1020
1021      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1022          internalGetFieldAccessorTable() {
1023        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_RequestInfoProto_fieldAccessorTable
1024            .ensureFieldAccessorsInitialized(
1025                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder.class);
1026      }
1027
1028      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder()
1029      private Builder() {
1030        maybeForceBuilderInitialization();
1031      }
1032
1033      private Builder(
1034          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1035        super(parent);
1036        maybeForceBuilderInitialization();
1037      }
1038      private void maybeForceBuilderInitialization() {
1039        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1040          getJournalIdFieldBuilder();
1041        }
1042      }
1043      private static Builder create() {
1044        return new Builder();
1045      }
1046
1047      public Builder clear() {
1048        super.clear();
1049        if (journalIdBuilder_ == null) {
1050          journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
1051        } else {
1052          journalIdBuilder_.clear();
1053        }
1054        bitField0_ = (bitField0_ & ~0x00000001);
1055        epoch_ = 0L;
1056        bitField0_ = (bitField0_ & ~0x00000002);
1057        ipcSerialNumber_ = 0L;
1058        bitField0_ = (bitField0_ & ~0x00000004);
1059        committedTxId_ = 0L;
1060        bitField0_ = (bitField0_ & ~0x00000008);
1061        return this;
1062      }
1063
1064      public Builder clone() {
1065        return create().mergeFrom(buildPartial());
1066      }
1067
1068      public com.google.protobuf.Descriptors.Descriptor
1069          getDescriptorForType() {
1070        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_RequestInfoProto_descriptor;
1071      }
1072
1073      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getDefaultInstanceForType() {
1074        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
1075      }
1076
1077      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto build() {
1078        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto result = buildPartial();
1079        if (!result.isInitialized()) {
1080          throw newUninitializedMessageException(result);
1081        }
1082        return result;
1083      }
1084
1085      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto buildPartial() {
1086        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto(this);
1087        int from_bitField0_ = bitField0_;
1088        int to_bitField0_ = 0;
1089        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
1090          to_bitField0_ |= 0x00000001;
1091        }
1092        if (journalIdBuilder_ == null) {
1093          result.journalId_ = journalId_;
1094        } else {
1095          result.journalId_ = journalIdBuilder_.build();
1096        }
1097        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
1098          to_bitField0_ |= 0x00000002;
1099        }
1100        result.epoch_ = epoch_;
1101        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
1102          to_bitField0_ |= 0x00000004;
1103        }
1104        result.ipcSerialNumber_ = ipcSerialNumber_;
1105        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
1106          to_bitField0_ |= 0x00000008;
1107        }
1108        result.committedTxId_ = committedTxId_;
1109        result.bitField0_ = to_bitField0_;
1110        onBuilt();
1111        return result;
1112      }
1113
1114      public Builder mergeFrom(com.google.protobuf.Message other) {
1115        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto) {
1116          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto)other);
1117        } else {
1118          super.mergeFrom(other);
1119          return this;
1120        }
1121      }
1122
1123      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto other) {
1124        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) return this;
1125        if (other.hasJournalId()) {
1126          mergeJournalId(other.getJournalId());
1127        }
1128        if (other.hasEpoch()) {
1129          setEpoch(other.getEpoch());
1130        }
1131        if (other.hasIpcSerialNumber()) {
1132          setIpcSerialNumber(other.getIpcSerialNumber());
1133        }
1134        if (other.hasCommittedTxId()) {
1135          setCommittedTxId(other.getCommittedTxId());
1136        }
1137        this.mergeUnknownFields(other.getUnknownFields());
1138        return this;
1139      }
1140
1141      public final boolean isInitialized() {
1142        if (!hasJournalId()) {
1143          
1144          return false;
1145        }
1146        if (!hasEpoch()) {
1147          
1148          return false;
1149        }
1150        if (!hasIpcSerialNumber()) {
1151          
1152          return false;
1153        }
1154        if (!getJournalId().isInitialized()) {
1155          
1156          return false;
1157        }
1158        return true;
1159      }
1160
1161      public Builder mergeFrom(
1162          com.google.protobuf.CodedInputStream input,
1163          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1164          throws java.io.IOException {
1165        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parsedMessage = null;
1166        try {
1167          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
1168        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1169          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto) e.getUnfinishedMessage();
1170          throw e;
1171        } finally {
1172          if (parsedMessage != null) {
1173            mergeFrom(parsedMessage);
1174          }
1175        }
1176        return this;
1177      }
1178      private int bitField0_;
1179
1180      // required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;
1181      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
1182      private com.google.protobuf.SingleFieldBuilder<
1183          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> journalIdBuilder_;
1184      /**
1185       * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
1186       */
1187      public boolean hasJournalId() {
1188        return ((bitField0_ & 0x00000001) == 0x00000001);
1189      }
1190      /**
1191       * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
1192       */
1193      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId() {
1194        if (journalIdBuilder_ == null) {
1195          return journalId_;
1196        } else {
1197          return journalIdBuilder_.getMessage();
1198        }
1199      }
1200      /**
1201       * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
1202       */
1203      public Builder setJournalId(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
1204        if (journalIdBuilder_ == null) {
1205          if (value == null) {
1206            throw new NullPointerException();
1207          }
1208          journalId_ = value;
1209          onChanged();
1210        } else {
1211          journalIdBuilder_.setMessage(value);
1212        }
1213        bitField0_ |= 0x00000001;
1214        return this;
1215      }
1216      /**
1217       * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
1218       */
1219      public Builder setJournalId(
1220          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
1221        if (journalIdBuilder_ == null) {
1222          journalId_ = builderForValue.build();
1223          onChanged();
1224        } else {
1225          journalIdBuilder_.setMessage(builderForValue.build());
1226        }
1227        bitField0_ |= 0x00000001;
1228        return this;
1229      }
1230      /**
1231       * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
1232       */
1233      public Builder mergeJournalId(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
1234        if (journalIdBuilder_ == null) {
1235          if (((bitField0_ & 0x00000001) == 0x00000001) &&
1236              journalId_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
1237            journalId_ =
1238              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(journalId_).mergeFrom(value).buildPartial();
1239          } else {
1240            journalId_ = value;
1241          }
1242          onChanged();
1243        } else {
1244          journalIdBuilder_.mergeFrom(value);
1245        }
1246        bitField0_ |= 0x00000001;
1247        return this;
1248      }
1249      /**
1250       * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
1251       */
1252      public Builder clearJournalId() {
1253        if (journalIdBuilder_ == null) {
1254          journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
1255          onChanged();
1256        } else {
1257          journalIdBuilder_.clear();
1258        }
1259        bitField0_ = (bitField0_ & ~0x00000001);
1260        return this;
1261      }
1262      /**
1263       * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
1264       */
1265      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJournalIdBuilder() {
1266        bitField0_ |= 0x00000001;
1267        onChanged();
1268        return getJournalIdFieldBuilder().getBuilder();
1269      }
1270      /**
1271       * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
1272       */
1273      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder() {
1274        if (journalIdBuilder_ != null) {
1275          return journalIdBuilder_.getMessageOrBuilder();
1276        } else {
1277          return journalId_;
1278        }
1279      }
1280      /**
1281       * <code>required .hadoop.hdfs.qjournal.JournalIdProto journalId = 1;</code>
1282       */
1283      private com.google.protobuf.SingleFieldBuilder<
1284          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
1285          getJournalIdFieldBuilder() {
1286        if (journalIdBuilder_ == null) {
1287          journalIdBuilder_ = new com.google.protobuf.SingleFieldBuilder<
1288              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
1289                  journalId_,
1290                  getParentForChildren(),
1291                  isClean());
1292          journalId_ = null;
1293        }
1294        return journalIdBuilder_;
1295      }
1296
1297      // required uint64 epoch = 2;
1298      private long epoch_ ;
1299      /**
1300       * <code>required uint64 epoch = 2;</code>
1301       */
1302      public boolean hasEpoch() {
1303        return ((bitField0_ & 0x00000002) == 0x00000002);
1304      }
1305      /**
1306       * <code>required uint64 epoch = 2;</code>
1307       */
1308      public long getEpoch() {
1309        return epoch_;
1310      }
1311      /**
1312       * <code>required uint64 epoch = 2;</code>
1313       */
1314      public Builder setEpoch(long value) {
1315        bitField0_ |= 0x00000002;
1316        epoch_ = value;
1317        onChanged();
1318        return this;
1319      }
1320      /**
1321       * <code>required uint64 epoch = 2;</code>
1322       */
1323      public Builder clearEpoch() {
1324        bitField0_ = (bitField0_ & ~0x00000002);
1325        epoch_ = 0L;
1326        onChanged();
1327        return this;
1328      }
1329
1330      // required uint64 ipcSerialNumber = 3;
1331      private long ipcSerialNumber_ ;
1332      /**
1333       * <code>required uint64 ipcSerialNumber = 3;</code>
1334       */
1335      public boolean hasIpcSerialNumber() {
1336        return ((bitField0_ & 0x00000004) == 0x00000004);
1337      }
1338      /**
1339       * <code>required uint64 ipcSerialNumber = 3;</code>
1340       */
1341      public long getIpcSerialNumber() {
1342        return ipcSerialNumber_;
1343      }
1344      /**
1345       * <code>required uint64 ipcSerialNumber = 3;</code>
1346       */
1347      public Builder setIpcSerialNumber(long value) {
1348        bitField0_ |= 0x00000004;
1349        ipcSerialNumber_ = value;
1350        onChanged();
1351        return this;
1352      }
1353      /**
1354       * <code>required uint64 ipcSerialNumber = 3;</code>
1355       */
1356      public Builder clearIpcSerialNumber() {
1357        bitField0_ = (bitField0_ & ~0x00000004);
1358        ipcSerialNumber_ = 0L;
1359        onChanged();
1360        return this;
1361      }
1362
1363      // optional uint64 committedTxId = 4;
1364      private long committedTxId_ ;
1365      /**
1366       * <code>optional uint64 committedTxId = 4;</code>
1367       *
1368       * <pre>
1369       * Whenever a writer makes a request, it informs
1370       * the node of the latest committed txid. This may
1371       * be higher than the transaction data included in the
1372       * request itself, eg in the case that the node has
1373       * fallen behind.
1374       * </pre>
1375       */
1376      public boolean hasCommittedTxId() {
1377        return ((bitField0_ & 0x00000008) == 0x00000008);
1378      }
1379      /**
1380       * <code>optional uint64 committedTxId = 4;</code>
1381       *
1382       * <pre>
1383       * Whenever a writer makes a request, it informs
1384       * the node of the latest committed txid. This may
1385       * be higher than the transaction data included in the
1386       * request itself, eg in the case that the node has
1387       * fallen behind.
1388       * </pre>
1389       */
1390      public long getCommittedTxId() {
1391        return committedTxId_;
1392      }
1393      /**
1394       * <code>optional uint64 committedTxId = 4;</code>
1395       *
1396       * <pre>
1397       * Whenever a writer makes a request, it informs
1398       * the node of the latest committed txid. This may
1399       * be higher than the transaction data included in the
1400       * request itself, eg in the case that the node has
1401       * fallen behind.
1402       * </pre>
1403       */
1404      public Builder setCommittedTxId(long value) {
1405        bitField0_ |= 0x00000008;
1406        committedTxId_ = value;
1407        onChanged();
1408        return this;
1409      }
1410      /**
1411       * <code>optional uint64 committedTxId = 4;</code>
1412       *
1413       * <pre>
1414       * Whenever a writer makes a request, it informs
1415       * the node of the latest committed txid. This may
1416       * be higher than the transaction data included in the
1417       * request itself, eg in the case that the node has
1418       * fallen behind.
1419       * </pre>
1420       */
1421      public Builder clearCommittedTxId() {
1422        bitField0_ = (bitField0_ & ~0x00000008);
1423        committedTxId_ = 0L;
1424        onChanged();
1425        return this;
1426      }
1427
1428      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.RequestInfoProto)
1429    }
1430
1431    static {
1432      defaultInstance = new RequestInfoProto(true);
1433      defaultInstance.initFields();
1434    }
1435
1436    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.RequestInfoProto)
1437  }
1438
1439  public interface SegmentStateProtoOrBuilder
1440      extends com.google.protobuf.MessageOrBuilder {
1441
1442    // required uint64 startTxId = 1;
1443    /**
1444     * <code>required uint64 startTxId = 1;</code>
1445     */
1446    boolean hasStartTxId();
1447    /**
1448     * <code>required uint64 startTxId = 1;</code>
1449     */
1450    long getStartTxId();
1451
1452    // required uint64 endTxId = 2;
1453    /**
1454     * <code>required uint64 endTxId = 2;</code>
1455     */
1456    boolean hasEndTxId();
1457    /**
1458     * <code>required uint64 endTxId = 2;</code>
1459     */
1460    long getEndTxId();
1461
1462    // required bool isInProgress = 3;
1463    /**
1464     * <code>required bool isInProgress = 3;</code>
1465     */
1466    boolean hasIsInProgress();
1467    /**
1468     * <code>required bool isInProgress = 3;</code>
1469     */
1470    boolean getIsInProgress();
1471  }
1472  /**
1473   * Protobuf type {@code hadoop.hdfs.qjournal.SegmentStateProto}
1474   */
1475  public static final class SegmentStateProto extends
1476      com.google.protobuf.GeneratedMessage
1477      implements SegmentStateProtoOrBuilder {
1478    // Use SegmentStateProto.newBuilder() to construct.
1479    private SegmentStateProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
1480      super(builder);
1481      this.unknownFields = builder.getUnknownFields();
1482    }
1483    private SegmentStateProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
1484
1485    private static final SegmentStateProto defaultInstance;
1486    public static SegmentStateProto getDefaultInstance() {
1487      return defaultInstance;
1488    }
1489
1490    public SegmentStateProto getDefaultInstanceForType() {
1491      return defaultInstance;
1492    }
1493
1494    private final com.google.protobuf.UnknownFieldSet unknownFields;
1495    @java.lang.Override
1496    public final com.google.protobuf.UnknownFieldSet
1497        getUnknownFields() {
1498      return this.unknownFields;
1499    }
1500    private SegmentStateProto(
1501        com.google.protobuf.CodedInputStream input,
1502        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1503        throws com.google.protobuf.InvalidProtocolBufferException {
1504      initFields();
1505      int mutable_bitField0_ = 0;
1506      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1507          com.google.protobuf.UnknownFieldSet.newBuilder();
1508      try {
1509        boolean done = false;
1510        while (!done) {
1511          int tag = input.readTag();
1512          switch (tag) {
1513            case 0:
1514              done = true;
1515              break;
1516            default: {
1517              if (!parseUnknownField(input, unknownFields,
1518                                     extensionRegistry, tag)) {
1519                done = true;
1520              }
1521              break;
1522            }
1523            case 8: {
1524              bitField0_ |= 0x00000001;
1525              startTxId_ = input.readUInt64();
1526              break;
1527            }
1528            case 16: {
1529              bitField0_ |= 0x00000002;
1530              endTxId_ = input.readUInt64();
1531              break;
1532            }
1533            case 24: {
1534              bitField0_ |= 0x00000004;
1535              isInProgress_ = input.readBool();
1536              break;
1537            }
1538          }
1539        }
1540      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1541        throw e.setUnfinishedMessage(this);
1542      } catch (java.io.IOException e) {
1543        throw new com.google.protobuf.InvalidProtocolBufferException(
1544            e.getMessage()).setUnfinishedMessage(this);
1545      } finally {
1546        this.unknownFields = unknownFields.build();
1547        makeExtensionsImmutable();
1548      }
1549    }
1550    public static final com.google.protobuf.Descriptors.Descriptor
1551        getDescriptor() {
1552      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_SegmentStateProto_descriptor;
1553    }
1554
1555    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1556        internalGetFieldAccessorTable() {
1557      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_SegmentStateProto_fieldAccessorTable
1558          .ensureFieldAccessorsInitialized(
1559              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder.class);
1560    }
1561
1562    public static com.google.protobuf.Parser<SegmentStateProto> PARSER =
1563        new com.google.protobuf.AbstractParser<SegmentStateProto>() {
1564      public SegmentStateProto parsePartialFrom(
1565          com.google.protobuf.CodedInputStream input,
1566          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1567          throws com.google.protobuf.InvalidProtocolBufferException {
1568        return new SegmentStateProto(input, extensionRegistry);
1569      }
1570    };
1571
1572    @java.lang.Override
1573    public com.google.protobuf.Parser<SegmentStateProto> getParserForType() {
1574      return PARSER;
1575    }
1576
1577    private int bitField0_;
1578    // required uint64 startTxId = 1;
1579    public static final int STARTTXID_FIELD_NUMBER = 1;
1580    private long startTxId_;
1581    /**
1582     * <code>required uint64 startTxId = 1;</code>
1583     */
1584    public boolean hasStartTxId() {
1585      return ((bitField0_ & 0x00000001) == 0x00000001);
1586    }
1587    /**
1588     * <code>required uint64 startTxId = 1;</code>
1589     */
1590    public long getStartTxId() {
1591      return startTxId_;
1592    }
1593
1594    // required uint64 endTxId = 2;
1595    public static final int ENDTXID_FIELD_NUMBER = 2;
1596    private long endTxId_;
1597    /**
1598     * <code>required uint64 endTxId = 2;</code>
1599     */
1600    public boolean hasEndTxId() {
1601      return ((bitField0_ & 0x00000002) == 0x00000002);
1602    }
1603    /**
1604     * <code>required uint64 endTxId = 2;</code>
1605     */
1606    public long getEndTxId() {
1607      return endTxId_;
1608    }
1609
1610    // required bool isInProgress = 3;
1611    public static final int ISINPROGRESS_FIELD_NUMBER = 3;
1612    private boolean isInProgress_;
1613    /**
1614     * <code>required bool isInProgress = 3;</code>
1615     */
1616    public boolean hasIsInProgress() {
1617      return ((bitField0_ & 0x00000004) == 0x00000004);
1618    }
1619    /**
1620     * <code>required bool isInProgress = 3;</code>
1621     */
1622    public boolean getIsInProgress() {
1623      return isInProgress_;
1624    }
1625
1626    private void initFields() {
1627      startTxId_ = 0L;
1628      endTxId_ = 0L;
1629      isInProgress_ = false;
1630    }
1631    private byte memoizedIsInitialized = -1;
1632    public final boolean isInitialized() {
1633      byte isInitialized = memoizedIsInitialized;
1634      if (isInitialized != -1) return isInitialized == 1;
1635
1636      if (!hasStartTxId()) {
1637        memoizedIsInitialized = 0;
1638        return false;
1639      }
1640      if (!hasEndTxId()) {
1641        memoizedIsInitialized = 0;
1642        return false;
1643      }
1644      if (!hasIsInProgress()) {
1645        memoizedIsInitialized = 0;
1646        return false;
1647      }
1648      memoizedIsInitialized = 1;
1649      return true;
1650    }
1651
1652    public void writeTo(com.google.protobuf.CodedOutputStream output)
1653                        throws java.io.IOException {
1654      getSerializedSize();
1655      if (((bitField0_ & 0x00000001) == 0x00000001)) {
1656        output.writeUInt64(1, startTxId_);
1657      }
1658      if (((bitField0_ & 0x00000002) == 0x00000002)) {
1659        output.writeUInt64(2, endTxId_);
1660      }
1661      if (((bitField0_ & 0x00000004) == 0x00000004)) {
1662        output.writeBool(3, isInProgress_);
1663      }
1664      getUnknownFields().writeTo(output);
1665    }
1666
1667    private int memoizedSerializedSize = -1;
1668    public int getSerializedSize() {
1669      int size = memoizedSerializedSize;
1670      if (size != -1) return size;
1671
1672      size = 0;
1673      if (((bitField0_ & 0x00000001) == 0x00000001)) {
1674        size += com.google.protobuf.CodedOutputStream
1675          .computeUInt64Size(1, startTxId_);
1676      }
1677      if (((bitField0_ & 0x00000002) == 0x00000002)) {
1678        size += com.google.protobuf.CodedOutputStream
1679          .computeUInt64Size(2, endTxId_);
1680      }
1681      if (((bitField0_ & 0x00000004) == 0x00000004)) {
1682        size += com.google.protobuf.CodedOutputStream
1683          .computeBoolSize(3, isInProgress_);
1684      }
1685      size += getUnknownFields().getSerializedSize();
1686      memoizedSerializedSize = size;
1687      return size;
1688    }
1689
1690    private static final long serialVersionUID = 0L;
1691    @java.lang.Override
1692    protected java.lang.Object writeReplace()
1693        throws java.io.ObjectStreamException {
1694      return super.writeReplace();
1695    }
1696
1697    @java.lang.Override
1698    public boolean equals(final java.lang.Object obj) {
1699      if (obj == this) {
1700       return true;
1701      }
1702      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto)) {
1703        return super.equals(obj);
1704      }
1705      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto) obj;
1706
1707      boolean result = true;
1708      result = result && (hasStartTxId() == other.hasStartTxId());
1709      if (hasStartTxId()) {
1710        result = result && (getStartTxId()
1711            == other.getStartTxId());
1712      }
1713      result = result && (hasEndTxId() == other.hasEndTxId());
1714      if (hasEndTxId()) {
1715        result = result && (getEndTxId()
1716            == other.getEndTxId());
1717      }
1718      result = result && (hasIsInProgress() == other.hasIsInProgress());
1719      if (hasIsInProgress()) {
1720        result = result && (getIsInProgress()
1721            == other.getIsInProgress());
1722      }
1723      result = result &&
1724          getUnknownFields().equals(other.getUnknownFields());
1725      return result;
1726    }
1727
1728    private int memoizedHashCode = 0;
1729    @java.lang.Override
1730    public int hashCode() {
1731      if (memoizedHashCode != 0) {
1732        return memoizedHashCode;
1733      }
1734      int hash = 41;
1735      hash = (19 * hash) + getDescriptorForType().hashCode();
1736      if (hasStartTxId()) {
1737        hash = (37 * hash) + STARTTXID_FIELD_NUMBER;
1738        hash = (53 * hash) + hashLong(getStartTxId());
1739      }
1740      if (hasEndTxId()) {
1741        hash = (37 * hash) + ENDTXID_FIELD_NUMBER;
1742        hash = (53 * hash) + hashLong(getEndTxId());
1743      }
1744      if (hasIsInProgress()) {
1745        hash = (37 * hash) + ISINPROGRESS_FIELD_NUMBER;
1746        hash = (53 * hash) + hashBoolean(getIsInProgress());
1747      }
1748      hash = (29 * hash) + getUnknownFields().hashCode();
1749      memoizedHashCode = hash;
1750      return hash;
1751    }
1752
1753    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1754        com.google.protobuf.ByteString data)
1755        throws com.google.protobuf.InvalidProtocolBufferException {
1756      return PARSER.parseFrom(data);
1757    }
1758    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1759        com.google.protobuf.ByteString data,
1760        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1761        throws com.google.protobuf.InvalidProtocolBufferException {
1762      return PARSER.parseFrom(data, extensionRegistry);
1763    }
1764    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(byte[] data)
1765        throws com.google.protobuf.InvalidProtocolBufferException {
1766      return PARSER.parseFrom(data);
1767    }
1768    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1769        byte[] data,
1770        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1771        throws com.google.protobuf.InvalidProtocolBufferException {
1772      return PARSER.parseFrom(data, extensionRegistry);
1773    }
1774    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(java.io.InputStream input)
1775        throws java.io.IOException {
1776      return PARSER.parseFrom(input);
1777    }
1778    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1779        java.io.InputStream input,
1780        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1781        throws java.io.IOException {
1782      return PARSER.parseFrom(input, extensionRegistry);
1783    }
1784    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseDelimitedFrom(java.io.InputStream input)
1785        throws java.io.IOException {
1786      return PARSER.parseDelimitedFrom(input);
1787    }
1788    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseDelimitedFrom(
1789        java.io.InputStream input,
1790        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1791        throws java.io.IOException {
1792      return PARSER.parseDelimitedFrom(input, extensionRegistry);
1793    }
1794    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1795        com.google.protobuf.CodedInputStream input)
1796        throws java.io.IOException {
1797      return PARSER.parseFrom(input);
1798    }
1799    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1800        com.google.protobuf.CodedInputStream input,
1801        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1802        throws java.io.IOException {
1803      return PARSER.parseFrom(input, extensionRegistry);
1804    }
1805
1806    public static Builder newBuilder() { return Builder.create(); }
1807    public Builder newBuilderForType() { return newBuilder(); }
1808    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto prototype) {
1809      return newBuilder().mergeFrom(prototype);
1810    }
1811    public Builder toBuilder() { return newBuilder(this); }
1812
1813    @java.lang.Override
1814    protected Builder newBuilderForType(
1815        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1816      Builder builder = new Builder(parent);
1817      return builder;
1818    }
1819    /**
1820     * Protobuf type {@code hadoop.hdfs.qjournal.SegmentStateProto}
1821     */
1822    public static final class Builder extends
1823        com.google.protobuf.GeneratedMessage.Builder<Builder>
1824       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder {
1825      public static final com.google.protobuf.Descriptors.Descriptor
1826          getDescriptor() {
1827        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_SegmentStateProto_descriptor;
1828      }
1829
1830      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1831          internalGetFieldAccessorTable() {
1832        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_SegmentStateProto_fieldAccessorTable
1833            .ensureFieldAccessorsInitialized(
1834                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder.class);
1835      }
1836
1837      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder()
1838      private Builder() {
1839        maybeForceBuilderInitialization();
1840      }
1841
1842      private Builder(
1843          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1844        super(parent);
1845        maybeForceBuilderInitialization();
1846      }
1847      private void maybeForceBuilderInitialization() {
1848        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1849        }
1850      }
1851      private static Builder create() {
1852        return new Builder();
1853      }
1854
1855      public Builder clear() {
1856        super.clear();
1857        startTxId_ = 0L;
1858        bitField0_ = (bitField0_ & ~0x00000001);
1859        endTxId_ = 0L;
1860        bitField0_ = (bitField0_ & ~0x00000002);
1861        isInProgress_ = false;
1862        bitField0_ = (bitField0_ & ~0x00000004);
1863        return this;
1864      }
1865
1866      public Builder clone() {
1867        return create().mergeFrom(buildPartial());
1868      }
1869
1870      public com.google.protobuf.Descriptors.Descriptor
1871          getDescriptorForType() {
1872        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_SegmentStateProto_descriptor;
1873      }
1874
1875      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getDefaultInstanceForType() {
1876        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
1877      }
1878
1879      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto build() {
1880        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto result = buildPartial();
1881        if (!result.isInitialized()) {
1882          throw newUninitializedMessageException(result);
1883        }
1884        return result;
1885      }
1886
1887      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto buildPartial() {
1888        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto(this);
1889        int from_bitField0_ = bitField0_;
1890        int to_bitField0_ = 0;
1891        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
1892          to_bitField0_ |= 0x00000001;
1893        }
1894        result.startTxId_ = startTxId_;
1895        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
1896          to_bitField0_ |= 0x00000002;
1897        }
1898        result.endTxId_ = endTxId_;
1899        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
1900          to_bitField0_ |= 0x00000004;
1901        }
1902        result.isInProgress_ = isInProgress_;
1903        result.bitField0_ = to_bitField0_;
1904        onBuilt();
1905        return result;
1906      }
1907
1908      public Builder mergeFrom(com.google.protobuf.Message other) {
1909        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto) {
1910          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto)other);
1911        } else {
1912          super.mergeFrom(other);
1913          return this;
1914        }
1915      }
1916
1917      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto other) {
1918        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) return this;
1919        if (other.hasStartTxId()) {
1920          setStartTxId(other.getStartTxId());
1921        }
1922        if (other.hasEndTxId()) {
1923          setEndTxId(other.getEndTxId());
1924        }
1925        if (other.hasIsInProgress()) {
1926          setIsInProgress(other.getIsInProgress());
1927        }
1928        this.mergeUnknownFields(other.getUnknownFields());
1929        return this;
1930      }
1931
1932      public final boolean isInitialized() {
1933        if (!hasStartTxId()) {
1934          
1935          return false;
1936        }
1937        if (!hasEndTxId()) {
1938          
1939          return false;
1940        }
1941        if (!hasIsInProgress()) {
1942          
1943          return false;
1944        }
1945        return true;
1946      }
1947
1948      public Builder mergeFrom(
1949          com.google.protobuf.CodedInputStream input,
1950          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1951          throws java.io.IOException {
1952        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parsedMessage = null;
1953        try {
1954          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
1955        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1956          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto) e.getUnfinishedMessage();
1957          throw e;
1958        } finally {
1959          if (parsedMessage != null) {
1960            mergeFrom(parsedMessage);
1961          }
1962        }
1963        return this;
1964      }
1965      private int bitField0_;
1966
1967      // required uint64 startTxId = 1;
1968      private long startTxId_ ;
1969      /**
1970       * <code>required uint64 startTxId = 1;</code>
1971       */
1972      public boolean hasStartTxId() {
1973        return ((bitField0_ & 0x00000001) == 0x00000001);
1974      }
1975      /**
1976       * <code>required uint64 startTxId = 1;</code>
1977       */
1978      public long getStartTxId() {
1979        return startTxId_;
1980      }
1981      /**
1982       * <code>required uint64 startTxId = 1;</code>
1983       */
1984      public Builder setStartTxId(long value) {
1985        bitField0_ |= 0x00000001;
1986        startTxId_ = value;
1987        onChanged();
1988        return this;
1989      }
1990      /**
1991       * <code>required uint64 startTxId = 1;</code>
1992       */
1993      public Builder clearStartTxId() {
1994        bitField0_ = (bitField0_ & ~0x00000001);
1995        startTxId_ = 0L;
1996        onChanged();
1997        return this;
1998      }
1999
2000      // required uint64 endTxId = 2;
2001      private long endTxId_ ;
2002      /**
2003       * <code>required uint64 endTxId = 2;</code>
2004       */
2005      public boolean hasEndTxId() {
2006        return ((bitField0_ & 0x00000002) == 0x00000002);
2007      }
2008      /**
2009       * <code>required uint64 endTxId = 2;</code>
2010       */
2011      public long getEndTxId() {
2012        return endTxId_;
2013      }
2014      /**
2015       * <code>required uint64 endTxId = 2;</code>
2016       */
2017      public Builder setEndTxId(long value) {
2018        bitField0_ |= 0x00000002;
2019        endTxId_ = value;
2020        onChanged();
2021        return this;
2022      }
2023      /**
2024       * <code>required uint64 endTxId = 2;</code>
2025       */
2026      public Builder clearEndTxId() {
2027        bitField0_ = (bitField0_ & ~0x00000002);
2028        endTxId_ = 0L;
2029        onChanged();
2030        return this;
2031      }
2032
2033      // required bool isInProgress = 3;
2034      private boolean isInProgress_ ;
2035      /**
2036       * <code>required bool isInProgress = 3;</code>
2037       */
2038      public boolean hasIsInProgress() {
2039        return ((bitField0_ & 0x00000004) == 0x00000004);
2040      }
2041      /**
2042       * <code>required bool isInProgress = 3;</code>
2043       */
2044      public boolean getIsInProgress() {
2045        return isInProgress_;
2046      }
2047      /**
2048       * <code>required bool isInProgress = 3;</code>
2049       */
2050      public Builder setIsInProgress(boolean value) {
2051        bitField0_ |= 0x00000004;
2052        isInProgress_ = value;
2053        onChanged();
2054        return this;
2055      }
2056      /**
2057       * <code>required bool isInProgress = 3;</code>
2058       */
2059      public Builder clearIsInProgress() {
2060        bitField0_ = (bitField0_ & ~0x00000004);
2061        isInProgress_ = false;
2062        onChanged();
2063        return this;
2064      }
2065
2066      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.SegmentStateProto)
2067    }
2068
2069    static {
2070      defaultInstance = new SegmentStateProto(true);
2071      defaultInstance.initFields();
2072    }
2073
2074    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.SegmentStateProto)
2075  }
2076
2077  public interface PersistedRecoveryPaxosDataOrBuilder
2078      extends com.google.protobuf.MessageOrBuilder {
2079
2080    // required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;
2081    /**
2082     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2083     */
2084    boolean hasSegmentState();
2085    /**
2086     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2087     */
2088    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState();
2089    /**
2090     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2091     */
2092    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder();
2093
2094    // required uint64 acceptedInEpoch = 2;
2095    /**
2096     * <code>required uint64 acceptedInEpoch = 2;</code>
2097     */
2098    boolean hasAcceptedInEpoch();
2099    /**
2100     * <code>required uint64 acceptedInEpoch = 2;</code>
2101     */
2102    long getAcceptedInEpoch();
2103  }
2104  /**
2105   * Protobuf type {@code hadoop.hdfs.qjournal.PersistedRecoveryPaxosData}
2106   *
2107   * <pre>
2108   **
2109   * The storage format used on local disk for previously
2110   * accepted decisions.
2111   * </pre>
2112   */
2113  public static final class PersistedRecoveryPaxosData extends
2114      com.google.protobuf.GeneratedMessage
2115      implements PersistedRecoveryPaxosDataOrBuilder {
2116    // Use PersistedRecoveryPaxosData.newBuilder() to construct.
2117    private PersistedRecoveryPaxosData(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
2118      super(builder);
2119      this.unknownFields = builder.getUnknownFields();
2120    }
2121    private PersistedRecoveryPaxosData(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
2122
2123    private static final PersistedRecoveryPaxosData defaultInstance;
2124    public static PersistedRecoveryPaxosData getDefaultInstance() {
2125      return defaultInstance;
2126    }
2127
2128    public PersistedRecoveryPaxosData getDefaultInstanceForType() {
2129      return defaultInstance;
2130    }
2131
2132    private final com.google.protobuf.UnknownFieldSet unknownFields;
2133    @java.lang.Override
2134    public final com.google.protobuf.UnknownFieldSet
2135        getUnknownFields() {
2136      return this.unknownFields;
2137    }
2138    private PersistedRecoveryPaxosData(
2139        com.google.protobuf.CodedInputStream input,
2140        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2141        throws com.google.protobuf.InvalidProtocolBufferException {
2142      initFields();
2143      int mutable_bitField0_ = 0;
2144      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2145          com.google.protobuf.UnknownFieldSet.newBuilder();
2146      try {
2147        boolean done = false;
2148        while (!done) {
2149          int tag = input.readTag();
2150          switch (tag) {
2151            case 0:
2152              done = true;
2153              break;
2154            default: {
2155              if (!parseUnknownField(input, unknownFields,
2156                                     extensionRegistry, tag)) {
2157                done = true;
2158              }
2159              break;
2160            }
2161            case 10: {
2162              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = null;
2163              if (((bitField0_ & 0x00000001) == 0x00000001)) {
2164                subBuilder = segmentState_.toBuilder();
2165              }
2166              segmentState_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.PARSER, extensionRegistry);
2167              if (subBuilder != null) {
2168                subBuilder.mergeFrom(segmentState_);
2169                segmentState_ = subBuilder.buildPartial();
2170              }
2171              bitField0_ |= 0x00000001;
2172              break;
2173            }
2174            case 16: {
2175              bitField0_ |= 0x00000002;
2176              acceptedInEpoch_ = input.readUInt64();
2177              break;
2178            }
2179          }
2180        }
2181      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
2182        throw e.setUnfinishedMessage(this);
2183      } catch (java.io.IOException e) {
2184        throw new com.google.protobuf.InvalidProtocolBufferException(
2185            e.getMessage()).setUnfinishedMessage(this);
2186      } finally {
2187        this.unknownFields = unknownFields.build();
2188        makeExtensionsImmutable();
2189      }
2190    }
2191    public static final com.google.protobuf.Descriptors.Descriptor
2192        getDescriptor() {
2193      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PersistedRecoveryPaxosData_descriptor;
2194    }
2195
2196    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2197        internalGetFieldAccessorTable() {
2198      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PersistedRecoveryPaxosData_fieldAccessorTable
2199          .ensureFieldAccessorsInitialized(
2200              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.Builder.class);
2201    }
2202
2203    public static com.google.protobuf.Parser<PersistedRecoveryPaxosData> PARSER =
2204        new com.google.protobuf.AbstractParser<PersistedRecoveryPaxosData>() {
2205      public PersistedRecoveryPaxosData parsePartialFrom(
2206          com.google.protobuf.CodedInputStream input,
2207          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2208          throws com.google.protobuf.InvalidProtocolBufferException {
2209        return new PersistedRecoveryPaxosData(input, extensionRegistry);
2210      }
2211    };
2212
2213    @java.lang.Override
2214    public com.google.protobuf.Parser<PersistedRecoveryPaxosData> getParserForType() {
2215      return PARSER;
2216    }
2217
2218    private int bitField0_;
2219    // required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;
2220    public static final int SEGMENTSTATE_FIELD_NUMBER = 1;
2221    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_;
2222    /**
2223     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2224     */
2225    public boolean hasSegmentState() {
2226      return ((bitField0_ & 0x00000001) == 0x00000001);
2227    }
2228    /**
2229     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2230     */
2231    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
2232      return segmentState_;
2233    }
2234    /**
2235     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2236     */
2237    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
2238      return segmentState_;
2239    }
2240
2241    // required uint64 acceptedInEpoch = 2;
2242    public static final int ACCEPTEDINEPOCH_FIELD_NUMBER = 2;
2243    private long acceptedInEpoch_;
2244    /**
2245     * <code>required uint64 acceptedInEpoch = 2;</code>
2246     */
2247    public boolean hasAcceptedInEpoch() {
2248      return ((bitField0_ & 0x00000002) == 0x00000002);
2249    }
2250    /**
2251     * <code>required uint64 acceptedInEpoch = 2;</code>
2252     */
2253    public long getAcceptedInEpoch() {
2254      return acceptedInEpoch_;
2255    }
2256
2257    private void initFields() {
2258      segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
2259      acceptedInEpoch_ = 0L;
2260    }
2261    private byte memoizedIsInitialized = -1;
2262    public final boolean isInitialized() {
2263      byte isInitialized = memoizedIsInitialized;
2264      if (isInitialized != -1) return isInitialized == 1;
2265
2266      if (!hasSegmentState()) {
2267        memoizedIsInitialized = 0;
2268        return false;
2269      }
2270      if (!hasAcceptedInEpoch()) {
2271        memoizedIsInitialized = 0;
2272        return false;
2273      }
2274      if (!getSegmentState().isInitialized()) {
2275        memoizedIsInitialized = 0;
2276        return false;
2277      }
2278      memoizedIsInitialized = 1;
2279      return true;
2280    }
2281
2282    public void writeTo(com.google.protobuf.CodedOutputStream output)
2283                        throws java.io.IOException {
2284      getSerializedSize();
2285      if (((bitField0_ & 0x00000001) == 0x00000001)) {
2286        output.writeMessage(1, segmentState_);
2287      }
2288      if (((bitField0_ & 0x00000002) == 0x00000002)) {
2289        output.writeUInt64(2, acceptedInEpoch_);
2290      }
2291      getUnknownFields().writeTo(output);
2292    }
2293
2294    private int memoizedSerializedSize = -1;
2295    public int getSerializedSize() {
2296      int size = memoizedSerializedSize;
2297      if (size != -1) return size;
2298
2299      size = 0;
2300      if (((bitField0_ & 0x00000001) == 0x00000001)) {
2301        size += com.google.protobuf.CodedOutputStream
2302          .computeMessageSize(1, segmentState_);
2303      }
2304      if (((bitField0_ & 0x00000002) == 0x00000002)) {
2305        size += com.google.protobuf.CodedOutputStream
2306          .computeUInt64Size(2, acceptedInEpoch_);
2307      }
2308      size += getUnknownFields().getSerializedSize();
2309      memoizedSerializedSize = size;
2310      return size;
2311    }
2312
2313    private static final long serialVersionUID = 0L;
2314    @java.lang.Override
2315    protected java.lang.Object writeReplace()
2316        throws java.io.ObjectStreamException {
2317      return super.writeReplace();
2318    }
2319
2320    @java.lang.Override
2321    public boolean equals(final java.lang.Object obj) {
2322      if (obj == this) {
2323       return true;
2324      }
2325      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData)) {
2326        return super.equals(obj);
2327      }
2328      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData) obj;
2329
2330      boolean result = true;
2331      result = result && (hasSegmentState() == other.hasSegmentState());
2332      if (hasSegmentState()) {
2333        result = result && getSegmentState()
2334            .equals(other.getSegmentState());
2335      }
2336      result = result && (hasAcceptedInEpoch() == other.hasAcceptedInEpoch());
2337      if (hasAcceptedInEpoch()) {
2338        result = result && (getAcceptedInEpoch()
2339            == other.getAcceptedInEpoch());
2340      }
2341      result = result &&
2342          getUnknownFields().equals(other.getUnknownFields());
2343      return result;
2344    }
2345
2346    private int memoizedHashCode = 0;
2347    @java.lang.Override
2348    public int hashCode() {
2349      if (memoizedHashCode != 0) {
2350        return memoizedHashCode;
2351      }
2352      int hash = 41;
2353      hash = (19 * hash) + getDescriptorForType().hashCode();
2354      if (hasSegmentState()) {
2355        hash = (37 * hash) + SEGMENTSTATE_FIELD_NUMBER;
2356        hash = (53 * hash) + getSegmentState().hashCode();
2357      }
2358      if (hasAcceptedInEpoch()) {
2359        hash = (37 * hash) + ACCEPTEDINEPOCH_FIELD_NUMBER;
2360        hash = (53 * hash) + hashLong(getAcceptedInEpoch());
2361      }
2362      hash = (29 * hash) + getUnknownFields().hashCode();
2363      memoizedHashCode = hash;
2364      return hash;
2365    }
2366
2367    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2368        com.google.protobuf.ByteString data)
2369        throws com.google.protobuf.InvalidProtocolBufferException {
2370      return PARSER.parseFrom(data);
2371    }
2372    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2373        com.google.protobuf.ByteString data,
2374        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2375        throws com.google.protobuf.InvalidProtocolBufferException {
2376      return PARSER.parseFrom(data, extensionRegistry);
2377    }
2378    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(byte[] data)
2379        throws com.google.protobuf.InvalidProtocolBufferException {
2380      return PARSER.parseFrom(data);
2381    }
2382    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2383        byte[] data,
2384        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2385        throws com.google.protobuf.InvalidProtocolBufferException {
2386      return PARSER.parseFrom(data, extensionRegistry);
2387    }
2388    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(java.io.InputStream input)
2389        throws java.io.IOException {
2390      return PARSER.parseFrom(input);
2391    }
2392    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2393        java.io.InputStream input,
2394        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2395        throws java.io.IOException {
2396      return PARSER.parseFrom(input, extensionRegistry);
2397    }
2398    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseDelimitedFrom(java.io.InputStream input)
2399        throws java.io.IOException {
2400      return PARSER.parseDelimitedFrom(input);
2401    }
2402    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseDelimitedFrom(
2403        java.io.InputStream input,
2404        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2405        throws java.io.IOException {
2406      return PARSER.parseDelimitedFrom(input, extensionRegistry);
2407    }
2408    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2409        com.google.protobuf.CodedInputStream input)
2410        throws java.io.IOException {
2411      return PARSER.parseFrom(input);
2412    }
2413    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2414        com.google.protobuf.CodedInputStream input,
2415        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2416        throws java.io.IOException {
2417      return PARSER.parseFrom(input, extensionRegistry);
2418    }
2419
2420    public static Builder newBuilder() { return Builder.create(); }
2421    public Builder newBuilderForType() { return newBuilder(); }
2422    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData prototype) {
2423      return newBuilder().mergeFrom(prototype);
2424    }
2425    public Builder toBuilder() { return newBuilder(this); }
2426
2427    @java.lang.Override
2428    protected Builder newBuilderForType(
2429        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2430      Builder builder = new Builder(parent);
2431      return builder;
2432    }
2433    /**
2434     * Protobuf type {@code hadoop.hdfs.qjournal.PersistedRecoveryPaxosData}
2435     *
2436     * <pre>
2437     **
2438     * The storage format used on local disk for previously
2439     * accepted decisions.
2440     * </pre>
2441     */
2442    public static final class Builder extends
2443        com.google.protobuf.GeneratedMessage.Builder<Builder>
2444       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosDataOrBuilder {
2445      public static final com.google.protobuf.Descriptors.Descriptor
2446          getDescriptor() {
2447        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PersistedRecoveryPaxosData_descriptor;
2448      }
2449
2450      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2451          internalGetFieldAccessorTable() {
2452        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PersistedRecoveryPaxosData_fieldAccessorTable
2453            .ensureFieldAccessorsInitialized(
2454                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.Builder.class);
2455      }
2456
2457      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.newBuilder()
2458      private Builder() {
2459        maybeForceBuilderInitialization();
2460      }
2461
2462      private Builder(
2463          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2464        super(parent);
2465        maybeForceBuilderInitialization();
2466      }
2467      private void maybeForceBuilderInitialization() {
2468        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
2469          getSegmentStateFieldBuilder();
2470        }
2471      }
2472      private static Builder create() {
2473        return new Builder();
2474      }
2475
2476      public Builder clear() {
2477        super.clear();
2478        if (segmentStateBuilder_ == null) {
2479          segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
2480        } else {
2481          segmentStateBuilder_.clear();
2482        }
2483        bitField0_ = (bitField0_ & ~0x00000001);
2484        acceptedInEpoch_ = 0L;
2485        bitField0_ = (bitField0_ & ~0x00000002);
2486        return this;
2487      }
2488
2489      public Builder clone() {
2490        return create().mergeFrom(buildPartial());
2491      }
2492
2493      public com.google.protobuf.Descriptors.Descriptor
2494          getDescriptorForType() {
2495        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PersistedRecoveryPaxosData_descriptor;
2496      }
2497
2498      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData getDefaultInstanceForType() {
2499        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.getDefaultInstance();
2500      }
2501
2502      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData build() {
2503        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData result = buildPartial();
2504        if (!result.isInitialized()) {
2505          throw newUninitializedMessageException(result);
2506        }
2507        return result;
2508      }
2509
2510      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData buildPartial() {
2511        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData(this);
2512        int from_bitField0_ = bitField0_;
2513        int to_bitField0_ = 0;
2514        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
2515          to_bitField0_ |= 0x00000001;
2516        }
2517        if (segmentStateBuilder_ == null) {
2518          result.segmentState_ = segmentState_;
2519        } else {
2520          result.segmentState_ = segmentStateBuilder_.build();
2521        }
2522        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
2523          to_bitField0_ |= 0x00000002;
2524        }
2525        result.acceptedInEpoch_ = acceptedInEpoch_;
2526        result.bitField0_ = to_bitField0_;
2527        onBuilt();
2528        return result;
2529      }
2530
2531      public Builder mergeFrom(com.google.protobuf.Message other) {
2532        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData) {
2533          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData)other);
2534        } else {
2535          super.mergeFrom(other);
2536          return this;
2537        }
2538      }
2539
2540      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData other) {
2541        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.getDefaultInstance()) return this;
2542        if (other.hasSegmentState()) {
2543          mergeSegmentState(other.getSegmentState());
2544        }
2545        if (other.hasAcceptedInEpoch()) {
2546          setAcceptedInEpoch(other.getAcceptedInEpoch());
2547        }
2548        this.mergeUnknownFields(other.getUnknownFields());
2549        return this;
2550      }
2551
2552      public final boolean isInitialized() {
2553        if (!hasSegmentState()) {
2554          
2555          return false;
2556        }
2557        if (!hasAcceptedInEpoch()) {
2558          
2559          return false;
2560        }
2561        if (!getSegmentState().isInitialized()) {
2562          
2563          return false;
2564        }
2565        return true;
2566      }
2567
2568      public Builder mergeFrom(
2569          com.google.protobuf.CodedInputStream input,
2570          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2571          throws java.io.IOException {
2572        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parsedMessage = null;
2573        try {
2574          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
2575        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
2576          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData) e.getUnfinishedMessage();
2577          throw e;
2578        } finally {
2579          if (parsedMessage != null) {
2580            mergeFrom(parsedMessage);
2581          }
2582        }
2583        return this;
2584      }
2585      private int bitField0_;
2586
2587      // required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;
2588      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
2589      private com.google.protobuf.SingleFieldBuilder<
2590          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> segmentStateBuilder_;
2591      /**
2592       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2593       */
2594      public boolean hasSegmentState() {
2595        return ((bitField0_ & 0x00000001) == 0x00000001);
2596      }
2597      /**
2598       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2599       */
2600      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
2601        if (segmentStateBuilder_ == null) {
2602          return segmentState_;
2603        } else {
2604          return segmentStateBuilder_.getMessage();
2605        }
2606      }
2607      /**
2608       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2609       */
2610      public Builder setSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
2611        if (segmentStateBuilder_ == null) {
2612          if (value == null) {
2613            throw new NullPointerException();
2614          }
2615          segmentState_ = value;
2616          onChanged();
2617        } else {
2618          segmentStateBuilder_.setMessage(value);
2619        }
2620        bitField0_ |= 0x00000001;
2621        return this;
2622      }
2623      /**
2624       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2625       */
2626      public Builder setSegmentState(
2627          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) {
2628        if (segmentStateBuilder_ == null) {
2629          segmentState_ = builderForValue.build();
2630          onChanged();
2631        } else {
2632          segmentStateBuilder_.setMessage(builderForValue.build());
2633        }
2634        bitField0_ |= 0x00000001;
2635        return this;
2636      }
2637      /**
2638       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2639       */
2640      public Builder mergeSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
2641        if (segmentStateBuilder_ == null) {
2642          if (((bitField0_ & 0x00000001) == 0x00000001) &&
2643              segmentState_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) {
2644            segmentState_ =
2645              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(segmentState_).mergeFrom(value).buildPartial();
2646          } else {
2647            segmentState_ = value;
2648          }
2649          onChanged();
2650        } else {
2651          segmentStateBuilder_.mergeFrom(value);
2652        }
2653        bitField0_ |= 0x00000001;
2654        return this;
2655      }
2656      /**
2657       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2658       */
2659      public Builder clearSegmentState() {
2660        if (segmentStateBuilder_ == null) {
2661          segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
2662          onChanged();
2663        } else {
2664          segmentStateBuilder_.clear();
2665        }
2666        bitField0_ = (bitField0_ & ~0x00000001);
2667        return this;
2668      }
2669      /**
2670       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2671       */
2672      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getSegmentStateBuilder() {
2673        bitField0_ |= 0x00000001;
2674        onChanged();
2675        return getSegmentStateFieldBuilder().getBuilder();
2676      }
2677      /**
2678       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2679       */
2680      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
2681        if (segmentStateBuilder_ != null) {
2682          return segmentStateBuilder_.getMessageOrBuilder();
2683        } else {
2684          return segmentState_;
2685        }
2686      }
2687      /**
2688       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
2689       */
2690      private com.google.protobuf.SingleFieldBuilder<
2691          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 
2692          getSegmentStateFieldBuilder() {
2693        if (segmentStateBuilder_ == null) {
2694          segmentStateBuilder_ = new com.google.protobuf.SingleFieldBuilder<
2695              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>(
2696                  segmentState_,
2697                  getParentForChildren(),
2698                  isClean());
2699          segmentState_ = null;
2700        }
2701        return segmentStateBuilder_;
2702      }
2703
2704      // required uint64 acceptedInEpoch = 2;
2705      private long acceptedInEpoch_ ;
2706      /**
2707       * <code>required uint64 acceptedInEpoch = 2;</code>
2708       */
2709      public boolean hasAcceptedInEpoch() {
2710        return ((bitField0_ & 0x00000002) == 0x00000002);
2711      }
2712      /**
2713       * <code>required uint64 acceptedInEpoch = 2;</code>
2714       */
2715      public long getAcceptedInEpoch() {
2716        return acceptedInEpoch_;
2717      }
2718      /**
2719       * <code>required uint64 acceptedInEpoch = 2;</code>
2720       */
2721      public Builder setAcceptedInEpoch(long value) {
2722        bitField0_ |= 0x00000002;
2723        acceptedInEpoch_ = value;
2724        onChanged();
2725        return this;
2726      }
2727      /**
2728       * <code>required uint64 acceptedInEpoch = 2;</code>
2729       */
2730      public Builder clearAcceptedInEpoch() {
2731        bitField0_ = (bitField0_ & ~0x00000002);
2732        acceptedInEpoch_ = 0L;
2733        onChanged();
2734        return this;
2735      }
2736
2737      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.PersistedRecoveryPaxosData)
2738    }
2739
2740    static {
2741      defaultInstance = new PersistedRecoveryPaxosData(true);
2742      defaultInstance.initFields();
2743    }
2744
2745    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.PersistedRecoveryPaxosData)
2746  }
2747
2748  public interface JournalRequestProtoOrBuilder
2749      extends com.google.protobuf.MessageOrBuilder {
2750
2751    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
2752    /**
2753     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
2754     */
2755    boolean hasReqInfo();
2756    /**
2757     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
2758     */
2759    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
2760    /**
2761     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
2762     */
2763    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
2764
2765    // required uint64 firstTxnId = 2;
2766    /**
2767     * <code>required uint64 firstTxnId = 2;</code>
2768     */
2769    boolean hasFirstTxnId();
2770    /**
2771     * <code>required uint64 firstTxnId = 2;</code>
2772     */
2773    long getFirstTxnId();
2774
2775    // required uint32 numTxns = 3;
2776    /**
2777     * <code>required uint32 numTxns = 3;</code>
2778     */
2779    boolean hasNumTxns();
2780    /**
2781     * <code>required uint32 numTxns = 3;</code>
2782     */
2783    int getNumTxns();
2784
2785    // required bytes records = 4;
2786    /**
2787     * <code>required bytes records = 4;</code>
2788     */
2789    boolean hasRecords();
2790    /**
2791     * <code>required bytes records = 4;</code>
2792     */
2793    com.google.protobuf.ByteString getRecords();
2794
2795    // required uint64 segmentTxnId = 5;
2796    /**
2797     * <code>required uint64 segmentTxnId = 5;</code>
2798     */
2799    boolean hasSegmentTxnId();
2800    /**
2801     * <code>required uint64 segmentTxnId = 5;</code>
2802     */
2803    long getSegmentTxnId();
2804  }
2805  /**
2806   * Protobuf type {@code hadoop.hdfs.qjournal.JournalRequestProto}
2807   */
2808  public static final class JournalRequestProto extends
2809      com.google.protobuf.GeneratedMessage
2810      implements JournalRequestProtoOrBuilder {
2811    // Use JournalRequestProto.newBuilder() to construct.
2812    private JournalRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
2813      super(builder);
2814      this.unknownFields = builder.getUnknownFields();
2815    }
2816    private JournalRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
2817
2818    private static final JournalRequestProto defaultInstance;
2819    public static JournalRequestProto getDefaultInstance() {
2820      return defaultInstance;
2821    }
2822
2823    public JournalRequestProto getDefaultInstanceForType() {
2824      return defaultInstance;
2825    }
2826
2827    private final com.google.protobuf.UnknownFieldSet unknownFields;
2828    @java.lang.Override
2829    public final com.google.protobuf.UnknownFieldSet
2830        getUnknownFields() {
2831      return this.unknownFields;
2832    }
2833    private JournalRequestProto(
2834        com.google.protobuf.CodedInputStream input,
2835        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2836        throws com.google.protobuf.InvalidProtocolBufferException {
2837      initFields();
2838      int mutable_bitField0_ = 0;
2839      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2840          com.google.protobuf.UnknownFieldSet.newBuilder();
2841      try {
2842        boolean done = false;
2843        while (!done) {
2844          int tag = input.readTag();
2845          switch (tag) {
2846            case 0:
2847              done = true;
2848              break;
2849            default: {
2850              if (!parseUnknownField(input, unknownFields,
2851                                     extensionRegistry, tag)) {
2852                done = true;
2853              }
2854              break;
2855            }
2856            case 10: {
2857              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
2858              if (((bitField0_ & 0x00000001) == 0x00000001)) {
2859                subBuilder = reqInfo_.toBuilder();
2860              }
2861              reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
2862              if (subBuilder != null) {
2863                subBuilder.mergeFrom(reqInfo_);
2864                reqInfo_ = subBuilder.buildPartial();
2865              }
2866              bitField0_ |= 0x00000001;
2867              break;
2868            }
2869            case 16: {
2870              bitField0_ |= 0x00000002;
2871              firstTxnId_ = input.readUInt64();
2872              break;
2873            }
2874            case 24: {
2875              bitField0_ |= 0x00000004;
2876              numTxns_ = input.readUInt32();
2877              break;
2878            }
2879            case 34: {
2880              bitField0_ |= 0x00000008;
2881              records_ = input.readBytes();
2882              break;
2883            }
2884            case 40: {
2885              bitField0_ |= 0x00000010;
2886              segmentTxnId_ = input.readUInt64();
2887              break;
2888            }
2889          }
2890        }
2891      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
2892        throw e.setUnfinishedMessage(this);
2893      } catch (java.io.IOException e) {
2894        throw new com.google.protobuf.InvalidProtocolBufferException(
2895            e.getMessage()).setUnfinishedMessage(this);
2896      } finally {
2897        this.unknownFields = unknownFields.build();
2898        makeExtensionsImmutable();
2899      }
2900    }
2901    public static final com.google.protobuf.Descriptors.Descriptor
2902        getDescriptor() {
2903      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalRequestProto_descriptor;
2904    }
2905
2906    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2907        internalGetFieldAccessorTable() {
2908      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalRequestProto_fieldAccessorTable
2909          .ensureFieldAccessorsInitialized(
2910              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.Builder.class);
2911    }
2912
2913    public static com.google.protobuf.Parser<JournalRequestProto> PARSER =
2914        new com.google.protobuf.AbstractParser<JournalRequestProto>() {
2915      public JournalRequestProto parsePartialFrom(
2916          com.google.protobuf.CodedInputStream input,
2917          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2918          throws com.google.protobuf.InvalidProtocolBufferException {
2919        return new JournalRequestProto(input, extensionRegistry);
2920      }
2921    };
2922
2923    @java.lang.Override
2924    public com.google.protobuf.Parser<JournalRequestProto> getParserForType() {
2925      return PARSER;
2926    }
2927
2928    private int bitField0_;
2929    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
2930    public static final int REQINFO_FIELD_NUMBER = 1;
2931    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
2932    /**
2933     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
2934     */
2935    public boolean hasReqInfo() {
2936      return ((bitField0_ & 0x00000001) == 0x00000001);
2937    }
2938    /**
2939     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
2940     */
2941    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
2942      return reqInfo_;
2943    }
2944    /**
2945     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
2946     */
2947    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
2948      return reqInfo_;
2949    }
2950
2951    // required uint64 firstTxnId = 2;
2952    public static final int FIRSTTXNID_FIELD_NUMBER = 2;
2953    private long firstTxnId_;
2954    /**
2955     * <code>required uint64 firstTxnId = 2;</code>
2956     */
2957    public boolean hasFirstTxnId() {
2958      return ((bitField0_ & 0x00000002) == 0x00000002);
2959    }
2960    /**
2961     * <code>required uint64 firstTxnId = 2;</code>
2962     */
2963    public long getFirstTxnId() {
2964      return firstTxnId_;
2965    }
2966
2967    // required uint32 numTxns = 3;
2968    public static final int NUMTXNS_FIELD_NUMBER = 3;
2969    private int numTxns_;
2970    /**
2971     * <code>required uint32 numTxns = 3;</code>
2972     */
2973    public boolean hasNumTxns() {
2974      return ((bitField0_ & 0x00000004) == 0x00000004);
2975    }
2976    /**
2977     * <code>required uint32 numTxns = 3;</code>
2978     */
2979    public int getNumTxns() {
2980      return numTxns_;
2981    }
2982
2983    // required bytes records = 4;
2984    public static final int RECORDS_FIELD_NUMBER = 4;
2985    private com.google.protobuf.ByteString records_;
2986    /**
2987     * <code>required bytes records = 4;</code>
2988     */
2989    public boolean hasRecords() {
2990      return ((bitField0_ & 0x00000008) == 0x00000008);
2991    }
2992    /**
2993     * <code>required bytes records = 4;</code>
2994     */
2995    public com.google.protobuf.ByteString getRecords() {
2996      return records_;
2997    }
2998
2999    // required uint64 segmentTxnId = 5;
3000    public static final int SEGMENTTXNID_FIELD_NUMBER = 5;
3001    private long segmentTxnId_;
3002    /**
3003     * <code>required uint64 segmentTxnId = 5;</code>
3004     */
3005    public boolean hasSegmentTxnId() {
3006      return ((bitField0_ & 0x00000010) == 0x00000010);
3007    }
3008    /**
3009     * <code>required uint64 segmentTxnId = 5;</code>
3010     */
3011    public long getSegmentTxnId() {
3012      return segmentTxnId_;
3013    }
3014
3015    private void initFields() {
3016      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3017      firstTxnId_ = 0L;
3018      numTxns_ = 0;
3019      records_ = com.google.protobuf.ByteString.EMPTY;
3020      segmentTxnId_ = 0L;
3021    }
3022    private byte memoizedIsInitialized = -1;
3023    public final boolean isInitialized() {
3024      byte isInitialized = memoizedIsInitialized;
3025      if (isInitialized != -1) return isInitialized == 1;
3026
3027      if (!hasReqInfo()) {
3028        memoizedIsInitialized = 0;
3029        return false;
3030      }
3031      if (!hasFirstTxnId()) {
3032        memoizedIsInitialized = 0;
3033        return false;
3034      }
3035      if (!hasNumTxns()) {
3036        memoizedIsInitialized = 0;
3037        return false;
3038      }
3039      if (!hasRecords()) {
3040        memoizedIsInitialized = 0;
3041        return false;
3042      }
3043      if (!hasSegmentTxnId()) {
3044        memoizedIsInitialized = 0;
3045        return false;
3046      }
3047      if (!getReqInfo().isInitialized()) {
3048        memoizedIsInitialized = 0;
3049        return false;
3050      }
3051      memoizedIsInitialized = 1;
3052      return true;
3053    }
3054
3055    public void writeTo(com.google.protobuf.CodedOutputStream output)
3056                        throws java.io.IOException {
3057      getSerializedSize();
3058      if (((bitField0_ & 0x00000001) == 0x00000001)) {
3059        output.writeMessage(1, reqInfo_);
3060      }
3061      if (((bitField0_ & 0x00000002) == 0x00000002)) {
3062        output.writeUInt64(2, firstTxnId_);
3063      }
3064      if (((bitField0_ & 0x00000004) == 0x00000004)) {
3065        output.writeUInt32(3, numTxns_);
3066      }
3067      if (((bitField0_ & 0x00000008) == 0x00000008)) {
3068        output.writeBytes(4, records_);
3069      }
3070      if (((bitField0_ & 0x00000010) == 0x00000010)) {
3071        output.writeUInt64(5, segmentTxnId_);
3072      }
3073      getUnknownFields().writeTo(output);
3074    }
3075
3076    private int memoizedSerializedSize = -1;
3077    public int getSerializedSize() {
3078      int size = memoizedSerializedSize;
3079      if (size != -1) return size;
3080
3081      size = 0;
3082      if (((bitField0_ & 0x00000001) == 0x00000001)) {
3083        size += com.google.protobuf.CodedOutputStream
3084          .computeMessageSize(1, reqInfo_);
3085      }
3086      if (((bitField0_ & 0x00000002) == 0x00000002)) {
3087        size += com.google.protobuf.CodedOutputStream
3088          .computeUInt64Size(2, firstTxnId_);
3089      }
3090      if (((bitField0_ & 0x00000004) == 0x00000004)) {
3091        size += com.google.protobuf.CodedOutputStream
3092          .computeUInt32Size(3, numTxns_);
3093      }
3094      if (((bitField0_ & 0x00000008) == 0x00000008)) {
3095        size += com.google.protobuf.CodedOutputStream
3096          .computeBytesSize(4, records_);
3097      }
3098      if (((bitField0_ & 0x00000010) == 0x00000010)) {
3099        size += com.google.protobuf.CodedOutputStream
3100          .computeUInt64Size(5, segmentTxnId_);
3101      }
3102      size += getUnknownFields().getSerializedSize();
3103      memoizedSerializedSize = size;
3104      return size;
3105    }
3106
3107    private static final long serialVersionUID = 0L;
3108    @java.lang.Override
3109    protected java.lang.Object writeReplace()
3110        throws java.io.ObjectStreamException {
3111      return super.writeReplace();
3112    }
3113
3114    @java.lang.Override
3115    public boolean equals(final java.lang.Object obj) {
3116      if (obj == this) {
3117       return true;
3118      }
3119      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)) {
3120        return super.equals(obj);
3121      }
3122      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto) obj;
3123
3124      boolean result = true;
3125      result = result && (hasReqInfo() == other.hasReqInfo());
3126      if (hasReqInfo()) {
3127        result = result && getReqInfo()
3128            .equals(other.getReqInfo());
3129      }
3130      result = result && (hasFirstTxnId() == other.hasFirstTxnId());
3131      if (hasFirstTxnId()) {
3132        result = result && (getFirstTxnId()
3133            == other.getFirstTxnId());
3134      }
3135      result = result && (hasNumTxns() == other.hasNumTxns());
3136      if (hasNumTxns()) {
3137        result = result && (getNumTxns()
3138            == other.getNumTxns());
3139      }
3140      result = result && (hasRecords() == other.hasRecords());
3141      if (hasRecords()) {
3142        result = result && getRecords()
3143            .equals(other.getRecords());
3144      }
3145      result = result && (hasSegmentTxnId() == other.hasSegmentTxnId());
3146      if (hasSegmentTxnId()) {
3147        result = result && (getSegmentTxnId()
3148            == other.getSegmentTxnId());
3149      }
3150      result = result &&
3151          getUnknownFields().equals(other.getUnknownFields());
3152      return result;
3153    }
3154
3155    private int memoizedHashCode = 0;
3156    @java.lang.Override
3157    public int hashCode() {
3158      if (memoizedHashCode != 0) {
3159        return memoizedHashCode;
3160      }
3161      int hash = 41;
3162      hash = (19 * hash) + getDescriptorForType().hashCode();
3163      if (hasReqInfo()) {
3164        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
3165        hash = (53 * hash) + getReqInfo().hashCode();
3166      }
3167      if (hasFirstTxnId()) {
3168        hash = (37 * hash) + FIRSTTXNID_FIELD_NUMBER;
3169        hash = (53 * hash) + hashLong(getFirstTxnId());
3170      }
3171      if (hasNumTxns()) {
3172        hash = (37 * hash) + NUMTXNS_FIELD_NUMBER;
3173        hash = (53 * hash) + getNumTxns();
3174      }
3175      if (hasRecords()) {
3176        hash = (37 * hash) + RECORDS_FIELD_NUMBER;
3177        hash = (53 * hash) + getRecords().hashCode();
3178      }
3179      if (hasSegmentTxnId()) {
3180        hash = (37 * hash) + SEGMENTTXNID_FIELD_NUMBER;
3181        hash = (53 * hash) + hashLong(getSegmentTxnId());
3182      }
3183      hash = (29 * hash) + getUnknownFields().hashCode();
3184      memoizedHashCode = hash;
3185      return hash;
3186    }
3187
3188    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3189        com.google.protobuf.ByteString data)
3190        throws com.google.protobuf.InvalidProtocolBufferException {
3191      return PARSER.parseFrom(data);
3192    }
3193    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3194        com.google.protobuf.ByteString data,
3195        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3196        throws com.google.protobuf.InvalidProtocolBufferException {
3197      return PARSER.parseFrom(data, extensionRegistry);
3198    }
3199    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(byte[] data)
3200        throws com.google.protobuf.InvalidProtocolBufferException {
3201      return PARSER.parseFrom(data);
3202    }
3203    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3204        byte[] data,
3205        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3206        throws com.google.protobuf.InvalidProtocolBufferException {
3207      return PARSER.parseFrom(data, extensionRegistry);
3208    }
3209    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(java.io.InputStream input)
3210        throws java.io.IOException {
3211      return PARSER.parseFrom(input);
3212    }
3213    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3214        java.io.InputStream input,
3215        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3216        throws java.io.IOException {
3217      return PARSER.parseFrom(input, extensionRegistry);
3218    }
3219    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseDelimitedFrom(java.io.InputStream input)
3220        throws java.io.IOException {
3221      return PARSER.parseDelimitedFrom(input);
3222    }
3223    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseDelimitedFrom(
3224        java.io.InputStream input,
3225        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3226        throws java.io.IOException {
3227      return PARSER.parseDelimitedFrom(input, extensionRegistry);
3228    }
3229    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3230        com.google.protobuf.CodedInputStream input)
3231        throws java.io.IOException {
3232      return PARSER.parseFrom(input);
3233    }
3234    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3235        com.google.protobuf.CodedInputStream input,
3236        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3237        throws java.io.IOException {
3238      return PARSER.parseFrom(input, extensionRegistry);
3239    }
3240
3241    public static Builder newBuilder() { return Builder.create(); }
3242    public Builder newBuilderForType() { return newBuilder(); }
3243    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto prototype) {
3244      return newBuilder().mergeFrom(prototype);
3245    }
3246    public Builder toBuilder() { return newBuilder(this); }
3247
3248    @java.lang.Override
3249    protected Builder newBuilderForType(
3250        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3251      Builder builder = new Builder(parent);
3252      return builder;
3253    }
3254    /**
3255     * Protobuf type {@code hadoop.hdfs.qjournal.JournalRequestProto}
3256     */
3257    public static final class Builder extends
3258        com.google.protobuf.GeneratedMessage.Builder<Builder>
3259       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProtoOrBuilder {
3260      public static final com.google.protobuf.Descriptors.Descriptor
3261          getDescriptor() {
3262        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalRequestProto_descriptor;
3263      }
3264
3265      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3266          internalGetFieldAccessorTable() {
3267        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalRequestProto_fieldAccessorTable
3268            .ensureFieldAccessorsInitialized(
3269                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.Builder.class);
3270      }
3271
3272      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.newBuilder()
3273      private Builder() {
3274        maybeForceBuilderInitialization();
3275      }
3276
3277      private Builder(
3278          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3279        super(parent);
3280        maybeForceBuilderInitialization();
3281      }
3282      private void maybeForceBuilderInitialization() {
3283        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
3284          getReqInfoFieldBuilder();
3285        }
3286      }
3287      private static Builder create() {
3288        return new Builder();
3289      }
3290
3291      public Builder clear() {
3292        super.clear();
3293        if (reqInfoBuilder_ == null) {
3294          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3295        } else {
3296          reqInfoBuilder_.clear();
3297        }
3298        bitField0_ = (bitField0_ & ~0x00000001);
3299        firstTxnId_ = 0L;
3300        bitField0_ = (bitField0_ & ~0x00000002);
3301        numTxns_ = 0;
3302        bitField0_ = (bitField0_ & ~0x00000004);
3303        records_ = com.google.protobuf.ByteString.EMPTY;
3304        bitField0_ = (bitField0_ & ~0x00000008);
3305        segmentTxnId_ = 0L;
3306        bitField0_ = (bitField0_ & ~0x00000010);
3307        return this;
3308      }
3309
3310      public Builder clone() {
3311        return create().mergeFrom(buildPartial());
3312      }
3313
3314      public com.google.protobuf.Descriptors.Descriptor
3315          getDescriptorForType() {
3316        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalRequestProto_descriptor;
3317      }
3318
3319      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto getDefaultInstanceForType() {
3320        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance();
3321      }
3322
3323      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto build() {
3324        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto result = buildPartial();
3325        if (!result.isInitialized()) {
3326          throw newUninitializedMessageException(result);
3327        }
3328        return result;
3329      }
3330
3331      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto buildPartial() {
3332        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto(this);
3333        int from_bitField0_ = bitField0_;
3334        int to_bitField0_ = 0;
3335        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
3336          to_bitField0_ |= 0x00000001;
3337        }
3338        if (reqInfoBuilder_ == null) {
3339          result.reqInfo_ = reqInfo_;
3340        } else {
3341          result.reqInfo_ = reqInfoBuilder_.build();
3342        }
3343        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
3344          to_bitField0_ |= 0x00000002;
3345        }
3346        result.firstTxnId_ = firstTxnId_;
3347        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
3348          to_bitField0_ |= 0x00000004;
3349        }
3350        result.numTxns_ = numTxns_;
3351        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
3352          to_bitField0_ |= 0x00000008;
3353        }
3354        result.records_ = records_;
3355        if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
3356          to_bitField0_ |= 0x00000010;
3357        }
3358        result.segmentTxnId_ = segmentTxnId_;
3359        result.bitField0_ = to_bitField0_;
3360        onBuilt();
3361        return result;
3362      }
3363
3364      public Builder mergeFrom(com.google.protobuf.Message other) {
3365        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto) {
3366          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)other);
3367        } else {
3368          super.mergeFrom(other);
3369          return this;
3370        }
3371      }
3372
3373      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto other) {
3374        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance()) return this;
3375        if (other.hasReqInfo()) {
3376          mergeReqInfo(other.getReqInfo());
3377        }
3378        if (other.hasFirstTxnId()) {
3379          setFirstTxnId(other.getFirstTxnId());
3380        }
3381        if (other.hasNumTxns()) {
3382          setNumTxns(other.getNumTxns());
3383        }
3384        if (other.hasRecords()) {
3385          setRecords(other.getRecords());
3386        }
3387        if (other.hasSegmentTxnId()) {
3388          setSegmentTxnId(other.getSegmentTxnId());
3389        }
3390        this.mergeUnknownFields(other.getUnknownFields());
3391        return this;
3392      }
3393
3394      public final boolean isInitialized() {
3395        if (!hasReqInfo()) {
3396          
3397          return false;
3398        }
3399        if (!hasFirstTxnId()) {
3400          
3401          return false;
3402        }
3403        if (!hasNumTxns()) {
3404          
3405          return false;
3406        }
3407        if (!hasRecords()) {
3408          
3409          return false;
3410        }
3411        if (!hasSegmentTxnId()) {
3412          
3413          return false;
3414        }
3415        if (!getReqInfo().isInitialized()) {
3416          
3417          return false;
3418        }
3419        return true;
3420      }
3421
3422      public Builder mergeFrom(
3423          com.google.protobuf.CodedInputStream input,
3424          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3425          throws java.io.IOException {
3426        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parsedMessage = null;
3427        try {
3428          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
3429        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
3430          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto) e.getUnfinishedMessage();
3431          throw e;
3432        } finally {
3433          if (parsedMessage != null) {
3434            mergeFrom(parsedMessage);
3435          }
3436        }
3437        return this;
3438      }
3439      private int bitField0_;
3440
3441      // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
3442      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3443      private com.google.protobuf.SingleFieldBuilder<
3444          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
3445      /**
3446       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
3447       */
3448      public boolean hasReqInfo() {
3449        return ((bitField0_ & 0x00000001) == 0x00000001);
3450      }
3451      /**
3452       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
3453       */
3454      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
3455        if (reqInfoBuilder_ == null) {
3456          return reqInfo_;
3457        } else {
3458          return reqInfoBuilder_.getMessage();
3459        }
3460      }
3461      /**
3462       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
3463       */
3464      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
3465        if (reqInfoBuilder_ == null) {
3466          if (value == null) {
3467            throw new NullPointerException();
3468          }
3469          reqInfo_ = value;
3470          onChanged();
3471        } else {
3472          reqInfoBuilder_.setMessage(value);
3473        }
3474        bitField0_ |= 0x00000001;
3475        return this;
3476      }
3477      /**
3478       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
3479       */
3480      public Builder setReqInfo(
3481          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
3482        if (reqInfoBuilder_ == null) {
3483          reqInfo_ = builderForValue.build();
3484          onChanged();
3485        } else {
3486          reqInfoBuilder_.setMessage(builderForValue.build());
3487        }
3488        bitField0_ |= 0x00000001;
3489        return this;
3490      }
3491      /**
3492       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
3493       */
3494      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
3495        if (reqInfoBuilder_ == null) {
3496          if (((bitField0_ & 0x00000001) == 0x00000001) &&
3497              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
3498            reqInfo_ =
3499              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
3500          } else {
3501            reqInfo_ = value;
3502          }
3503          onChanged();
3504        } else {
3505          reqInfoBuilder_.mergeFrom(value);
3506        }
3507        bitField0_ |= 0x00000001;
3508        return this;
3509      }
3510      /**
3511       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
3512       */
3513      public Builder clearReqInfo() {
3514        if (reqInfoBuilder_ == null) {
3515          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3516          onChanged();
3517        } else {
3518          reqInfoBuilder_.clear();
3519        }
3520        bitField0_ = (bitField0_ & ~0x00000001);
3521        return this;
3522      }
3523      /**
3524       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
3525       */
3526      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
3527        bitField0_ |= 0x00000001;
3528        onChanged();
3529        return getReqInfoFieldBuilder().getBuilder();
3530      }
3531      /**
3532       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
3533       */
3534      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
3535        if (reqInfoBuilder_ != null) {
3536          return reqInfoBuilder_.getMessageOrBuilder();
3537        } else {
3538          return reqInfo_;
3539        }
3540      }
3541      /**
3542       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
3543       */
3544      private com.google.protobuf.SingleFieldBuilder<
3545          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
3546          getReqInfoFieldBuilder() {
3547        if (reqInfoBuilder_ == null) {
3548          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
3549              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
3550                  reqInfo_,
3551                  getParentForChildren(),
3552                  isClean());
3553          reqInfo_ = null;
3554        }
3555        return reqInfoBuilder_;
3556      }
3557
3558      // required uint64 firstTxnId = 2;
3559      private long firstTxnId_ ;
3560      /**
3561       * <code>required uint64 firstTxnId = 2;</code>
3562       */
3563      public boolean hasFirstTxnId() {
3564        return ((bitField0_ & 0x00000002) == 0x00000002);
3565      }
3566      /**
3567       * <code>required uint64 firstTxnId = 2;</code>
3568       */
3569      public long getFirstTxnId() {
3570        return firstTxnId_;
3571      }
3572      /**
3573       * <code>required uint64 firstTxnId = 2;</code>
3574       */
3575      public Builder setFirstTxnId(long value) {
3576        bitField0_ |= 0x00000002;
3577        firstTxnId_ = value;
3578        onChanged();
3579        return this;
3580      }
3581      /**
3582       * <code>required uint64 firstTxnId = 2;</code>
3583       */
3584      public Builder clearFirstTxnId() {
3585        bitField0_ = (bitField0_ & ~0x00000002);
3586        firstTxnId_ = 0L;
3587        onChanged();
3588        return this;
3589      }
3590
3591      // required uint32 numTxns = 3;
3592      private int numTxns_ ;
3593      /**
3594       * <code>required uint32 numTxns = 3;</code>
3595       */
3596      public boolean hasNumTxns() {
3597        return ((bitField0_ & 0x00000004) == 0x00000004);
3598      }
3599      /**
3600       * <code>required uint32 numTxns = 3;</code>
3601       */
3602      public int getNumTxns() {
3603        return numTxns_;
3604      }
3605      /**
3606       * <code>required uint32 numTxns = 3;</code>
3607       */
3608      public Builder setNumTxns(int value) {
3609        bitField0_ |= 0x00000004;
3610        numTxns_ = value;
3611        onChanged();
3612        return this;
3613      }
3614      /**
3615       * <code>required uint32 numTxns = 3;</code>
3616       */
3617      public Builder clearNumTxns() {
3618        bitField0_ = (bitField0_ & ~0x00000004);
3619        numTxns_ = 0;
3620        onChanged();
3621        return this;
3622      }
3623
3624      // required bytes records = 4;
3625      private com.google.protobuf.ByteString records_ = com.google.protobuf.ByteString.EMPTY;
3626      /**
3627       * <code>required bytes records = 4;</code>
3628       */
3629      public boolean hasRecords() {
3630        return ((bitField0_ & 0x00000008) == 0x00000008);
3631      }
3632      /**
3633       * <code>required bytes records = 4;</code>
3634       */
3635      public com.google.protobuf.ByteString getRecords() {
3636        return records_;
3637      }
3638      /**
3639       * <code>required bytes records = 4;</code>
3640       */
3641      public Builder setRecords(com.google.protobuf.ByteString value) {
3642        if (value == null) {
3643    throw new NullPointerException();
3644  }
3645  bitField0_ |= 0x00000008;
3646        records_ = value;
3647        onChanged();
3648        return this;
3649      }
3650      /**
3651       * <code>required bytes records = 4;</code>
3652       */
3653      public Builder clearRecords() {
3654        bitField0_ = (bitField0_ & ~0x00000008);
3655        records_ = getDefaultInstance().getRecords();
3656        onChanged();
3657        return this;
3658      }
3659
3660      // required uint64 segmentTxnId = 5;
3661      private long segmentTxnId_ ;
3662      /**
3663       * <code>required uint64 segmentTxnId = 5;</code>
3664       */
3665      public boolean hasSegmentTxnId() {
3666        return ((bitField0_ & 0x00000010) == 0x00000010);
3667      }
3668      /**
3669       * <code>required uint64 segmentTxnId = 5;</code>
3670       */
3671      public long getSegmentTxnId() {
3672        return segmentTxnId_;
3673      }
3674      /**
3675       * <code>required uint64 segmentTxnId = 5;</code>
3676       */
3677      public Builder setSegmentTxnId(long value) {
3678        bitField0_ |= 0x00000010;
3679        segmentTxnId_ = value;
3680        onChanged();
3681        return this;
3682      }
3683      /**
3684       * <code>required uint64 segmentTxnId = 5;</code>
3685       */
3686      public Builder clearSegmentTxnId() {
3687        bitField0_ = (bitField0_ & ~0x00000010);
3688        segmentTxnId_ = 0L;
3689        onChanged();
3690        return this;
3691      }
3692
3693      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.JournalRequestProto)
3694    }
3695
3696    static {
3697      defaultInstance = new JournalRequestProto(true);
3698      defaultInstance.initFields();
3699    }
3700
3701    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.JournalRequestProto)
3702  }
3703
3704  public interface JournalResponseProtoOrBuilder
3705      extends com.google.protobuf.MessageOrBuilder {
3706  }
3707  /**
3708   * Protobuf type {@code hadoop.hdfs.qjournal.JournalResponseProto}
3709   */
3710  public static final class JournalResponseProto extends
3711      com.google.protobuf.GeneratedMessage
3712      implements JournalResponseProtoOrBuilder {
3713    // Use JournalResponseProto.newBuilder() to construct.
3714    private JournalResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
3715      super(builder);
3716      this.unknownFields = builder.getUnknownFields();
3717    }
3718    private JournalResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
3719
3720    private static final JournalResponseProto defaultInstance;
3721    public static JournalResponseProto getDefaultInstance() {
3722      return defaultInstance;
3723    }
3724
3725    public JournalResponseProto getDefaultInstanceForType() {
3726      return defaultInstance;
3727    }
3728
3729    private final com.google.protobuf.UnknownFieldSet unknownFields;
3730    @java.lang.Override
3731    public final com.google.protobuf.UnknownFieldSet
3732        getUnknownFields() {
3733      return this.unknownFields;
3734    }
3735    private JournalResponseProto(
3736        com.google.protobuf.CodedInputStream input,
3737        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3738        throws com.google.protobuf.InvalidProtocolBufferException {
3739      initFields();
3740      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
3741          com.google.protobuf.UnknownFieldSet.newBuilder();
3742      try {
3743        boolean done = false;
3744        while (!done) {
3745          int tag = input.readTag();
3746          switch (tag) {
3747            case 0:
3748              done = true;
3749              break;
3750            default: {
3751              if (!parseUnknownField(input, unknownFields,
3752                                     extensionRegistry, tag)) {
3753                done = true;
3754              }
3755              break;
3756            }
3757          }
3758        }
3759      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
3760        throw e.setUnfinishedMessage(this);
3761      } catch (java.io.IOException e) {
3762        throw new com.google.protobuf.InvalidProtocolBufferException(
3763            e.getMessage()).setUnfinishedMessage(this);
3764      } finally {
3765        this.unknownFields = unknownFields.build();
3766        makeExtensionsImmutable();
3767      }
3768    }
3769    public static final com.google.protobuf.Descriptors.Descriptor
3770        getDescriptor() {
3771      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalResponseProto_descriptor;
3772    }
3773
3774    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3775        internalGetFieldAccessorTable() {
3776      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalResponseProto_fieldAccessorTable
3777          .ensureFieldAccessorsInitialized(
3778              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.Builder.class);
3779    }
3780
3781    public static com.google.protobuf.Parser<JournalResponseProto> PARSER =
3782        new com.google.protobuf.AbstractParser<JournalResponseProto>() {
3783      public JournalResponseProto parsePartialFrom(
3784          com.google.protobuf.CodedInputStream input,
3785          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3786          throws com.google.protobuf.InvalidProtocolBufferException {
3787        return new JournalResponseProto(input, extensionRegistry);
3788      }
3789    };
3790
3791    @java.lang.Override
3792    public com.google.protobuf.Parser<JournalResponseProto> getParserForType() {
3793      return PARSER;
3794    }
3795
3796    private void initFields() {
3797    }
3798    private byte memoizedIsInitialized = -1;
3799    public final boolean isInitialized() {
3800      byte isInitialized = memoizedIsInitialized;
3801      if (isInitialized != -1) return isInitialized == 1;
3802
3803      memoizedIsInitialized = 1;
3804      return true;
3805    }
3806
3807    public void writeTo(com.google.protobuf.CodedOutputStream output)
3808                        throws java.io.IOException {
3809      getSerializedSize();
3810      getUnknownFields().writeTo(output);
3811    }
3812
3813    private int memoizedSerializedSize = -1;
3814    public int getSerializedSize() {
3815      int size = memoizedSerializedSize;
3816      if (size != -1) return size;
3817
3818      size = 0;
3819      size += getUnknownFields().getSerializedSize();
3820      memoizedSerializedSize = size;
3821      return size;
3822    }
3823
3824    private static final long serialVersionUID = 0L;
3825    @java.lang.Override
3826    protected java.lang.Object writeReplace()
3827        throws java.io.ObjectStreamException {
3828      return super.writeReplace();
3829    }
3830
3831    @java.lang.Override
3832    public boolean equals(final java.lang.Object obj) {
3833      if (obj == this) {
3834       return true;
3835      }
3836      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto)) {
3837        return super.equals(obj);
3838      }
3839      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) obj;
3840
3841      boolean result = true;
3842      result = result &&
3843          getUnknownFields().equals(other.getUnknownFields());
3844      return result;
3845    }
3846
3847    private int memoizedHashCode = 0;
3848    @java.lang.Override
3849    public int hashCode() {
3850      if (memoizedHashCode != 0) {
3851        return memoizedHashCode;
3852      }
3853      int hash = 41;
3854      hash = (19 * hash) + getDescriptorForType().hashCode();
3855      hash = (29 * hash) + getUnknownFields().hashCode();
3856      memoizedHashCode = hash;
3857      return hash;
3858    }
3859
3860    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3861        com.google.protobuf.ByteString data)
3862        throws com.google.protobuf.InvalidProtocolBufferException {
3863      return PARSER.parseFrom(data);
3864    }
3865    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3866        com.google.protobuf.ByteString data,
3867        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3868        throws com.google.protobuf.InvalidProtocolBufferException {
3869      return PARSER.parseFrom(data, extensionRegistry);
3870    }
3871    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(byte[] data)
3872        throws com.google.protobuf.InvalidProtocolBufferException {
3873      return PARSER.parseFrom(data);
3874    }
3875    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3876        byte[] data,
3877        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3878        throws com.google.protobuf.InvalidProtocolBufferException {
3879      return PARSER.parseFrom(data, extensionRegistry);
3880    }
3881    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(java.io.InputStream input)
3882        throws java.io.IOException {
3883      return PARSER.parseFrom(input);
3884    }
3885    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3886        java.io.InputStream input,
3887        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3888        throws java.io.IOException {
3889      return PARSER.parseFrom(input, extensionRegistry);
3890    }
3891    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseDelimitedFrom(java.io.InputStream input)
3892        throws java.io.IOException {
3893      return PARSER.parseDelimitedFrom(input);
3894    }
3895    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseDelimitedFrom(
3896        java.io.InputStream input,
3897        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3898        throws java.io.IOException {
3899      return PARSER.parseDelimitedFrom(input, extensionRegistry);
3900    }
3901    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3902        com.google.protobuf.CodedInputStream input)
3903        throws java.io.IOException {
3904      return PARSER.parseFrom(input);
3905    }
3906    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3907        com.google.protobuf.CodedInputStream input,
3908        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3909        throws java.io.IOException {
3910      return PARSER.parseFrom(input, extensionRegistry);
3911    }
3912
3913    public static Builder newBuilder() { return Builder.create(); }
3914    public Builder newBuilderForType() { return newBuilder(); }
3915    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto prototype) {
3916      return newBuilder().mergeFrom(prototype);
3917    }
3918    public Builder toBuilder() { return newBuilder(this); }
3919
3920    @java.lang.Override
3921    protected Builder newBuilderForType(
3922        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3923      Builder builder = new Builder(parent);
3924      return builder;
3925    }
3926    /**
3927     * Protobuf type {@code hadoop.hdfs.qjournal.JournalResponseProto}
3928     */
3929    public static final class Builder extends
3930        com.google.protobuf.GeneratedMessage.Builder<Builder>
3931       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProtoOrBuilder {
3932      public static final com.google.protobuf.Descriptors.Descriptor
3933          getDescriptor() {
3934        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalResponseProto_descriptor;
3935      }
3936
3937      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3938          internalGetFieldAccessorTable() {
3939        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalResponseProto_fieldAccessorTable
3940            .ensureFieldAccessorsInitialized(
3941                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.Builder.class);
3942      }
3943
3944      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.newBuilder()
3945      private Builder() {
3946        maybeForceBuilderInitialization();
3947      }
3948
3949      private Builder(
3950          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3951        super(parent);
3952        maybeForceBuilderInitialization();
3953      }
3954      private void maybeForceBuilderInitialization() {
3955        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
3956        }
3957      }
3958      private static Builder create() {
3959        return new Builder();
3960      }
3961
3962      public Builder clear() {
3963        super.clear();
3964        return this;
3965      }
3966
3967      public Builder clone() {
3968        return create().mergeFrom(buildPartial());
3969      }
3970
3971      public com.google.protobuf.Descriptors.Descriptor
3972          getDescriptorForType() {
3973        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_JournalResponseProto_descriptor;
3974      }
3975
3976      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto getDefaultInstanceForType() {
3977        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance();
3978      }
3979
3980      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto build() {
3981        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto result = buildPartial();
3982        if (!result.isInitialized()) {
3983          throw newUninitializedMessageException(result);
3984        }
3985        return result;
3986      }
3987
3988      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto buildPartial() {
3989        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto(this);
3990        onBuilt();
3991        return result;
3992      }
3993
3994      public Builder mergeFrom(com.google.protobuf.Message other) {
3995        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) {
3996          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto)other);
3997        } else {
3998          super.mergeFrom(other);
3999          return this;
4000        }
4001      }
4002
4003      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto other) {
4004        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance()) return this;
4005        this.mergeUnknownFields(other.getUnknownFields());
4006        return this;
4007      }
4008
4009      public final boolean isInitialized() {
4010        return true;
4011      }
4012
4013      public Builder mergeFrom(
4014          com.google.protobuf.CodedInputStream input,
4015          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4016          throws java.io.IOException {
4017        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parsedMessage = null;
4018        try {
4019          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
4020        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
4021          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) e.getUnfinishedMessage();
4022          throw e;
4023        } finally {
4024          if (parsedMessage != null) {
4025            mergeFrom(parsedMessage);
4026          }
4027        }
4028        return this;
4029      }
4030
4031      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.JournalResponseProto)
4032    }
4033
4034    static {
4035      defaultInstance = new JournalResponseProto(true);
4036      defaultInstance.initFields();
4037    }
4038
4039    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.JournalResponseProto)
4040  }
4041
4042  public interface HeartbeatRequestProtoOrBuilder
4043      extends com.google.protobuf.MessageOrBuilder {
4044
4045    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
4046    /**
4047     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4048     */
4049    boolean hasReqInfo();
4050    /**
4051     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4052     */
4053    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
4054    /**
4055     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4056     */
4057    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
4058  }
4059  /**
4060   * Protobuf type {@code hadoop.hdfs.qjournal.HeartbeatRequestProto}
4061   */
4062  public static final class HeartbeatRequestProto extends
4063      com.google.protobuf.GeneratedMessage
4064      implements HeartbeatRequestProtoOrBuilder {
4065    // Use HeartbeatRequestProto.newBuilder() to construct.
4066    private HeartbeatRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
4067      super(builder);
4068      this.unknownFields = builder.getUnknownFields();
4069    }
4070    private HeartbeatRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
4071
4072    private static final HeartbeatRequestProto defaultInstance;
4073    public static HeartbeatRequestProto getDefaultInstance() {
4074      return defaultInstance;
4075    }
4076
4077    public HeartbeatRequestProto getDefaultInstanceForType() {
4078      return defaultInstance;
4079    }
4080
4081    private final com.google.protobuf.UnknownFieldSet unknownFields;
4082    @java.lang.Override
4083    public final com.google.protobuf.UnknownFieldSet
4084        getUnknownFields() {
4085      return this.unknownFields;
4086    }
4087    private HeartbeatRequestProto(
4088        com.google.protobuf.CodedInputStream input,
4089        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4090        throws com.google.protobuf.InvalidProtocolBufferException {
4091      initFields();
4092      int mutable_bitField0_ = 0;
4093      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
4094          com.google.protobuf.UnknownFieldSet.newBuilder();
4095      try {
4096        boolean done = false;
4097        while (!done) {
4098          int tag = input.readTag();
4099          switch (tag) {
4100            case 0:
4101              done = true;
4102              break;
4103            default: {
4104              if (!parseUnknownField(input, unknownFields,
4105                                     extensionRegistry, tag)) {
4106                done = true;
4107              }
4108              break;
4109            }
4110            case 10: {
4111              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
4112              if (((bitField0_ & 0x00000001) == 0x00000001)) {
4113                subBuilder = reqInfo_.toBuilder();
4114              }
4115              reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
4116              if (subBuilder != null) {
4117                subBuilder.mergeFrom(reqInfo_);
4118                reqInfo_ = subBuilder.buildPartial();
4119              }
4120              bitField0_ |= 0x00000001;
4121              break;
4122            }
4123          }
4124        }
4125      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
4126        throw e.setUnfinishedMessage(this);
4127      } catch (java.io.IOException e) {
4128        throw new com.google.protobuf.InvalidProtocolBufferException(
4129            e.getMessage()).setUnfinishedMessage(this);
4130      } finally {
4131        this.unknownFields = unknownFields.build();
4132        makeExtensionsImmutable();
4133      }
4134    }
4135    public static final com.google.protobuf.Descriptors.Descriptor
4136        getDescriptor() {
4137      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_HeartbeatRequestProto_descriptor;
4138    }
4139
4140    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4141        internalGetFieldAccessorTable() {
4142      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_HeartbeatRequestProto_fieldAccessorTable
4143          .ensureFieldAccessorsInitialized(
4144              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.Builder.class);
4145    }
4146
4147    public static com.google.protobuf.Parser<HeartbeatRequestProto> PARSER =
4148        new com.google.protobuf.AbstractParser<HeartbeatRequestProto>() {
4149      public HeartbeatRequestProto parsePartialFrom(
4150          com.google.protobuf.CodedInputStream input,
4151          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4152          throws com.google.protobuf.InvalidProtocolBufferException {
4153        return new HeartbeatRequestProto(input, extensionRegistry);
4154      }
4155    };
4156
4157    @java.lang.Override
4158    public com.google.protobuf.Parser<HeartbeatRequestProto> getParserForType() {
4159      return PARSER;
4160    }
4161
4162    private int bitField0_;
4163    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
4164    public static final int REQINFO_FIELD_NUMBER = 1;
4165    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
4166    /**
4167     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4168     */
4169    public boolean hasReqInfo() {
4170      return ((bitField0_ & 0x00000001) == 0x00000001);
4171    }
4172    /**
4173     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4174     */
4175    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
4176      return reqInfo_;
4177    }
4178    /**
4179     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4180     */
4181    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
4182      return reqInfo_;
4183    }
4184
4185    private void initFields() {
4186      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4187    }
4188    private byte memoizedIsInitialized = -1;
4189    public final boolean isInitialized() {
4190      byte isInitialized = memoizedIsInitialized;
4191      if (isInitialized != -1) return isInitialized == 1;
4192
4193      if (!hasReqInfo()) {
4194        memoizedIsInitialized = 0;
4195        return false;
4196      }
4197      if (!getReqInfo().isInitialized()) {
4198        memoizedIsInitialized = 0;
4199        return false;
4200      }
4201      memoizedIsInitialized = 1;
4202      return true;
4203    }
4204
4205    public void writeTo(com.google.protobuf.CodedOutputStream output)
4206                        throws java.io.IOException {
4207      getSerializedSize();
4208      if (((bitField0_ & 0x00000001) == 0x00000001)) {
4209        output.writeMessage(1, reqInfo_);
4210      }
4211      getUnknownFields().writeTo(output);
4212    }
4213
4214    private int memoizedSerializedSize = -1;
4215    public int getSerializedSize() {
4216      int size = memoizedSerializedSize;
4217      if (size != -1) return size;
4218
4219      size = 0;
4220      if (((bitField0_ & 0x00000001) == 0x00000001)) {
4221        size += com.google.protobuf.CodedOutputStream
4222          .computeMessageSize(1, reqInfo_);
4223      }
4224      size += getUnknownFields().getSerializedSize();
4225      memoizedSerializedSize = size;
4226      return size;
4227    }
4228
4229    private static final long serialVersionUID = 0L;
4230    @java.lang.Override
4231    protected java.lang.Object writeReplace()
4232        throws java.io.ObjectStreamException {
4233      return super.writeReplace();
4234    }
4235
4236    @java.lang.Override
4237    public boolean equals(final java.lang.Object obj) {
4238      if (obj == this) {
4239       return true;
4240      }
4241      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)) {
4242        return super.equals(obj);
4243      }
4244      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto) obj;
4245
4246      boolean result = true;
4247      result = result && (hasReqInfo() == other.hasReqInfo());
4248      if (hasReqInfo()) {
4249        result = result && getReqInfo()
4250            .equals(other.getReqInfo());
4251      }
4252      result = result &&
4253          getUnknownFields().equals(other.getUnknownFields());
4254      return result;
4255    }
4256
4257    private int memoizedHashCode = 0;
4258    @java.lang.Override
4259    public int hashCode() {
4260      if (memoizedHashCode != 0) {
4261        return memoizedHashCode;
4262      }
4263      int hash = 41;
4264      hash = (19 * hash) + getDescriptorForType().hashCode();
4265      if (hasReqInfo()) {
4266        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
4267        hash = (53 * hash) + getReqInfo().hashCode();
4268      }
4269      hash = (29 * hash) + getUnknownFields().hashCode();
4270      memoizedHashCode = hash;
4271      return hash;
4272    }
4273
4274    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4275        com.google.protobuf.ByteString data)
4276        throws com.google.protobuf.InvalidProtocolBufferException {
4277      return PARSER.parseFrom(data);
4278    }
4279    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4280        com.google.protobuf.ByteString data,
4281        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4282        throws com.google.protobuf.InvalidProtocolBufferException {
4283      return PARSER.parseFrom(data, extensionRegistry);
4284    }
4285    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(byte[] data)
4286        throws com.google.protobuf.InvalidProtocolBufferException {
4287      return PARSER.parseFrom(data);
4288    }
4289    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4290        byte[] data,
4291        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4292        throws com.google.protobuf.InvalidProtocolBufferException {
4293      return PARSER.parseFrom(data, extensionRegistry);
4294    }
4295    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(java.io.InputStream input)
4296        throws java.io.IOException {
4297      return PARSER.parseFrom(input);
4298    }
4299    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4300        java.io.InputStream input,
4301        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4302        throws java.io.IOException {
4303      return PARSER.parseFrom(input, extensionRegistry);
4304    }
4305    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseDelimitedFrom(java.io.InputStream input)
4306        throws java.io.IOException {
4307      return PARSER.parseDelimitedFrom(input);
4308    }
4309    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseDelimitedFrom(
4310        java.io.InputStream input,
4311        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4312        throws java.io.IOException {
4313      return PARSER.parseDelimitedFrom(input, extensionRegistry);
4314    }
4315    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4316        com.google.protobuf.CodedInputStream input)
4317        throws java.io.IOException {
4318      return PARSER.parseFrom(input);
4319    }
4320    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4321        com.google.protobuf.CodedInputStream input,
4322        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4323        throws java.io.IOException {
4324      return PARSER.parseFrom(input, extensionRegistry);
4325    }
4326
4327    public static Builder newBuilder() { return Builder.create(); }
4328    public Builder newBuilderForType() { return newBuilder(); }
4329    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto prototype) {
4330      return newBuilder().mergeFrom(prototype);
4331    }
4332    public Builder toBuilder() { return newBuilder(this); }
4333
4334    @java.lang.Override
4335    protected Builder newBuilderForType(
4336        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
4337      Builder builder = new Builder(parent);
4338      return builder;
4339    }
4340    /**
4341     * Protobuf type {@code hadoop.hdfs.qjournal.HeartbeatRequestProto}
4342     */
4343    public static final class Builder extends
4344        com.google.protobuf.GeneratedMessage.Builder<Builder>
4345       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProtoOrBuilder {
4346      public static final com.google.protobuf.Descriptors.Descriptor
4347          getDescriptor() {
4348        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_HeartbeatRequestProto_descriptor;
4349      }
4350
4351      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4352          internalGetFieldAccessorTable() {
4353        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_HeartbeatRequestProto_fieldAccessorTable
4354            .ensureFieldAccessorsInitialized(
4355                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.Builder.class);
4356      }
4357
4358      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.newBuilder()
4359      private Builder() {
4360        maybeForceBuilderInitialization();
4361      }
4362
4363      private Builder(
4364          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
4365        super(parent);
4366        maybeForceBuilderInitialization();
4367      }
4368      private void maybeForceBuilderInitialization() {
4369        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
4370          getReqInfoFieldBuilder();
4371        }
4372      }
4373      private static Builder create() {
4374        return new Builder();
4375      }
4376
4377      public Builder clear() {
4378        super.clear();
4379        if (reqInfoBuilder_ == null) {
4380          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4381        } else {
4382          reqInfoBuilder_.clear();
4383        }
4384        bitField0_ = (bitField0_ & ~0x00000001);
4385        return this;
4386      }
4387
4388      public Builder clone() {
4389        return create().mergeFrom(buildPartial());
4390      }
4391
4392      public com.google.protobuf.Descriptors.Descriptor
4393          getDescriptorForType() {
4394        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_HeartbeatRequestProto_descriptor;
4395      }
4396
4397      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto getDefaultInstanceForType() {
4398        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance();
4399      }
4400
4401      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto build() {
4402        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto result = buildPartial();
4403        if (!result.isInitialized()) {
4404          throw newUninitializedMessageException(result);
4405        }
4406        return result;
4407      }
4408
4409      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto buildPartial() {
4410        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto(this);
4411        int from_bitField0_ = bitField0_;
4412        int to_bitField0_ = 0;
4413        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
4414          to_bitField0_ |= 0x00000001;
4415        }
4416        if (reqInfoBuilder_ == null) {
4417          result.reqInfo_ = reqInfo_;
4418        } else {
4419          result.reqInfo_ = reqInfoBuilder_.build();
4420        }
4421        result.bitField0_ = to_bitField0_;
4422        onBuilt();
4423        return result;
4424      }
4425
4426      public Builder mergeFrom(com.google.protobuf.Message other) {
4427        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto) {
4428          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)other);
4429        } else {
4430          super.mergeFrom(other);
4431          return this;
4432        }
4433      }
4434
4435      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto other) {
4436        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance()) return this;
4437        if (other.hasReqInfo()) {
4438          mergeReqInfo(other.getReqInfo());
4439        }
4440        this.mergeUnknownFields(other.getUnknownFields());
4441        return this;
4442      }
4443
4444      public final boolean isInitialized() {
4445        if (!hasReqInfo()) {
4446          
4447          return false;
4448        }
4449        if (!getReqInfo().isInitialized()) {
4450          
4451          return false;
4452        }
4453        return true;
4454      }
4455
4456      public Builder mergeFrom(
4457          com.google.protobuf.CodedInputStream input,
4458          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4459          throws java.io.IOException {
4460        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parsedMessage = null;
4461        try {
4462          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
4463        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
4464          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto) e.getUnfinishedMessage();
4465          throw e;
4466        } finally {
4467          if (parsedMessage != null) {
4468            mergeFrom(parsedMessage);
4469          }
4470        }
4471        return this;
4472      }
4473      private int bitField0_;
4474
4475      // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
4476      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4477      private com.google.protobuf.SingleFieldBuilder<
4478          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
4479      /**
4480       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4481       */
4482      public boolean hasReqInfo() {
4483        return ((bitField0_ & 0x00000001) == 0x00000001);
4484      }
4485      /**
4486       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4487       */
4488      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
4489        if (reqInfoBuilder_ == null) {
4490          return reqInfo_;
4491        } else {
4492          return reqInfoBuilder_.getMessage();
4493        }
4494      }
4495      /**
4496       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4497       */
4498      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
4499        if (reqInfoBuilder_ == null) {
4500          if (value == null) {
4501            throw new NullPointerException();
4502          }
4503          reqInfo_ = value;
4504          onChanged();
4505        } else {
4506          reqInfoBuilder_.setMessage(value);
4507        }
4508        bitField0_ |= 0x00000001;
4509        return this;
4510      }
4511      /**
4512       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4513       */
4514      public Builder setReqInfo(
4515          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
4516        if (reqInfoBuilder_ == null) {
4517          reqInfo_ = builderForValue.build();
4518          onChanged();
4519        } else {
4520          reqInfoBuilder_.setMessage(builderForValue.build());
4521        }
4522        bitField0_ |= 0x00000001;
4523        return this;
4524      }
4525      /**
4526       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4527       */
4528      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
4529        if (reqInfoBuilder_ == null) {
4530          if (((bitField0_ & 0x00000001) == 0x00000001) &&
4531              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
4532            reqInfo_ =
4533              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
4534          } else {
4535            reqInfo_ = value;
4536          }
4537          onChanged();
4538        } else {
4539          reqInfoBuilder_.mergeFrom(value);
4540        }
4541        bitField0_ |= 0x00000001;
4542        return this;
4543      }
4544      /**
4545       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4546       */
4547      public Builder clearReqInfo() {
4548        if (reqInfoBuilder_ == null) {
4549          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4550          onChanged();
4551        } else {
4552          reqInfoBuilder_.clear();
4553        }
4554        bitField0_ = (bitField0_ & ~0x00000001);
4555        return this;
4556      }
4557      /**
4558       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4559       */
4560      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
4561        bitField0_ |= 0x00000001;
4562        onChanged();
4563        return getReqInfoFieldBuilder().getBuilder();
4564      }
4565      /**
4566       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4567       */
4568      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
4569        if (reqInfoBuilder_ != null) {
4570          return reqInfoBuilder_.getMessageOrBuilder();
4571        } else {
4572          return reqInfo_;
4573        }
4574      }
4575      /**
4576       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4577       */
4578      private com.google.protobuf.SingleFieldBuilder<
4579          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
4580          getReqInfoFieldBuilder() {
4581        if (reqInfoBuilder_ == null) {
4582          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
4583              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
4584                  reqInfo_,
4585                  getParentForChildren(),
4586                  isClean());
4587          reqInfo_ = null;
4588        }
4589        return reqInfoBuilder_;
4590      }
4591
4592      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.HeartbeatRequestProto)
4593    }
4594
4595    static {
4596      defaultInstance = new HeartbeatRequestProto(true);
4597      defaultInstance.initFields();
4598    }
4599
4600    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.HeartbeatRequestProto)
4601  }
4602
4603  public interface HeartbeatResponseProtoOrBuilder
4604      extends com.google.protobuf.MessageOrBuilder {
4605  }
4606  /**
4607   * Protobuf type {@code hadoop.hdfs.qjournal.HeartbeatResponseProto}
4608   *
4609   * <pre>
4610   * void response
4611   * </pre>
4612   */
4613  public static final class HeartbeatResponseProto extends
4614      com.google.protobuf.GeneratedMessage
4615      implements HeartbeatResponseProtoOrBuilder {
4616    // Use HeartbeatResponseProto.newBuilder() to construct.
4617    private HeartbeatResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
4618      super(builder);
4619      this.unknownFields = builder.getUnknownFields();
4620    }
4621    private HeartbeatResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
4622
4623    private static final HeartbeatResponseProto defaultInstance;
4624    public static HeartbeatResponseProto getDefaultInstance() {
4625      return defaultInstance;
4626    }
4627
4628    public HeartbeatResponseProto getDefaultInstanceForType() {
4629      return defaultInstance;
4630    }
4631
4632    private final com.google.protobuf.UnknownFieldSet unknownFields;
4633    @java.lang.Override
4634    public final com.google.protobuf.UnknownFieldSet
4635        getUnknownFields() {
4636      return this.unknownFields;
4637    }
4638    private HeartbeatResponseProto(
4639        com.google.protobuf.CodedInputStream input,
4640        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4641        throws com.google.protobuf.InvalidProtocolBufferException {
4642      initFields();
4643      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
4644          com.google.protobuf.UnknownFieldSet.newBuilder();
4645      try {
4646        boolean done = false;
4647        while (!done) {
4648          int tag = input.readTag();
4649          switch (tag) {
4650            case 0:
4651              done = true;
4652              break;
4653            default: {
4654              if (!parseUnknownField(input, unknownFields,
4655                                     extensionRegistry, tag)) {
4656                done = true;
4657              }
4658              break;
4659            }
4660          }
4661        }
4662      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
4663        throw e.setUnfinishedMessage(this);
4664      } catch (java.io.IOException e) {
4665        throw new com.google.protobuf.InvalidProtocolBufferException(
4666            e.getMessage()).setUnfinishedMessage(this);
4667      } finally {
4668        this.unknownFields = unknownFields.build();
4669        makeExtensionsImmutable();
4670      }
4671    }
4672    public static final com.google.protobuf.Descriptors.Descriptor
4673        getDescriptor() {
4674      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_HeartbeatResponseProto_descriptor;
4675    }
4676
4677    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4678        internalGetFieldAccessorTable() {
4679      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_HeartbeatResponseProto_fieldAccessorTable
4680          .ensureFieldAccessorsInitialized(
4681              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.Builder.class);
4682    }
4683
4684    public static com.google.protobuf.Parser<HeartbeatResponseProto> PARSER =
4685        new com.google.protobuf.AbstractParser<HeartbeatResponseProto>() {
4686      public HeartbeatResponseProto parsePartialFrom(
4687          com.google.protobuf.CodedInputStream input,
4688          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4689          throws com.google.protobuf.InvalidProtocolBufferException {
4690        return new HeartbeatResponseProto(input, extensionRegistry);
4691      }
4692    };
4693
4694    @java.lang.Override
4695    public com.google.protobuf.Parser<HeartbeatResponseProto> getParserForType() {
4696      return PARSER;
4697    }
4698
4699    private void initFields() {
4700    }
4701    private byte memoizedIsInitialized = -1;
4702    public final boolean isInitialized() {
4703      byte isInitialized = memoizedIsInitialized;
4704      if (isInitialized != -1) return isInitialized == 1;
4705
4706      memoizedIsInitialized = 1;
4707      return true;
4708    }
4709
4710    public void writeTo(com.google.protobuf.CodedOutputStream output)
4711                        throws java.io.IOException {
4712      getSerializedSize();
4713      getUnknownFields().writeTo(output);
4714    }
4715
4716    private int memoizedSerializedSize = -1;
4717    public int getSerializedSize() {
4718      int size = memoizedSerializedSize;
4719      if (size != -1) return size;
4720
4721      size = 0;
4722      size += getUnknownFields().getSerializedSize();
4723      memoizedSerializedSize = size;
4724      return size;
4725    }
4726
4727    private static final long serialVersionUID = 0L;
4728    @java.lang.Override
4729    protected java.lang.Object writeReplace()
4730        throws java.io.ObjectStreamException {
4731      return super.writeReplace();
4732    }
4733
4734    @java.lang.Override
4735    public boolean equals(final java.lang.Object obj) {
4736      if (obj == this) {
4737       return true;
4738      }
4739      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto)) {
4740        return super.equals(obj);
4741      }
4742      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) obj;
4743
4744      boolean result = true;
4745      result = result &&
4746          getUnknownFields().equals(other.getUnknownFields());
4747      return result;
4748    }
4749
4750    private int memoizedHashCode = 0;
4751    @java.lang.Override
4752    public int hashCode() {
4753      if (memoizedHashCode != 0) {
4754        return memoizedHashCode;
4755      }
4756      int hash = 41;
4757      hash = (19 * hash) + getDescriptorForType().hashCode();
4758      hash = (29 * hash) + getUnknownFields().hashCode();
4759      memoizedHashCode = hash;
4760      return hash;
4761    }
4762
4763    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4764        com.google.protobuf.ByteString data)
4765        throws com.google.protobuf.InvalidProtocolBufferException {
4766      return PARSER.parseFrom(data);
4767    }
4768    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4769        com.google.protobuf.ByteString data,
4770        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4771        throws com.google.protobuf.InvalidProtocolBufferException {
4772      return PARSER.parseFrom(data, extensionRegistry);
4773    }
4774    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(byte[] data)
4775        throws com.google.protobuf.InvalidProtocolBufferException {
4776      return PARSER.parseFrom(data);
4777    }
4778    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4779        byte[] data,
4780        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4781        throws com.google.protobuf.InvalidProtocolBufferException {
4782      return PARSER.parseFrom(data, extensionRegistry);
4783    }
4784    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(java.io.InputStream input)
4785        throws java.io.IOException {
4786      return PARSER.parseFrom(input);
4787    }
4788    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4789        java.io.InputStream input,
4790        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4791        throws java.io.IOException {
4792      return PARSER.parseFrom(input, extensionRegistry);
4793    }
4794    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseDelimitedFrom(java.io.InputStream input)
4795        throws java.io.IOException {
4796      return PARSER.parseDelimitedFrom(input);
4797    }
4798    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseDelimitedFrom(
4799        java.io.InputStream input,
4800        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4801        throws java.io.IOException {
4802      return PARSER.parseDelimitedFrom(input, extensionRegistry);
4803    }
4804    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4805        com.google.protobuf.CodedInputStream input)
4806        throws java.io.IOException {
4807      return PARSER.parseFrom(input);
4808    }
4809    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4810        com.google.protobuf.CodedInputStream input,
4811        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4812        throws java.io.IOException {
4813      return PARSER.parseFrom(input, extensionRegistry);
4814    }
4815
4816    public static Builder newBuilder() { return Builder.create(); }
4817    public Builder newBuilderForType() { return newBuilder(); }
4818    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto prototype) {
4819      return newBuilder().mergeFrom(prototype);
4820    }
4821    public Builder toBuilder() { return newBuilder(this); }
4822
4823    @java.lang.Override
4824    protected Builder newBuilderForType(
4825        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
4826      Builder builder = new Builder(parent);
4827      return builder;
4828    }
4829    /**
4830     * Protobuf type {@code hadoop.hdfs.qjournal.HeartbeatResponseProto}
4831     *
4832     * <pre>
4833     * void response
4834     * </pre>
4835     */
4836    public static final class Builder extends
4837        com.google.protobuf.GeneratedMessage.Builder<Builder>
4838       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProtoOrBuilder {
4839      public static final com.google.protobuf.Descriptors.Descriptor
4840          getDescriptor() {
4841        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_HeartbeatResponseProto_descriptor;
4842      }
4843
4844      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4845          internalGetFieldAccessorTable() {
4846        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_HeartbeatResponseProto_fieldAccessorTable
4847            .ensureFieldAccessorsInitialized(
4848                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.Builder.class);
4849      }
4850
4851      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.newBuilder()
4852      private Builder() {
4853        maybeForceBuilderInitialization();
4854      }
4855
4856      private Builder(
4857          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
4858        super(parent);
4859        maybeForceBuilderInitialization();
4860      }
4861      private void maybeForceBuilderInitialization() {
4862        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
4863        }
4864      }
4865      private static Builder create() {
4866        return new Builder();
4867      }
4868
4869      public Builder clear() {
4870        super.clear();
4871        return this;
4872      }
4873
4874      public Builder clone() {
4875        return create().mergeFrom(buildPartial());
4876      }
4877
4878      public com.google.protobuf.Descriptors.Descriptor
4879          getDescriptorForType() {
4880        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_HeartbeatResponseProto_descriptor;
4881      }
4882
4883      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto getDefaultInstanceForType() {
4884        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance();
4885      }
4886
4887      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto build() {
4888        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto result = buildPartial();
4889        if (!result.isInitialized()) {
4890          throw newUninitializedMessageException(result);
4891        }
4892        return result;
4893      }
4894
4895      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto buildPartial() {
4896        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto(this);
4897        onBuilt();
4898        return result;
4899      }
4900
4901      public Builder mergeFrom(com.google.protobuf.Message other) {
4902        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) {
4903          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto)other);
4904        } else {
4905          super.mergeFrom(other);
4906          return this;
4907        }
4908      }
4909
4910      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto other) {
4911        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance()) return this;
4912        this.mergeUnknownFields(other.getUnknownFields());
4913        return this;
4914      }
4915
4916      public final boolean isInitialized() {
4917        return true;
4918      }
4919
4920      public Builder mergeFrom(
4921          com.google.protobuf.CodedInputStream input,
4922          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4923          throws java.io.IOException {
4924        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parsedMessage = null;
4925        try {
4926          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
4927        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
4928          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) e.getUnfinishedMessage();
4929          throw e;
4930        } finally {
4931          if (parsedMessage != null) {
4932            mergeFrom(parsedMessage);
4933          }
4934        }
4935        return this;
4936      }
4937
4938      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.HeartbeatResponseProto)
4939    }
4940
4941    static {
4942      defaultInstance = new HeartbeatResponseProto(true);
4943      defaultInstance.initFields();
4944    }
4945
4946    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.HeartbeatResponseProto)
4947  }
4948
4949  public interface StartLogSegmentRequestProtoOrBuilder
4950      extends com.google.protobuf.MessageOrBuilder {
4951
4952    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
4953    /**
4954     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4955     */
4956    boolean hasReqInfo();
4957    /**
4958     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4959     */
4960    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
4961    /**
4962     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
4963     */
4964    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
4965
4966    // required uint64 txid = 2;
4967    /**
4968     * <code>required uint64 txid = 2;</code>
4969     *
4970     * <pre>
4971     * Transaction ID
4972     * </pre>
4973     */
4974    boolean hasTxid();
4975    /**
4976     * <code>required uint64 txid = 2;</code>
4977     *
4978     * <pre>
4979     * Transaction ID
4980     * </pre>
4981     */
4982    long getTxid();
4983
4984    // optional sint32 layoutVersion = 3;
4985    /**
4986     * <code>optional sint32 layoutVersion = 3;</code>
4987     *
4988     * <pre>
4989     * the LayoutVersion in the client
4990     * </pre>
4991     */
4992    boolean hasLayoutVersion();
4993    /**
4994     * <code>optional sint32 layoutVersion = 3;</code>
4995     *
4996     * <pre>
4997     * the LayoutVersion in the client
4998     * </pre>
4999     */
5000    int getLayoutVersion();
5001  }
5002  /**
5003   * Protobuf type {@code hadoop.hdfs.qjournal.StartLogSegmentRequestProto}
5004   *
5005   * <pre>
5006   **
5007   * startLogSegment()
5008   * </pre>
5009   */
5010  public static final class StartLogSegmentRequestProto extends
5011      com.google.protobuf.GeneratedMessage
5012      implements StartLogSegmentRequestProtoOrBuilder {
5013    // Use StartLogSegmentRequestProto.newBuilder() to construct.
5014    private StartLogSegmentRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
5015      super(builder);
5016      this.unknownFields = builder.getUnknownFields();
5017    }
5018    private StartLogSegmentRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
5019
5020    private static final StartLogSegmentRequestProto defaultInstance;
5021    public static StartLogSegmentRequestProto getDefaultInstance() {
5022      return defaultInstance;
5023    }
5024
5025    public StartLogSegmentRequestProto getDefaultInstanceForType() {
5026      return defaultInstance;
5027    }
5028
5029    private final com.google.protobuf.UnknownFieldSet unknownFields;
5030    @java.lang.Override
5031    public final com.google.protobuf.UnknownFieldSet
5032        getUnknownFields() {
5033      return this.unknownFields;
5034    }
5035    private StartLogSegmentRequestProto(
5036        com.google.protobuf.CodedInputStream input,
5037        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5038        throws com.google.protobuf.InvalidProtocolBufferException {
5039      initFields();
5040      int mutable_bitField0_ = 0;
5041      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
5042          com.google.protobuf.UnknownFieldSet.newBuilder();
5043      try {
5044        boolean done = false;
5045        while (!done) {
5046          int tag = input.readTag();
5047          switch (tag) {
5048            case 0:
5049              done = true;
5050              break;
5051            default: {
5052              if (!parseUnknownField(input, unknownFields,
5053                                     extensionRegistry, tag)) {
5054                done = true;
5055              }
5056              break;
5057            }
5058            case 10: {
5059              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
5060              if (((bitField0_ & 0x00000001) == 0x00000001)) {
5061                subBuilder = reqInfo_.toBuilder();
5062              }
5063              reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
5064              if (subBuilder != null) {
5065                subBuilder.mergeFrom(reqInfo_);
5066                reqInfo_ = subBuilder.buildPartial();
5067              }
5068              bitField0_ |= 0x00000001;
5069              break;
5070            }
5071            case 16: {
5072              bitField0_ |= 0x00000002;
5073              txid_ = input.readUInt64();
5074              break;
5075            }
5076            case 24: {
5077              bitField0_ |= 0x00000004;
5078              layoutVersion_ = input.readSInt32();
5079              break;
5080            }
5081          }
5082        }
5083      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
5084        throw e.setUnfinishedMessage(this);
5085      } catch (java.io.IOException e) {
5086        throw new com.google.protobuf.InvalidProtocolBufferException(
5087            e.getMessage()).setUnfinishedMessage(this);
5088      } finally {
5089        this.unknownFields = unknownFields.build();
5090        makeExtensionsImmutable();
5091      }
5092    }
5093    public static final com.google.protobuf.Descriptors.Descriptor
5094        getDescriptor() {
5095      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_StartLogSegmentRequestProto_descriptor;
5096    }
5097
5098    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
5099        internalGetFieldAccessorTable() {
5100      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_StartLogSegmentRequestProto_fieldAccessorTable
5101          .ensureFieldAccessorsInitialized(
5102              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.Builder.class);
5103    }
5104
5105    public static com.google.protobuf.Parser<StartLogSegmentRequestProto> PARSER =
5106        new com.google.protobuf.AbstractParser<StartLogSegmentRequestProto>() {
5107      public StartLogSegmentRequestProto parsePartialFrom(
5108          com.google.protobuf.CodedInputStream input,
5109          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5110          throws com.google.protobuf.InvalidProtocolBufferException {
5111        return new StartLogSegmentRequestProto(input, extensionRegistry);
5112      }
5113    };
5114
5115    @java.lang.Override
5116    public com.google.protobuf.Parser<StartLogSegmentRequestProto> getParserForType() {
5117      return PARSER;
5118    }
5119
5120    private int bitField0_;
5121    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
5122    public static final int REQINFO_FIELD_NUMBER = 1;
5123    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
5124    /**
5125     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5126     */
5127    public boolean hasReqInfo() {
5128      return ((bitField0_ & 0x00000001) == 0x00000001);
5129    }
5130    /**
5131     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5132     */
5133    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
5134      return reqInfo_;
5135    }
5136    /**
5137     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5138     */
5139    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
5140      return reqInfo_;
5141    }
5142
5143    // required uint64 txid = 2;
5144    public static final int TXID_FIELD_NUMBER = 2;
5145    private long txid_;
5146    /**
5147     * <code>required uint64 txid = 2;</code>
5148     *
5149     * <pre>
5150     * Transaction ID
5151     * </pre>
5152     */
5153    public boolean hasTxid() {
5154      return ((bitField0_ & 0x00000002) == 0x00000002);
5155    }
5156    /**
5157     * <code>required uint64 txid = 2;</code>
5158     *
5159     * <pre>
5160     * Transaction ID
5161     * </pre>
5162     */
5163    public long getTxid() {
5164      return txid_;
5165    }
5166
5167    // optional sint32 layoutVersion = 3;
5168    public static final int LAYOUTVERSION_FIELD_NUMBER = 3;
5169    private int layoutVersion_;
5170    /**
5171     * <code>optional sint32 layoutVersion = 3;</code>
5172     *
5173     * <pre>
5174     * the LayoutVersion in the client
5175     * </pre>
5176     */
5177    public boolean hasLayoutVersion() {
5178      return ((bitField0_ & 0x00000004) == 0x00000004);
5179    }
5180    /**
5181     * <code>optional sint32 layoutVersion = 3;</code>
5182     *
5183     * <pre>
5184     * the LayoutVersion in the client
5185     * </pre>
5186     */
5187    public int getLayoutVersion() {
5188      return layoutVersion_;
5189    }
5190
5191    private void initFields() {
5192      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5193      txid_ = 0L;
5194      layoutVersion_ = 0;
5195    }
5196    private byte memoizedIsInitialized = -1;
5197    public final boolean isInitialized() {
5198      byte isInitialized = memoizedIsInitialized;
5199      if (isInitialized != -1) return isInitialized == 1;
5200
5201      if (!hasReqInfo()) {
5202        memoizedIsInitialized = 0;
5203        return false;
5204      }
5205      if (!hasTxid()) {
5206        memoizedIsInitialized = 0;
5207        return false;
5208      }
5209      if (!getReqInfo().isInitialized()) {
5210        memoizedIsInitialized = 0;
5211        return false;
5212      }
5213      memoizedIsInitialized = 1;
5214      return true;
5215    }
5216
5217    public void writeTo(com.google.protobuf.CodedOutputStream output)
5218                        throws java.io.IOException {
5219      getSerializedSize();
5220      if (((bitField0_ & 0x00000001) == 0x00000001)) {
5221        output.writeMessage(1, reqInfo_);
5222      }
5223      if (((bitField0_ & 0x00000002) == 0x00000002)) {
5224        output.writeUInt64(2, txid_);
5225      }
5226      if (((bitField0_ & 0x00000004) == 0x00000004)) {
5227        output.writeSInt32(3, layoutVersion_);
5228      }
5229      getUnknownFields().writeTo(output);
5230    }
5231
5232    private int memoizedSerializedSize = -1;
5233    public int getSerializedSize() {
5234      int size = memoizedSerializedSize;
5235      if (size != -1) return size;
5236
5237      size = 0;
5238      if (((bitField0_ & 0x00000001) == 0x00000001)) {
5239        size += com.google.protobuf.CodedOutputStream
5240          .computeMessageSize(1, reqInfo_);
5241      }
5242      if (((bitField0_ & 0x00000002) == 0x00000002)) {
5243        size += com.google.protobuf.CodedOutputStream
5244          .computeUInt64Size(2, txid_);
5245      }
5246      if (((bitField0_ & 0x00000004) == 0x00000004)) {
5247        size += com.google.protobuf.CodedOutputStream
5248          .computeSInt32Size(3, layoutVersion_);
5249      }
5250      size += getUnknownFields().getSerializedSize();
5251      memoizedSerializedSize = size;
5252      return size;
5253    }
5254
5255    private static final long serialVersionUID = 0L;
5256    @java.lang.Override
5257    protected java.lang.Object writeReplace()
5258        throws java.io.ObjectStreamException {
5259      return super.writeReplace();
5260    }
5261
5262    @java.lang.Override
5263    public boolean equals(final java.lang.Object obj) {
5264      if (obj == this) {
5265       return true;
5266      }
5267      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)) {
5268        return super.equals(obj);
5269      }
5270      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto) obj;
5271
5272      boolean result = true;
5273      result = result && (hasReqInfo() == other.hasReqInfo());
5274      if (hasReqInfo()) {
5275        result = result && getReqInfo()
5276            .equals(other.getReqInfo());
5277      }
5278      result = result && (hasTxid() == other.hasTxid());
5279      if (hasTxid()) {
5280        result = result && (getTxid()
5281            == other.getTxid());
5282      }
5283      result = result && (hasLayoutVersion() == other.hasLayoutVersion());
5284      if (hasLayoutVersion()) {
5285        result = result && (getLayoutVersion()
5286            == other.getLayoutVersion());
5287      }
5288      result = result &&
5289          getUnknownFields().equals(other.getUnknownFields());
5290      return result;
5291    }
5292
5293    private int memoizedHashCode = 0;
5294    @java.lang.Override
5295    public int hashCode() {
5296      if (memoizedHashCode != 0) {
5297        return memoizedHashCode;
5298      }
5299      int hash = 41;
5300      hash = (19 * hash) + getDescriptorForType().hashCode();
5301      if (hasReqInfo()) {
5302        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
5303        hash = (53 * hash) + getReqInfo().hashCode();
5304      }
5305      if (hasTxid()) {
5306        hash = (37 * hash) + TXID_FIELD_NUMBER;
5307        hash = (53 * hash) + hashLong(getTxid());
5308      }
5309      if (hasLayoutVersion()) {
5310        hash = (37 * hash) + LAYOUTVERSION_FIELD_NUMBER;
5311        hash = (53 * hash) + getLayoutVersion();
5312      }
5313      hash = (29 * hash) + getUnknownFields().hashCode();
5314      memoizedHashCode = hash;
5315      return hash;
5316    }
5317
5318    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5319        com.google.protobuf.ByteString data)
5320        throws com.google.protobuf.InvalidProtocolBufferException {
5321      return PARSER.parseFrom(data);
5322    }
5323    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5324        com.google.protobuf.ByteString data,
5325        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5326        throws com.google.protobuf.InvalidProtocolBufferException {
5327      return PARSER.parseFrom(data, extensionRegistry);
5328    }
5329    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(byte[] data)
5330        throws com.google.protobuf.InvalidProtocolBufferException {
5331      return PARSER.parseFrom(data);
5332    }
5333    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5334        byte[] data,
5335        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5336        throws com.google.protobuf.InvalidProtocolBufferException {
5337      return PARSER.parseFrom(data, extensionRegistry);
5338    }
5339    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(java.io.InputStream input)
5340        throws java.io.IOException {
5341      return PARSER.parseFrom(input);
5342    }
5343    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5344        java.io.InputStream input,
5345        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5346        throws java.io.IOException {
5347      return PARSER.parseFrom(input, extensionRegistry);
5348    }
5349    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseDelimitedFrom(java.io.InputStream input)
5350        throws java.io.IOException {
5351      return PARSER.parseDelimitedFrom(input);
5352    }
5353    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseDelimitedFrom(
5354        java.io.InputStream input,
5355        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5356        throws java.io.IOException {
5357      return PARSER.parseDelimitedFrom(input, extensionRegistry);
5358    }
5359    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5360        com.google.protobuf.CodedInputStream input)
5361        throws java.io.IOException {
5362      return PARSER.parseFrom(input);
5363    }
5364    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5365        com.google.protobuf.CodedInputStream input,
5366        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5367        throws java.io.IOException {
5368      return PARSER.parseFrom(input, extensionRegistry);
5369    }
5370
5371    public static Builder newBuilder() { return Builder.create(); }
5372    public Builder newBuilderForType() { return newBuilder(); }
5373    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto prototype) {
5374      return newBuilder().mergeFrom(prototype);
5375    }
5376    public Builder toBuilder() { return newBuilder(this); }
5377
5378    @java.lang.Override
5379    protected Builder newBuilderForType(
5380        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
5381      Builder builder = new Builder(parent);
5382      return builder;
5383    }
5384    /**
5385     * Protobuf type {@code hadoop.hdfs.qjournal.StartLogSegmentRequestProto}
5386     *
5387     * <pre>
5388     **
5389     * startLogSegment()
5390     * </pre>
5391     */
5392    public static final class Builder extends
5393        com.google.protobuf.GeneratedMessage.Builder<Builder>
5394       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProtoOrBuilder {
5395      public static final com.google.protobuf.Descriptors.Descriptor
5396          getDescriptor() {
5397        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_StartLogSegmentRequestProto_descriptor;
5398      }
5399
5400      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
5401          internalGetFieldAccessorTable() {
5402        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_StartLogSegmentRequestProto_fieldAccessorTable
5403            .ensureFieldAccessorsInitialized(
5404                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.Builder.class);
5405      }
5406
5407      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.newBuilder()
5408      private Builder() {
5409        maybeForceBuilderInitialization();
5410      }
5411
5412      private Builder(
5413          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
5414        super(parent);
5415        maybeForceBuilderInitialization();
5416      }
5417      private void maybeForceBuilderInitialization() {
5418        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
5419          getReqInfoFieldBuilder();
5420        }
5421      }
5422      private static Builder create() {
5423        return new Builder();
5424      }
5425
5426      public Builder clear() {
5427        super.clear();
5428        if (reqInfoBuilder_ == null) {
5429          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5430        } else {
5431          reqInfoBuilder_.clear();
5432        }
5433        bitField0_ = (bitField0_ & ~0x00000001);
5434        txid_ = 0L;
5435        bitField0_ = (bitField0_ & ~0x00000002);
5436        layoutVersion_ = 0;
5437        bitField0_ = (bitField0_ & ~0x00000004);
5438        return this;
5439      }
5440
5441      public Builder clone() {
5442        return create().mergeFrom(buildPartial());
5443      }
5444
5445      public com.google.protobuf.Descriptors.Descriptor
5446          getDescriptorForType() {
5447        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_StartLogSegmentRequestProto_descriptor;
5448      }
5449
5450      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto getDefaultInstanceForType() {
5451        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance();
5452      }
5453
5454      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto build() {
5455        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto result = buildPartial();
5456        if (!result.isInitialized()) {
5457          throw newUninitializedMessageException(result);
5458        }
5459        return result;
5460      }
5461
5462      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto buildPartial() {
5463        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto(this);
5464        int from_bitField0_ = bitField0_;
5465        int to_bitField0_ = 0;
5466        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
5467          to_bitField0_ |= 0x00000001;
5468        }
5469        if (reqInfoBuilder_ == null) {
5470          result.reqInfo_ = reqInfo_;
5471        } else {
5472          result.reqInfo_ = reqInfoBuilder_.build();
5473        }
5474        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
5475          to_bitField0_ |= 0x00000002;
5476        }
5477        result.txid_ = txid_;
5478        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
5479          to_bitField0_ |= 0x00000004;
5480        }
5481        result.layoutVersion_ = layoutVersion_;
5482        result.bitField0_ = to_bitField0_;
5483        onBuilt();
5484        return result;
5485      }
5486
5487      public Builder mergeFrom(com.google.protobuf.Message other) {
5488        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto) {
5489          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)other);
5490        } else {
5491          super.mergeFrom(other);
5492          return this;
5493        }
5494      }
5495
5496      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto other) {
5497        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance()) return this;
5498        if (other.hasReqInfo()) {
5499          mergeReqInfo(other.getReqInfo());
5500        }
5501        if (other.hasTxid()) {
5502          setTxid(other.getTxid());
5503        }
5504        if (other.hasLayoutVersion()) {
5505          setLayoutVersion(other.getLayoutVersion());
5506        }
5507        this.mergeUnknownFields(other.getUnknownFields());
5508        return this;
5509      }
5510
5511      public final boolean isInitialized() {
5512        if (!hasReqInfo()) {
5513          
5514          return false;
5515        }
5516        if (!hasTxid()) {
5517          
5518          return false;
5519        }
5520        if (!getReqInfo().isInitialized()) {
5521          
5522          return false;
5523        }
5524        return true;
5525      }
5526
5527      public Builder mergeFrom(
5528          com.google.protobuf.CodedInputStream input,
5529          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5530          throws java.io.IOException {
5531        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parsedMessage = null;
5532        try {
5533          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
5534        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
5535          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto) e.getUnfinishedMessage();
5536          throw e;
5537        } finally {
5538          if (parsedMessage != null) {
5539            mergeFrom(parsedMessage);
5540          }
5541        }
5542        return this;
5543      }
5544      private int bitField0_;
5545
5546      // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
5547      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5548      private com.google.protobuf.SingleFieldBuilder<
5549          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
5550      /**
5551       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5552       */
5553      public boolean hasReqInfo() {
5554        return ((bitField0_ & 0x00000001) == 0x00000001);
5555      }
5556      /**
5557       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5558       */
5559      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
5560        if (reqInfoBuilder_ == null) {
5561          return reqInfo_;
5562        } else {
5563          return reqInfoBuilder_.getMessage();
5564        }
5565      }
5566      /**
5567       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5568       */
5569      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
5570        if (reqInfoBuilder_ == null) {
5571          if (value == null) {
5572            throw new NullPointerException();
5573          }
5574          reqInfo_ = value;
5575          onChanged();
5576        } else {
5577          reqInfoBuilder_.setMessage(value);
5578        }
5579        bitField0_ |= 0x00000001;
5580        return this;
5581      }
5582      /**
5583       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5584       */
5585      public Builder setReqInfo(
5586          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
5587        if (reqInfoBuilder_ == null) {
5588          reqInfo_ = builderForValue.build();
5589          onChanged();
5590        } else {
5591          reqInfoBuilder_.setMessage(builderForValue.build());
5592        }
5593        bitField0_ |= 0x00000001;
5594        return this;
5595      }
5596      /**
5597       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5598       */
5599      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
5600        if (reqInfoBuilder_ == null) {
5601          if (((bitField0_ & 0x00000001) == 0x00000001) &&
5602              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
5603            reqInfo_ =
5604              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
5605          } else {
5606            reqInfo_ = value;
5607          }
5608          onChanged();
5609        } else {
5610          reqInfoBuilder_.mergeFrom(value);
5611        }
5612        bitField0_ |= 0x00000001;
5613        return this;
5614      }
5615      /**
5616       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5617       */
5618      public Builder clearReqInfo() {
5619        if (reqInfoBuilder_ == null) {
5620          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5621          onChanged();
5622        } else {
5623          reqInfoBuilder_.clear();
5624        }
5625        bitField0_ = (bitField0_ & ~0x00000001);
5626        return this;
5627      }
5628      /**
5629       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5630       */
5631      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
5632        bitField0_ |= 0x00000001;
5633        onChanged();
5634        return getReqInfoFieldBuilder().getBuilder();
5635      }
5636      /**
5637       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5638       */
5639      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
5640        if (reqInfoBuilder_ != null) {
5641          return reqInfoBuilder_.getMessageOrBuilder();
5642        } else {
5643          return reqInfo_;
5644        }
5645      }
5646      /**
5647       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
5648       */
5649      private com.google.protobuf.SingleFieldBuilder<
5650          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
5651          getReqInfoFieldBuilder() {
5652        if (reqInfoBuilder_ == null) {
5653          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
5654              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
5655                  reqInfo_,
5656                  getParentForChildren(),
5657                  isClean());
5658          reqInfo_ = null;
5659        }
5660        return reqInfoBuilder_;
5661      }
5662
5663      // required uint64 txid = 2;
5664      private long txid_ ;
5665      /**
5666       * <code>required uint64 txid = 2;</code>
5667       *
5668       * <pre>
5669       * Transaction ID
5670       * </pre>
5671       */
5672      public boolean hasTxid() {
5673        return ((bitField0_ & 0x00000002) == 0x00000002);
5674      }
5675      /**
5676       * <code>required uint64 txid = 2;</code>
5677       *
5678       * <pre>
5679       * Transaction ID
5680       * </pre>
5681       */
5682      public long getTxid() {
5683        return txid_;
5684      }
5685      /**
5686       * <code>required uint64 txid = 2;</code>
5687       *
5688       * <pre>
5689       * Transaction ID
5690       * </pre>
5691       */
5692      public Builder setTxid(long value) {
5693        bitField0_ |= 0x00000002;
5694        txid_ = value;
5695        onChanged();
5696        return this;
5697      }
5698      /**
5699       * <code>required uint64 txid = 2;</code>
5700       *
5701       * <pre>
5702       * Transaction ID
5703       * </pre>
5704       */
5705      public Builder clearTxid() {
5706        bitField0_ = (bitField0_ & ~0x00000002);
5707        txid_ = 0L;
5708        onChanged();
5709        return this;
5710      }
5711
5712      // optional sint32 layoutVersion = 3;
5713      private int layoutVersion_ ;
5714      /**
5715       * <code>optional sint32 layoutVersion = 3;</code>
5716       *
5717       * <pre>
5718       * the LayoutVersion in the client
5719       * </pre>
5720       */
5721      public boolean hasLayoutVersion() {
5722        return ((bitField0_ & 0x00000004) == 0x00000004);
5723      }
5724      /**
5725       * <code>optional sint32 layoutVersion = 3;</code>
5726       *
5727       * <pre>
5728       * the LayoutVersion in the client
5729       * </pre>
5730       */
5731      public int getLayoutVersion() {
5732        return layoutVersion_;
5733      }
5734      /**
5735       * <code>optional sint32 layoutVersion = 3;</code>
5736       *
5737       * <pre>
5738       * the LayoutVersion in the client
5739       * </pre>
5740       */
5741      public Builder setLayoutVersion(int value) {
5742        bitField0_ |= 0x00000004;
5743        layoutVersion_ = value;
5744        onChanged();
5745        return this;
5746      }
5747      /**
5748       * <code>optional sint32 layoutVersion = 3;</code>
5749       *
5750       * <pre>
5751       * the LayoutVersion in the client
5752       * </pre>
5753       */
5754      public Builder clearLayoutVersion() {
5755        bitField0_ = (bitField0_ & ~0x00000004);
5756        layoutVersion_ = 0;
5757        onChanged();
5758        return this;
5759      }
5760
5761      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.StartLogSegmentRequestProto)
5762    }
5763
5764    static {
5765      defaultInstance = new StartLogSegmentRequestProto(true);
5766      defaultInstance.initFields();
5767    }
5768
5769    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.StartLogSegmentRequestProto)
5770  }
5771
5772  public interface StartLogSegmentResponseProtoOrBuilder
5773      extends com.google.protobuf.MessageOrBuilder {
5774  }
5775  /**
5776   * Protobuf type {@code hadoop.hdfs.qjournal.StartLogSegmentResponseProto}
5777   */
5778  public static final class StartLogSegmentResponseProto extends
5779      com.google.protobuf.GeneratedMessage
5780      implements StartLogSegmentResponseProtoOrBuilder {
5781    // Use StartLogSegmentResponseProto.newBuilder() to construct.
5782    private StartLogSegmentResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
5783      super(builder);
5784      this.unknownFields = builder.getUnknownFields();
5785    }
5786    private StartLogSegmentResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
5787
5788    private static final StartLogSegmentResponseProto defaultInstance;
5789    public static StartLogSegmentResponseProto getDefaultInstance() {
5790      return defaultInstance;
5791    }
5792
5793    public StartLogSegmentResponseProto getDefaultInstanceForType() {
5794      return defaultInstance;
5795    }
5796
5797    private final com.google.protobuf.UnknownFieldSet unknownFields;
5798    @java.lang.Override
5799    public final com.google.protobuf.UnknownFieldSet
5800        getUnknownFields() {
5801      return this.unknownFields;
5802    }
5803    private StartLogSegmentResponseProto(
5804        com.google.protobuf.CodedInputStream input,
5805        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5806        throws com.google.protobuf.InvalidProtocolBufferException {
5807      initFields();
5808      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
5809          com.google.protobuf.UnknownFieldSet.newBuilder();
5810      try {
5811        boolean done = false;
5812        while (!done) {
5813          int tag = input.readTag();
5814          switch (tag) {
5815            case 0:
5816              done = true;
5817              break;
5818            default: {
5819              if (!parseUnknownField(input, unknownFields,
5820                                     extensionRegistry, tag)) {
5821                done = true;
5822              }
5823              break;
5824            }
5825          }
5826        }
5827      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
5828        throw e.setUnfinishedMessage(this);
5829      } catch (java.io.IOException e) {
5830        throw new com.google.protobuf.InvalidProtocolBufferException(
5831            e.getMessage()).setUnfinishedMessage(this);
5832      } finally {
5833        this.unknownFields = unknownFields.build();
5834        makeExtensionsImmutable();
5835      }
5836    }
5837    public static final com.google.protobuf.Descriptors.Descriptor
5838        getDescriptor() {
5839      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_StartLogSegmentResponseProto_descriptor;
5840    }
5841
5842    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
5843        internalGetFieldAccessorTable() {
5844      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_StartLogSegmentResponseProto_fieldAccessorTable
5845          .ensureFieldAccessorsInitialized(
5846              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.Builder.class);
5847    }
5848
5849    public static com.google.protobuf.Parser<StartLogSegmentResponseProto> PARSER =
5850        new com.google.protobuf.AbstractParser<StartLogSegmentResponseProto>() {
5851      public StartLogSegmentResponseProto parsePartialFrom(
5852          com.google.protobuf.CodedInputStream input,
5853          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5854          throws com.google.protobuf.InvalidProtocolBufferException {
5855        return new StartLogSegmentResponseProto(input, extensionRegistry);
5856      }
5857    };
5858
5859    @java.lang.Override
5860    public com.google.protobuf.Parser<StartLogSegmentResponseProto> getParserForType() {
5861      return PARSER;
5862    }
5863
5864    private void initFields() {
5865    }
5866    private byte memoizedIsInitialized = -1;
5867    public final boolean isInitialized() {
5868      byte isInitialized = memoizedIsInitialized;
5869      if (isInitialized != -1) return isInitialized == 1;
5870
5871      memoizedIsInitialized = 1;
5872      return true;
5873    }
5874
5875    public void writeTo(com.google.protobuf.CodedOutputStream output)
5876                        throws java.io.IOException {
5877      getSerializedSize();
5878      getUnknownFields().writeTo(output);
5879    }
5880
5881    private int memoizedSerializedSize = -1;
5882    public int getSerializedSize() {
5883      int size = memoizedSerializedSize;
5884      if (size != -1) return size;
5885
5886      size = 0;
5887      size += getUnknownFields().getSerializedSize();
5888      memoizedSerializedSize = size;
5889      return size;
5890    }
5891
5892    private static final long serialVersionUID = 0L;
5893    @java.lang.Override
5894    protected java.lang.Object writeReplace()
5895        throws java.io.ObjectStreamException {
5896      return super.writeReplace();
5897    }
5898
5899    @java.lang.Override
5900    public boolean equals(final java.lang.Object obj) {
5901      if (obj == this) {
5902       return true;
5903      }
5904      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto)) {
5905        return super.equals(obj);
5906      }
5907      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) obj;
5908
5909      boolean result = true;
5910      result = result &&
5911          getUnknownFields().equals(other.getUnknownFields());
5912      return result;
5913    }
5914
5915    private int memoizedHashCode = 0;
5916    @java.lang.Override
5917    public int hashCode() {
5918      if (memoizedHashCode != 0) {
5919        return memoizedHashCode;
5920      }
5921      int hash = 41;
5922      hash = (19 * hash) + getDescriptorForType().hashCode();
5923      hash = (29 * hash) + getUnknownFields().hashCode();
5924      memoizedHashCode = hash;
5925      return hash;
5926    }
5927
5928    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5929        com.google.protobuf.ByteString data)
5930        throws com.google.protobuf.InvalidProtocolBufferException {
5931      return PARSER.parseFrom(data);
5932    }
5933    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5934        com.google.protobuf.ByteString data,
5935        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5936        throws com.google.protobuf.InvalidProtocolBufferException {
5937      return PARSER.parseFrom(data, extensionRegistry);
5938    }
5939    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(byte[] data)
5940        throws com.google.protobuf.InvalidProtocolBufferException {
5941      return PARSER.parseFrom(data);
5942    }
5943    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5944        byte[] data,
5945        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5946        throws com.google.protobuf.InvalidProtocolBufferException {
5947      return PARSER.parseFrom(data, extensionRegistry);
5948    }
5949    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(java.io.InputStream input)
5950        throws java.io.IOException {
5951      return PARSER.parseFrom(input);
5952    }
5953    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5954        java.io.InputStream input,
5955        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5956        throws java.io.IOException {
5957      return PARSER.parseFrom(input, extensionRegistry);
5958    }
5959    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseDelimitedFrom(java.io.InputStream input)
5960        throws java.io.IOException {
5961      return PARSER.parseDelimitedFrom(input);
5962    }
5963    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseDelimitedFrom(
5964        java.io.InputStream input,
5965        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5966        throws java.io.IOException {
5967      return PARSER.parseDelimitedFrom(input, extensionRegistry);
5968    }
5969    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5970        com.google.protobuf.CodedInputStream input)
5971        throws java.io.IOException {
5972      return PARSER.parseFrom(input);
5973    }
5974    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5975        com.google.protobuf.CodedInputStream input,
5976        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5977        throws java.io.IOException {
5978      return PARSER.parseFrom(input, extensionRegistry);
5979    }
5980
5981    public static Builder newBuilder() { return Builder.create(); }
5982    public Builder newBuilderForType() { return newBuilder(); }
5983    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto prototype) {
5984      return newBuilder().mergeFrom(prototype);
5985    }
5986    public Builder toBuilder() { return newBuilder(this); }
5987
5988    @java.lang.Override
5989    protected Builder newBuilderForType(
5990        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
5991      Builder builder = new Builder(parent);
5992      return builder;
5993    }
5994    /**
5995     * Protobuf type {@code hadoop.hdfs.qjournal.StartLogSegmentResponseProto}
5996     */
5997    public static final class Builder extends
5998        com.google.protobuf.GeneratedMessage.Builder<Builder>
5999       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProtoOrBuilder {
6000      public static final com.google.protobuf.Descriptors.Descriptor
6001          getDescriptor() {
6002        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_StartLogSegmentResponseProto_descriptor;
6003      }
6004
6005      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6006          internalGetFieldAccessorTable() {
6007        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_StartLogSegmentResponseProto_fieldAccessorTable
6008            .ensureFieldAccessorsInitialized(
6009                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.Builder.class);
6010      }
6011
6012      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.newBuilder()
6013      private Builder() {
6014        maybeForceBuilderInitialization();
6015      }
6016
6017      private Builder(
6018          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
6019        super(parent);
6020        maybeForceBuilderInitialization();
6021      }
6022      private void maybeForceBuilderInitialization() {
6023        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
6024        }
6025      }
6026      private static Builder create() {
6027        return new Builder();
6028      }
6029
6030      public Builder clear() {
6031        super.clear();
6032        return this;
6033      }
6034
6035      public Builder clone() {
6036        return create().mergeFrom(buildPartial());
6037      }
6038
6039      public com.google.protobuf.Descriptors.Descriptor
6040          getDescriptorForType() {
6041        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_StartLogSegmentResponseProto_descriptor;
6042      }
6043
6044      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto getDefaultInstanceForType() {
6045        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance();
6046      }
6047
6048      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto build() {
6049        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto result = buildPartial();
6050        if (!result.isInitialized()) {
6051          throw newUninitializedMessageException(result);
6052        }
6053        return result;
6054      }
6055
6056      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto buildPartial() {
6057        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto(this);
6058        onBuilt();
6059        return result;
6060      }
6061
6062      public Builder mergeFrom(com.google.protobuf.Message other) {
6063        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) {
6064          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto)other);
6065        } else {
6066          super.mergeFrom(other);
6067          return this;
6068        }
6069      }
6070
6071      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto other) {
6072        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance()) return this;
6073        this.mergeUnknownFields(other.getUnknownFields());
6074        return this;
6075      }
6076
6077      public final boolean isInitialized() {
6078        return true;
6079      }
6080
6081      public Builder mergeFrom(
6082          com.google.protobuf.CodedInputStream input,
6083          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6084          throws java.io.IOException {
6085        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parsedMessage = null;
6086        try {
6087          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
6088        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
6089          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) e.getUnfinishedMessage();
6090          throw e;
6091        } finally {
6092          if (parsedMessage != null) {
6093            mergeFrom(parsedMessage);
6094          }
6095        }
6096        return this;
6097      }
6098
6099      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.StartLogSegmentResponseProto)
6100    }
6101
6102    static {
6103      defaultInstance = new StartLogSegmentResponseProto(true);
6104      defaultInstance.initFields();
6105    }
6106
6107    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.StartLogSegmentResponseProto)
6108  }
6109
6110  public interface FinalizeLogSegmentRequestProtoOrBuilder
6111      extends com.google.protobuf.MessageOrBuilder {
6112
6113    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
6114    /**
6115     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6116     */
6117    boolean hasReqInfo();
6118    /**
6119     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6120     */
6121    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
6122    /**
6123     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6124     */
6125    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
6126
6127    // required uint64 startTxId = 2;
6128    /**
6129     * <code>required uint64 startTxId = 2;</code>
6130     */
6131    boolean hasStartTxId();
6132    /**
6133     * <code>required uint64 startTxId = 2;</code>
6134     */
6135    long getStartTxId();
6136
6137    // required uint64 endTxId = 3;
6138    /**
6139     * <code>required uint64 endTxId = 3;</code>
6140     */
6141    boolean hasEndTxId();
6142    /**
6143     * <code>required uint64 endTxId = 3;</code>
6144     */
6145    long getEndTxId();
6146  }
6147  /**
6148   * Protobuf type {@code hadoop.hdfs.qjournal.FinalizeLogSegmentRequestProto}
6149   *
6150   * <pre>
6151   **
6152   * finalizeLogSegment()
6153   * </pre>
6154   */
6155  public static final class FinalizeLogSegmentRequestProto extends
6156      com.google.protobuf.GeneratedMessage
6157      implements FinalizeLogSegmentRequestProtoOrBuilder {
6158    // Use FinalizeLogSegmentRequestProto.newBuilder() to construct.
6159    private FinalizeLogSegmentRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
6160      super(builder);
6161      this.unknownFields = builder.getUnknownFields();
6162    }
6163    private FinalizeLogSegmentRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
6164
6165    private static final FinalizeLogSegmentRequestProto defaultInstance;
6166    public static FinalizeLogSegmentRequestProto getDefaultInstance() {
6167      return defaultInstance;
6168    }
6169
6170    public FinalizeLogSegmentRequestProto getDefaultInstanceForType() {
6171      return defaultInstance;
6172    }
6173
6174    private final com.google.protobuf.UnknownFieldSet unknownFields;
6175    @java.lang.Override
6176    public final com.google.protobuf.UnknownFieldSet
6177        getUnknownFields() {
6178      return this.unknownFields;
6179    }
6180    private FinalizeLogSegmentRequestProto(
6181        com.google.protobuf.CodedInputStream input,
6182        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6183        throws com.google.protobuf.InvalidProtocolBufferException {
6184      initFields();
6185      int mutable_bitField0_ = 0;
6186      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
6187          com.google.protobuf.UnknownFieldSet.newBuilder();
6188      try {
6189        boolean done = false;
6190        while (!done) {
6191          int tag = input.readTag();
6192          switch (tag) {
6193            case 0:
6194              done = true;
6195              break;
6196            default: {
6197              if (!parseUnknownField(input, unknownFields,
6198                                     extensionRegistry, tag)) {
6199                done = true;
6200              }
6201              break;
6202            }
6203            case 10: {
6204              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
6205              if (((bitField0_ & 0x00000001) == 0x00000001)) {
6206                subBuilder = reqInfo_.toBuilder();
6207              }
6208              reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
6209              if (subBuilder != null) {
6210                subBuilder.mergeFrom(reqInfo_);
6211                reqInfo_ = subBuilder.buildPartial();
6212              }
6213              bitField0_ |= 0x00000001;
6214              break;
6215            }
6216            case 16: {
6217              bitField0_ |= 0x00000002;
6218              startTxId_ = input.readUInt64();
6219              break;
6220            }
6221            case 24: {
6222              bitField0_ |= 0x00000004;
6223              endTxId_ = input.readUInt64();
6224              break;
6225            }
6226          }
6227        }
6228      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
6229        throw e.setUnfinishedMessage(this);
6230      } catch (java.io.IOException e) {
6231        throw new com.google.protobuf.InvalidProtocolBufferException(
6232            e.getMessage()).setUnfinishedMessage(this);
6233      } finally {
6234        this.unknownFields = unknownFields.build();
6235        makeExtensionsImmutable();
6236      }
6237    }
6238    public static final com.google.protobuf.Descriptors.Descriptor
6239        getDescriptor() {
6240      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentRequestProto_descriptor;
6241    }
6242
6243    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6244        internalGetFieldAccessorTable() {
6245      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentRequestProto_fieldAccessorTable
6246          .ensureFieldAccessorsInitialized(
6247              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.Builder.class);
6248    }
6249
6250    public static com.google.protobuf.Parser<FinalizeLogSegmentRequestProto> PARSER =
6251        new com.google.protobuf.AbstractParser<FinalizeLogSegmentRequestProto>() {
6252      public FinalizeLogSegmentRequestProto parsePartialFrom(
6253          com.google.protobuf.CodedInputStream input,
6254          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6255          throws com.google.protobuf.InvalidProtocolBufferException {
6256        return new FinalizeLogSegmentRequestProto(input, extensionRegistry);
6257      }
6258    };
6259
6260    @java.lang.Override
6261    public com.google.protobuf.Parser<FinalizeLogSegmentRequestProto> getParserForType() {
6262      return PARSER;
6263    }
6264
6265    private int bitField0_;
6266    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
6267    public static final int REQINFO_FIELD_NUMBER = 1;
6268    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
6269    /**
6270     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6271     */
6272    public boolean hasReqInfo() {
6273      return ((bitField0_ & 0x00000001) == 0x00000001);
6274    }
6275    /**
6276     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6277     */
6278    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
6279      return reqInfo_;
6280    }
6281    /**
6282     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6283     */
6284    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
6285      return reqInfo_;
6286    }
6287
6288    // required uint64 startTxId = 2;
6289    public static final int STARTTXID_FIELD_NUMBER = 2;
6290    private long startTxId_;
6291    /**
6292     * <code>required uint64 startTxId = 2;</code>
6293     */
6294    public boolean hasStartTxId() {
6295      return ((bitField0_ & 0x00000002) == 0x00000002);
6296    }
6297    /**
6298     * <code>required uint64 startTxId = 2;</code>
6299     */
6300    public long getStartTxId() {
6301      return startTxId_;
6302    }
6303
6304    // required uint64 endTxId = 3;
6305    public static final int ENDTXID_FIELD_NUMBER = 3;
6306    private long endTxId_;
6307    /**
6308     * <code>required uint64 endTxId = 3;</code>
6309     */
6310    public boolean hasEndTxId() {
6311      return ((bitField0_ & 0x00000004) == 0x00000004);
6312    }
6313    /**
6314     * <code>required uint64 endTxId = 3;</code>
6315     */
6316    public long getEndTxId() {
6317      return endTxId_;
6318    }
6319
6320    private void initFields() {
6321      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
6322      startTxId_ = 0L;
6323      endTxId_ = 0L;
6324    }
6325    private byte memoizedIsInitialized = -1;
6326    public final boolean isInitialized() {
6327      byte isInitialized = memoizedIsInitialized;
6328      if (isInitialized != -1) return isInitialized == 1;
6329
6330      if (!hasReqInfo()) {
6331        memoizedIsInitialized = 0;
6332        return false;
6333      }
6334      if (!hasStartTxId()) {
6335        memoizedIsInitialized = 0;
6336        return false;
6337      }
6338      if (!hasEndTxId()) {
6339        memoizedIsInitialized = 0;
6340        return false;
6341      }
6342      if (!getReqInfo().isInitialized()) {
6343        memoizedIsInitialized = 0;
6344        return false;
6345      }
6346      memoizedIsInitialized = 1;
6347      return true;
6348    }
6349
6350    public void writeTo(com.google.protobuf.CodedOutputStream output)
6351                        throws java.io.IOException {
6352      getSerializedSize();
6353      if (((bitField0_ & 0x00000001) == 0x00000001)) {
6354        output.writeMessage(1, reqInfo_);
6355      }
6356      if (((bitField0_ & 0x00000002) == 0x00000002)) {
6357        output.writeUInt64(2, startTxId_);
6358      }
6359      if (((bitField0_ & 0x00000004) == 0x00000004)) {
6360        output.writeUInt64(3, endTxId_);
6361      }
6362      getUnknownFields().writeTo(output);
6363    }
6364
6365    private int memoizedSerializedSize = -1;
6366    public int getSerializedSize() {
6367      int size = memoizedSerializedSize;
6368      if (size != -1) return size;
6369
6370      size = 0;
6371      if (((bitField0_ & 0x00000001) == 0x00000001)) {
6372        size += com.google.protobuf.CodedOutputStream
6373          .computeMessageSize(1, reqInfo_);
6374      }
6375      if (((bitField0_ & 0x00000002) == 0x00000002)) {
6376        size += com.google.protobuf.CodedOutputStream
6377          .computeUInt64Size(2, startTxId_);
6378      }
6379      if (((bitField0_ & 0x00000004) == 0x00000004)) {
6380        size += com.google.protobuf.CodedOutputStream
6381          .computeUInt64Size(3, endTxId_);
6382      }
6383      size += getUnknownFields().getSerializedSize();
6384      memoizedSerializedSize = size;
6385      return size;
6386    }
6387
6388    private static final long serialVersionUID = 0L;
6389    @java.lang.Override
6390    protected java.lang.Object writeReplace()
6391        throws java.io.ObjectStreamException {
6392      return super.writeReplace();
6393    }
6394
6395    @java.lang.Override
6396    public boolean equals(final java.lang.Object obj) {
6397      if (obj == this) {
6398       return true;
6399      }
6400      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)) {
6401        return super.equals(obj);
6402      }
6403      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto) obj;
6404
6405      boolean result = true;
6406      result = result && (hasReqInfo() == other.hasReqInfo());
6407      if (hasReqInfo()) {
6408        result = result && getReqInfo()
6409            .equals(other.getReqInfo());
6410      }
6411      result = result && (hasStartTxId() == other.hasStartTxId());
6412      if (hasStartTxId()) {
6413        result = result && (getStartTxId()
6414            == other.getStartTxId());
6415      }
6416      result = result && (hasEndTxId() == other.hasEndTxId());
6417      if (hasEndTxId()) {
6418        result = result && (getEndTxId()
6419            == other.getEndTxId());
6420      }
6421      result = result &&
6422          getUnknownFields().equals(other.getUnknownFields());
6423      return result;
6424    }
6425
6426    private int memoizedHashCode = 0;
6427    @java.lang.Override
6428    public int hashCode() {
6429      if (memoizedHashCode != 0) {
6430        return memoizedHashCode;
6431      }
6432      int hash = 41;
6433      hash = (19 * hash) + getDescriptorForType().hashCode();
6434      if (hasReqInfo()) {
6435        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
6436        hash = (53 * hash) + getReqInfo().hashCode();
6437      }
6438      if (hasStartTxId()) {
6439        hash = (37 * hash) + STARTTXID_FIELD_NUMBER;
6440        hash = (53 * hash) + hashLong(getStartTxId());
6441      }
6442      if (hasEndTxId()) {
6443        hash = (37 * hash) + ENDTXID_FIELD_NUMBER;
6444        hash = (53 * hash) + hashLong(getEndTxId());
6445      }
6446      hash = (29 * hash) + getUnknownFields().hashCode();
6447      memoizedHashCode = hash;
6448      return hash;
6449    }
6450
6451    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6452        com.google.protobuf.ByteString data)
6453        throws com.google.protobuf.InvalidProtocolBufferException {
6454      return PARSER.parseFrom(data);
6455    }
6456    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6457        com.google.protobuf.ByteString data,
6458        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6459        throws com.google.protobuf.InvalidProtocolBufferException {
6460      return PARSER.parseFrom(data, extensionRegistry);
6461    }
6462    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(byte[] data)
6463        throws com.google.protobuf.InvalidProtocolBufferException {
6464      return PARSER.parseFrom(data);
6465    }
6466    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6467        byte[] data,
6468        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6469        throws com.google.protobuf.InvalidProtocolBufferException {
6470      return PARSER.parseFrom(data, extensionRegistry);
6471    }
6472    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(java.io.InputStream input)
6473        throws java.io.IOException {
6474      return PARSER.parseFrom(input);
6475    }
6476    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6477        java.io.InputStream input,
6478        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6479        throws java.io.IOException {
6480      return PARSER.parseFrom(input, extensionRegistry);
6481    }
6482    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseDelimitedFrom(java.io.InputStream input)
6483        throws java.io.IOException {
6484      return PARSER.parseDelimitedFrom(input);
6485    }
6486    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseDelimitedFrom(
6487        java.io.InputStream input,
6488        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6489        throws java.io.IOException {
6490      return PARSER.parseDelimitedFrom(input, extensionRegistry);
6491    }
6492    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6493        com.google.protobuf.CodedInputStream input)
6494        throws java.io.IOException {
6495      return PARSER.parseFrom(input);
6496    }
6497    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6498        com.google.protobuf.CodedInputStream input,
6499        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6500        throws java.io.IOException {
6501      return PARSER.parseFrom(input, extensionRegistry);
6502    }
6503
6504    public static Builder newBuilder() { return Builder.create(); }
6505    public Builder newBuilderForType() { return newBuilder(); }
6506    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto prototype) {
6507      return newBuilder().mergeFrom(prototype);
6508    }
6509    public Builder toBuilder() { return newBuilder(this); }
6510
6511    @java.lang.Override
6512    protected Builder newBuilderForType(
6513        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
6514      Builder builder = new Builder(parent);
6515      return builder;
6516    }
6517    /**
6518     * Protobuf type {@code hadoop.hdfs.qjournal.FinalizeLogSegmentRequestProto}
6519     *
6520     * <pre>
6521     **
6522     * finalizeLogSegment()
6523     * </pre>
6524     */
6525    public static final class Builder extends
6526        com.google.protobuf.GeneratedMessage.Builder<Builder>
6527       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProtoOrBuilder {
6528      public static final com.google.protobuf.Descriptors.Descriptor
6529          getDescriptor() {
6530        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentRequestProto_descriptor;
6531      }
6532
6533      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6534          internalGetFieldAccessorTable() {
6535        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentRequestProto_fieldAccessorTable
6536            .ensureFieldAccessorsInitialized(
6537                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.Builder.class);
6538      }
6539
6540      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.newBuilder()
6541      private Builder() {
6542        maybeForceBuilderInitialization();
6543      }
6544
6545      private Builder(
6546          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
6547        super(parent);
6548        maybeForceBuilderInitialization();
6549      }
6550      private void maybeForceBuilderInitialization() {
6551        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
6552          getReqInfoFieldBuilder();
6553        }
6554      }
6555      private static Builder create() {
6556        return new Builder();
6557      }
6558
6559      public Builder clear() {
6560        super.clear();
6561        if (reqInfoBuilder_ == null) {
6562          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
6563        } else {
6564          reqInfoBuilder_.clear();
6565        }
6566        bitField0_ = (bitField0_ & ~0x00000001);
6567        startTxId_ = 0L;
6568        bitField0_ = (bitField0_ & ~0x00000002);
6569        endTxId_ = 0L;
6570        bitField0_ = (bitField0_ & ~0x00000004);
6571        return this;
6572      }
6573
6574      public Builder clone() {
6575        return create().mergeFrom(buildPartial());
6576      }
6577
6578      public com.google.protobuf.Descriptors.Descriptor
6579          getDescriptorForType() {
6580        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentRequestProto_descriptor;
6581      }
6582
6583      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto getDefaultInstanceForType() {
6584        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance();
6585      }
6586
6587      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto build() {
6588        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto result = buildPartial();
6589        if (!result.isInitialized()) {
6590          throw newUninitializedMessageException(result);
6591        }
6592        return result;
6593      }
6594
6595      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto buildPartial() {
6596        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto(this);
6597        int from_bitField0_ = bitField0_;
6598        int to_bitField0_ = 0;
6599        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
6600          to_bitField0_ |= 0x00000001;
6601        }
6602        if (reqInfoBuilder_ == null) {
6603          result.reqInfo_ = reqInfo_;
6604        } else {
6605          result.reqInfo_ = reqInfoBuilder_.build();
6606        }
6607        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
6608          to_bitField0_ |= 0x00000002;
6609        }
6610        result.startTxId_ = startTxId_;
6611        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
6612          to_bitField0_ |= 0x00000004;
6613        }
6614        result.endTxId_ = endTxId_;
6615        result.bitField0_ = to_bitField0_;
6616        onBuilt();
6617        return result;
6618      }
6619
6620      public Builder mergeFrom(com.google.protobuf.Message other) {
6621        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto) {
6622          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)other);
6623        } else {
6624          super.mergeFrom(other);
6625          return this;
6626        }
6627      }
6628
6629      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto other) {
6630        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance()) return this;
6631        if (other.hasReqInfo()) {
6632          mergeReqInfo(other.getReqInfo());
6633        }
6634        if (other.hasStartTxId()) {
6635          setStartTxId(other.getStartTxId());
6636        }
6637        if (other.hasEndTxId()) {
6638          setEndTxId(other.getEndTxId());
6639        }
6640        this.mergeUnknownFields(other.getUnknownFields());
6641        return this;
6642      }
6643
6644      public final boolean isInitialized() {
6645        if (!hasReqInfo()) {
6646          
6647          return false;
6648        }
6649        if (!hasStartTxId()) {
6650          
6651          return false;
6652        }
6653        if (!hasEndTxId()) {
6654          
6655          return false;
6656        }
6657        if (!getReqInfo().isInitialized()) {
6658          
6659          return false;
6660        }
6661        return true;
6662      }
6663
6664      public Builder mergeFrom(
6665          com.google.protobuf.CodedInputStream input,
6666          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6667          throws java.io.IOException {
6668        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parsedMessage = null;
6669        try {
6670          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
6671        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
6672          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto) e.getUnfinishedMessage();
6673          throw e;
6674        } finally {
6675          if (parsedMessage != null) {
6676            mergeFrom(parsedMessage);
6677          }
6678        }
6679        return this;
6680      }
6681      private int bitField0_;
6682
6683      // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
6684      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
6685      private com.google.protobuf.SingleFieldBuilder<
6686          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
6687      /**
6688       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6689       */
6690      public boolean hasReqInfo() {
6691        return ((bitField0_ & 0x00000001) == 0x00000001);
6692      }
6693      /**
6694       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6695       */
6696      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
6697        if (reqInfoBuilder_ == null) {
6698          return reqInfo_;
6699        } else {
6700          return reqInfoBuilder_.getMessage();
6701        }
6702      }
6703      /**
6704       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6705       */
6706      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
6707        if (reqInfoBuilder_ == null) {
6708          if (value == null) {
6709            throw new NullPointerException();
6710          }
6711          reqInfo_ = value;
6712          onChanged();
6713        } else {
6714          reqInfoBuilder_.setMessage(value);
6715        }
6716        bitField0_ |= 0x00000001;
6717        return this;
6718      }
6719      /**
6720       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6721       */
6722      public Builder setReqInfo(
6723          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
6724        if (reqInfoBuilder_ == null) {
6725          reqInfo_ = builderForValue.build();
6726          onChanged();
6727        } else {
6728          reqInfoBuilder_.setMessage(builderForValue.build());
6729        }
6730        bitField0_ |= 0x00000001;
6731        return this;
6732      }
6733      /**
6734       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6735       */
6736      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
6737        if (reqInfoBuilder_ == null) {
6738          if (((bitField0_ & 0x00000001) == 0x00000001) &&
6739              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
6740            reqInfo_ =
6741              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
6742          } else {
6743            reqInfo_ = value;
6744          }
6745          onChanged();
6746        } else {
6747          reqInfoBuilder_.mergeFrom(value);
6748        }
6749        bitField0_ |= 0x00000001;
6750        return this;
6751      }
6752      /**
6753       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6754       */
6755      public Builder clearReqInfo() {
6756        if (reqInfoBuilder_ == null) {
6757          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
6758          onChanged();
6759        } else {
6760          reqInfoBuilder_.clear();
6761        }
6762        bitField0_ = (bitField0_ & ~0x00000001);
6763        return this;
6764      }
6765      /**
6766       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6767       */
6768      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
6769        bitField0_ |= 0x00000001;
6770        onChanged();
6771        return getReqInfoFieldBuilder().getBuilder();
6772      }
6773      /**
6774       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6775       */
6776      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
6777        if (reqInfoBuilder_ != null) {
6778          return reqInfoBuilder_.getMessageOrBuilder();
6779        } else {
6780          return reqInfo_;
6781        }
6782      }
6783      /**
6784       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
6785       */
6786      private com.google.protobuf.SingleFieldBuilder<
6787          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
6788          getReqInfoFieldBuilder() {
6789        if (reqInfoBuilder_ == null) {
6790          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
6791              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
6792                  reqInfo_,
6793                  getParentForChildren(),
6794                  isClean());
6795          reqInfo_ = null;
6796        }
6797        return reqInfoBuilder_;
6798      }
6799
6800      // required uint64 startTxId = 2;
6801      private long startTxId_ ;
6802      /**
6803       * <code>required uint64 startTxId = 2;</code>
6804       */
6805      public boolean hasStartTxId() {
6806        return ((bitField0_ & 0x00000002) == 0x00000002);
6807      }
6808      /**
6809       * <code>required uint64 startTxId = 2;</code>
6810       */
6811      public long getStartTxId() {
6812        return startTxId_;
6813      }
6814      /**
6815       * <code>required uint64 startTxId = 2;</code>
6816       */
6817      public Builder setStartTxId(long value) {
6818        bitField0_ |= 0x00000002;
6819        startTxId_ = value;
6820        onChanged();
6821        return this;
6822      }
6823      /**
6824       * <code>required uint64 startTxId = 2;</code>
6825       */
6826      public Builder clearStartTxId() {
6827        bitField0_ = (bitField0_ & ~0x00000002);
6828        startTxId_ = 0L;
6829        onChanged();
6830        return this;
6831      }
6832
6833      // required uint64 endTxId = 3;
6834      private long endTxId_ ;
6835      /**
6836       * <code>required uint64 endTxId = 3;</code>
6837       */
6838      public boolean hasEndTxId() {
6839        return ((bitField0_ & 0x00000004) == 0x00000004);
6840      }
6841      /**
6842       * <code>required uint64 endTxId = 3;</code>
6843       */
6844      public long getEndTxId() {
6845        return endTxId_;
6846      }
6847      /**
6848       * <code>required uint64 endTxId = 3;</code>
6849       */
6850      public Builder setEndTxId(long value) {
6851        bitField0_ |= 0x00000004;
6852        endTxId_ = value;
6853        onChanged();
6854        return this;
6855      }
6856      /**
6857       * <code>required uint64 endTxId = 3;</code>
6858       */
6859      public Builder clearEndTxId() {
6860        bitField0_ = (bitField0_ & ~0x00000004);
6861        endTxId_ = 0L;
6862        onChanged();
6863        return this;
6864      }
6865
6866      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.FinalizeLogSegmentRequestProto)
6867    }
6868
6869    static {
6870      defaultInstance = new FinalizeLogSegmentRequestProto(true);
6871      defaultInstance.initFields();
6872    }
6873
6874    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.FinalizeLogSegmentRequestProto)
6875  }
6876
6877  public interface FinalizeLogSegmentResponseProtoOrBuilder
6878      extends com.google.protobuf.MessageOrBuilder {
6879  }
6880  /**
6881   * Protobuf type {@code hadoop.hdfs.qjournal.FinalizeLogSegmentResponseProto}
6882   */
6883  public static final class FinalizeLogSegmentResponseProto extends
6884      com.google.protobuf.GeneratedMessage
6885      implements FinalizeLogSegmentResponseProtoOrBuilder {
6886    // Use FinalizeLogSegmentResponseProto.newBuilder() to construct.
6887    private FinalizeLogSegmentResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
6888      super(builder);
6889      this.unknownFields = builder.getUnknownFields();
6890    }
6891    private FinalizeLogSegmentResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
6892
6893    private static final FinalizeLogSegmentResponseProto defaultInstance;
6894    public static FinalizeLogSegmentResponseProto getDefaultInstance() {
6895      return defaultInstance;
6896    }
6897
6898    public FinalizeLogSegmentResponseProto getDefaultInstanceForType() {
6899      return defaultInstance;
6900    }
6901
6902    private final com.google.protobuf.UnknownFieldSet unknownFields;
6903    @java.lang.Override
6904    public final com.google.protobuf.UnknownFieldSet
6905        getUnknownFields() {
6906      return this.unknownFields;
6907    }
6908    private FinalizeLogSegmentResponseProto(
6909        com.google.protobuf.CodedInputStream input,
6910        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6911        throws com.google.protobuf.InvalidProtocolBufferException {
6912      initFields();
6913      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
6914          com.google.protobuf.UnknownFieldSet.newBuilder();
6915      try {
6916        boolean done = false;
6917        while (!done) {
6918          int tag = input.readTag();
6919          switch (tag) {
6920            case 0:
6921              done = true;
6922              break;
6923            default: {
6924              if (!parseUnknownField(input, unknownFields,
6925                                     extensionRegistry, tag)) {
6926                done = true;
6927              }
6928              break;
6929            }
6930          }
6931        }
6932      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
6933        throw e.setUnfinishedMessage(this);
6934      } catch (java.io.IOException e) {
6935        throw new com.google.protobuf.InvalidProtocolBufferException(
6936            e.getMessage()).setUnfinishedMessage(this);
6937      } finally {
6938        this.unknownFields = unknownFields.build();
6939        makeExtensionsImmutable();
6940      }
6941    }
6942    public static final com.google.protobuf.Descriptors.Descriptor
6943        getDescriptor() {
6944      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentResponseProto_descriptor;
6945    }
6946
6947    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6948        internalGetFieldAccessorTable() {
6949      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentResponseProto_fieldAccessorTable
6950          .ensureFieldAccessorsInitialized(
6951              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.Builder.class);
6952    }
6953
6954    public static com.google.protobuf.Parser<FinalizeLogSegmentResponseProto> PARSER =
6955        new com.google.protobuf.AbstractParser<FinalizeLogSegmentResponseProto>() {
6956      public FinalizeLogSegmentResponseProto parsePartialFrom(
6957          com.google.protobuf.CodedInputStream input,
6958          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6959          throws com.google.protobuf.InvalidProtocolBufferException {
6960        return new FinalizeLogSegmentResponseProto(input, extensionRegistry);
6961      }
6962    };
6963
6964    @java.lang.Override
6965    public com.google.protobuf.Parser<FinalizeLogSegmentResponseProto> getParserForType() {
6966      return PARSER;
6967    }
6968
6969    private void initFields() {
6970    }
6971    private byte memoizedIsInitialized = -1;
6972    public final boolean isInitialized() {
6973      byte isInitialized = memoizedIsInitialized;
6974      if (isInitialized != -1) return isInitialized == 1;
6975
6976      memoizedIsInitialized = 1;
6977      return true;
6978    }
6979
6980    public void writeTo(com.google.protobuf.CodedOutputStream output)
6981                        throws java.io.IOException {
6982      getSerializedSize();
6983      getUnknownFields().writeTo(output);
6984    }
6985
6986    private int memoizedSerializedSize = -1;
6987    public int getSerializedSize() {
6988      int size = memoizedSerializedSize;
6989      if (size != -1) return size;
6990
6991      size = 0;
6992      size += getUnknownFields().getSerializedSize();
6993      memoizedSerializedSize = size;
6994      return size;
6995    }
6996
6997    private static final long serialVersionUID = 0L;
6998    @java.lang.Override
6999    protected java.lang.Object writeReplace()
7000        throws java.io.ObjectStreamException {
7001      return super.writeReplace();
7002    }
7003
7004    @java.lang.Override
7005    public boolean equals(final java.lang.Object obj) {
7006      if (obj == this) {
7007       return true;
7008      }
7009      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto)) {
7010        return super.equals(obj);
7011      }
7012      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) obj;
7013
7014      boolean result = true;
7015      result = result &&
7016          getUnknownFields().equals(other.getUnknownFields());
7017      return result;
7018    }
7019
7020    private int memoizedHashCode = 0;
7021    @java.lang.Override
7022    public int hashCode() {
7023      if (memoizedHashCode != 0) {
7024        return memoizedHashCode;
7025      }
7026      int hash = 41;
7027      hash = (19 * hash) + getDescriptorForType().hashCode();
7028      hash = (29 * hash) + getUnknownFields().hashCode();
7029      memoizedHashCode = hash;
7030      return hash;
7031    }
7032
7033    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
7034        com.google.protobuf.ByteString data)
7035        throws com.google.protobuf.InvalidProtocolBufferException {
7036      return PARSER.parseFrom(data);
7037    }
7038    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
7039        com.google.protobuf.ByteString data,
7040        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7041        throws com.google.protobuf.InvalidProtocolBufferException {
7042      return PARSER.parseFrom(data, extensionRegistry);
7043    }
7044    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(byte[] data)
7045        throws com.google.protobuf.InvalidProtocolBufferException {
7046      return PARSER.parseFrom(data);
7047    }
7048    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
7049        byte[] data,
7050        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7051        throws com.google.protobuf.InvalidProtocolBufferException {
7052      return PARSER.parseFrom(data, extensionRegistry);
7053    }
7054    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(java.io.InputStream input)
7055        throws java.io.IOException {
7056      return PARSER.parseFrom(input);
7057    }
7058    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
7059        java.io.InputStream input,
7060        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7061        throws java.io.IOException {
7062      return PARSER.parseFrom(input, extensionRegistry);
7063    }
7064    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseDelimitedFrom(java.io.InputStream input)
7065        throws java.io.IOException {
7066      return PARSER.parseDelimitedFrom(input);
7067    }
7068    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseDelimitedFrom(
7069        java.io.InputStream input,
7070        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7071        throws java.io.IOException {
7072      return PARSER.parseDelimitedFrom(input, extensionRegistry);
7073    }
7074    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
7075        com.google.protobuf.CodedInputStream input)
7076        throws java.io.IOException {
7077      return PARSER.parseFrom(input);
7078    }
7079    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
7080        com.google.protobuf.CodedInputStream input,
7081        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7082        throws java.io.IOException {
7083      return PARSER.parseFrom(input, extensionRegistry);
7084    }
7085
7086    public static Builder newBuilder() { return Builder.create(); }
7087    public Builder newBuilderForType() { return newBuilder(); }
7088    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto prototype) {
7089      return newBuilder().mergeFrom(prototype);
7090    }
7091    public Builder toBuilder() { return newBuilder(this); }
7092
7093    @java.lang.Override
7094    protected Builder newBuilderForType(
7095        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
7096      Builder builder = new Builder(parent);
7097      return builder;
7098    }
7099    /**
7100     * Protobuf type {@code hadoop.hdfs.qjournal.FinalizeLogSegmentResponseProto}
7101     */
7102    public static final class Builder extends
7103        com.google.protobuf.GeneratedMessage.Builder<Builder>
7104       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProtoOrBuilder {
7105      public static final com.google.protobuf.Descriptors.Descriptor
7106          getDescriptor() {
7107        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentResponseProto_descriptor;
7108      }
7109
7110      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7111          internalGetFieldAccessorTable() {
7112        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentResponseProto_fieldAccessorTable
7113            .ensureFieldAccessorsInitialized(
7114                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.Builder.class);
7115      }
7116
7117      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.newBuilder()
7118      private Builder() {
7119        maybeForceBuilderInitialization();
7120      }
7121
7122      private Builder(
7123          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
7124        super(parent);
7125        maybeForceBuilderInitialization();
7126      }
7127      private void maybeForceBuilderInitialization() {
7128        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
7129        }
7130      }
7131      private static Builder create() {
7132        return new Builder();
7133      }
7134
7135      public Builder clear() {
7136        super.clear();
7137        return this;
7138      }
7139
7140      public Builder clone() {
7141        return create().mergeFrom(buildPartial());
7142      }
7143
7144      public com.google.protobuf.Descriptors.Descriptor
7145          getDescriptorForType() {
7146        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentResponseProto_descriptor;
7147      }
7148
7149      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto getDefaultInstanceForType() {
7150        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance();
7151      }
7152
7153      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto build() {
7154        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto result = buildPartial();
7155        if (!result.isInitialized()) {
7156          throw newUninitializedMessageException(result);
7157        }
7158        return result;
7159      }
7160
7161      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto buildPartial() {
7162        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto(this);
7163        onBuilt();
7164        return result;
7165      }
7166
7167      public Builder mergeFrom(com.google.protobuf.Message other) {
7168        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) {
7169          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto)other);
7170        } else {
7171          super.mergeFrom(other);
7172          return this;
7173        }
7174      }
7175
7176      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto other) {
7177        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance()) return this;
7178        this.mergeUnknownFields(other.getUnknownFields());
7179        return this;
7180      }
7181
7182      public final boolean isInitialized() {
7183        return true;
7184      }
7185
7186      public Builder mergeFrom(
7187          com.google.protobuf.CodedInputStream input,
7188          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7189          throws java.io.IOException {
7190        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parsedMessage = null;
7191        try {
7192          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
7193        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
7194          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) e.getUnfinishedMessage();
7195          throw e;
7196        } finally {
7197          if (parsedMessage != null) {
7198            mergeFrom(parsedMessage);
7199          }
7200        }
7201        return this;
7202      }
7203
7204      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.FinalizeLogSegmentResponseProto)
7205    }
7206
7207    static {
7208      defaultInstance = new FinalizeLogSegmentResponseProto(true);
7209      defaultInstance.initFields();
7210    }
7211
7212    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.FinalizeLogSegmentResponseProto)
7213  }
7214
7215  public interface PurgeLogsRequestProtoOrBuilder
7216      extends com.google.protobuf.MessageOrBuilder {
7217
7218    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
7219    /**
7220     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7221     */
7222    boolean hasReqInfo();
7223    /**
7224     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7225     */
7226    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
7227    /**
7228     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7229     */
7230    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
7231
7232    // required uint64 minTxIdToKeep = 2;
7233    /**
7234     * <code>required uint64 minTxIdToKeep = 2;</code>
7235     */
7236    boolean hasMinTxIdToKeep();
7237    /**
7238     * <code>required uint64 minTxIdToKeep = 2;</code>
7239     */
7240    long getMinTxIdToKeep();
7241  }
7242  /**
7243   * Protobuf type {@code hadoop.hdfs.qjournal.PurgeLogsRequestProto}
7244   *
7245   * <pre>
7246   **
7247   * purgeLogs()
7248   * </pre>
7249   */
7250  public static final class PurgeLogsRequestProto extends
7251      com.google.protobuf.GeneratedMessage
7252      implements PurgeLogsRequestProtoOrBuilder {
7253    // Use PurgeLogsRequestProto.newBuilder() to construct.
7254    private PurgeLogsRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
7255      super(builder);
7256      this.unknownFields = builder.getUnknownFields();
7257    }
7258    private PurgeLogsRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
7259
7260    private static final PurgeLogsRequestProto defaultInstance;
7261    public static PurgeLogsRequestProto getDefaultInstance() {
7262      return defaultInstance;
7263    }
7264
7265    public PurgeLogsRequestProto getDefaultInstanceForType() {
7266      return defaultInstance;
7267    }
7268
7269    private final com.google.protobuf.UnknownFieldSet unknownFields;
7270    @java.lang.Override
7271    public final com.google.protobuf.UnknownFieldSet
7272        getUnknownFields() {
7273      return this.unknownFields;
7274    }
7275    private PurgeLogsRequestProto(
7276        com.google.protobuf.CodedInputStream input,
7277        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7278        throws com.google.protobuf.InvalidProtocolBufferException {
7279      initFields();
7280      int mutable_bitField0_ = 0;
7281      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
7282          com.google.protobuf.UnknownFieldSet.newBuilder();
7283      try {
7284        boolean done = false;
7285        while (!done) {
7286          int tag = input.readTag();
7287          switch (tag) {
7288            case 0:
7289              done = true;
7290              break;
7291            default: {
7292              if (!parseUnknownField(input, unknownFields,
7293                                     extensionRegistry, tag)) {
7294                done = true;
7295              }
7296              break;
7297            }
7298            case 10: {
7299              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
7300              if (((bitField0_ & 0x00000001) == 0x00000001)) {
7301                subBuilder = reqInfo_.toBuilder();
7302              }
7303              reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
7304              if (subBuilder != null) {
7305                subBuilder.mergeFrom(reqInfo_);
7306                reqInfo_ = subBuilder.buildPartial();
7307              }
7308              bitField0_ |= 0x00000001;
7309              break;
7310            }
7311            case 16: {
7312              bitField0_ |= 0x00000002;
7313              minTxIdToKeep_ = input.readUInt64();
7314              break;
7315            }
7316          }
7317        }
7318      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
7319        throw e.setUnfinishedMessage(this);
7320      } catch (java.io.IOException e) {
7321        throw new com.google.protobuf.InvalidProtocolBufferException(
7322            e.getMessage()).setUnfinishedMessage(this);
7323      } finally {
7324        this.unknownFields = unknownFields.build();
7325        makeExtensionsImmutable();
7326      }
7327    }
7328    public static final com.google.protobuf.Descriptors.Descriptor
7329        getDescriptor() {
7330      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PurgeLogsRequestProto_descriptor;
7331    }
7332
7333    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7334        internalGetFieldAccessorTable() {
7335      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PurgeLogsRequestProto_fieldAccessorTable
7336          .ensureFieldAccessorsInitialized(
7337              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.Builder.class);
7338    }
7339
7340    public static com.google.protobuf.Parser<PurgeLogsRequestProto> PARSER =
7341        new com.google.protobuf.AbstractParser<PurgeLogsRequestProto>() {
7342      public PurgeLogsRequestProto parsePartialFrom(
7343          com.google.protobuf.CodedInputStream input,
7344          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7345          throws com.google.protobuf.InvalidProtocolBufferException {
7346        return new PurgeLogsRequestProto(input, extensionRegistry);
7347      }
7348    };
7349
7350    @java.lang.Override
7351    public com.google.protobuf.Parser<PurgeLogsRequestProto> getParserForType() {
7352      return PARSER;
7353    }
7354
7355    private int bitField0_;
7356    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
7357    public static final int REQINFO_FIELD_NUMBER = 1;
7358    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
7359    /**
7360     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7361     */
7362    public boolean hasReqInfo() {
7363      return ((bitField0_ & 0x00000001) == 0x00000001);
7364    }
7365    /**
7366     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7367     */
7368    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
7369      return reqInfo_;
7370    }
7371    /**
7372     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7373     */
7374    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
7375      return reqInfo_;
7376    }
7377
7378    // required uint64 minTxIdToKeep = 2;
7379    public static final int MINTXIDTOKEEP_FIELD_NUMBER = 2;
7380    private long minTxIdToKeep_;
7381    /**
7382     * <code>required uint64 minTxIdToKeep = 2;</code>
7383     */
7384    public boolean hasMinTxIdToKeep() {
7385      return ((bitField0_ & 0x00000002) == 0x00000002);
7386    }
7387    /**
7388     * <code>required uint64 minTxIdToKeep = 2;</code>
7389     */
7390    public long getMinTxIdToKeep() {
7391      return minTxIdToKeep_;
7392    }
7393
7394    private void initFields() {
7395      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
7396      minTxIdToKeep_ = 0L;
7397    }
7398    private byte memoizedIsInitialized = -1;
7399    public final boolean isInitialized() {
7400      byte isInitialized = memoizedIsInitialized;
7401      if (isInitialized != -1) return isInitialized == 1;
7402
7403      if (!hasReqInfo()) {
7404        memoizedIsInitialized = 0;
7405        return false;
7406      }
7407      if (!hasMinTxIdToKeep()) {
7408        memoizedIsInitialized = 0;
7409        return false;
7410      }
7411      if (!getReqInfo().isInitialized()) {
7412        memoizedIsInitialized = 0;
7413        return false;
7414      }
7415      memoizedIsInitialized = 1;
7416      return true;
7417    }
7418
7419    public void writeTo(com.google.protobuf.CodedOutputStream output)
7420                        throws java.io.IOException {
7421      getSerializedSize();
7422      if (((bitField0_ & 0x00000001) == 0x00000001)) {
7423        output.writeMessage(1, reqInfo_);
7424      }
7425      if (((bitField0_ & 0x00000002) == 0x00000002)) {
7426        output.writeUInt64(2, minTxIdToKeep_);
7427      }
7428      getUnknownFields().writeTo(output);
7429    }
7430
7431    private int memoizedSerializedSize = -1;
7432    public int getSerializedSize() {
7433      int size = memoizedSerializedSize;
7434      if (size != -1) return size;
7435
7436      size = 0;
7437      if (((bitField0_ & 0x00000001) == 0x00000001)) {
7438        size += com.google.protobuf.CodedOutputStream
7439          .computeMessageSize(1, reqInfo_);
7440      }
7441      if (((bitField0_ & 0x00000002) == 0x00000002)) {
7442        size += com.google.protobuf.CodedOutputStream
7443          .computeUInt64Size(2, minTxIdToKeep_);
7444      }
7445      size += getUnknownFields().getSerializedSize();
7446      memoizedSerializedSize = size;
7447      return size;
7448    }
7449
7450    private static final long serialVersionUID = 0L;
7451    @java.lang.Override
7452    protected java.lang.Object writeReplace()
7453        throws java.io.ObjectStreamException {
7454      return super.writeReplace();
7455    }
7456
7457    @java.lang.Override
7458    public boolean equals(final java.lang.Object obj) {
7459      if (obj == this) {
7460       return true;
7461      }
7462      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)) {
7463        return super.equals(obj);
7464      }
7465      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto) obj;
7466
7467      boolean result = true;
7468      result = result && (hasReqInfo() == other.hasReqInfo());
7469      if (hasReqInfo()) {
7470        result = result && getReqInfo()
7471            .equals(other.getReqInfo());
7472      }
7473      result = result && (hasMinTxIdToKeep() == other.hasMinTxIdToKeep());
7474      if (hasMinTxIdToKeep()) {
7475        result = result && (getMinTxIdToKeep()
7476            == other.getMinTxIdToKeep());
7477      }
7478      result = result &&
7479          getUnknownFields().equals(other.getUnknownFields());
7480      return result;
7481    }
7482
7483    private int memoizedHashCode = 0;
7484    @java.lang.Override
7485    public int hashCode() {
7486      if (memoizedHashCode != 0) {
7487        return memoizedHashCode;
7488      }
7489      int hash = 41;
7490      hash = (19 * hash) + getDescriptorForType().hashCode();
7491      if (hasReqInfo()) {
7492        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
7493        hash = (53 * hash) + getReqInfo().hashCode();
7494      }
7495      if (hasMinTxIdToKeep()) {
7496        hash = (37 * hash) + MINTXIDTOKEEP_FIELD_NUMBER;
7497        hash = (53 * hash) + hashLong(getMinTxIdToKeep());
7498      }
7499      hash = (29 * hash) + getUnknownFields().hashCode();
7500      memoizedHashCode = hash;
7501      return hash;
7502    }
7503
7504    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7505        com.google.protobuf.ByteString data)
7506        throws com.google.protobuf.InvalidProtocolBufferException {
7507      return PARSER.parseFrom(data);
7508    }
7509    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7510        com.google.protobuf.ByteString data,
7511        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7512        throws com.google.protobuf.InvalidProtocolBufferException {
7513      return PARSER.parseFrom(data, extensionRegistry);
7514    }
7515    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(byte[] data)
7516        throws com.google.protobuf.InvalidProtocolBufferException {
7517      return PARSER.parseFrom(data);
7518    }
7519    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7520        byte[] data,
7521        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7522        throws com.google.protobuf.InvalidProtocolBufferException {
7523      return PARSER.parseFrom(data, extensionRegistry);
7524    }
7525    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(java.io.InputStream input)
7526        throws java.io.IOException {
7527      return PARSER.parseFrom(input);
7528    }
7529    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7530        java.io.InputStream input,
7531        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7532        throws java.io.IOException {
7533      return PARSER.parseFrom(input, extensionRegistry);
7534    }
7535    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseDelimitedFrom(java.io.InputStream input)
7536        throws java.io.IOException {
7537      return PARSER.parseDelimitedFrom(input);
7538    }
7539    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseDelimitedFrom(
7540        java.io.InputStream input,
7541        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7542        throws java.io.IOException {
7543      return PARSER.parseDelimitedFrom(input, extensionRegistry);
7544    }
7545    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7546        com.google.protobuf.CodedInputStream input)
7547        throws java.io.IOException {
7548      return PARSER.parseFrom(input);
7549    }
7550    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7551        com.google.protobuf.CodedInputStream input,
7552        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7553        throws java.io.IOException {
7554      return PARSER.parseFrom(input, extensionRegistry);
7555    }
7556
7557    public static Builder newBuilder() { return Builder.create(); }
7558    public Builder newBuilderForType() { return newBuilder(); }
7559    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto prototype) {
7560      return newBuilder().mergeFrom(prototype);
7561    }
7562    public Builder toBuilder() { return newBuilder(this); }
7563
7564    @java.lang.Override
7565    protected Builder newBuilderForType(
7566        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
7567      Builder builder = new Builder(parent);
7568      return builder;
7569    }
7570    /**
7571     * Protobuf type {@code hadoop.hdfs.qjournal.PurgeLogsRequestProto}
7572     *
7573     * <pre>
7574     **
7575     * purgeLogs()
7576     * </pre>
7577     */
7578    public static final class Builder extends
7579        com.google.protobuf.GeneratedMessage.Builder<Builder>
7580       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProtoOrBuilder {
7581      public static final com.google.protobuf.Descriptors.Descriptor
7582          getDescriptor() {
7583        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PurgeLogsRequestProto_descriptor;
7584      }
7585
7586      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7587          internalGetFieldAccessorTable() {
7588        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PurgeLogsRequestProto_fieldAccessorTable
7589            .ensureFieldAccessorsInitialized(
7590                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.Builder.class);
7591      }
7592
7593      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.newBuilder()
7594      private Builder() {
7595        maybeForceBuilderInitialization();
7596      }
7597
7598      private Builder(
7599          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
7600        super(parent);
7601        maybeForceBuilderInitialization();
7602      }
7603      private void maybeForceBuilderInitialization() {
7604        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
7605          getReqInfoFieldBuilder();
7606        }
7607      }
7608      private static Builder create() {
7609        return new Builder();
7610      }
7611
7612      public Builder clear() {
7613        super.clear();
7614        if (reqInfoBuilder_ == null) {
7615          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
7616        } else {
7617          reqInfoBuilder_.clear();
7618        }
7619        bitField0_ = (bitField0_ & ~0x00000001);
7620        minTxIdToKeep_ = 0L;
7621        bitField0_ = (bitField0_ & ~0x00000002);
7622        return this;
7623      }
7624
7625      public Builder clone() {
7626        return create().mergeFrom(buildPartial());
7627      }
7628
7629      public com.google.protobuf.Descriptors.Descriptor
7630          getDescriptorForType() {
7631        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PurgeLogsRequestProto_descriptor;
7632      }
7633
7634      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto getDefaultInstanceForType() {
7635        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance();
7636      }
7637
7638      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto build() {
7639        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto result = buildPartial();
7640        if (!result.isInitialized()) {
7641          throw newUninitializedMessageException(result);
7642        }
7643        return result;
7644      }
7645
7646      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto buildPartial() {
7647        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto(this);
7648        int from_bitField0_ = bitField0_;
7649        int to_bitField0_ = 0;
7650        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
7651          to_bitField0_ |= 0x00000001;
7652        }
7653        if (reqInfoBuilder_ == null) {
7654          result.reqInfo_ = reqInfo_;
7655        } else {
7656          result.reqInfo_ = reqInfoBuilder_.build();
7657        }
7658        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
7659          to_bitField0_ |= 0x00000002;
7660        }
7661        result.minTxIdToKeep_ = minTxIdToKeep_;
7662        result.bitField0_ = to_bitField0_;
7663        onBuilt();
7664        return result;
7665      }
7666
7667      public Builder mergeFrom(com.google.protobuf.Message other) {
7668        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto) {
7669          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)other);
7670        } else {
7671          super.mergeFrom(other);
7672          return this;
7673        }
7674      }
7675
7676      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto other) {
7677        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance()) return this;
7678        if (other.hasReqInfo()) {
7679          mergeReqInfo(other.getReqInfo());
7680        }
7681        if (other.hasMinTxIdToKeep()) {
7682          setMinTxIdToKeep(other.getMinTxIdToKeep());
7683        }
7684        this.mergeUnknownFields(other.getUnknownFields());
7685        return this;
7686      }
7687
7688      public final boolean isInitialized() {
7689        if (!hasReqInfo()) {
7690          
7691          return false;
7692        }
7693        if (!hasMinTxIdToKeep()) {
7694          
7695          return false;
7696        }
7697        if (!getReqInfo().isInitialized()) {
7698          
7699          return false;
7700        }
7701        return true;
7702      }
7703
7704      public Builder mergeFrom(
7705          com.google.protobuf.CodedInputStream input,
7706          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7707          throws java.io.IOException {
7708        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parsedMessage = null;
7709        try {
7710          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
7711        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
7712          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto) e.getUnfinishedMessage();
7713          throw e;
7714        } finally {
7715          if (parsedMessage != null) {
7716            mergeFrom(parsedMessage);
7717          }
7718        }
7719        return this;
7720      }
7721      private int bitField0_;
7722
7723      // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
7724      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
7725      private com.google.protobuf.SingleFieldBuilder<
7726          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
7727      /**
7728       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7729       */
7730      public boolean hasReqInfo() {
7731        return ((bitField0_ & 0x00000001) == 0x00000001);
7732      }
7733      /**
7734       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7735       */
7736      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
7737        if (reqInfoBuilder_ == null) {
7738          return reqInfo_;
7739        } else {
7740          return reqInfoBuilder_.getMessage();
7741        }
7742      }
7743      /**
7744       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7745       */
7746      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
7747        if (reqInfoBuilder_ == null) {
7748          if (value == null) {
7749            throw new NullPointerException();
7750          }
7751          reqInfo_ = value;
7752          onChanged();
7753        } else {
7754          reqInfoBuilder_.setMessage(value);
7755        }
7756        bitField0_ |= 0x00000001;
7757        return this;
7758      }
7759      /**
7760       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7761       */
7762      public Builder setReqInfo(
7763          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
7764        if (reqInfoBuilder_ == null) {
7765          reqInfo_ = builderForValue.build();
7766          onChanged();
7767        } else {
7768          reqInfoBuilder_.setMessage(builderForValue.build());
7769        }
7770        bitField0_ |= 0x00000001;
7771        return this;
7772      }
7773      /**
7774       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7775       */
7776      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
7777        if (reqInfoBuilder_ == null) {
7778          if (((bitField0_ & 0x00000001) == 0x00000001) &&
7779              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
7780            reqInfo_ =
7781              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
7782          } else {
7783            reqInfo_ = value;
7784          }
7785          onChanged();
7786        } else {
7787          reqInfoBuilder_.mergeFrom(value);
7788        }
7789        bitField0_ |= 0x00000001;
7790        return this;
7791      }
7792      /**
7793       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7794       */
7795      public Builder clearReqInfo() {
7796        if (reqInfoBuilder_ == null) {
7797          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
7798          onChanged();
7799        } else {
7800          reqInfoBuilder_.clear();
7801        }
7802        bitField0_ = (bitField0_ & ~0x00000001);
7803        return this;
7804      }
7805      /**
7806       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7807       */
7808      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
7809        bitField0_ |= 0x00000001;
7810        onChanged();
7811        return getReqInfoFieldBuilder().getBuilder();
7812      }
7813      /**
7814       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7815       */
7816      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
7817        if (reqInfoBuilder_ != null) {
7818          return reqInfoBuilder_.getMessageOrBuilder();
7819        } else {
7820          return reqInfo_;
7821        }
7822      }
7823      /**
7824       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
7825       */
7826      private com.google.protobuf.SingleFieldBuilder<
7827          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
7828          getReqInfoFieldBuilder() {
7829        if (reqInfoBuilder_ == null) {
7830          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
7831              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
7832                  reqInfo_,
7833                  getParentForChildren(),
7834                  isClean());
7835          reqInfo_ = null;
7836        }
7837        return reqInfoBuilder_;
7838      }
7839
7840      // required uint64 minTxIdToKeep = 2;
7841      private long minTxIdToKeep_ ;
7842      /**
7843       * <code>required uint64 minTxIdToKeep = 2;</code>
7844       */
7845      public boolean hasMinTxIdToKeep() {
7846        return ((bitField0_ & 0x00000002) == 0x00000002);
7847      }
7848      /**
7849       * <code>required uint64 minTxIdToKeep = 2;</code>
7850       */
7851      public long getMinTxIdToKeep() {
7852        return minTxIdToKeep_;
7853      }
7854      /**
7855       * <code>required uint64 minTxIdToKeep = 2;</code>
7856       */
7857      public Builder setMinTxIdToKeep(long value) {
7858        bitField0_ |= 0x00000002;
7859        minTxIdToKeep_ = value;
7860        onChanged();
7861        return this;
7862      }
7863      /**
7864       * <code>required uint64 minTxIdToKeep = 2;</code>
7865       */
7866      public Builder clearMinTxIdToKeep() {
7867        bitField0_ = (bitField0_ & ~0x00000002);
7868        minTxIdToKeep_ = 0L;
7869        onChanged();
7870        return this;
7871      }
7872
7873      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.PurgeLogsRequestProto)
7874    }
7875
7876    static {
7877      defaultInstance = new PurgeLogsRequestProto(true);
7878      defaultInstance.initFields();
7879    }
7880
7881    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.PurgeLogsRequestProto)
7882  }
7883
7884  public interface PurgeLogsResponseProtoOrBuilder
7885      extends com.google.protobuf.MessageOrBuilder {
7886  }
7887  /**
7888   * Protobuf type {@code hadoop.hdfs.qjournal.PurgeLogsResponseProto}
7889   */
7890  public static final class PurgeLogsResponseProto extends
7891      com.google.protobuf.GeneratedMessage
7892      implements PurgeLogsResponseProtoOrBuilder {
7893    // Use PurgeLogsResponseProto.newBuilder() to construct.
7894    private PurgeLogsResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
7895      super(builder);
7896      this.unknownFields = builder.getUnknownFields();
7897    }
7898    private PurgeLogsResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
7899
7900    private static final PurgeLogsResponseProto defaultInstance;
7901    public static PurgeLogsResponseProto getDefaultInstance() {
7902      return defaultInstance;
7903    }
7904
7905    public PurgeLogsResponseProto getDefaultInstanceForType() {
7906      return defaultInstance;
7907    }
7908
7909    private final com.google.protobuf.UnknownFieldSet unknownFields;
7910    @java.lang.Override
7911    public final com.google.protobuf.UnknownFieldSet
7912        getUnknownFields() {
7913      return this.unknownFields;
7914    }
7915    private PurgeLogsResponseProto(
7916        com.google.protobuf.CodedInputStream input,
7917        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7918        throws com.google.protobuf.InvalidProtocolBufferException {
7919      initFields();
7920      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
7921          com.google.protobuf.UnknownFieldSet.newBuilder();
7922      try {
7923        boolean done = false;
7924        while (!done) {
7925          int tag = input.readTag();
7926          switch (tag) {
7927            case 0:
7928              done = true;
7929              break;
7930            default: {
7931              if (!parseUnknownField(input, unknownFields,
7932                                     extensionRegistry, tag)) {
7933                done = true;
7934              }
7935              break;
7936            }
7937          }
7938        }
7939      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
7940        throw e.setUnfinishedMessage(this);
7941      } catch (java.io.IOException e) {
7942        throw new com.google.protobuf.InvalidProtocolBufferException(
7943            e.getMessage()).setUnfinishedMessage(this);
7944      } finally {
7945        this.unknownFields = unknownFields.build();
7946        makeExtensionsImmutable();
7947      }
7948    }
7949    public static final com.google.protobuf.Descriptors.Descriptor
7950        getDescriptor() {
7951      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PurgeLogsResponseProto_descriptor;
7952    }
7953
7954    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7955        internalGetFieldAccessorTable() {
7956      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PurgeLogsResponseProto_fieldAccessorTable
7957          .ensureFieldAccessorsInitialized(
7958              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.Builder.class);
7959    }
7960
7961    public static com.google.protobuf.Parser<PurgeLogsResponseProto> PARSER =
7962        new com.google.protobuf.AbstractParser<PurgeLogsResponseProto>() {
7963      public PurgeLogsResponseProto parsePartialFrom(
7964          com.google.protobuf.CodedInputStream input,
7965          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7966          throws com.google.protobuf.InvalidProtocolBufferException {
7967        return new PurgeLogsResponseProto(input, extensionRegistry);
7968      }
7969    };
7970
7971    @java.lang.Override
7972    public com.google.protobuf.Parser<PurgeLogsResponseProto> getParserForType() {
7973      return PARSER;
7974    }
7975
7976    private void initFields() {
7977    }
7978    private byte memoizedIsInitialized = -1;
7979    public final boolean isInitialized() {
7980      byte isInitialized = memoizedIsInitialized;
7981      if (isInitialized != -1) return isInitialized == 1;
7982
7983      memoizedIsInitialized = 1;
7984      return true;
7985    }
7986
7987    public void writeTo(com.google.protobuf.CodedOutputStream output)
7988                        throws java.io.IOException {
7989      getSerializedSize();
7990      getUnknownFields().writeTo(output);
7991    }
7992
7993    private int memoizedSerializedSize = -1;
7994    public int getSerializedSize() {
7995      int size = memoizedSerializedSize;
7996      if (size != -1) return size;
7997
7998      size = 0;
7999      size += getUnknownFields().getSerializedSize();
8000      memoizedSerializedSize = size;
8001      return size;
8002    }
8003
8004    private static final long serialVersionUID = 0L;
8005    @java.lang.Override
8006    protected java.lang.Object writeReplace()
8007        throws java.io.ObjectStreamException {
8008      return super.writeReplace();
8009    }
8010
8011    @java.lang.Override
8012    public boolean equals(final java.lang.Object obj) {
8013      if (obj == this) {
8014       return true;
8015      }
8016      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto)) {
8017        return super.equals(obj);
8018      }
8019      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) obj;
8020
8021      boolean result = true;
8022      result = result &&
8023          getUnknownFields().equals(other.getUnknownFields());
8024      return result;
8025    }
8026
8027    private int memoizedHashCode = 0;
8028    @java.lang.Override
8029    public int hashCode() {
8030      if (memoizedHashCode != 0) {
8031        return memoizedHashCode;
8032      }
8033      int hash = 41;
8034      hash = (19 * hash) + getDescriptorForType().hashCode();
8035      hash = (29 * hash) + getUnknownFields().hashCode();
8036      memoizedHashCode = hash;
8037      return hash;
8038    }
8039
8040    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
8041        com.google.protobuf.ByteString data)
8042        throws com.google.protobuf.InvalidProtocolBufferException {
8043      return PARSER.parseFrom(data);
8044    }
8045    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
8046        com.google.protobuf.ByteString data,
8047        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8048        throws com.google.protobuf.InvalidProtocolBufferException {
8049      return PARSER.parseFrom(data, extensionRegistry);
8050    }
8051    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(byte[] data)
8052        throws com.google.protobuf.InvalidProtocolBufferException {
8053      return PARSER.parseFrom(data);
8054    }
8055    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
8056        byte[] data,
8057        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8058        throws com.google.protobuf.InvalidProtocolBufferException {
8059      return PARSER.parseFrom(data, extensionRegistry);
8060    }
8061    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(java.io.InputStream input)
8062        throws java.io.IOException {
8063      return PARSER.parseFrom(input);
8064    }
8065    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
8066        java.io.InputStream input,
8067        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8068        throws java.io.IOException {
8069      return PARSER.parseFrom(input, extensionRegistry);
8070    }
8071    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseDelimitedFrom(java.io.InputStream input)
8072        throws java.io.IOException {
8073      return PARSER.parseDelimitedFrom(input);
8074    }
8075    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseDelimitedFrom(
8076        java.io.InputStream input,
8077        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8078        throws java.io.IOException {
8079      return PARSER.parseDelimitedFrom(input, extensionRegistry);
8080    }
8081    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
8082        com.google.protobuf.CodedInputStream input)
8083        throws java.io.IOException {
8084      return PARSER.parseFrom(input);
8085    }
8086    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
8087        com.google.protobuf.CodedInputStream input,
8088        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8089        throws java.io.IOException {
8090      return PARSER.parseFrom(input, extensionRegistry);
8091    }
8092
8093    public static Builder newBuilder() { return Builder.create(); }
8094    public Builder newBuilderForType() { return newBuilder(); }
8095    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto prototype) {
8096      return newBuilder().mergeFrom(prototype);
8097    }
8098    public Builder toBuilder() { return newBuilder(this); }
8099
8100    @java.lang.Override
8101    protected Builder newBuilderForType(
8102        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
8103      Builder builder = new Builder(parent);
8104      return builder;
8105    }
8106    /**
8107     * Protobuf type {@code hadoop.hdfs.qjournal.PurgeLogsResponseProto}
8108     */
8109    public static final class Builder extends
8110        com.google.protobuf.GeneratedMessage.Builder<Builder>
8111       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProtoOrBuilder {
8112      public static final com.google.protobuf.Descriptors.Descriptor
8113          getDescriptor() {
8114        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PurgeLogsResponseProto_descriptor;
8115      }
8116
8117      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8118          internalGetFieldAccessorTable() {
8119        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PurgeLogsResponseProto_fieldAccessorTable
8120            .ensureFieldAccessorsInitialized(
8121                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.Builder.class);
8122      }
8123
8124      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.newBuilder()
8125      private Builder() {
8126        maybeForceBuilderInitialization();
8127      }
8128
8129      private Builder(
8130          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
8131        super(parent);
8132        maybeForceBuilderInitialization();
8133      }
8134      private void maybeForceBuilderInitialization() {
8135        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
8136        }
8137      }
8138      private static Builder create() {
8139        return new Builder();
8140      }
8141
8142      public Builder clear() {
8143        super.clear();
8144        return this;
8145      }
8146
8147      public Builder clone() {
8148        return create().mergeFrom(buildPartial());
8149      }
8150
8151      public com.google.protobuf.Descriptors.Descriptor
8152          getDescriptorForType() {
8153        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PurgeLogsResponseProto_descriptor;
8154      }
8155
8156      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto getDefaultInstanceForType() {
8157        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance();
8158      }
8159
8160      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto build() {
8161        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto result = buildPartial();
8162        if (!result.isInitialized()) {
8163          throw newUninitializedMessageException(result);
8164        }
8165        return result;
8166      }
8167
8168      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto buildPartial() {
8169        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto(this);
8170        onBuilt();
8171        return result;
8172      }
8173
8174      public Builder mergeFrom(com.google.protobuf.Message other) {
8175        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) {
8176          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto)other);
8177        } else {
8178          super.mergeFrom(other);
8179          return this;
8180        }
8181      }
8182
8183      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto other) {
8184        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance()) return this;
8185        this.mergeUnknownFields(other.getUnknownFields());
8186        return this;
8187      }
8188
8189      public final boolean isInitialized() {
8190        return true;
8191      }
8192
8193      public Builder mergeFrom(
8194          com.google.protobuf.CodedInputStream input,
8195          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8196          throws java.io.IOException {
8197        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parsedMessage = null;
8198        try {
8199          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
8200        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
8201          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) e.getUnfinishedMessage();
8202          throw e;
8203        } finally {
8204          if (parsedMessage != null) {
8205            mergeFrom(parsedMessage);
8206          }
8207        }
8208        return this;
8209      }
8210
8211      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.PurgeLogsResponseProto)
8212    }
8213
8214    static {
8215      defaultInstance = new PurgeLogsResponseProto(true);
8216      defaultInstance.initFields();
8217    }
8218
8219    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.PurgeLogsResponseProto)
8220  }
8221
8222  public interface IsFormattedRequestProtoOrBuilder
8223      extends com.google.protobuf.MessageOrBuilder {
8224
8225    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
8226    /**
8227     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8228     */
8229    boolean hasJid();
8230    /**
8231     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8232     */
8233    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
8234    /**
8235     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8236     */
8237    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
8238  }
8239  /**
8240   * Protobuf type {@code hadoop.hdfs.qjournal.IsFormattedRequestProto}
8241   *
8242   * <pre>
8243   **
8244   * isFormatted()
8245   * </pre>
8246   */
8247  public static final class IsFormattedRequestProto extends
8248      com.google.protobuf.GeneratedMessage
8249      implements IsFormattedRequestProtoOrBuilder {
8250    // Use IsFormattedRequestProto.newBuilder() to construct.
8251    private IsFormattedRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
8252      super(builder);
8253      this.unknownFields = builder.getUnknownFields();
8254    }
8255    private IsFormattedRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
8256
8257    private static final IsFormattedRequestProto defaultInstance;
8258    public static IsFormattedRequestProto getDefaultInstance() {
8259      return defaultInstance;
8260    }
8261
8262    public IsFormattedRequestProto getDefaultInstanceForType() {
8263      return defaultInstance;
8264    }
8265
8266    private final com.google.protobuf.UnknownFieldSet unknownFields;
8267    @java.lang.Override
8268    public final com.google.protobuf.UnknownFieldSet
8269        getUnknownFields() {
8270      return this.unknownFields;
8271    }
8272    private IsFormattedRequestProto(
8273        com.google.protobuf.CodedInputStream input,
8274        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8275        throws com.google.protobuf.InvalidProtocolBufferException {
8276      initFields();
8277      int mutable_bitField0_ = 0;
8278      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
8279          com.google.protobuf.UnknownFieldSet.newBuilder();
8280      try {
8281        boolean done = false;
8282        while (!done) {
8283          int tag = input.readTag();
8284          switch (tag) {
8285            case 0:
8286              done = true;
8287              break;
8288            default: {
8289              if (!parseUnknownField(input, unknownFields,
8290                                     extensionRegistry, tag)) {
8291                done = true;
8292              }
8293              break;
8294            }
8295            case 10: {
8296              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
8297              if (((bitField0_ & 0x00000001) == 0x00000001)) {
8298                subBuilder = jid_.toBuilder();
8299              }
8300              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
8301              if (subBuilder != null) {
8302                subBuilder.mergeFrom(jid_);
8303                jid_ = subBuilder.buildPartial();
8304              }
8305              bitField0_ |= 0x00000001;
8306              break;
8307            }
8308          }
8309        }
8310      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
8311        throw e.setUnfinishedMessage(this);
8312      } catch (java.io.IOException e) {
8313        throw new com.google.protobuf.InvalidProtocolBufferException(
8314            e.getMessage()).setUnfinishedMessage(this);
8315      } finally {
8316        this.unknownFields = unknownFields.build();
8317        makeExtensionsImmutable();
8318      }
8319    }
8320    public static final com.google.protobuf.Descriptors.Descriptor
8321        getDescriptor() {
8322      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_IsFormattedRequestProto_descriptor;
8323    }
8324
8325    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8326        internalGetFieldAccessorTable() {
8327      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_IsFormattedRequestProto_fieldAccessorTable
8328          .ensureFieldAccessorsInitialized(
8329              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.Builder.class);
8330    }
8331
8332    public static com.google.protobuf.Parser<IsFormattedRequestProto> PARSER =
8333        new com.google.protobuf.AbstractParser<IsFormattedRequestProto>() {
8334      public IsFormattedRequestProto parsePartialFrom(
8335          com.google.protobuf.CodedInputStream input,
8336          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8337          throws com.google.protobuf.InvalidProtocolBufferException {
8338        return new IsFormattedRequestProto(input, extensionRegistry);
8339      }
8340    };
8341
8342    @java.lang.Override
8343    public com.google.protobuf.Parser<IsFormattedRequestProto> getParserForType() {
8344      return PARSER;
8345    }
8346
8347    private int bitField0_;
8348    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
8349    public static final int JID_FIELD_NUMBER = 1;
8350    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
8351    /**
8352     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8353     */
8354    public boolean hasJid() {
8355      return ((bitField0_ & 0x00000001) == 0x00000001);
8356    }
8357    /**
8358     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8359     */
8360    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
8361      return jid_;
8362    }
8363    /**
8364     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8365     */
8366    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
8367      return jid_;
8368    }
8369
8370    private void initFields() {
8371      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8372    }
8373    private byte memoizedIsInitialized = -1;
8374    public final boolean isInitialized() {
8375      byte isInitialized = memoizedIsInitialized;
8376      if (isInitialized != -1) return isInitialized == 1;
8377
8378      if (!hasJid()) {
8379        memoizedIsInitialized = 0;
8380        return false;
8381      }
8382      if (!getJid().isInitialized()) {
8383        memoizedIsInitialized = 0;
8384        return false;
8385      }
8386      memoizedIsInitialized = 1;
8387      return true;
8388    }
8389
8390    public void writeTo(com.google.protobuf.CodedOutputStream output)
8391                        throws java.io.IOException {
8392      getSerializedSize();
8393      if (((bitField0_ & 0x00000001) == 0x00000001)) {
8394        output.writeMessage(1, jid_);
8395      }
8396      getUnknownFields().writeTo(output);
8397    }
8398
8399    private int memoizedSerializedSize = -1;
8400    public int getSerializedSize() {
8401      int size = memoizedSerializedSize;
8402      if (size != -1) return size;
8403
8404      size = 0;
8405      if (((bitField0_ & 0x00000001) == 0x00000001)) {
8406        size += com.google.protobuf.CodedOutputStream
8407          .computeMessageSize(1, jid_);
8408      }
8409      size += getUnknownFields().getSerializedSize();
8410      memoizedSerializedSize = size;
8411      return size;
8412    }
8413
8414    private static final long serialVersionUID = 0L;
8415    @java.lang.Override
8416    protected java.lang.Object writeReplace()
8417        throws java.io.ObjectStreamException {
8418      return super.writeReplace();
8419    }
8420
8421    @java.lang.Override
8422    public boolean equals(final java.lang.Object obj) {
8423      if (obj == this) {
8424       return true;
8425      }
8426      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)) {
8427        return super.equals(obj);
8428      }
8429      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto) obj;
8430
8431      boolean result = true;
8432      result = result && (hasJid() == other.hasJid());
8433      if (hasJid()) {
8434        result = result && getJid()
8435            .equals(other.getJid());
8436      }
8437      result = result &&
8438          getUnknownFields().equals(other.getUnknownFields());
8439      return result;
8440    }
8441
8442    private int memoizedHashCode = 0;
8443    @java.lang.Override
8444    public int hashCode() {
8445      if (memoizedHashCode != 0) {
8446        return memoizedHashCode;
8447      }
8448      int hash = 41;
8449      hash = (19 * hash) + getDescriptorForType().hashCode();
8450      if (hasJid()) {
8451        hash = (37 * hash) + JID_FIELD_NUMBER;
8452        hash = (53 * hash) + getJid().hashCode();
8453      }
8454      hash = (29 * hash) + getUnknownFields().hashCode();
8455      memoizedHashCode = hash;
8456      return hash;
8457    }
8458
8459    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8460        com.google.protobuf.ByteString data)
8461        throws com.google.protobuf.InvalidProtocolBufferException {
8462      return PARSER.parseFrom(data);
8463    }
8464    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8465        com.google.protobuf.ByteString data,
8466        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8467        throws com.google.protobuf.InvalidProtocolBufferException {
8468      return PARSER.parseFrom(data, extensionRegistry);
8469    }
8470    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(byte[] data)
8471        throws com.google.protobuf.InvalidProtocolBufferException {
8472      return PARSER.parseFrom(data);
8473    }
8474    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8475        byte[] data,
8476        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8477        throws com.google.protobuf.InvalidProtocolBufferException {
8478      return PARSER.parseFrom(data, extensionRegistry);
8479    }
8480    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(java.io.InputStream input)
8481        throws java.io.IOException {
8482      return PARSER.parseFrom(input);
8483    }
8484    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8485        java.io.InputStream input,
8486        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8487        throws java.io.IOException {
8488      return PARSER.parseFrom(input, extensionRegistry);
8489    }
8490    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseDelimitedFrom(java.io.InputStream input)
8491        throws java.io.IOException {
8492      return PARSER.parseDelimitedFrom(input);
8493    }
8494    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseDelimitedFrom(
8495        java.io.InputStream input,
8496        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8497        throws java.io.IOException {
8498      return PARSER.parseDelimitedFrom(input, extensionRegistry);
8499    }
8500    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8501        com.google.protobuf.CodedInputStream input)
8502        throws java.io.IOException {
8503      return PARSER.parseFrom(input);
8504    }
8505    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8506        com.google.protobuf.CodedInputStream input,
8507        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8508        throws java.io.IOException {
8509      return PARSER.parseFrom(input, extensionRegistry);
8510    }
8511
8512    public static Builder newBuilder() { return Builder.create(); }
8513    public Builder newBuilderForType() { return newBuilder(); }
8514    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto prototype) {
8515      return newBuilder().mergeFrom(prototype);
8516    }
8517    public Builder toBuilder() { return newBuilder(this); }
8518
8519    @java.lang.Override
8520    protected Builder newBuilderForType(
8521        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
8522      Builder builder = new Builder(parent);
8523      return builder;
8524    }
8525    /**
8526     * Protobuf type {@code hadoop.hdfs.qjournal.IsFormattedRequestProto}
8527     *
8528     * <pre>
8529     **
8530     * isFormatted()
8531     * </pre>
8532     */
8533    public static final class Builder extends
8534        com.google.protobuf.GeneratedMessage.Builder<Builder>
8535       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProtoOrBuilder {
8536      public static final com.google.protobuf.Descriptors.Descriptor
8537          getDescriptor() {
8538        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_IsFormattedRequestProto_descriptor;
8539      }
8540
8541      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8542          internalGetFieldAccessorTable() {
8543        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_IsFormattedRequestProto_fieldAccessorTable
8544            .ensureFieldAccessorsInitialized(
8545                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.Builder.class);
8546      }
8547
8548      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.newBuilder()
8549      private Builder() {
8550        maybeForceBuilderInitialization();
8551      }
8552
8553      private Builder(
8554          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
8555        super(parent);
8556        maybeForceBuilderInitialization();
8557      }
8558      private void maybeForceBuilderInitialization() {
8559        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
8560          getJidFieldBuilder();
8561        }
8562      }
8563      private static Builder create() {
8564        return new Builder();
8565      }
8566
8567      public Builder clear() {
8568        super.clear();
8569        if (jidBuilder_ == null) {
8570          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8571        } else {
8572          jidBuilder_.clear();
8573        }
8574        bitField0_ = (bitField0_ & ~0x00000001);
8575        return this;
8576      }
8577
8578      public Builder clone() {
8579        return create().mergeFrom(buildPartial());
8580      }
8581
8582      public com.google.protobuf.Descriptors.Descriptor
8583          getDescriptorForType() {
8584        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_IsFormattedRequestProto_descriptor;
8585      }
8586
8587      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto getDefaultInstanceForType() {
8588        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance();
8589      }
8590
8591      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto build() {
8592        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto result = buildPartial();
8593        if (!result.isInitialized()) {
8594          throw newUninitializedMessageException(result);
8595        }
8596        return result;
8597      }
8598
8599      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto buildPartial() {
8600        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto(this);
8601        int from_bitField0_ = bitField0_;
8602        int to_bitField0_ = 0;
8603        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
8604          to_bitField0_ |= 0x00000001;
8605        }
8606        if (jidBuilder_ == null) {
8607          result.jid_ = jid_;
8608        } else {
8609          result.jid_ = jidBuilder_.build();
8610        }
8611        result.bitField0_ = to_bitField0_;
8612        onBuilt();
8613        return result;
8614      }
8615
8616      public Builder mergeFrom(com.google.protobuf.Message other) {
8617        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto) {
8618          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)other);
8619        } else {
8620          super.mergeFrom(other);
8621          return this;
8622        }
8623      }
8624
8625      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto other) {
8626        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance()) return this;
8627        if (other.hasJid()) {
8628          mergeJid(other.getJid());
8629        }
8630        this.mergeUnknownFields(other.getUnknownFields());
8631        return this;
8632      }
8633
8634      public final boolean isInitialized() {
8635        if (!hasJid()) {
8636          
8637          return false;
8638        }
8639        if (!getJid().isInitialized()) {
8640          
8641          return false;
8642        }
8643        return true;
8644      }
8645
8646      public Builder mergeFrom(
8647          com.google.protobuf.CodedInputStream input,
8648          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8649          throws java.io.IOException {
8650        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parsedMessage = null;
8651        try {
8652          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
8653        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
8654          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto) e.getUnfinishedMessage();
8655          throw e;
8656        } finally {
8657          if (parsedMessage != null) {
8658            mergeFrom(parsedMessage);
8659          }
8660        }
8661        return this;
8662      }
8663      private int bitField0_;
8664
8665      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
8666      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8667      private com.google.protobuf.SingleFieldBuilder<
8668          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
8669      /**
8670       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8671       */
8672      public boolean hasJid() {
8673        return ((bitField0_ & 0x00000001) == 0x00000001);
8674      }
8675      /**
8676       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8677       */
8678      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
8679        if (jidBuilder_ == null) {
8680          return jid_;
8681        } else {
8682          return jidBuilder_.getMessage();
8683        }
8684      }
8685      /**
8686       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8687       */
8688      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
8689        if (jidBuilder_ == null) {
8690          if (value == null) {
8691            throw new NullPointerException();
8692          }
8693          jid_ = value;
8694          onChanged();
8695        } else {
8696          jidBuilder_.setMessage(value);
8697        }
8698        bitField0_ |= 0x00000001;
8699        return this;
8700      }
8701      /**
8702       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8703       */
8704      public Builder setJid(
8705          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
8706        if (jidBuilder_ == null) {
8707          jid_ = builderForValue.build();
8708          onChanged();
8709        } else {
8710          jidBuilder_.setMessage(builderForValue.build());
8711        }
8712        bitField0_ |= 0x00000001;
8713        return this;
8714      }
8715      /**
8716       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8717       */
8718      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
8719        if (jidBuilder_ == null) {
8720          if (((bitField0_ & 0x00000001) == 0x00000001) &&
8721              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
8722            jid_ =
8723              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
8724          } else {
8725            jid_ = value;
8726          }
8727          onChanged();
8728        } else {
8729          jidBuilder_.mergeFrom(value);
8730        }
8731        bitField0_ |= 0x00000001;
8732        return this;
8733      }
8734      /**
8735       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8736       */
8737      public Builder clearJid() {
8738        if (jidBuilder_ == null) {
8739          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8740          onChanged();
8741        } else {
8742          jidBuilder_.clear();
8743        }
8744        bitField0_ = (bitField0_ & ~0x00000001);
8745        return this;
8746      }
8747      /**
8748       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8749       */
8750      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
8751        bitField0_ |= 0x00000001;
8752        onChanged();
8753        return getJidFieldBuilder().getBuilder();
8754      }
8755      /**
8756       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8757       */
8758      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
8759        if (jidBuilder_ != null) {
8760          return jidBuilder_.getMessageOrBuilder();
8761        } else {
8762          return jid_;
8763        }
8764      }
8765      /**
8766       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
8767       */
8768      private com.google.protobuf.SingleFieldBuilder<
8769          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
8770          getJidFieldBuilder() {
8771        if (jidBuilder_ == null) {
8772          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
8773              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
8774                  jid_,
8775                  getParentForChildren(),
8776                  isClean());
8777          jid_ = null;
8778        }
8779        return jidBuilder_;
8780      }
8781
8782      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.IsFormattedRequestProto)
8783    }
8784
8785    static {
8786      defaultInstance = new IsFormattedRequestProto(true);
8787      defaultInstance.initFields();
8788    }
8789
8790    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.IsFormattedRequestProto)
8791  }
8792
8793  public interface IsFormattedResponseProtoOrBuilder
8794      extends com.google.protobuf.MessageOrBuilder {
8795
8796    // required bool isFormatted = 1;
8797    /**
8798     * <code>required bool isFormatted = 1;</code>
8799     */
8800    boolean hasIsFormatted();
8801    /**
8802     * <code>required bool isFormatted = 1;</code>
8803     */
8804    boolean getIsFormatted();
8805  }
8806  /**
8807   * Protobuf type {@code hadoop.hdfs.qjournal.IsFormattedResponseProto}
8808   */
8809  public static final class IsFormattedResponseProto extends
8810      com.google.protobuf.GeneratedMessage
8811      implements IsFormattedResponseProtoOrBuilder {
8812    // Use IsFormattedResponseProto.newBuilder() to construct.
8813    private IsFormattedResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
8814      super(builder);
8815      this.unknownFields = builder.getUnknownFields();
8816    }
8817    private IsFormattedResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
8818
8819    private static final IsFormattedResponseProto defaultInstance;
8820    public static IsFormattedResponseProto getDefaultInstance() {
8821      return defaultInstance;
8822    }
8823
8824    public IsFormattedResponseProto getDefaultInstanceForType() {
8825      return defaultInstance;
8826    }
8827
8828    private final com.google.protobuf.UnknownFieldSet unknownFields;
8829    @java.lang.Override
8830    public final com.google.protobuf.UnknownFieldSet
8831        getUnknownFields() {
8832      return this.unknownFields;
8833    }
8834    private IsFormattedResponseProto(
8835        com.google.protobuf.CodedInputStream input,
8836        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8837        throws com.google.protobuf.InvalidProtocolBufferException {
8838      initFields();
8839      int mutable_bitField0_ = 0;
8840      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
8841          com.google.protobuf.UnknownFieldSet.newBuilder();
8842      try {
8843        boolean done = false;
8844        while (!done) {
8845          int tag = input.readTag();
8846          switch (tag) {
8847            case 0:
8848              done = true;
8849              break;
8850            default: {
8851              if (!parseUnknownField(input, unknownFields,
8852                                     extensionRegistry, tag)) {
8853                done = true;
8854              }
8855              break;
8856            }
8857            case 8: {
8858              bitField0_ |= 0x00000001;
8859              isFormatted_ = input.readBool();
8860              break;
8861            }
8862          }
8863        }
8864      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
8865        throw e.setUnfinishedMessage(this);
8866      } catch (java.io.IOException e) {
8867        throw new com.google.protobuf.InvalidProtocolBufferException(
8868            e.getMessage()).setUnfinishedMessage(this);
8869      } finally {
8870        this.unknownFields = unknownFields.build();
8871        makeExtensionsImmutable();
8872      }
8873    }
8874    public static final com.google.protobuf.Descriptors.Descriptor
8875        getDescriptor() {
8876      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_IsFormattedResponseProto_descriptor;
8877    }
8878
8879    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8880        internalGetFieldAccessorTable() {
8881      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_IsFormattedResponseProto_fieldAccessorTable
8882          .ensureFieldAccessorsInitialized(
8883              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.Builder.class);
8884    }
8885
8886    public static com.google.protobuf.Parser<IsFormattedResponseProto> PARSER =
8887        new com.google.protobuf.AbstractParser<IsFormattedResponseProto>() {
8888      public IsFormattedResponseProto parsePartialFrom(
8889          com.google.protobuf.CodedInputStream input,
8890          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8891          throws com.google.protobuf.InvalidProtocolBufferException {
8892        return new IsFormattedResponseProto(input, extensionRegistry);
8893      }
8894    };
8895
8896    @java.lang.Override
8897    public com.google.protobuf.Parser<IsFormattedResponseProto> getParserForType() {
8898      return PARSER;
8899    }
8900
8901    private int bitField0_;
8902    // required bool isFormatted = 1;
8903    public static final int ISFORMATTED_FIELD_NUMBER = 1;
8904    private boolean isFormatted_;
8905    /**
8906     * <code>required bool isFormatted = 1;</code>
8907     */
8908    public boolean hasIsFormatted() {
8909      return ((bitField0_ & 0x00000001) == 0x00000001);
8910    }
8911    /**
8912     * <code>required bool isFormatted = 1;</code>
8913     */
8914    public boolean getIsFormatted() {
8915      return isFormatted_;
8916    }
8917
8918    private void initFields() {
8919      isFormatted_ = false;
8920    }
8921    private byte memoizedIsInitialized = -1;
8922    public final boolean isInitialized() {
8923      byte isInitialized = memoizedIsInitialized;
8924      if (isInitialized != -1) return isInitialized == 1;
8925
8926      if (!hasIsFormatted()) {
8927        memoizedIsInitialized = 0;
8928        return false;
8929      }
8930      memoizedIsInitialized = 1;
8931      return true;
8932    }
8933
8934    public void writeTo(com.google.protobuf.CodedOutputStream output)
8935                        throws java.io.IOException {
8936      getSerializedSize();
8937      if (((bitField0_ & 0x00000001) == 0x00000001)) {
8938        output.writeBool(1, isFormatted_);
8939      }
8940      getUnknownFields().writeTo(output);
8941    }
8942
8943    private int memoizedSerializedSize = -1;
8944    public int getSerializedSize() {
8945      int size = memoizedSerializedSize;
8946      if (size != -1) return size;
8947
8948      size = 0;
8949      if (((bitField0_ & 0x00000001) == 0x00000001)) {
8950        size += com.google.protobuf.CodedOutputStream
8951          .computeBoolSize(1, isFormatted_);
8952      }
8953      size += getUnknownFields().getSerializedSize();
8954      memoizedSerializedSize = size;
8955      return size;
8956    }
8957
8958    private static final long serialVersionUID = 0L;
8959    @java.lang.Override
8960    protected java.lang.Object writeReplace()
8961        throws java.io.ObjectStreamException {
8962      return super.writeReplace();
8963    }
8964
8965    @java.lang.Override
8966    public boolean equals(final java.lang.Object obj) {
8967      if (obj == this) {
8968       return true;
8969      }
8970      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto)) {
8971        return super.equals(obj);
8972      }
8973      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) obj;
8974
8975      boolean result = true;
8976      result = result && (hasIsFormatted() == other.hasIsFormatted());
8977      if (hasIsFormatted()) {
8978        result = result && (getIsFormatted()
8979            == other.getIsFormatted());
8980      }
8981      result = result &&
8982          getUnknownFields().equals(other.getUnknownFields());
8983      return result;
8984    }
8985
8986    private int memoizedHashCode = 0;
8987    @java.lang.Override
8988    public int hashCode() {
8989      if (memoizedHashCode != 0) {
8990        return memoizedHashCode;
8991      }
8992      int hash = 41;
8993      hash = (19 * hash) + getDescriptorForType().hashCode();
8994      if (hasIsFormatted()) {
8995        hash = (37 * hash) + ISFORMATTED_FIELD_NUMBER;
8996        hash = (53 * hash) + hashBoolean(getIsFormatted());
8997      }
8998      hash = (29 * hash) + getUnknownFields().hashCode();
8999      memoizedHashCode = hash;
9000      return hash;
9001    }
9002
9003    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
9004        com.google.protobuf.ByteString data)
9005        throws com.google.protobuf.InvalidProtocolBufferException {
9006      return PARSER.parseFrom(data);
9007    }
9008    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
9009        com.google.protobuf.ByteString data,
9010        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9011        throws com.google.protobuf.InvalidProtocolBufferException {
9012      return PARSER.parseFrom(data, extensionRegistry);
9013    }
9014    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(byte[] data)
9015        throws com.google.protobuf.InvalidProtocolBufferException {
9016      return PARSER.parseFrom(data);
9017    }
9018    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
9019        byte[] data,
9020        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9021        throws com.google.protobuf.InvalidProtocolBufferException {
9022      return PARSER.parseFrom(data, extensionRegistry);
9023    }
9024    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(java.io.InputStream input)
9025        throws java.io.IOException {
9026      return PARSER.parseFrom(input);
9027    }
9028    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
9029        java.io.InputStream input,
9030        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9031        throws java.io.IOException {
9032      return PARSER.parseFrom(input, extensionRegistry);
9033    }
9034    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseDelimitedFrom(java.io.InputStream input)
9035        throws java.io.IOException {
9036      return PARSER.parseDelimitedFrom(input);
9037    }
9038    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseDelimitedFrom(
9039        java.io.InputStream input,
9040        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9041        throws java.io.IOException {
9042      return PARSER.parseDelimitedFrom(input, extensionRegistry);
9043    }
9044    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
9045        com.google.protobuf.CodedInputStream input)
9046        throws java.io.IOException {
9047      return PARSER.parseFrom(input);
9048    }
9049    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
9050        com.google.protobuf.CodedInputStream input,
9051        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9052        throws java.io.IOException {
9053      return PARSER.parseFrom(input, extensionRegistry);
9054    }
9055
9056    public static Builder newBuilder() { return Builder.create(); }
9057    public Builder newBuilderForType() { return newBuilder(); }
9058    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto prototype) {
9059      return newBuilder().mergeFrom(prototype);
9060    }
9061    public Builder toBuilder() { return newBuilder(this); }
9062
9063    @java.lang.Override
9064    protected Builder newBuilderForType(
9065        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
9066      Builder builder = new Builder(parent);
9067      return builder;
9068    }
9069    /**
9070     * Protobuf type {@code hadoop.hdfs.qjournal.IsFormattedResponseProto}
9071     */
9072    public static final class Builder extends
9073        com.google.protobuf.GeneratedMessage.Builder<Builder>
9074       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProtoOrBuilder {
9075      public static final com.google.protobuf.Descriptors.Descriptor
9076          getDescriptor() {
9077        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_IsFormattedResponseProto_descriptor;
9078      }
9079
9080      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
9081          internalGetFieldAccessorTable() {
9082        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_IsFormattedResponseProto_fieldAccessorTable
9083            .ensureFieldAccessorsInitialized(
9084                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.Builder.class);
9085      }
9086
9087      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.newBuilder()
9088      private Builder() {
9089        maybeForceBuilderInitialization();
9090      }
9091
9092      private Builder(
9093          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
9094        super(parent);
9095        maybeForceBuilderInitialization();
9096      }
9097      private void maybeForceBuilderInitialization() {
9098        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
9099        }
9100      }
9101      private static Builder create() {
9102        return new Builder();
9103      }
9104
9105      public Builder clear() {
9106        super.clear();
9107        isFormatted_ = false;
9108        bitField0_ = (bitField0_ & ~0x00000001);
9109        return this;
9110      }
9111
9112      public Builder clone() {
9113        return create().mergeFrom(buildPartial());
9114      }
9115
9116      public com.google.protobuf.Descriptors.Descriptor
9117          getDescriptorForType() {
9118        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_IsFormattedResponseProto_descriptor;
9119      }
9120
9121      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto getDefaultInstanceForType() {
9122        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance();
9123      }
9124
9125      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto build() {
9126        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto result = buildPartial();
9127        if (!result.isInitialized()) {
9128          throw newUninitializedMessageException(result);
9129        }
9130        return result;
9131      }
9132
9133      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto buildPartial() {
9134        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto(this);
9135        int from_bitField0_ = bitField0_;
9136        int to_bitField0_ = 0;
9137        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
9138          to_bitField0_ |= 0x00000001;
9139        }
9140        result.isFormatted_ = isFormatted_;
9141        result.bitField0_ = to_bitField0_;
9142        onBuilt();
9143        return result;
9144      }
9145
9146      public Builder mergeFrom(com.google.protobuf.Message other) {
9147        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) {
9148          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto)other);
9149        } else {
9150          super.mergeFrom(other);
9151          return this;
9152        }
9153      }
9154
9155      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto other) {
9156        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance()) return this;
9157        if (other.hasIsFormatted()) {
9158          setIsFormatted(other.getIsFormatted());
9159        }
9160        this.mergeUnknownFields(other.getUnknownFields());
9161        return this;
9162      }
9163
9164      public final boolean isInitialized() {
9165        if (!hasIsFormatted()) {
9166          
9167          return false;
9168        }
9169        return true;
9170      }
9171
9172      public Builder mergeFrom(
9173          com.google.protobuf.CodedInputStream input,
9174          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9175          throws java.io.IOException {
9176        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parsedMessage = null;
9177        try {
9178          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
9179        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
9180          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) e.getUnfinishedMessage();
9181          throw e;
9182        } finally {
9183          if (parsedMessage != null) {
9184            mergeFrom(parsedMessage);
9185          }
9186        }
9187        return this;
9188      }
9189      private int bitField0_;
9190
9191      // required bool isFormatted = 1;
9192      private boolean isFormatted_ ;
9193      /**
9194       * <code>required bool isFormatted = 1;</code>
9195       */
9196      public boolean hasIsFormatted() {
9197        return ((bitField0_ & 0x00000001) == 0x00000001);
9198      }
9199      /**
9200       * <code>required bool isFormatted = 1;</code>
9201       */
9202      public boolean getIsFormatted() {
9203        return isFormatted_;
9204      }
9205      /**
9206       * <code>required bool isFormatted = 1;</code>
9207       */
9208      public Builder setIsFormatted(boolean value) {
9209        bitField0_ |= 0x00000001;
9210        isFormatted_ = value;
9211        onChanged();
9212        return this;
9213      }
9214      /**
9215       * <code>required bool isFormatted = 1;</code>
9216       */
9217      public Builder clearIsFormatted() {
9218        bitField0_ = (bitField0_ & ~0x00000001);
9219        isFormatted_ = false;
9220        onChanged();
9221        return this;
9222      }
9223
9224      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.IsFormattedResponseProto)
9225    }
9226
9227    static {
9228      defaultInstance = new IsFormattedResponseProto(true);
9229      defaultInstance.initFields();
9230    }
9231
9232    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.IsFormattedResponseProto)
9233  }
9234
9235  public interface DiscardSegmentsRequestProtoOrBuilder
9236      extends com.google.protobuf.MessageOrBuilder {
9237
9238    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
9239    /**
9240     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9241     */
9242    boolean hasJid();
9243    /**
9244     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9245     */
9246    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
9247    /**
9248     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9249     */
9250    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
9251
9252    // required uint64 startTxId = 2;
9253    /**
9254     * <code>required uint64 startTxId = 2;</code>
9255     */
9256    boolean hasStartTxId();
9257    /**
9258     * <code>required uint64 startTxId = 2;</code>
9259     */
9260    long getStartTxId();
9261  }
9262  /**
9263   * Protobuf type {@code hadoop.hdfs.qjournal.DiscardSegmentsRequestProto}
9264   *
9265   * <pre>
9266   **
9267   * discardSegments()
9268   * </pre>
9269   */
9270  public static final class DiscardSegmentsRequestProto extends
9271      com.google.protobuf.GeneratedMessage
9272      implements DiscardSegmentsRequestProtoOrBuilder {
9273    // Use DiscardSegmentsRequestProto.newBuilder() to construct.
9274    private DiscardSegmentsRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
9275      super(builder);
9276      this.unknownFields = builder.getUnknownFields();
9277    }
9278    private DiscardSegmentsRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
9279
9280    private static final DiscardSegmentsRequestProto defaultInstance;
9281    public static DiscardSegmentsRequestProto getDefaultInstance() {
9282      return defaultInstance;
9283    }
9284
9285    public DiscardSegmentsRequestProto getDefaultInstanceForType() {
9286      return defaultInstance;
9287    }
9288
9289    private final com.google.protobuf.UnknownFieldSet unknownFields;
9290    @java.lang.Override
9291    public final com.google.protobuf.UnknownFieldSet
9292        getUnknownFields() {
9293      return this.unknownFields;
9294    }
9295    private DiscardSegmentsRequestProto(
9296        com.google.protobuf.CodedInputStream input,
9297        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9298        throws com.google.protobuf.InvalidProtocolBufferException {
9299      initFields();
9300      int mutable_bitField0_ = 0;
9301      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
9302          com.google.protobuf.UnknownFieldSet.newBuilder();
9303      try {
9304        boolean done = false;
9305        while (!done) {
9306          int tag = input.readTag();
9307          switch (tag) {
9308            case 0:
9309              done = true;
9310              break;
9311            default: {
9312              if (!parseUnknownField(input, unknownFields,
9313                                     extensionRegistry, tag)) {
9314                done = true;
9315              }
9316              break;
9317            }
9318            case 10: {
9319              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
9320              if (((bitField0_ & 0x00000001) == 0x00000001)) {
9321                subBuilder = jid_.toBuilder();
9322              }
9323              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
9324              if (subBuilder != null) {
9325                subBuilder.mergeFrom(jid_);
9326                jid_ = subBuilder.buildPartial();
9327              }
9328              bitField0_ |= 0x00000001;
9329              break;
9330            }
9331            case 16: {
9332              bitField0_ |= 0x00000002;
9333              startTxId_ = input.readUInt64();
9334              break;
9335            }
9336          }
9337        }
9338      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
9339        throw e.setUnfinishedMessage(this);
9340      } catch (java.io.IOException e) {
9341        throw new com.google.protobuf.InvalidProtocolBufferException(
9342            e.getMessage()).setUnfinishedMessage(this);
9343      } finally {
9344        this.unknownFields = unknownFields.build();
9345        makeExtensionsImmutable();
9346      }
9347    }
9348    public static final com.google.protobuf.Descriptors.Descriptor
9349        getDescriptor() {
9350      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DiscardSegmentsRequestProto_descriptor;
9351    }
9352
9353    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
9354        internalGetFieldAccessorTable() {
9355      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DiscardSegmentsRequestProto_fieldAccessorTable
9356          .ensureFieldAccessorsInitialized(
9357              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto.Builder.class);
9358    }
9359
9360    public static com.google.protobuf.Parser<DiscardSegmentsRequestProto> PARSER =
9361        new com.google.protobuf.AbstractParser<DiscardSegmentsRequestProto>() {
9362      public DiscardSegmentsRequestProto parsePartialFrom(
9363          com.google.protobuf.CodedInputStream input,
9364          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9365          throws com.google.protobuf.InvalidProtocolBufferException {
9366        return new DiscardSegmentsRequestProto(input, extensionRegistry);
9367      }
9368    };
9369
9370    @java.lang.Override
9371    public com.google.protobuf.Parser<DiscardSegmentsRequestProto> getParserForType() {
9372      return PARSER;
9373    }
9374
9375    private int bitField0_;
9376    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
9377    public static final int JID_FIELD_NUMBER = 1;
9378    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
9379    /**
9380     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9381     */
9382    public boolean hasJid() {
9383      return ((bitField0_ & 0x00000001) == 0x00000001);
9384    }
9385    /**
9386     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9387     */
9388    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
9389      return jid_;
9390    }
9391    /**
9392     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9393     */
9394    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
9395      return jid_;
9396    }
9397
9398    // required uint64 startTxId = 2;
9399    public static final int STARTTXID_FIELD_NUMBER = 2;
9400    private long startTxId_;
9401    /**
9402     * <code>required uint64 startTxId = 2;</code>
9403     */
9404    public boolean hasStartTxId() {
9405      return ((bitField0_ & 0x00000002) == 0x00000002);
9406    }
9407    /**
9408     * <code>required uint64 startTxId = 2;</code>
9409     */
9410    public long getStartTxId() {
9411      return startTxId_;
9412    }
9413
9414    private void initFields() {
9415      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9416      startTxId_ = 0L;
9417    }
9418    private byte memoizedIsInitialized = -1;
9419    public final boolean isInitialized() {
9420      byte isInitialized = memoizedIsInitialized;
9421      if (isInitialized != -1) return isInitialized == 1;
9422
9423      if (!hasJid()) {
9424        memoizedIsInitialized = 0;
9425        return false;
9426      }
9427      if (!hasStartTxId()) {
9428        memoizedIsInitialized = 0;
9429        return false;
9430      }
9431      if (!getJid().isInitialized()) {
9432        memoizedIsInitialized = 0;
9433        return false;
9434      }
9435      memoizedIsInitialized = 1;
9436      return true;
9437    }
9438
9439    public void writeTo(com.google.protobuf.CodedOutputStream output)
9440                        throws java.io.IOException {
9441      getSerializedSize();
9442      if (((bitField0_ & 0x00000001) == 0x00000001)) {
9443        output.writeMessage(1, jid_);
9444      }
9445      if (((bitField0_ & 0x00000002) == 0x00000002)) {
9446        output.writeUInt64(2, startTxId_);
9447      }
9448      getUnknownFields().writeTo(output);
9449    }
9450
9451    private int memoizedSerializedSize = -1;
9452    public int getSerializedSize() {
9453      int size = memoizedSerializedSize;
9454      if (size != -1) return size;
9455
9456      size = 0;
9457      if (((bitField0_ & 0x00000001) == 0x00000001)) {
9458        size += com.google.protobuf.CodedOutputStream
9459          .computeMessageSize(1, jid_);
9460      }
9461      if (((bitField0_ & 0x00000002) == 0x00000002)) {
9462        size += com.google.protobuf.CodedOutputStream
9463          .computeUInt64Size(2, startTxId_);
9464      }
9465      size += getUnknownFields().getSerializedSize();
9466      memoizedSerializedSize = size;
9467      return size;
9468    }
9469
9470    private static final long serialVersionUID = 0L;
9471    @java.lang.Override
9472    protected java.lang.Object writeReplace()
9473        throws java.io.ObjectStreamException {
9474      return super.writeReplace();
9475    }
9476
9477    @java.lang.Override
9478    public boolean equals(final java.lang.Object obj) {
9479      if (obj == this) {
9480       return true;
9481      }
9482      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto)) {
9483        return super.equals(obj);
9484      }
9485      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto) obj;
9486
9487      boolean result = true;
9488      result = result && (hasJid() == other.hasJid());
9489      if (hasJid()) {
9490        result = result && getJid()
9491            .equals(other.getJid());
9492      }
9493      result = result && (hasStartTxId() == other.hasStartTxId());
9494      if (hasStartTxId()) {
9495        result = result && (getStartTxId()
9496            == other.getStartTxId());
9497      }
9498      result = result &&
9499          getUnknownFields().equals(other.getUnknownFields());
9500      return result;
9501    }
9502
9503    private int memoizedHashCode = 0;
9504    @java.lang.Override
9505    public int hashCode() {
9506      if (memoizedHashCode != 0) {
9507        return memoizedHashCode;
9508      }
9509      int hash = 41;
9510      hash = (19 * hash) + getDescriptorForType().hashCode();
9511      if (hasJid()) {
9512        hash = (37 * hash) + JID_FIELD_NUMBER;
9513        hash = (53 * hash) + getJid().hashCode();
9514      }
9515      if (hasStartTxId()) {
9516        hash = (37 * hash) + STARTTXID_FIELD_NUMBER;
9517        hash = (53 * hash) + hashLong(getStartTxId());
9518      }
9519      hash = (29 * hash) + getUnknownFields().hashCode();
9520      memoizedHashCode = hash;
9521      return hash;
9522    }
9523
9524    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto parseFrom(
9525        com.google.protobuf.ByteString data)
9526        throws com.google.protobuf.InvalidProtocolBufferException {
9527      return PARSER.parseFrom(data);
9528    }
9529    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto parseFrom(
9530        com.google.protobuf.ByteString data,
9531        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9532        throws com.google.protobuf.InvalidProtocolBufferException {
9533      return PARSER.parseFrom(data, extensionRegistry);
9534    }
9535    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto parseFrom(byte[] data)
9536        throws com.google.protobuf.InvalidProtocolBufferException {
9537      return PARSER.parseFrom(data);
9538    }
9539    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto parseFrom(
9540        byte[] data,
9541        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9542        throws com.google.protobuf.InvalidProtocolBufferException {
9543      return PARSER.parseFrom(data, extensionRegistry);
9544    }
9545    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto parseFrom(java.io.InputStream input)
9546        throws java.io.IOException {
9547      return PARSER.parseFrom(input);
9548    }
9549    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto parseFrom(
9550        java.io.InputStream input,
9551        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9552        throws java.io.IOException {
9553      return PARSER.parseFrom(input, extensionRegistry);
9554    }
9555    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto parseDelimitedFrom(java.io.InputStream input)
9556        throws java.io.IOException {
9557      return PARSER.parseDelimitedFrom(input);
9558    }
9559    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto parseDelimitedFrom(
9560        java.io.InputStream input,
9561        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9562        throws java.io.IOException {
9563      return PARSER.parseDelimitedFrom(input, extensionRegistry);
9564    }
9565    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto parseFrom(
9566        com.google.protobuf.CodedInputStream input)
9567        throws java.io.IOException {
9568      return PARSER.parseFrom(input);
9569    }
9570    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto parseFrom(
9571        com.google.protobuf.CodedInputStream input,
9572        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9573        throws java.io.IOException {
9574      return PARSER.parseFrom(input, extensionRegistry);
9575    }
9576
9577    public static Builder newBuilder() { return Builder.create(); }
9578    public Builder newBuilderForType() { return newBuilder(); }
9579    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto prototype) {
9580      return newBuilder().mergeFrom(prototype);
9581    }
9582    public Builder toBuilder() { return newBuilder(this); }
9583
9584    @java.lang.Override
9585    protected Builder newBuilderForType(
9586        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
9587      Builder builder = new Builder(parent);
9588      return builder;
9589    }
9590    /**
9591     * Protobuf type {@code hadoop.hdfs.qjournal.DiscardSegmentsRequestProto}
9592     *
9593     * <pre>
9594     **
9595     * discardSegments()
9596     * </pre>
9597     */
9598    public static final class Builder extends
9599        com.google.protobuf.GeneratedMessage.Builder<Builder>
9600       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProtoOrBuilder {
9601      public static final com.google.protobuf.Descriptors.Descriptor
9602          getDescriptor() {
9603        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DiscardSegmentsRequestProto_descriptor;
9604      }
9605
9606      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
9607          internalGetFieldAccessorTable() {
9608        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DiscardSegmentsRequestProto_fieldAccessorTable
9609            .ensureFieldAccessorsInitialized(
9610                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto.Builder.class);
9611      }
9612
9613      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto.newBuilder()
9614      private Builder() {
9615        maybeForceBuilderInitialization();
9616      }
9617
9618      private Builder(
9619          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
9620        super(parent);
9621        maybeForceBuilderInitialization();
9622      }
9623      private void maybeForceBuilderInitialization() {
9624        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
9625          getJidFieldBuilder();
9626        }
9627      }
9628      private static Builder create() {
9629        return new Builder();
9630      }
9631
9632      public Builder clear() {
9633        super.clear();
9634        if (jidBuilder_ == null) {
9635          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9636        } else {
9637          jidBuilder_.clear();
9638        }
9639        bitField0_ = (bitField0_ & ~0x00000001);
9640        startTxId_ = 0L;
9641        bitField0_ = (bitField0_ & ~0x00000002);
9642        return this;
9643      }
9644
9645      public Builder clone() {
9646        return create().mergeFrom(buildPartial());
9647      }
9648
9649      public com.google.protobuf.Descriptors.Descriptor
9650          getDescriptorForType() {
9651        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DiscardSegmentsRequestProto_descriptor;
9652      }
9653
9654      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto getDefaultInstanceForType() {
9655        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto.getDefaultInstance();
9656      }
9657
9658      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto build() {
9659        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto result = buildPartial();
9660        if (!result.isInitialized()) {
9661          throw newUninitializedMessageException(result);
9662        }
9663        return result;
9664      }
9665
9666      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto buildPartial() {
9667        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto(this);
9668        int from_bitField0_ = bitField0_;
9669        int to_bitField0_ = 0;
9670        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
9671          to_bitField0_ |= 0x00000001;
9672        }
9673        if (jidBuilder_ == null) {
9674          result.jid_ = jid_;
9675        } else {
9676          result.jid_ = jidBuilder_.build();
9677        }
9678        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
9679          to_bitField0_ |= 0x00000002;
9680        }
9681        result.startTxId_ = startTxId_;
9682        result.bitField0_ = to_bitField0_;
9683        onBuilt();
9684        return result;
9685      }
9686
9687      public Builder mergeFrom(com.google.protobuf.Message other) {
9688        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto) {
9689          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto)other);
9690        } else {
9691          super.mergeFrom(other);
9692          return this;
9693        }
9694      }
9695
9696      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto other) {
9697        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto.getDefaultInstance()) return this;
9698        if (other.hasJid()) {
9699          mergeJid(other.getJid());
9700        }
9701        if (other.hasStartTxId()) {
9702          setStartTxId(other.getStartTxId());
9703        }
9704        this.mergeUnknownFields(other.getUnknownFields());
9705        return this;
9706      }
9707
9708      public final boolean isInitialized() {
9709        if (!hasJid()) {
9710          
9711          return false;
9712        }
9713        if (!hasStartTxId()) {
9714          
9715          return false;
9716        }
9717        if (!getJid().isInitialized()) {
9718          
9719          return false;
9720        }
9721        return true;
9722      }
9723
9724      public Builder mergeFrom(
9725          com.google.protobuf.CodedInputStream input,
9726          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9727          throws java.io.IOException {
9728        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto parsedMessage = null;
9729        try {
9730          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
9731        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
9732          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto) e.getUnfinishedMessage();
9733          throw e;
9734        } finally {
9735          if (parsedMessage != null) {
9736            mergeFrom(parsedMessage);
9737          }
9738        }
9739        return this;
9740      }
9741      private int bitField0_;
9742
9743      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
9744      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9745      private com.google.protobuf.SingleFieldBuilder<
9746          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
9747      /**
9748       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9749       */
9750      public boolean hasJid() {
9751        return ((bitField0_ & 0x00000001) == 0x00000001);
9752      }
9753      /**
9754       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9755       */
9756      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
9757        if (jidBuilder_ == null) {
9758          return jid_;
9759        } else {
9760          return jidBuilder_.getMessage();
9761        }
9762      }
9763      /**
9764       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9765       */
9766      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
9767        if (jidBuilder_ == null) {
9768          if (value == null) {
9769            throw new NullPointerException();
9770          }
9771          jid_ = value;
9772          onChanged();
9773        } else {
9774          jidBuilder_.setMessage(value);
9775        }
9776        bitField0_ |= 0x00000001;
9777        return this;
9778      }
9779      /**
9780       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9781       */
9782      public Builder setJid(
9783          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
9784        if (jidBuilder_ == null) {
9785          jid_ = builderForValue.build();
9786          onChanged();
9787        } else {
9788          jidBuilder_.setMessage(builderForValue.build());
9789        }
9790        bitField0_ |= 0x00000001;
9791        return this;
9792      }
9793      /**
9794       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9795       */
9796      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
9797        if (jidBuilder_ == null) {
9798          if (((bitField0_ & 0x00000001) == 0x00000001) &&
9799              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
9800            jid_ =
9801              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
9802          } else {
9803            jid_ = value;
9804          }
9805          onChanged();
9806        } else {
9807          jidBuilder_.mergeFrom(value);
9808        }
9809        bitField0_ |= 0x00000001;
9810        return this;
9811      }
9812      /**
9813       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9814       */
9815      public Builder clearJid() {
9816        if (jidBuilder_ == null) {
9817          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9818          onChanged();
9819        } else {
9820          jidBuilder_.clear();
9821        }
9822        bitField0_ = (bitField0_ & ~0x00000001);
9823        return this;
9824      }
9825      /**
9826       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9827       */
9828      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
9829        bitField0_ |= 0x00000001;
9830        onChanged();
9831        return getJidFieldBuilder().getBuilder();
9832      }
9833      /**
9834       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9835       */
9836      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
9837        if (jidBuilder_ != null) {
9838          return jidBuilder_.getMessageOrBuilder();
9839        } else {
9840          return jid_;
9841        }
9842      }
9843      /**
9844       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
9845       */
9846      private com.google.protobuf.SingleFieldBuilder<
9847          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
9848          getJidFieldBuilder() {
9849        if (jidBuilder_ == null) {
9850          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
9851              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
9852                  jid_,
9853                  getParentForChildren(),
9854                  isClean());
9855          jid_ = null;
9856        }
9857        return jidBuilder_;
9858      }
9859
9860      // required uint64 startTxId = 2;
9861      private long startTxId_ ;
9862      /**
9863       * <code>required uint64 startTxId = 2;</code>
9864       */
9865      public boolean hasStartTxId() {
9866        return ((bitField0_ & 0x00000002) == 0x00000002);
9867      }
9868      /**
9869       * <code>required uint64 startTxId = 2;</code>
9870       */
9871      public long getStartTxId() {
9872        return startTxId_;
9873      }
9874      /**
9875       * <code>required uint64 startTxId = 2;</code>
9876       */
9877      public Builder setStartTxId(long value) {
9878        bitField0_ |= 0x00000002;
9879        startTxId_ = value;
9880        onChanged();
9881        return this;
9882      }
9883      /**
9884       * <code>required uint64 startTxId = 2;</code>
9885       */
9886      public Builder clearStartTxId() {
9887        bitField0_ = (bitField0_ & ~0x00000002);
9888        startTxId_ = 0L;
9889        onChanged();
9890        return this;
9891      }
9892
9893      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.DiscardSegmentsRequestProto)
9894    }
9895
9896    static {
9897      defaultInstance = new DiscardSegmentsRequestProto(true);
9898      defaultInstance.initFields();
9899    }
9900
9901    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.DiscardSegmentsRequestProto)
9902  }
9903
9904  public interface DiscardSegmentsResponseProtoOrBuilder
9905      extends com.google.protobuf.MessageOrBuilder {
9906  }
9907  /**
9908   * Protobuf type {@code hadoop.hdfs.qjournal.DiscardSegmentsResponseProto}
9909   */
9910  public static final class DiscardSegmentsResponseProto extends
9911      com.google.protobuf.GeneratedMessage
9912      implements DiscardSegmentsResponseProtoOrBuilder {
9913    // Use DiscardSegmentsResponseProto.newBuilder() to construct.
9914    private DiscardSegmentsResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
9915      super(builder);
9916      this.unknownFields = builder.getUnknownFields();
9917    }
9918    private DiscardSegmentsResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
9919
9920    private static final DiscardSegmentsResponseProto defaultInstance;
9921    public static DiscardSegmentsResponseProto getDefaultInstance() {
9922      return defaultInstance;
9923    }
9924
9925    public DiscardSegmentsResponseProto getDefaultInstanceForType() {
9926      return defaultInstance;
9927    }
9928
9929    private final com.google.protobuf.UnknownFieldSet unknownFields;
9930    @java.lang.Override
9931    public final com.google.protobuf.UnknownFieldSet
9932        getUnknownFields() {
9933      return this.unknownFields;
9934    }
9935    private DiscardSegmentsResponseProto(
9936        com.google.protobuf.CodedInputStream input,
9937        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9938        throws com.google.protobuf.InvalidProtocolBufferException {
9939      initFields();
9940      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
9941          com.google.protobuf.UnknownFieldSet.newBuilder();
9942      try {
9943        boolean done = false;
9944        while (!done) {
9945          int tag = input.readTag();
9946          switch (tag) {
9947            case 0:
9948              done = true;
9949              break;
9950            default: {
9951              if (!parseUnknownField(input, unknownFields,
9952                                     extensionRegistry, tag)) {
9953                done = true;
9954              }
9955              break;
9956            }
9957          }
9958        }
9959      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
9960        throw e.setUnfinishedMessage(this);
9961      } catch (java.io.IOException e) {
9962        throw new com.google.protobuf.InvalidProtocolBufferException(
9963            e.getMessage()).setUnfinishedMessage(this);
9964      } finally {
9965        this.unknownFields = unknownFields.build();
9966        makeExtensionsImmutable();
9967      }
9968    }
9969    public static final com.google.protobuf.Descriptors.Descriptor
9970        getDescriptor() {
9971      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DiscardSegmentsResponseProto_descriptor;
9972    }
9973
9974    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
9975        internalGetFieldAccessorTable() {
9976      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DiscardSegmentsResponseProto_fieldAccessorTable
9977          .ensureFieldAccessorsInitialized(
9978              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.Builder.class);
9979    }
9980
9981    public static com.google.protobuf.Parser<DiscardSegmentsResponseProto> PARSER =
9982        new com.google.protobuf.AbstractParser<DiscardSegmentsResponseProto>() {
9983      public DiscardSegmentsResponseProto parsePartialFrom(
9984          com.google.protobuf.CodedInputStream input,
9985          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9986          throws com.google.protobuf.InvalidProtocolBufferException {
9987        return new DiscardSegmentsResponseProto(input, extensionRegistry);
9988      }
9989    };
9990
9991    @java.lang.Override
9992    public com.google.protobuf.Parser<DiscardSegmentsResponseProto> getParserForType() {
9993      return PARSER;
9994    }
9995
9996    private void initFields() {
9997    }
9998    private byte memoizedIsInitialized = -1;
9999    public final boolean isInitialized() {
10000      byte isInitialized = memoizedIsInitialized;
10001      if (isInitialized != -1) return isInitialized == 1;
10002
10003      memoizedIsInitialized = 1;
10004      return true;
10005    }
10006
10007    public void writeTo(com.google.protobuf.CodedOutputStream output)
10008                        throws java.io.IOException {
10009      getSerializedSize();
10010      getUnknownFields().writeTo(output);
10011    }
10012
10013    private int memoizedSerializedSize = -1;
10014    public int getSerializedSize() {
10015      int size = memoizedSerializedSize;
10016      if (size != -1) return size;
10017
10018      size = 0;
10019      size += getUnknownFields().getSerializedSize();
10020      memoizedSerializedSize = size;
10021      return size;
10022    }
10023
10024    private static final long serialVersionUID = 0L;
10025    @java.lang.Override
10026    protected java.lang.Object writeReplace()
10027        throws java.io.ObjectStreamException {
10028      return super.writeReplace();
10029    }
10030
10031    @java.lang.Override
10032    public boolean equals(final java.lang.Object obj) {
10033      if (obj == this) {
10034       return true;
10035      }
10036      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto)) {
10037        return super.equals(obj);
10038      }
10039      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto) obj;
10040
10041      boolean result = true;
10042      result = result &&
10043          getUnknownFields().equals(other.getUnknownFields());
10044      return result;
10045    }
10046
10047    private int memoizedHashCode = 0;
10048    @java.lang.Override
10049    public int hashCode() {
10050      if (memoizedHashCode != 0) {
10051        return memoizedHashCode;
10052      }
10053      int hash = 41;
10054      hash = (19 * hash) + getDescriptorForType().hashCode();
10055      hash = (29 * hash) + getUnknownFields().hashCode();
10056      memoizedHashCode = hash;
10057      return hash;
10058    }
10059
10060    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto parseFrom(
10061        com.google.protobuf.ByteString data)
10062        throws com.google.protobuf.InvalidProtocolBufferException {
10063      return PARSER.parseFrom(data);
10064    }
10065    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto parseFrom(
10066        com.google.protobuf.ByteString data,
10067        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10068        throws com.google.protobuf.InvalidProtocolBufferException {
10069      return PARSER.parseFrom(data, extensionRegistry);
10070    }
10071    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto parseFrom(byte[] data)
10072        throws com.google.protobuf.InvalidProtocolBufferException {
10073      return PARSER.parseFrom(data);
10074    }
10075    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto parseFrom(
10076        byte[] data,
10077        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10078        throws com.google.protobuf.InvalidProtocolBufferException {
10079      return PARSER.parseFrom(data, extensionRegistry);
10080    }
10081    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto parseFrom(java.io.InputStream input)
10082        throws java.io.IOException {
10083      return PARSER.parseFrom(input);
10084    }
10085    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto parseFrom(
10086        java.io.InputStream input,
10087        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10088        throws java.io.IOException {
10089      return PARSER.parseFrom(input, extensionRegistry);
10090    }
10091    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto parseDelimitedFrom(java.io.InputStream input)
10092        throws java.io.IOException {
10093      return PARSER.parseDelimitedFrom(input);
10094    }
10095    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto parseDelimitedFrom(
10096        java.io.InputStream input,
10097        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10098        throws java.io.IOException {
10099      return PARSER.parseDelimitedFrom(input, extensionRegistry);
10100    }
10101    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto parseFrom(
10102        com.google.protobuf.CodedInputStream input)
10103        throws java.io.IOException {
10104      return PARSER.parseFrom(input);
10105    }
10106    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto parseFrom(
10107        com.google.protobuf.CodedInputStream input,
10108        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10109        throws java.io.IOException {
10110      return PARSER.parseFrom(input, extensionRegistry);
10111    }
10112
10113    public static Builder newBuilder() { return Builder.create(); }
10114    public Builder newBuilderForType() { return newBuilder(); }
10115    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto prototype) {
10116      return newBuilder().mergeFrom(prototype);
10117    }
10118    public Builder toBuilder() { return newBuilder(this); }
10119
10120    @java.lang.Override
10121    protected Builder newBuilderForType(
10122        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
10123      Builder builder = new Builder(parent);
10124      return builder;
10125    }
10126    /**
10127     * Protobuf type {@code hadoop.hdfs.qjournal.DiscardSegmentsResponseProto}
10128     */
10129    public static final class Builder extends
10130        com.google.protobuf.GeneratedMessage.Builder<Builder>
10131       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProtoOrBuilder {
10132      public static final com.google.protobuf.Descriptors.Descriptor
10133          getDescriptor() {
10134        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DiscardSegmentsResponseProto_descriptor;
10135      }
10136
10137      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
10138          internalGetFieldAccessorTable() {
10139        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DiscardSegmentsResponseProto_fieldAccessorTable
10140            .ensureFieldAccessorsInitialized(
10141                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.Builder.class);
10142      }
10143
10144      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.newBuilder()
10145      private Builder() {
10146        maybeForceBuilderInitialization();
10147      }
10148
10149      private Builder(
10150          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
10151        super(parent);
10152        maybeForceBuilderInitialization();
10153      }
10154      private void maybeForceBuilderInitialization() {
10155        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
10156        }
10157      }
10158      private static Builder create() {
10159        return new Builder();
10160      }
10161
10162      public Builder clear() {
10163        super.clear();
10164        return this;
10165      }
10166
10167      public Builder clone() {
10168        return create().mergeFrom(buildPartial());
10169      }
10170
10171      public com.google.protobuf.Descriptors.Descriptor
10172          getDescriptorForType() {
10173        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DiscardSegmentsResponseProto_descriptor;
10174      }
10175
10176      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto getDefaultInstanceForType() {
10177        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.getDefaultInstance();
10178      }
10179
10180      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto build() {
10181        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto result = buildPartial();
10182        if (!result.isInitialized()) {
10183          throw newUninitializedMessageException(result);
10184        }
10185        return result;
10186      }
10187
10188      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto buildPartial() {
10189        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto(this);
10190        onBuilt();
10191        return result;
10192      }
10193
10194      public Builder mergeFrom(com.google.protobuf.Message other) {
10195        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto) {
10196          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto)other);
10197        } else {
10198          super.mergeFrom(other);
10199          return this;
10200        }
10201      }
10202
10203      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto other) {
10204        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.getDefaultInstance()) return this;
10205        this.mergeUnknownFields(other.getUnknownFields());
10206        return this;
10207      }
10208
10209      public final boolean isInitialized() {
10210        return true;
10211      }
10212
10213      public Builder mergeFrom(
10214          com.google.protobuf.CodedInputStream input,
10215          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10216          throws java.io.IOException {
10217        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto parsedMessage = null;
10218        try {
10219          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
10220        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
10221          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto) e.getUnfinishedMessage();
10222          throw e;
10223        } finally {
10224          if (parsedMessage != null) {
10225            mergeFrom(parsedMessage);
10226          }
10227        }
10228        return this;
10229      }
10230
10231      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.DiscardSegmentsResponseProto)
10232    }
10233
10234    static {
10235      defaultInstance = new DiscardSegmentsResponseProto(true);
10236      defaultInstance.initFields();
10237    }
10238
10239    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.DiscardSegmentsResponseProto)
10240  }
10241
10242  public interface GetJournalCTimeRequestProtoOrBuilder
10243      extends com.google.protobuf.MessageOrBuilder {
10244
10245    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
10246    /**
10247     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10248     */
10249    boolean hasJid();
10250    /**
10251     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10252     */
10253    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
10254    /**
10255     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10256     */
10257    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
10258  }
10259  /**
10260   * Protobuf type {@code hadoop.hdfs.qjournal.GetJournalCTimeRequestProto}
10261   *
10262   * <pre>
10263   **
10264   * getJournalCTime()
10265   * </pre>
10266   */
10267  public static final class GetJournalCTimeRequestProto extends
10268      com.google.protobuf.GeneratedMessage
10269      implements GetJournalCTimeRequestProtoOrBuilder {
10270    // Use GetJournalCTimeRequestProto.newBuilder() to construct.
10271    private GetJournalCTimeRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
10272      super(builder);
10273      this.unknownFields = builder.getUnknownFields();
10274    }
10275    private GetJournalCTimeRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
10276
10277    private static final GetJournalCTimeRequestProto defaultInstance;
10278    public static GetJournalCTimeRequestProto getDefaultInstance() {
10279      return defaultInstance;
10280    }
10281
10282    public GetJournalCTimeRequestProto getDefaultInstanceForType() {
10283      return defaultInstance;
10284    }
10285
10286    private final com.google.protobuf.UnknownFieldSet unknownFields;
10287    @java.lang.Override
10288    public final com.google.protobuf.UnknownFieldSet
10289        getUnknownFields() {
10290      return this.unknownFields;
10291    }
10292    private GetJournalCTimeRequestProto(
10293        com.google.protobuf.CodedInputStream input,
10294        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10295        throws com.google.protobuf.InvalidProtocolBufferException {
10296      initFields();
10297      int mutable_bitField0_ = 0;
10298      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
10299          com.google.protobuf.UnknownFieldSet.newBuilder();
10300      try {
10301        boolean done = false;
10302        while (!done) {
10303          int tag = input.readTag();
10304          switch (tag) {
10305            case 0:
10306              done = true;
10307              break;
10308            default: {
10309              if (!parseUnknownField(input, unknownFields,
10310                                     extensionRegistry, tag)) {
10311                done = true;
10312              }
10313              break;
10314            }
10315            case 10: {
10316              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
10317              if (((bitField0_ & 0x00000001) == 0x00000001)) {
10318                subBuilder = jid_.toBuilder();
10319              }
10320              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
10321              if (subBuilder != null) {
10322                subBuilder.mergeFrom(jid_);
10323                jid_ = subBuilder.buildPartial();
10324              }
10325              bitField0_ |= 0x00000001;
10326              break;
10327            }
10328          }
10329        }
10330      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
10331        throw e.setUnfinishedMessage(this);
10332      } catch (java.io.IOException e) {
10333        throw new com.google.protobuf.InvalidProtocolBufferException(
10334            e.getMessage()).setUnfinishedMessage(this);
10335      } finally {
10336        this.unknownFields = unknownFields.build();
10337        makeExtensionsImmutable();
10338      }
10339    }
10340    public static final com.google.protobuf.Descriptors.Descriptor
10341        getDescriptor() {
10342      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalCTimeRequestProto_descriptor;
10343    }
10344
10345    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
10346        internalGetFieldAccessorTable() {
10347      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalCTimeRequestProto_fieldAccessorTable
10348          .ensureFieldAccessorsInitialized(
10349              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto.Builder.class);
10350    }
10351
10352    public static com.google.protobuf.Parser<GetJournalCTimeRequestProto> PARSER =
10353        new com.google.protobuf.AbstractParser<GetJournalCTimeRequestProto>() {
10354      public GetJournalCTimeRequestProto parsePartialFrom(
10355          com.google.protobuf.CodedInputStream input,
10356          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10357          throws com.google.protobuf.InvalidProtocolBufferException {
10358        return new GetJournalCTimeRequestProto(input, extensionRegistry);
10359      }
10360    };
10361
10362    @java.lang.Override
10363    public com.google.protobuf.Parser<GetJournalCTimeRequestProto> getParserForType() {
10364      return PARSER;
10365    }
10366
10367    private int bitField0_;
10368    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
10369    public static final int JID_FIELD_NUMBER = 1;
10370    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
10371    /**
10372     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10373     */
10374    public boolean hasJid() {
10375      return ((bitField0_ & 0x00000001) == 0x00000001);
10376    }
10377    /**
10378     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10379     */
10380    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
10381      return jid_;
10382    }
10383    /**
10384     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10385     */
10386    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
10387      return jid_;
10388    }
10389
10390    private void initFields() {
10391      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10392    }
10393    private byte memoizedIsInitialized = -1;
10394    public final boolean isInitialized() {
10395      byte isInitialized = memoizedIsInitialized;
10396      if (isInitialized != -1) return isInitialized == 1;
10397
10398      if (!hasJid()) {
10399        memoizedIsInitialized = 0;
10400        return false;
10401      }
10402      if (!getJid().isInitialized()) {
10403        memoizedIsInitialized = 0;
10404        return false;
10405      }
10406      memoizedIsInitialized = 1;
10407      return true;
10408    }
10409
10410    public void writeTo(com.google.protobuf.CodedOutputStream output)
10411                        throws java.io.IOException {
10412      getSerializedSize();
10413      if (((bitField0_ & 0x00000001) == 0x00000001)) {
10414        output.writeMessage(1, jid_);
10415      }
10416      getUnknownFields().writeTo(output);
10417    }
10418
10419    private int memoizedSerializedSize = -1;
10420    public int getSerializedSize() {
10421      int size = memoizedSerializedSize;
10422      if (size != -1) return size;
10423
10424      size = 0;
10425      if (((bitField0_ & 0x00000001) == 0x00000001)) {
10426        size += com.google.protobuf.CodedOutputStream
10427          .computeMessageSize(1, jid_);
10428      }
10429      size += getUnknownFields().getSerializedSize();
10430      memoizedSerializedSize = size;
10431      return size;
10432    }
10433
10434    private static final long serialVersionUID = 0L;
10435    @java.lang.Override
10436    protected java.lang.Object writeReplace()
10437        throws java.io.ObjectStreamException {
10438      return super.writeReplace();
10439    }
10440
10441    @java.lang.Override
10442    public boolean equals(final java.lang.Object obj) {
10443      if (obj == this) {
10444       return true;
10445      }
10446      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto)) {
10447        return super.equals(obj);
10448      }
10449      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto) obj;
10450
10451      boolean result = true;
10452      result = result && (hasJid() == other.hasJid());
10453      if (hasJid()) {
10454        result = result && getJid()
10455            .equals(other.getJid());
10456      }
10457      result = result &&
10458          getUnknownFields().equals(other.getUnknownFields());
10459      return result;
10460    }
10461
10462    private int memoizedHashCode = 0;
10463    @java.lang.Override
10464    public int hashCode() {
10465      if (memoizedHashCode != 0) {
10466        return memoizedHashCode;
10467      }
10468      int hash = 41;
10469      hash = (19 * hash) + getDescriptorForType().hashCode();
10470      if (hasJid()) {
10471        hash = (37 * hash) + JID_FIELD_NUMBER;
10472        hash = (53 * hash) + getJid().hashCode();
10473      }
10474      hash = (29 * hash) + getUnknownFields().hashCode();
10475      memoizedHashCode = hash;
10476      return hash;
10477    }
10478
10479    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto parseFrom(
10480        com.google.protobuf.ByteString data)
10481        throws com.google.protobuf.InvalidProtocolBufferException {
10482      return PARSER.parseFrom(data);
10483    }
10484    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto parseFrom(
10485        com.google.protobuf.ByteString data,
10486        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10487        throws com.google.protobuf.InvalidProtocolBufferException {
10488      return PARSER.parseFrom(data, extensionRegistry);
10489    }
10490    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto parseFrom(byte[] data)
10491        throws com.google.protobuf.InvalidProtocolBufferException {
10492      return PARSER.parseFrom(data);
10493    }
10494    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto parseFrom(
10495        byte[] data,
10496        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10497        throws com.google.protobuf.InvalidProtocolBufferException {
10498      return PARSER.parseFrom(data, extensionRegistry);
10499    }
10500    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto parseFrom(java.io.InputStream input)
10501        throws java.io.IOException {
10502      return PARSER.parseFrom(input);
10503    }
10504    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto parseFrom(
10505        java.io.InputStream input,
10506        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10507        throws java.io.IOException {
10508      return PARSER.parseFrom(input, extensionRegistry);
10509    }
10510    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto parseDelimitedFrom(java.io.InputStream input)
10511        throws java.io.IOException {
10512      return PARSER.parseDelimitedFrom(input);
10513    }
10514    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto parseDelimitedFrom(
10515        java.io.InputStream input,
10516        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10517        throws java.io.IOException {
10518      return PARSER.parseDelimitedFrom(input, extensionRegistry);
10519    }
10520    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto parseFrom(
10521        com.google.protobuf.CodedInputStream input)
10522        throws java.io.IOException {
10523      return PARSER.parseFrom(input);
10524    }
10525    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto parseFrom(
10526        com.google.protobuf.CodedInputStream input,
10527        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10528        throws java.io.IOException {
10529      return PARSER.parseFrom(input, extensionRegistry);
10530    }
10531
10532    public static Builder newBuilder() { return Builder.create(); }
10533    public Builder newBuilderForType() { return newBuilder(); }
10534    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto prototype) {
10535      return newBuilder().mergeFrom(prototype);
10536    }
10537    public Builder toBuilder() { return newBuilder(this); }
10538
10539    @java.lang.Override
10540    protected Builder newBuilderForType(
10541        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
10542      Builder builder = new Builder(parent);
10543      return builder;
10544    }
10545    /**
10546     * Protobuf type {@code hadoop.hdfs.qjournal.GetJournalCTimeRequestProto}
10547     *
10548     * <pre>
10549     **
10550     * getJournalCTime()
10551     * </pre>
10552     */
10553    public static final class Builder extends
10554        com.google.protobuf.GeneratedMessage.Builder<Builder>
10555       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProtoOrBuilder {
10556      public static final com.google.protobuf.Descriptors.Descriptor
10557          getDescriptor() {
10558        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalCTimeRequestProto_descriptor;
10559      }
10560
10561      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
10562          internalGetFieldAccessorTable() {
10563        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalCTimeRequestProto_fieldAccessorTable
10564            .ensureFieldAccessorsInitialized(
10565                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto.Builder.class);
10566      }
10567
10568      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto.newBuilder()
10569      private Builder() {
10570        maybeForceBuilderInitialization();
10571      }
10572
10573      private Builder(
10574          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
10575        super(parent);
10576        maybeForceBuilderInitialization();
10577      }
10578      private void maybeForceBuilderInitialization() {
10579        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
10580          getJidFieldBuilder();
10581        }
10582      }
10583      private static Builder create() {
10584        return new Builder();
10585      }
10586
10587      public Builder clear() {
10588        super.clear();
10589        if (jidBuilder_ == null) {
10590          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10591        } else {
10592          jidBuilder_.clear();
10593        }
10594        bitField0_ = (bitField0_ & ~0x00000001);
10595        return this;
10596      }
10597
10598      public Builder clone() {
10599        return create().mergeFrom(buildPartial());
10600      }
10601
10602      public com.google.protobuf.Descriptors.Descriptor
10603          getDescriptorForType() {
10604        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalCTimeRequestProto_descriptor;
10605      }
10606
10607      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto getDefaultInstanceForType() {
10608        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto.getDefaultInstance();
10609      }
10610
10611      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto build() {
10612        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto result = buildPartial();
10613        if (!result.isInitialized()) {
10614          throw newUninitializedMessageException(result);
10615        }
10616        return result;
10617      }
10618
10619      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto buildPartial() {
10620        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto(this);
10621        int from_bitField0_ = bitField0_;
10622        int to_bitField0_ = 0;
10623        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
10624          to_bitField0_ |= 0x00000001;
10625        }
10626        if (jidBuilder_ == null) {
10627          result.jid_ = jid_;
10628        } else {
10629          result.jid_ = jidBuilder_.build();
10630        }
10631        result.bitField0_ = to_bitField0_;
10632        onBuilt();
10633        return result;
10634      }
10635
10636      public Builder mergeFrom(com.google.protobuf.Message other) {
10637        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto) {
10638          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto)other);
10639        } else {
10640          super.mergeFrom(other);
10641          return this;
10642        }
10643      }
10644
10645      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto other) {
10646        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto.getDefaultInstance()) return this;
10647        if (other.hasJid()) {
10648          mergeJid(other.getJid());
10649        }
10650        this.mergeUnknownFields(other.getUnknownFields());
10651        return this;
10652      }
10653
10654      public final boolean isInitialized() {
10655        if (!hasJid()) {
10656          
10657          return false;
10658        }
10659        if (!getJid().isInitialized()) {
10660          
10661          return false;
10662        }
10663        return true;
10664      }
10665
10666      public Builder mergeFrom(
10667          com.google.protobuf.CodedInputStream input,
10668          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10669          throws java.io.IOException {
10670        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto parsedMessage = null;
10671        try {
10672          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
10673        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
10674          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto) e.getUnfinishedMessage();
10675          throw e;
10676        } finally {
10677          if (parsedMessage != null) {
10678            mergeFrom(parsedMessage);
10679          }
10680        }
10681        return this;
10682      }
10683      private int bitField0_;
10684
10685      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
10686      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10687      private com.google.protobuf.SingleFieldBuilder<
10688          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
10689      /**
10690       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10691       */
10692      public boolean hasJid() {
10693        return ((bitField0_ & 0x00000001) == 0x00000001);
10694      }
10695      /**
10696       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10697       */
10698      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
10699        if (jidBuilder_ == null) {
10700          return jid_;
10701        } else {
10702          return jidBuilder_.getMessage();
10703        }
10704      }
10705      /**
10706       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10707       */
10708      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
10709        if (jidBuilder_ == null) {
10710          if (value == null) {
10711            throw new NullPointerException();
10712          }
10713          jid_ = value;
10714          onChanged();
10715        } else {
10716          jidBuilder_.setMessage(value);
10717        }
10718        bitField0_ |= 0x00000001;
10719        return this;
10720      }
10721      /**
10722       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10723       */
10724      public Builder setJid(
10725          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
10726        if (jidBuilder_ == null) {
10727          jid_ = builderForValue.build();
10728          onChanged();
10729        } else {
10730          jidBuilder_.setMessage(builderForValue.build());
10731        }
10732        bitField0_ |= 0x00000001;
10733        return this;
10734      }
10735      /**
10736       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10737       */
10738      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
10739        if (jidBuilder_ == null) {
10740          if (((bitField0_ & 0x00000001) == 0x00000001) &&
10741              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
10742            jid_ =
10743              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
10744          } else {
10745            jid_ = value;
10746          }
10747          onChanged();
10748        } else {
10749          jidBuilder_.mergeFrom(value);
10750        }
10751        bitField0_ |= 0x00000001;
10752        return this;
10753      }
10754      /**
10755       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10756       */
10757      public Builder clearJid() {
10758        if (jidBuilder_ == null) {
10759          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10760          onChanged();
10761        } else {
10762          jidBuilder_.clear();
10763        }
10764        bitField0_ = (bitField0_ & ~0x00000001);
10765        return this;
10766      }
10767      /**
10768       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10769       */
10770      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
10771        bitField0_ |= 0x00000001;
10772        onChanged();
10773        return getJidFieldBuilder().getBuilder();
10774      }
10775      /**
10776       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10777       */
10778      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
10779        if (jidBuilder_ != null) {
10780          return jidBuilder_.getMessageOrBuilder();
10781        } else {
10782          return jid_;
10783        }
10784      }
10785      /**
10786       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
10787       */
10788      private com.google.protobuf.SingleFieldBuilder<
10789          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
10790          getJidFieldBuilder() {
10791        if (jidBuilder_ == null) {
10792          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
10793              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
10794                  jid_,
10795                  getParentForChildren(),
10796                  isClean());
10797          jid_ = null;
10798        }
10799        return jidBuilder_;
10800      }
10801
10802      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.GetJournalCTimeRequestProto)
10803    }
10804
10805    static {
10806      defaultInstance = new GetJournalCTimeRequestProto(true);
10807      defaultInstance.initFields();
10808    }
10809
10810    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.GetJournalCTimeRequestProto)
10811  }
10812
10813  public interface GetJournalCTimeResponseProtoOrBuilder
10814      extends com.google.protobuf.MessageOrBuilder {
10815
10816    // required int64 resultCTime = 1;
10817    /**
10818     * <code>required int64 resultCTime = 1;</code>
10819     */
10820    boolean hasResultCTime();
10821    /**
10822     * <code>required int64 resultCTime = 1;</code>
10823     */
10824    long getResultCTime();
10825  }
10826  /**
10827   * Protobuf type {@code hadoop.hdfs.qjournal.GetJournalCTimeResponseProto}
10828   */
10829  public static final class GetJournalCTimeResponseProto extends
10830      com.google.protobuf.GeneratedMessage
10831      implements GetJournalCTimeResponseProtoOrBuilder {
10832    // Use GetJournalCTimeResponseProto.newBuilder() to construct.
10833    private GetJournalCTimeResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
10834      super(builder);
10835      this.unknownFields = builder.getUnknownFields();
10836    }
10837    private GetJournalCTimeResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
10838
10839    private static final GetJournalCTimeResponseProto defaultInstance;
10840    public static GetJournalCTimeResponseProto getDefaultInstance() {
10841      return defaultInstance;
10842    }
10843
10844    public GetJournalCTimeResponseProto getDefaultInstanceForType() {
10845      return defaultInstance;
10846    }
10847
10848    private final com.google.protobuf.UnknownFieldSet unknownFields;
10849    @java.lang.Override
10850    public final com.google.protobuf.UnknownFieldSet
10851        getUnknownFields() {
10852      return this.unknownFields;
10853    }
10854    private GetJournalCTimeResponseProto(
10855        com.google.protobuf.CodedInputStream input,
10856        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10857        throws com.google.protobuf.InvalidProtocolBufferException {
10858      initFields();
10859      int mutable_bitField0_ = 0;
10860      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
10861          com.google.protobuf.UnknownFieldSet.newBuilder();
10862      try {
10863        boolean done = false;
10864        while (!done) {
10865          int tag = input.readTag();
10866          switch (tag) {
10867            case 0:
10868              done = true;
10869              break;
10870            default: {
10871              if (!parseUnknownField(input, unknownFields,
10872                                     extensionRegistry, tag)) {
10873                done = true;
10874              }
10875              break;
10876            }
10877            case 8: {
10878              bitField0_ |= 0x00000001;
10879              resultCTime_ = input.readInt64();
10880              break;
10881            }
10882          }
10883        }
10884      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
10885        throw e.setUnfinishedMessage(this);
10886      } catch (java.io.IOException e) {
10887        throw new com.google.protobuf.InvalidProtocolBufferException(
10888            e.getMessage()).setUnfinishedMessage(this);
10889      } finally {
10890        this.unknownFields = unknownFields.build();
10891        makeExtensionsImmutable();
10892      }
10893    }
10894    public static final com.google.protobuf.Descriptors.Descriptor
10895        getDescriptor() {
10896      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalCTimeResponseProto_descriptor;
10897    }
10898
10899    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
10900        internalGetFieldAccessorTable() {
10901      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalCTimeResponseProto_fieldAccessorTable
10902          .ensureFieldAccessorsInitialized(
10903              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.Builder.class);
10904    }
10905
10906    public static com.google.protobuf.Parser<GetJournalCTimeResponseProto> PARSER =
10907        new com.google.protobuf.AbstractParser<GetJournalCTimeResponseProto>() {
10908      public GetJournalCTimeResponseProto parsePartialFrom(
10909          com.google.protobuf.CodedInputStream input,
10910          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10911          throws com.google.protobuf.InvalidProtocolBufferException {
10912        return new GetJournalCTimeResponseProto(input, extensionRegistry);
10913      }
10914    };
10915
10916    @java.lang.Override
10917    public com.google.protobuf.Parser<GetJournalCTimeResponseProto> getParserForType() {
10918      return PARSER;
10919    }
10920
10921    private int bitField0_;
10922    // required int64 resultCTime = 1;
10923    public static final int RESULTCTIME_FIELD_NUMBER = 1;
10924    private long resultCTime_;
10925    /**
10926     * <code>required int64 resultCTime = 1;</code>
10927     */
10928    public boolean hasResultCTime() {
10929      return ((bitField0_ & 0x00000001) == 0x00000001);
10930    }
10931    /**
10932     * <code>required int64 resultCTime = 1;</code>
10933     */
10934    public long getResultCTime() {
10935      return resultCTime_;
10936    }
10937
10938    private void initFields() {
10939      resultCTime_ = 0L;
10940    }
10941    private byte memoizedIsInitialized = -1;
10942    public final boolean isInitialized() {
10943      byte isInitialized = memoizedIsInitialized;
10944      if (isInitialized != -1) return isInitialized == 1;
10945
10946      if (!hasResultCTime()) {
10947        memoizedIsInitialized = 0;
10948        return false;
10949      }
10950      memoizedIsInitialized = 1;
10951      return true;
10952    }
10953
10954    public void writeTo(com.google.protobuf.CodedOutputStream output)
10955                        throws java.io.IOException {
10956      getSerializedSize();
10957      if (((bitField0_ & 0x00000001) == 0x00000001)) {
10958        output.writeInt64(1, resultCTime_);
10959      }
10960      getUnknownFields().writeTo(output);
10961    }
10962
10963    private int memoizedSerializedSize = -1;
10964    public int getSerializedSize() {
10965      int size = memoizedSerializedSize;
10966      if (size != -1) return size;
10967
10968      size = 0;
10969      if (((bitField0_ & 0x00000001) == 0x00000001)) {
10970        size += com.google.protobuf.CodedOutputStream
10971          .computeInt64Size(1, resultCTime_);
10972      }
10973      size += getUnknownFields().getSerializedSize();
10974      memoizedSerializedSize = size;
10975      return size;
10976    }
10977
10978    private static final long serialVersionUID = 0L;
10979    @java.lang.Override
10980    protected java.lang.Object writeReplace()
10981        throws java.io.ObjectStreamException {
10982      return super.writeReplace();
10983    }
10984
10985    @java.lang.Override
10986    public boolean equals(final java.lang.Object obj) {
10987      if (obj == this) {
10988       return true;
10989      }
10990      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto)) {
10991        return super.equals(obj);
10992      }
10993      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto) obj;
10994
10995      boolean result = true;
10996      result = result && (hasResultCTime() == other.hasResultCTime());
10997      if (hasResultCTime()) {
10998        result = result && (getResultCTime()
10999            == other.getResultCTime());
11000      }
11001      result = result &&
11002          getUnknownFields().equals(other.getUnknownFields());
11003      return result;
11004    }
11005
11006    private int memoizedHashCode = 0;
11007    @java.lang.Override
11008    public int hashCode() {
11009      if (memoizedHashCode != 0) {
11010        return memoizedHashCode;
11011      }
11012      int hash = 41;
11013      hash = (19 * hash) + getDescriptorForType().hashCode();
11014      if (hasResultCTime()) {
11015        hash = (37 * hash) + RESULTCTIME_FIELD_NUMBER;
11016        hash = (53 * hash) + hashLong(getResultCTime());
11017      }
11018      hash = (29 * hash) + getUnknownFields().hashCode();
11019      memoizedHashCode = hash;
11020      return hash;
11021    }
11022
11023    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto parseFrom(
11024        com.google.protobuf.ByteString data)
11025        throws com.google.protobuf.InvalidProtocolBufferException {
11026      return PARSER.parseFrom(data);
11027    }
11028    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto parseFrom(
11029        com.google.protobuf.ByteString data,
11030        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11031        throws com.google.protobuf.InvalidProtocolBufferException {
11032      return PARSER.parseFrom(data, extensionRegistry);
11033    }
11034    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto parseFrom(byte[] data)
11035        throws com.google.protobuf.InvalidProtocolBufferException {
11036      return PARSER.parseFrom(data);
11037    }
11038    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto parseFrom(
11039        byte[] data,
11040        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11041        throws com.google.protobuf.InvalidProtocolBufferException {
11042      return PARSER.parseFrom(data, extensionRegistry);
11043    }
11044    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto parseFrom(java.io.InputStream input)
11045        throws java.io.IOException {
11046      return PARSER.parseFrom(input);
11047    }
11048    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto parseFrom(
11049        java.io.InputStream input,
11050        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11051        throws java.io.IOException {
11052      return PARSER.parseFrom(input, extensionRegistry);
11053    }
11054    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto parseDelimitedFrom(java.io.InputStream input)
11055        throws java.io.IOException {
11056      return PARSER.parseDelimitedFrom(input);
11057    }
11058    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto parseDelimitedFrom(
11059        java.io.InputStream input,
11060        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11061        throws java.io.IOException {
11062      return PARSER.parseDelimitedFrom(input, extensionRegistry);
11063    }
11064    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto parseFrom(
11065        com.google.protobuf.CodedInputStream input)
11066        throws java.io.IOException {
11067      return PARSER.parseFrom(input);
11068    }
11069    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto parseFrom(
11070        com.google.protobuf.CodedInputStream input,
11071        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11072        throws java.io.IOException {
11073      return PARSER.parseFrom(input, extensionRegistry);
11074    }
11075
11076    public static Builder newBuilder() { return Builder.create(); }
11077    public Builder newBuilderForType() { return newBuilder(); }
11078    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto prototype) {
11079      return newBuilder().mergeFrom(prototype);
11080    }
11081    public Builder toBuilder() { return newBuilder(this); }
11082
11083    @java.lang.Override
11084    protected Builder newBuilderForType(
11085        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
11086      Builder builder = new Builder(parent);
11087      return builder;
11088    }
11089    /**
11090     * Protobuf type {@code hadoop.hdfs.qjournal.GetJournalCTimeResponseProto}
11091     */
11092    public static final class Builder extends
11093        com.google.protobuf.GeneratedMessage.Builder<Builder>
11094       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProtoOrBuilder {
11095      public static final com.google.protobuf.Descriptors.Descriptor
11096          getDescriptor() {
11097        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalCTimeResponseProto_descriptor;
11098      }
11099
11100      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11101          internalGetFieldAccessorTable() {
11102        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalCTimeResponseProto_fieldAccessorTable
11103            .ensureFieldAccessorsInitialized(
11104                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.Builder.class);
11105      }
11106
11107      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.newBuilder()
11108      private Builder() {
11109        maybeForceBuilderInitialization();
11110      }
11111
11112      private Builder(
11113          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
11114        super(parent);
11115        maybeForceBuilderInitialization();
11116      }
11117      private void maybeForceBuilderInitialization() {
11118        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
11119        }
11120      }
11121      private static Builder create() {
11122        return new Builder();
11123      }
11124
11125      public Builder clear() {
11126        super.clear();
11127        resultCTime_ = 0L;
11128        bitField0_ = (bitField0_ & ~0x00000001);
11129        return this;
11130      }
11131
11132      public Builder clone() {
11133        return create().mergeFrom(buildPartial());
11134      }
11135
11136      public com.google.protobuf.Descriptors.Descriptor
11137          getDescriptorForType() {
11138        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalCTimeResponseProto_descriptor;
11139      }
11140
11141      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto getDefaultInstanceForType() {
11142        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.getDefaultInstance();
11143      }
11144
11145      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto build() {
11146        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto result = buildPartial();
11147        if (!result.isInitialized()) {
11148          throw newUninitializedMessageException(result);
11149        }
11150        return result;
11151      }
11152
11153      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto buildPartial() {
11154        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto(this);
11155        int from_bitField0_ = bitField0_;
11156        int to_bitField0_ = 0;
11157        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
11158          to_bitField0_ |= 0x00000001;
11159        }
11160        result.resultCTime_ = resultCTime_;
11161        result.bitField0_ = to_bitField0_;
11162        onBuilt();
11163        return result;
11164      }
11165
11166      public Builder mergeFrom(com.google.protobuf.Message other) {
11167        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto) {
11168          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto)other);
11169        } else {
11170          super.mergeFrom(other);
11171          return this;
11172        }
11173      }
11174
11175      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto other) {
11176        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.getDefaultInstance()) return this;
11177        if (other.hasResultCTime()) {
11178          setResultCTime(other.getResultCTime());
11179        }
11180        this.mergeUnknownFields(other.getUnknownFields());
11181        return this;
11182      }
11183
11184      public final boolean isInitialized() {
11185        if (!hasResultCTime()) {
11186          
11187          return false;
11188        }
11189        return true;
11190      }
11191
11192      public Builder mergeFrom(
11193          com.google.protobuf.CodedInputStream input,
11194          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11195          throws java.io.IOException {
11196        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto parsedMessage = null;
11197        try {
11198          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
11199        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
11200          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto) e.getUnfinishedMessage();
11201          throw e;
11202        } finally {
11203          if (parsedMessage != null) {
11204            mergeFrom(parsedMessage);
11205          }
11206        }
11207        return this;
11208      }
11209      private int bitField0_;
11210
11211      // required int64 resultCTime = 1;
11212      private long resultCTime_ ;
11213      /**
11214       * <code>required int64 resultCTime = 1;</code>
11215       */
11216      public boolean hasResultCTime() {
11217        return ((bitField0_ & 0x00000001) == 0x00000001);
11218      }
11219      /**
11220       * <code>required int64 resultCTime = 1;</code>
11221       */
11222      public long getResultCTime() {
11223        return resultCTime_;
11224      }
11225      /**
11226       * <code>required int64 resultCTime = 1;</code>
11227       */
11228      public Builder setResultCTime(long value) {
11229        bitField0_ |= 0x00000001;
11230        resultCTime_ = value;
11231        onChanged();
11232        return this;
11233      }
11234      /**
11235       * <code>required int64 resultCTime = 1;</code>
11236       */
11237      public Builder clearResultCTime() {
11238        bitField0_ = (bitField0_ & ~0x00000001);
11239        resultCTime_ = 0L;
11240        onChanged();
11241        return this;
11242      }
11243
11244      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.GetJournalCTimeResponseProto)
11245    }
11246
11247    static {
11248      defaultInstance = new GetJournalCTimeResponseProto(true);
11249      defaultInstance.initFields();
11250    }
11251
11252    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.GetJournalCTimeResponseProto)
11253  }
11254
11255  public interface DoPreUpgradeRequestProtoOrBuilder
11256      extends com.google.protobuf.MessageOrBuilder {
11257
11258    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
11259    /**
11260     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11261     */
11262    boolean hasJid();
11263    /**
11264     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11265     */
11266    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
11267    /**
11268     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11269     */
11270    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
11271  }
11272  /**
11273   * Protobuf type {@code hadoop.hdfs.qjournal.DoPreUpgradeRequestProto}
11274   *
11275   * <pre>
11276   **
11277   * doPreUpgrade()
11278   * </pre>
11279   */
11280  public static final class DoPreUpgradeRequestProto extends
11281      com.google.protobuf.GeneratedMessage
11282      implements DoPreUpgradeRequestProtoOrBuilder {
11283    // Use DoPreUpgradeRequestProto.newBuilder() to construct.
11284    private DoPreUpgradeRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
11285      super(builder);
11286      this.unknownFields = builder.getUnknownFields();
11287    }
11288    private DoPreUpgradeRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
11289
11290    private static final DoPreUpgradeRequestProto defaultInstance;
11291    public static DoPreUpgradeRequestProto getDefaultInstance() {
11292      return defaultInstance;
11293    }
11294
11295    public DoPreUpgradeRequestProto getDefaultInstanceForType() {
11296      return defaultInstance;
11297    }
11298
11299    private final com.google.protobuf.UnknownFieldSet unknownFields;
11300    @java.lang.Override
11301    public final com.google.protobuf.UnknownFieldSet
11302        getUnknownFields() {
11303      return this.unknownFields;
11304    }
11305    private DoPreUpgradeRequestProto(
11306        com.google.protobuf.CodedInputStream input,
11307        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11308        throws com.google.protobuf.InvalidProtocolBufferException {
11309      initFields();
11310      int mutable_bitField0_ = 0;
11311      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
11312          com.google.protobuf.UnknownFieldSet.newBuilder();
11313      try {
11314        boolean done = false;
11315        while (!done) {
11316          int tag = input.readTag();
11317          switch (tag) {
11318            case 0:
11319              done = true;
11320              break;
11321            default: {
11322              if (!parseUnknownField(input, unknownFields,
11323                                     extensionRegistry, tag)) {
11324                done = true;
11325              }
11326              break;
11327            }
11328            case 10: {
11329              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
11330              if (((bitField0_ & 0x00000001) == 0x00000001)) {
11331                subBuilder = jid_.toBuilder();
11332              }
11333              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
11334              if (subBuilder != null) {
11335                subBuilder.mergeFrom(jid_);
11336                jid_ = subBuilder.buildPartial();
11337              }
11338              bitField0_ |= 0x00000001;
11339              break;
11340            }
11341          }
11342        }
11343      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
11344        throw e.setUnfinishedMessage(this);
11345      } catch (java.io.IOException e) {
11346        throw new com.google.protobuf.InvalidProtocolBufferException(
11347            e.getMessage()).setUnfinishedMessage(this);
11348      } finally {
11349        this.unknownFields = unknownFields.build();
11350        makeExtensionsImmutable();
11351      }
11352    }
11353    public static final com.google.protobuf.Descriptors.Descriptor
11354        getDescriptor() {
11355      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoPreUpgradeRequestProto_descriptor;
11356    }
11357
11358    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11359        internalGetFieldAccessorTable() {
11360      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoPreUpgradeRequestProto_fieldAccessorTable
11361          .ensureFieldAccessorsInitialized(
11362              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto.Builder.class);
11363    }
11364
11365    public static com.google.protobuf.Parser<DoPreUpgradeRequestProto> PARSER =
11366        new com.google.protobuf.AbstractParser<DoPreUpgradeRequestProto>() {
11367      public DoPreUpgradeRequestProto parsePartialFrom(
11368          com.google.protobuf.CodedInputStream input,
11369          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11370          throws com.google.protobuf.InvalidProtocolBufferException {
11371        return new DoPreUpgradeRequestProto(input, extensionRegistry);
11372      }
11373    };
11374
11375    @java.lang.Override
11376    public com.google.protobuf.Parser<DoPreUpgradeRequestProto> getParserForType() {
11377      return PARSER;
11378    }
11379
11380    private int bitField0_;
11381    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
11382    public static final int JID_FIELD_NUMBER = 1;
11383    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
11384    /**
11385     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11386     */
11387    public boolean hasJid() {
11388      return ((bitField0_ & 0x00000001) == 0x00000001);
11389    }
11390    /**
11391     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11392     */
11393    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
11394      return jid_;
11395    }
11396    /**
11397     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11398     */
11399    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
11400      return jid_;
11401    }
11402
11403    private void initFields() {
11404      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
11405    }
11406    private byte memoizedIsInitialized = -1;
11407    public final boolean isInitialized() {
11408      byte isInitialized = memoizedIsInitialized;
11409      if (isInitialized != -1) return isInitialized == 1;
11410
11411      if (!hasJid()) {
11412        memoizedIsInitialized = 0;
11413        return false;
11414      }
11415      if (!getJid().isInitialized()) {
11416        memoizedIsInitialized = 0;
11417        return false;
11418      }
11419      memoizedIsInitialized = 1;
11420      return true;
11421    }
11422
11423    public void writeTo(com.google.protobuf.CodedOutputStream output)
11424                        throws java.io.IOException {
11425      getSerializedSize();
11426      if (((bitField0_ & 0x00000001) == 0x00000001)) {
11427        output.writeMessage(1, jid_);
11428      }
11429      getUnknownFields().writeTo(output);
11430    }
11431
11432    private int memoizedSerializedSize = -1;
11433    public int getSerializedSize() {
11434      int size = memoizedSerializedSize;
11435      if (size != -1) return size;
11436
11437      size = 0;
11438      if (((bitField0_ & 0x00000001) == 0x00000001)) {
11439        size += com.google.protobuf.CodedOutputStream
11440          .computeMessageSize(1, jid_);
11441      }
11442      size += getUnknownFields().getSerializedSize();
11443      memoizedSerializedSize = size;
11444      return size;
11445    }
11446
11447    private static final long serialVersionUID = 0L;
11448    @java.lang.Override
11449    protected java.lang.Object writeReplace()
11450        throws java.io.ObjectStreamException {
11451      return super.writeReplace();
11452    }
11453
11454    @java.lang.Override
11455    public boolean equals(final java.lang.Object obj) {
11456      if (obj == this) {
11457       return true;
11458      }
11459      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto)) {
11460        return super.equals(obj);
11461      }
11462      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto) obj;
11463
11464      boolean result = true;
11465      result = result && (hasJid() == other.hasJid());
11466      if (hasJid()) {
11467        result = result && getJid()
11468            .equals(other.getJid());
11469      }
11470      result = result &&
11471          getUnknownFields().equals(other.getUnknownFields());
11472      return result;
11473    }
11474
11475    private int memoizedHashCode = 0;
11476    @java.lang.Override
11477    public int hashCode() {
11478      if (memoizedHashCode != 0) {
11479        return memoizedHashCode;
11480      }
11481      int hash = 41;
11482      hash = (19 * hash) + getDescriptorForType().hashCode();
11483      if (hasJid()) {
11484        hash = (37 * hash) + JID_FIELD_NUMBER;
11485        hash = (53 * hash) + getJid().hashCode();
11486      }
11487      hash = (29 * hash) + getUnknownFields().hashCode();
11488      memoizedHashCode = hash;
11489      return hash;
11490    }
11491
11492    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto parseFrom(
11493        com.google.protobuf.ByteString data)
11494        throws com.google.protobuf.InvalidProtocolBufferException {
11495      return PARSER.parseFrom(data);
11496    }
11497    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto parseFrom(
11498        com.google.protobuf.ByteString data,
11499        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11500        throws com.google.protobuf.InvalidProtocolBufferException {
11501      return PARSER.parseFrom(data, extensionRegistry);
11502    }
11503    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto parseFrom(byte[] data)
11504        throws com.google.protobuf.InvalidProtocolBufferException {
11505      return PARSER.parseFrom(data);
11506    }
11507    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto parseFrom(
11508        byte[] data,
11509        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11510        throws com.google.protobuf.InvalidProtocolBufferException {
11511      return PARSER.parseFrom(data, extensionRegistry);
11512    }
11513    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto parseFrom(java.io.InputStream input)
11514        throws java.io.IOException {
11515      return PARSER.parseFrom(input);
11516    }
11517    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto parseFrom(
11518        java.io.InputStream input,
11519        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11520        throws java.io.IOException {
11521      return PARSER.parseFrom(input, extensionRegistry);
11522    }
11523    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto parseDelimitedFrom(java.io.InputStream input)
11524        throws java.io.IOException {
11525      return PARSER.parseDelimitedFrom(input);
11526    }
11527    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto parseDelimitedFrom(
11528        java.io.InputStream input,
11529        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11530        throws java.io.IOException {
11531      return PARSER.parseDelimitedFrom(input, extensionRegistry);
11532    }
11533    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto parseFrom(
11534        com.google.protobuf.CodedInputStream input)
11535        throws java.io.IOException {
11536      return PARSER.parseFrom(input);
11537    }
11538    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto parseFrom(
11539        com.google.protobuf.CodedInputStream input,
11540        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11541        throws java.io.IOException {
11542      return PARSER.parseFrom(input, extensionRegistry);
11543    }
11544
11545    public static Builder newBuilder() { return Builder.create(); }
11546    public Builder newBuilderForType() { return newBuilder(); }
11547    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto prototype) {
11548      return newBuilder().mergeFrom(prototype);
11549    }
11550    public Builder toBuilder() { return newBuilder(this); }
11551
11552    @java.lang.Override
11553    protected Builder newBuilderForType(
11554        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
11555      Builder builder = new Builder(parent);
11556      return builder;
11557    }
11558    /**
11559     * Protobuf type {@code hadoop.hdfs.qjournal.DoPreUpgradeRequestProto}
11560     *
11561     * <pre>
11562     **
11563     * doPreUpgrade()
11564     * </pre>
11565     */
11566    public static final class Builder extends
11567        com.google.protobuf.GeneratedMessage.Builder<Builder>
11568       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProtoOrBuilder {
11569      public static final com.google.protobuf.Descriptors.Descriptor
11570          getDescriptor() {
11571        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoPreUpgradeRequestProto_descriptor;
11572      }
11573
11574      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11575          internalGetFieldAccessorTable() {
11576        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoPreUpgradeRequestProto_fieldAccessorTable
11577            .ensureFieldAccessorsInitialized(
11578                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto.Builder.class);
11579      }
11580
11581      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto.newBuilder()
11582      private Builder() {
11583        maybeForceBuilderInitialization();
11584      }
11585
11586      private Builder(
11587          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
11588        super(parent);
11589        maybeForceBuilderInitialization();
11590      }
11591      private void maybeForceBuilderInitialization() {
11592        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
11593          getJidFieldBuilder();
11594        }
11595      }
11596      private static Builder create() {
11597        return new Builder();
11598      }
11599
11600      public Builder clear() {
11601        super.clear();
11602        if (jidBuilder_ == null) {
11603          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
11604        } else {
11605          jidBuilder_.clear();
11606        }
11607        bitField0_ = (bitField0_ & ~0x00000001);
11608        return this;
11609      }
11610
11611      public Builder clone() {
11612        return create().mergeFrom(buildPartial());
11613      }
11614
11615      public com.google.protobuf.Descriptors.Descriptor
11616          getDescriptorForType() {
11617        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoPreUpgradeRequestProto_descriptor;
11618      }
11619
11620      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto getDefaultInstanceForType() {
11621        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto.getDefaultInstance();
11622      }
11623
11624      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto build() {
11625        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto result = buildPartial();
11626        if (!result.isInitialized()) {
11627          throw newUninitializedMessageException(result);
11628        }
11629        return result;
11630      }
11631
11632      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto buildPartial() {
11633        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto(this);
11634        int from_bitField0_ = bitField0_;
11635        int to_bitField0_ = 0;
11636        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
11637          to_bitField0_ |= 0x00000001;
11638        }
11639        if (jidBuilder_ == null) {
11640          result.jid_ = jid_;
11641        } else {
11642          result.jid_ = jidBuilder_.build();
11643        }
11644        result.bitField0_ = to_bitField0_;
11645        onBuilt();
11646        return result;
11647      }
11648
11649      public Builder mergeFrom(com.google.protobuf.Message other) {
11650        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto) {
11651          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto)other);
11652        } else {
11653          super.mergeFrom(other);
11654          return this;
11655        }
11656      }
11657
11658      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto other) {
11659        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto.getDefaultInstance()) return this;
11660        if (other.hasJid()) {
11661          mergeJid(other.getJid());
11662        }
11663        this.mergeUnknownFields(other.getUnknownFields());
11664        return this;
11665      }
11666
11667      public final boolean isInitialized() {
11668        if (!hasJid()) {
11669          
11670          return false;
11671        }
11672        if (!getJid().isInitialized()) {
11673          
11674          return false;
11675        }
11676        return true;
11677      }
11678
11679      public Builder mergeFrom(
11680          com.google.protobuf.CodedInputStream input,
11681          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11682          throws java.io.IOException {
11683        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto parsedMessage = null;
11684        try {
11685          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
11686        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
11687          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto) e.getUnfinishedMessage();
11688          throw e;
11689        } finally {
11690          if (parsedMessage != null) {
11691            mergeFrom(parsedMessage);
11692          }
11693        }
11694        return this;
11695      }
11696      private int bitField0_;
11697
11698      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
11699      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
11700      private com.google.protobuf.SingleFieldBuilder<
11701          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
11702      /**
11703       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11704       */
11705      public boolean hasJid() {
11706        return ((bitField0_ & 0x00000001) == 0x00000001);
11707      }
11708      /**
11709       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11710       */
11711      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
11712        if (jidBuilder_ == null) {
11713          return jid_;
11714        } else {
11715          return jidBuilder_.getMessage();
11716        }
11717      }
11718      /**
11719       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11720       */
11721      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
11722        if (jidBuilder_ == null) {
11723          if (value == null) {
11724            throw new NullPointerException();
11725          }
11726          jid_ = value;
11727          onChanged();
11728        } else {
11729          jidBuilder_.setMessage(value);
11730        }
11731        bitField0_ |= 0x00000001;
11732        return this;
11733      }
11734      /**
11735       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11736       */
11737      public Builder setJid(
11738          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
11739        if (jidBuilder_ == null) {
11740          jid_ = builderForValue.build();
11741          onChanged();
11742        } else {
11743          jidBuilder_.setMessage(builderForValue.build());
11744        }
11745        bitField0_ |= 0x00000001;
11746        return this;
11747      }
11748      /**
11749       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11750       */
11751      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
11752        if (jidBuilder_ == null) {
11753          if (((bitField0_ & 0x00000001) == 0x00000001) &&
11754              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
11755            jid_ =
11756              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
11757          } else {
11758            jid_ = value;
11759          }
11760          onChanged();
11761        } else {
11762          jidBuilder_.mergeFrom(value);
11763        }
11764        bitField0_ |= 0x00000001;
11765        return this;
11766      }
11767      /**
11768       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11769       */
11770      public Builder clearJid() {
11771        if (jidBuilder_ == null) {
11772          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
11773          onChanged();
11774        } else {
11775          jidBuilder_.clear();
11776        }
11777        bitField0_ = (bitField0_ & ~0x00000001);
11778        return this;
11779      }
11780      /**
11781       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11782       */
11783      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
11784        bitField0_ |= 0x00000001;
11785        onChanged();
11786        return getJidFieldBuilder().getBuilder();
11787      }
11788      /**
11789       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11790       */
11791      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
11792        if (jidBuilder_ != null) {
11793          return jidBuilder_.getMessageOrBuilder();
11794        } else {
11795          return jid_;
11796        }
11797      }
11798      /**
11799       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
11800       */
11801      private com.google.protobuf.SingleFieldBuilder<
11802          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
11803          getJidFieldBuilder() {
11804        if (jidBuilder_ == null) {
11805          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
11806              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
11807                  jid_,
11808                  getParentForChildren(),
11809                  isClean());
11810          jid_ = null;
11811        }
11812        return jidBuilder_;
11813      }
11814
11815      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.DoPreUpgradeRequestProto)
11816    }
11817
11818    static {
11819      defaultInstance = new DoPreUpgradeRequestProto(true);
11820      defaultInstance.initFields();
11821    }
11822
11823    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.DoPreUpgradeRequestProto)
11824  }
11825
11826  public interface DoPreUpgradeResponseProtoOrBuilder
11827      extends com.google.protobuf.MessageOrBuilder {
11828  }
11829  /**
11830   * Protobuf type {@code hadoop.hdfs.qjournal.DoPreUpgradeResponseProto}
11831   */
11832  public static final class DoPreUpgradeResponseProto extends
11833      com.google.protobuf.GeneratedMessage
11834      implements DoPreUpgradeResponseProtoOrBuilder {
11835    // Use DoPreUpgradeResponseProto.newBuilder() to construct.
11836    private DoPreUpgradeResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
11837      super(builder);
11838      this.unknownFields = builder.getUnknownFields();
11839    }
11840    private DoPreUpgradeResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
11841
11842    private static final DoPreUpgradeResponseProto defaultInstance;
11843    public static DoPreUpgradeResponseProto getDefaultInstance() {
11844      return defaultInstance;
11845    }
11846
11847    public DoPreUpgradeResponseProto getDefaultInstanceForType() {
11848      return defaultInstance;
11849    }
11850
11851    private final com.google.protobuf.UnknownFieldSet unknownFields;
11852    @java.lang.Override
11853    public final com.google.protobuf.UnknownFieldSet
11854        getUnknownFields() {
11855      return this.unknownFields;
11856    }
11857    private DoPreUpgradeResponseProto(
11858        com.google.protobuf.CodedInputStream input,
11859        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11860        throws com.google.protobuf.InvalidProtocolBufferException {
11861      initFields();
11862      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
11863          com.google.protobuf.UnknownFieldSet.newBuilder();
11864      try {
11865        boolean done = false;
11866        while (!done) {
11867          int tag = input.readTag();
11868          switch (tag) {
11869            case 0:
11870              done = true;
11871              break;
11872            default: {
11873              if (!parseUnknownField(input, unknownFields,
11874                                     extensionRegistry, tag)) {
11875                done = true;
11876              }
11877              break;
11878            }
11879          }
11880        }
11881      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
11882        throw e.setUnfinishedMessage(this);
11883      } catch (java.io.IOException e) {
11884        throw new com.google.protobuf.InvalidProtocolBufferException(
11885            e.getMessage()).setUnfinishedMessage(this);
11886      } finally {
11887        this.unknownFields = unknownFields.build();
11888        makeExtensionsImmutable();
11889      }
11890    }
11891    public static final com.google.protobuf.Descriptors.Descriptor
11892        getDescriptor() {
11893      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoPreUpgradeResponseProto_descriptor;
11894    }
11895
11896    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11897        internalGetFieldAccessorTable() {
11898      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoPreUpgradeResponseProto_fieldAccessorTable
11899          .ensureFieldAccessorsInitialized(
11900              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.Builder.class);
11901    }
11902
11903    public static com.google.protobuf.Parser<DoPreUpgradeResponseProto> PARSER =
11904        new com.google.protobuf.AbstractParser<DoPreUpgradeResponseProto>() {
11905      public DoPreUpgradeResponseProto parsePartialFrom(
11906          com.google.protobuf.CodedInputStream input,
11907          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11908          throws com.google.protobuf.InvalidProtocolBufferException {
11909        return new DoPreUpgradeResponseProto(input, extensionRegistry);
11910      }
11911    };
11912
11913    @java.lang.Override
11914    public com.google.protobuf.Parser<DoPreUpgradeResponseProto> getParserForType() {
11915      return PARSER;
11916    }
11917
11918    private void initFields() {
11919    }
11920    private byte memoizedIsInitialized = -1;
11921    public final boolean isInitialized() {
11922      byte isInitialized = memoizedIsInitialized;
11923      if (isInitialized != -1) return isInitialized == 1;
11924
11925      memoizedIsInitialized = 1;
11926      return true;
11927    }
11928
11929    public void writeTo(com.google.protobuf.CodedOutputStream output)
11930                        throws java.io.IOException {
11931      getSerializedSize();
11932      getUnknownFields().writeTo(output);
11933    }
11934
11935    private int memoizedSerializedSize = -1;
11936    public int getSerializedSize() {
11937      int size = memoizedSerializedSize;
11938      if (size != -1) return size;
11939
11940      size = 0;
11941      size += getUnknownFields().getSerializedSize();
11942      memoizedSerializedSize = size;
11943      return size;
11944    }
11945
11946    private static final long serialVersionUID = 0L;
11947    @java.lang.Override
11948    protected java.lang.Object writeReplace()
11949        throws java.io.ObjectStreamException {
11950      return super.writeReplace();
11951    }
11952
11953    @java.lang.Override
11954    public boolean equals(final java.lang.Object obj) {
11955      if (obj == this) {
11956       return true;
11957      }
11958      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto)) {
11959        return super.equals(obj);
11960      }
11961      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto) obj;
11962
11963      boolean result = true;
11964      result = result &&
11965          getUnknownFields().equals(other.getUnknownFields());
11966      return result;
11967    }
11968
11969    private int memoizedHashCode = 0;
11970    @java.lang.Override
11971    public int hashCode() {
11972      if (memoizedHashCode != 0) {
11973        return memoizedHashCode;
11974      }
11975      int hash = 41;
11976      hash = (19 * hash) + getDescriptorForType().hashCode();
11977      hash = (29 * hash) + getUnknownFields().hashCode();
11978      memoizedHashCode = hash;
11979      return hash;
11980    }
11981
11982    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto parseFrom(
11983        com.google.protobuf.ByteString data)
11984        throws com.google.protobuf.InvalidProtocolBufferException {
11985      return PARSER.parseFrom(data);
11986    }
11987    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto parseFrom(
11988        com.google.protobuf.ByteString data,
11989        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11990        throws com.google.protobuf.InvalidProtocolBufferException {
11991      return PARSER.parseFrom(data, extensionRegistry);
11992    }
11993    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto parseFrom(byte[] data)
11994        throws com.google.protobuf.InvalidProtocolBufferException {
11995      return PARSER.parseFrom(data);
11996    }
11997    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto parseFrom(
11998        byte[] data,
11999        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12000        throws com.google.protobuf.InvalidProtocolBufferException {
12001      return PARSER.parseFrom(data, extensionRegistry);
12002    }
12003    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto parseFrom(java.io.InputStream input)
12004        throws java.io.IOException {
12005      return PARSER.parseFrom(input);
12006    }
12007    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto parseFrom(
12008        java.io.InputStream input,
12009        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12010        throws java.io.IOException {
12011      return PARSER.parseFrom(input, extensionRegistry);
12012    }
12013    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto parseDelimitedFrom(java.io.InputStream input)
12014        throws java.io.IOException {
12015      return PARSER.parseDelimitedFrom(input);
12016    }
12017    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto parseDelimitedFrom(
12018        java.io.InputStream input,
12019        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12020        throws java.io.IOException {
12021      return PARSER.parseDelimitedFrom(input, extensionRegistry);
12022    }
12023    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto parseFrom(
12024        com.google.protobuf.CodedInputStream input)
12025        throws java.io.IOException {
12026      return PARSER.parseFrom(input);
12027    }
12028    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto parseFrom(
12029        com.google.protobuf.CodedInputStream input,
12030        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12031        throws java.io.IOException {
12032      return PARSER.parseFrom(input, extensionRegistry);
12033    }
12034
12035    public static Builder newBuilder() { return Builder.create(); }
12036    public Builder newBuilderForType() { return newBuilder(); }
12037    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto prototype) {
12038      return newBuilder().mergeFrom(prototype);
12039    }
12040    public Builder toBuilder() { return newBuilder(this); }
12041
12042    @java.lang.Override
12043    protected Builder newBuilderForType(
12044        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
12045      Builder builder = new Builder(parent);
12046      return builder;
12047    }
12048    /**
12049     * Protobuf type {@code hadoop.hdfs.qjournal.DoPreUpgradeResponseProto}
12050     */
12051    public static final class Builder extends
12052        com.google.protobuf.GeneratedMessage.Builder<Builder>
12053       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProtoOrBuilder {
12054      public static final com.google.protobuf.Descriptors.Descriptor
12055          getDescriptor() {
12056        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoPreUpgradeResponseProto_descriptor;
12057      }
12058
12059      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
12060          internalGetFieldAccessorTable() {
12061        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoPreUpgradeResponseProto_fieldAccessorTable
12062            .ensureFieldAccessorsInitialized(
12063                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.Builder.class);
12064      }
12065
12066      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.newBuilder()
12067      private Builder() {
12068        maybeForceBuilderInitialization();
12069      }
12070
12071      private Builder(
12072          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
12073        super(parent);
12074        maybeForceBuilderInitialization();
12075      }
12076      private void maybeForceBuilderInitialization() {
12077        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
12078        }
12079      }
12080      private static Builder create() {
12081        return new Builder();
12082      }
12083
12084      public Builder clear() {
12085        super.clear();
12086        return this;
12087      }
12088
12089      public Builder clone() {
12090        return create().mergeFrom(buildPartial());
12091      }
12092
12093      public com.google.protobuf.Descriptors.Descriptor
12094          getDescriptorForType() {
12095        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoPreUpgradeResponseProto_descriptor;
12096      }
12097
12098      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto getDefaultInstanceForType() {
12099        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.getDefaultInstance();
12100      }
12101
12102      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto build() {
12103        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto result = buildPartial();
12104        if (!result.isInitialized()) {
12105          throw newUninitializedMessageException(result);
12106        }
12107        return result;
12108      }
12109
12110      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto buildPartial() {
12111        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto(this);
12112        onBuilt();
12113        return result;
12114      }
12115
12116      public Builder mergeFrom(com.google.protobuf.Message other) {
12117        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto) {
12118          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto)other);
12119        } else {
12120          super.mergeFrom(other);
12121          return this;
12122        }
12123      }
12124
12125      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto other) {
12126        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.getDefaultInstance()) return this;
12127        this.mergeUnknownFields(other.getUnknownFields());
12128        return this;
12129      }
12130
12131      public final boolean isInitialized() {
12132        return true;
12133      }
12134
12135      public Builder mergeFrom(
12136          com.google.protobuf.CodedInputStream input,
12137          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12138          throws java.io.IOException {
12139        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto parsedMessage = null;
12140        try {
12141          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
12142        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
12143          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto) e.getUnfinishedMessage();
12144          throw e;
12145        } finally {
12146          if (parsedMessage != null) {
12147            mergeFrom(parsedMessage);
12148          }
12149        }
12150        return this;
12151      }
12152
12153      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.DoPreUpgradeResponseProto)
12154    }
12155
12156    static {
12157      defaultInstance = new DoPreUpgradeResponseProto(true);
12158      defaultInstance.initFields();
12159    }
12160
12161    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.DoPreUpgradeResponseProto)
12162  }
12163
12164  public interface DoUpgradeRequestProtoOrBuilder
12165      extends com.google.protobuf.MessageOrBuilder {
12166
12167    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
12168    /**
12169     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12170     */
12171    boolean hasJid();
12172    /**
12173     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12174     */
12175    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
12176    /**
12177     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12178     */
12179    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
12180
12181    // required .hadoop.hdfs.StorageInfoProto sInfo = 2;
12182    /**
12183     * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12184     */
12185    boolean hasSInfo();
12186    /**
12187     * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12188     */
12189    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getSInfo();
12190    /**
12191     * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12192     */
12193    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getSInfoOrBuilder();
12194  }
12195  /**
12196   * Protobuf type {@code hadoop.hdfs.qjournal.DoUpgradeRequestProto}
12197   *
12198   * <pre>
12199   **
12200   * doUpgrade()
12201   * </pre>
12202   */
12203  public static final class DoUpgradeRequestProto extends
12204      com.google.protobuf.GeneratedMessage
12205      implements DoUpgradeRequestProtoOrBuilder {
12206    // Use DoUpgradeRequestProto.newBuilder() to construct.
12207    private DoUpgradeRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
12208      super(builder);
12209      this.unknownFields = builder.getUnknownFields();
12210    }
12211    private DoUpgradeRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
12212
12213    private static final DoUpgradeRequestProto defaultInstance;
12214    public static DoUpgradeRequestProto getDefaultInstance() {
12215      return defaultInstance;
12216    }
12217
12218    public DoUpgradeRequestProto getDefaultInstanceForType() {
12219      return defaultInstance;
12220    }
12221
12222    private final com.google.protobuf.UnknownFieldSet unknownFields;
12223    @java.lang.Override
12224    public final com.google.protobuf.UnknownFieldSet
12225        getUnknownFields() {
12226      return this.unknownFields;
12227    }
12228    private DoUpgradeRequestProto(
12229        com.google.protobuf.CodedInputStream input,
12230        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12231        throws com.google.protobuf.InvalidProtocolBufferException {
12232      initFields();
12233      int mutable_bitField0_ = 0;
12234      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
12235          com.google.protobuf.UnknownFieldSet.newBuilder();
12236      try {
12237        boolean done = false;
12238        while (!done) {
12239          int tag = input.readTag();
12240          switch (tag) {
12241            case 0:
12242              done = true;
12243              break;
12244            default: {
12245              if (!parseUnknownField(input, unknownFields,
12246                                     extensionRegistry, tag)) {
12247                done = true;
12248              }
12249              break;
12250            }
12251            case 10: {
12252              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
12253              if (((bitField0_ & 0x00000001) == 0x00000001)) {
12254                subBuilder = jid_.toBuilder();
12255              }
12256              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
12257              if (subBuilder != null) {
12258                subBuilder.mergeFrom(jid_);
12259                jid_ = subBuilder.buildPartial();
12260              }
12261              bitField0_ |= 0x00000001;
12262              break;
12263            }
12264            case 18: {
12265              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder subBuilder = null;
12266              if (((bitField0_ & 0x00000002) == 0x00000002)) {
12267                subBuilder = sInfo_.toBuilder();
12268              }
12269              sInfo_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.PARSER, extensionRegistry);
12270              if (subBuilder != null) {
12271                subBuilder.mergeFrom(sInfo_);
12272                sInfo_ = subBuilder.buildPartial();
12273              }
12274              bitField0_ |= 0x00000002;
12275              break;
12276            }
12277          }
12278        }
12279      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
12280        throw e.setUnfinishedMessage(this);
12281      } catch (java.io.IOException e) {
12282        throw new com.google.protobuf.InvalidProtocolBufferException(
12283            e.getMessage()).setUnfinishedMessage(this);
12284      } finally {
12285        this.unknownFields = unknownFields.build();
12286        makeExtensionsImmutable();
12287      }
12288    }
12289    public static final com.google.protobuf.Descriptors.Descriptor
12290        getDescriptor() {
12291      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoUpgradeRequestProto_descriptor;
12292    }
12293
12294    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
12295        internalGetFieldAccessorTable() {
12296      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoUpgradeRequestProto_fieldAccessorTable
12297          .ensureFieldAccessorsInitialized(
12298              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto.Builder.class);
12299    }
12300
12301    public static com.google.protobuf.Parser<DoUpgradeRequestProto> PARSER =
12302        new com.google.protobuf.AbstractParser<DoUpgradeRequestProto>() {
12303      public DoUpgradeRequestProto parsePartialFrom(
12304          com.google.protobuf.CodedInputStream input,
12305          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12306          throws com.google.protobuf.InvalidProtocolBufferException {
12307        return new DoUpgradeRequestProto(input, extensionRegistry);
12308      }
12309    };
12310
12311    @java.lang.Override
12312    public com.google.protobuf.Parser<DoUpgradeRequestProto> getParserForType() {
12313      return PARSER;
12314    }
12315
12316    private int bitField0_;
12317    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
12318    public static final int JID_FIELD_NUMBER = 1;
12319    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
12320    /**
12321     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12322     */
12323    public boolean hasJid() {
12324      return ((bitField0_ & 0x00000001) == 0x00000001);
12325    }
12326    /**
12327     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12328     */
12329    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
12330      return jid_;
12331    }
12332    /**
12333     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12334     */
12335    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
12336      return jid_;
12337    }
12338
12339    // required .hadoop.hdfs.StorageInfoProto sInfo = 2;
12340    public static final int SINFO_FIELD_NUMBER = 2;
12341    private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto sInfo_;
12342    /**
12343     * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12344     */
12345    public boolean hasSInfo() {
12346      return ((bitField0_ & 0x00000002) == 0x00000002);
12347    }
12348    /**
12349     * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12350     */
12351    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getSInfo() {
12352      return sInfo_;
12353    }
12354    /**
12355     * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12356     */
12357    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getSInfoOrBuilder() {
12358      return sInfo_;
12359    }
12360
12361    private void initFields() {
12362      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
12363      sInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
12364    }
12365    private byte memoizedIsInitialized = -1;
12366    public final boolean isInitialized() {
12367      byte isInitialized = memoizedIsInitialized;
12368      if (isInitialized != -1) return isInitialized == 1;
12369
12370      if (!hasJid()) {
12371        memoizedIsInitialized = 0;
12372        return false;
12373      }
12374      if (!hasSInfo()) {
12375        memoizedIsInitialized = 0;
12376        return false;
12377      }
12378      if (!getJid().isInitialized()) {
12379        memoizedIsInitialized = 0;
12380        return false;
12381      }
12382      if (!getSInfo().isInitialized()) {
12383        memoizedIsInitialized = 0;
12384        return false;
12385      }
12386      memoizedIsInitialized = 1;
12387      return true;
12388    }
12389
12390    public void writeTo(com.google.protobuf.CodedOutputStream output)
12391                        throws java.io.IOException {
12392      getSerializedSize();
12393      if (((bitField0_ & 0x00000001) == 0x00000001)) {
12394        output.writeMessage(1, jid_);
12395      }
12396      if (((bitField0_ & 0x00000002) == 0x00000002)) {
12397        output.writeMessage(2, sInfo_);
12398      }
12399      getUnknownFields().writeTo(output);
12400    }
12401
12402    private int memoizedSerializedSize = -1;
12403    public int getSerializedSize() {
12404      int size = memoizedSerializedSize;
12405      if (size != -1) return size;
12406
12407      size = 0;
12408      if (((bitField0_ & 0x00000001) == 0x00000001)) {
12409        size += com.google.protobuf.CodedOutputStream
12410          .computeMessageSize(1, jid_);
12411      }
12412      if (((bitField0_ & 0x00000002) == 0x00000002)) {
12413        size += com.google.protobuf.CodedOutputStream
12414          .computeMessageSize(2, sInfo_);
12415      }
12416      size += getUnknownFields().getSerializedSize();
12417      memoizedSerializedSize = size;
12418      return size;
12419    }
12420
12421    private static final long serialVersionUID = 0L;
12422    @java.lang.Override
12423    protected java.lang.Object writeReplace()
12424        throws java.io.ObjectStreamException {
12425      return super.writeReplace();
12426    }
12427
12428    @java.lang.Override
12429    public boolean equals(final java.lang.Object obj) {
12430      if (obj == this) {
12431       return true;
12432      }
12433      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto)) {
12434        return super.equals(obj);
12435      }
12436      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto) obj;
12437
12438      boolean result = true;
12439      result = result && (hasJid() == other.hasJid());
12440      if (hasJid()) {
12441        result = result && getJid()
12442            .equals(other.getJid());
12443      }
12444      result = result && (hasSInfo() == other.hasSInfo());
12445      if (hasSInfo()) {
12446        result = result && getSInfo()
12447            .equals(other.getSInfo());
12448      }
12449      result = result &&
12450          getUnknownFields().equals(other.getUnknownFields());
12451      return result;
12452    }
12453
12454    private int memoizedHashCode = 0;
12455    @java.lang.Override
12456    public int hashCode() {
12457      if (memoizedHashCode != 0) {
12458        return memoizedHashCode;
12459      }
12460      int hash = 41;
12461      hash = (19 * hash) + getDescriptorForType().hashCode();
12462      if (hasJid()) {
12463        hash = (37 * hash) + JID_FIELD_NUMBER;
12464        hash = (53 * hash) + getJid().hashCode();
12465      }
12466      if (hasSInfo()) {
12467        hash = (37 * hash) + SINFO_FIELD_NUMBER;
12468        hash = (53 * hash) + getSInfo().hashCode();
12469      }
12470      hash = (29 * hash) + getUnknownFields().hashCode();
12471      memoizedHashCode = hash;
12472      return hash;
12473    }
12474
12475    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto parseFrom(
12476        com.google.protobuf.ByteString data)
12477        throws com.google.protobuf.InvalidProtocolBufferException {
12478      return PARSER.parseFrom(data);
12479    }
12480    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto parseFrom(
12481        com.google.protobuf.ByteString data,
12482        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12483        throws com.google.protobuf.InvalidProtocolBufferException {
12484      return PARSER.parseFrom(data, extensionRegistry);
12485    }
12486    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto parseFrom(byte[] data)
12487        throws com.google.protobuf.InvalidProtocolBufferException {
12488      return PARSER.parseFrom(data);
12489    }
12490    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto parseFrom(
12491        byte[] data,
12492        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12493        throws com.google.protobuf.InvalidProtocolBufferException {
12494      return PARSER.parseFrom(data, extensionRegistry);
12495    }
12496    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto parseFrom(java.io.InputStream input)
12497        throws java.io.IOException {
12498      return PARSER.parseFrom(input);
12499    }
12500    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto parseFrom(
12501        java.io.InputStream input,
12502        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12503        throws java.io.IOException {
12504      return PARSER.parseFrom(input, extensionRegistry);
12505    }
12506    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto parseDelimitedFrom(java.io.InputStream input)
12507        throws java.io.IOException {
12508      return PARSER.parseDelimitedFrom(input);
12509    }
12510    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto parseDelimitedFrom(
12511        java.io.InputStream input,
12512        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12513        throws java.io.IOException {
12514      return PARSER.parseDelimitedFrom(input, extensionRegistry);
12515    }
12516    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto parseFrom(
12517        com.google.protobuf.CodedInputStream input)
12518        throws java.io.IOException {
12519      return PARSER.parseFrom(input);
12520    }
12521    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto parseFrom(
12522        com.google.protobuf.CodedInputStream input,
12523        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12524        throws java.io.IOException {
12525      return PARSER.parseFrom(input, extensionRegistry);
12526    }
12527
12528    public static Builder newBuilder() { return Builder.create(); }
12529    public Builder newBuilderForType() { return newBuilder(); }
12530    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto prototype) {
12531      return newBuilder().mergeFrom(prototype);
12532    }
12533    public Builder toBuilder() { return newBuilder(this); }
12534
12535    @java.lang.Override
12536    protected Builder newBuilderForType(
12537        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
12538      Builder builder = new Builder(parent);
12539      return builder;
12540    }
12541    /**
12542     * Protobuf type {@code hadoop.hdfs.qjournal.DoUpgradeRequestProto}
12543     *
12544     * <pre>
12545     **
12546     * doUpgrade()
12547     * </pre>
12548     */
12549    public static final class Builder extends
12550        com.google.protobuf.GeneratedMessage.Builder<Builder>
12551       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProtoOrBuilder {
12552      public static final com.google.protobuf.Descriptors.Descriptor
12553          getDescriptor() {
12554        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoUpgradeRequestProto_descriptor;
12555      }
12556
12557      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
12558          internalGetFieldAccessorTable() {
12559        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoUpgradeRequestProto_fieldAccessorTable
12560            .ensureFieldAccessorsInitialized(
12561                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto.Builder.class);
12562      }
12563
12564      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto.newBuilder()
12565      private Builder() {
12566        maybeForceBuilderInitialization();
12567      }
12568
12569      private Builder(
12570          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
12571        super(parent);
12572        maybeForceBuilderInitialization();
12573      }
12574      private void maybeForceBuilderInitialization() {
12575        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
12576          getJidFieldBuilder();
12577          getSInfoFieldBuilder();
12578        }
12579      }
12580      private static Builder create() {
12581        return new Builder();
12582      }
12583
12584      public Builder clear() {
12585        super.clear();
12586        if (jidBuilder_ == null) {
12587          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
12588        } else {
12589          jidBuilder_.clear();
12590        }
12591        bitField0_ = (bitField0_ & ~0x00000001);
12592        if (sInfoBuilder_ == null) {
12593          sInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
12594        } else {
12595          sInfoBuilder_.clear();
12596        }
12597        bitField0_ = (bitField0_ & ~0x00000002);
12598        return this;
12599      }
12600
12601      public Builder clone() {
12602        return create().mergeFrom(buildPartial());
12603      }
12604
12605      public com.google.protobuf.Descriptors.Descriptor
12606          getDescriptorForType() {
12607        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoUpgradeRequestProto_descriptor;
12608      }
12609
12610      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto getDefaultInstanceForType() {
12611        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto.getDefaultInstance();
12612      }
12613
12614      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto build() {
12615        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto result = buildPartial();
12616        if (!result.isInitialized()) {
12617          throw newUninitializedMessageException(result);
12618        }
12619        return result;
12620      }
12621
12622      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto buildPartial() {
12623        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto(this);
12624        int from_bitField0_ = bitField0_;
12625        int to_bitField0_ = 0;
12626        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
12627          to_bitField0_ |= 0x00000001;
12628        }
12629        if (jidBuilder_ == null) {
12630          result.jid_ = jid_;
12631        } else {
12632          result.jid_ = jidBuilder_.build();
12633        }
12634        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
12635          to_bitField0_ |= 0x00000002;
12636        }
12637        if (sInfoBuilder_ == null) {
12638          result.sInfo_ = sInfo_;
12639        } else {
12640          result.sInfo_ = sInfoBuilder_.build();
12641        }
12642        result.bitField0_ = to_bitField0_;
12643        onBuilt();
12644        return result;
12645      }
12646
12647      public Builder mergeFrom(com.google.protobuf.Message other) {
12648        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto) {
12649          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto)other);
12650        } else {
12651          super.mergeFrom(other);
12652          return this;
12653        }
12654      }
12655
12656      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto other) {
12657        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto.getDefaultInstance()) return this;
12658        if (other.hasJid()) {
12659          mergeJid(other.getJid());
12660        }
12661        if (other.hasSInfo()) {
12662          mergeSInfo(other.getSInfo());
12663        }
12664        this.mergeUnknownFields(other.getUnknownFields());
12665        return this;
12666      }
12667
12668      public final boolean isInitialized() {
12669        if (!hasJid()) {
12670          
12671          return false;
12672        }
12673        if (!hasSInfo()) {
12674          
12675          return false;
12676        }
12677        if (!getJid().isInitialized()) {
12678          
12679          return false;
12680        }
12681        if (!getSInfo().isInitialized()) {
12682          
12683          return false;
12684        }
12685        return true;
12686      }
12687
12688      public Builder mergeFrom(
12689          com.google.protobuf.CodedInputStream input,
12690          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12691          throws java.io.IOException {
12692        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto parsedMessage = null;
12693        try {
12694          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
12695        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
12696          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto) e.getUnfinishedMessage();
12697          throw e;
12698        } finally {
12699          if (parsedMessage != null) {
12700            mergeFrom(parsedMessage);
12701          }
12702        }
12703        return this;
12704      }
12705      private int bitField0_;
12706
12707      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
12708      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
12709      private com.google.protobuf.SingleFieldBuilder<
12710          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
12711      /**
12712       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12713       */
12714      public boolean hasJid() {
12715        return ((bitField0_ & 0x00000001) == 0x00000001);
12716      }
12717      /**
12718       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12719       */
12720      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
12721        if (jidBuilder_ == null) {
12722          return jid_;
12723        } else {
12724          return jidBuilder_.getMessage();
12725        }
12726      }
12727      /**
12728       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12729       */
12730      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
12731        if (jidBuilder_ == null) {
12732          if (value == null) {
12733            throw new NullPointerException();
12734          }
12735          jid_ = value;
12736          onChanged();
12737        } else {
12738          jidBuilder_.setMessage(value);
12739        }
12740        bitField0_ |= 0x00000001;
12741        return this;
12742      }
12743      /**
12744       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12745       */
12746      public Builder setJid(
12747          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
12748        if (jidBuilder_ == null) {
12749          jid_ = builderForValue.build();
12750          onChanged();
12751        } else {
12752          jidBuilder_.setMessage(builderForValue.build());
12753        }
12754        bitField0_ |= 0x00000001;
12755        return this;
12756      }
12757      /**
12758       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12759       */
12760      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
12761        if (jidBuilder_ == null) {
12762          if (((bitField0_ & 0x00000001) == 0x00000001) &&
12763              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
12764            jid_ =
12765              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
12766          } else {
12767            jid_ = value;
12768          }
12769          onChanged();
12770        } else {
12771          jidBuilder_.mergeFrom(value);
12772        }
12773        bitField0_ |= 0x00000001;
12774        return this;
12775      }
12776      /**
12777       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12778       */
12779      public Builder clearJid() {
12780        if (jidBuilder_ == null) {
12781          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
12782          onChanged();
12783        } else {
12784          jidBuilder_.clear();
12785        }
12786        bitField0_ = (bitField0_ & ~0x00000001);
12787        return this;
12788      }
12789      /**
12790       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12791       */
12792      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
12793        bitField0_ |= 0x00000001;
12794        onChanged();
12795        return getJidFieldBuilder().getBuilder();
12796      }
12797      /**
12798       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12799       */
12800      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
12801        if (jidBuilder_ != null) {
12802          return jidBuilder_.getMessageOrBuilder();
12803        } else {
12804          return jid_;
12805        }
12806      }
12807      /**
12808       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
12809       */
12810      private com.google.protobuf.SingleFieldBuilder<
12811          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
12812          getJidFieldBuilder() {
12813        if (jidBuilder_ == null) {
12814          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
12815              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
12816                  jid_,
12817                  getParentForChildren(),
12818                  isClean());
12819          jid_ = null;
12820        }
12821        return jidBuilder_;
12822      }
12823
12824      // required .hadoop.hdfs.StorageInfoProto sInfo = 2;
12825      private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto sInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
12826      private com.google.protobuf.SingleFieldBuilder<
12827          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder> sInfoBuilder_;
12828      /**
12829       * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12830       */
12831      public boolean hasSInfo() {
12832        return ((bitField0_ & 0x00000002) == 0x00000002);
12833      }
12834      /**
12835       * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12836       */
12837      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getSInfo() {
12838        if (sInfoBuilder_ == null) {
12839          return sInfo_;
12840        } else {
12841          return sInfoBuilder_.getMessage();
12842        }
12843      }
12844      /**
12845       * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12846       */
12847      public Builder setSInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto value) {
12848        if (sInfoBuilder_ == null) {
12849          if (value == null) {
12850            throw new NullPointerException();
12851          }
12852          sInfo_ = value;
12853          onChanged();
12854        } else {
12855          sInfoBuilder_.setMessage(value);
12856        }
12857        bitField0_ |= 0x00000002;
12858        return this;
12859      }
12860      /**
12861       * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12862       */
12863      public Builder setSInfo(
12864          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder builderForValue) {
12865        if (sInfoBuilder_ == null) {
12866          sInfo_ = builderForValue.build();
12867          onChanged();
12868        } else {
12869          sInfoBuilder_.setMessage(builderForValue.build());
12870        }
12871        bitField0_ |= 0x00000002;
12872        return this;
12873      }
12874      /**
12875       * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12876       */
12877      public Builder mergeSInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto value) {
12878        if (sInfoBuilder_ == null) {
12879          if (((bitField0_ & 0x00000002) == 0x00000002) &&
12880              sInfo_ != org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance()) {
12881            sInfo_ =
12882              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.newBuilder(sInfo_).mergeFrom(value).buildPartial();
12883          } else {
12884            sInfo_ = value;
12885          }
12886          onChanged();
12887        } else {
12888          sInfoBuilder_.mergeFrom(value);
12889        }
12890        bitField0_ |= 0x00000002;
12891        return this;
12892      }
12893      /**
12894       * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12895       */
12896      public Builder clearSInfo() {
12897        if (sInfoBuilder_ == null) {
12898          sInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
12899          onChanged();
12900        } else {
12901          sInfoBuilder_.clear();
12902        }
12903        bitField0_ = (bitField0_ & ~0x00000002);
12904        return this;
12905      }
12906      /**
12907       * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12908       */
12909      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder getSInfoBuilder() {
12910        bitField0_ |= 0x00000002;
12911        onChanged();
12912        return getSInfoFieldBuilder().getBuilder();
12913      }
12914      /**
12915       * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12916       */
12917      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getSInfoOrBuilder() {
12918        if (sInfoBuilder_ != null) {
12919          return sInfoBuilder_.getMessageOrBuilder();
12920        } else {
12921          return sInfo_;
12922        }
12923      }
12924      /**
12925       * <code>required .hadoop.hdfs.StorageInfoProto sInfo = 2;</code>
12926       */
12927      private com.google.protobuf.SingleFieldBuilder<
12928          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder> 
12929          getSInfoFieldBuilder() {
12930        if (sInfoBuilder_ == null) {
12931          sInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
12932              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder>(
12933                  sInfo_,
12934                  getParentForChildren(),
12935                  isClean());
12936          sInfo_ = null;
12937        }
12938        return sInfoBuilder_;
12939      }
12940
12941      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.DoUpgradeRequestProto)
12942    }
12943
12944    static {
12945      defaultInstance = new DoUpgradeRequestProto(true);
12946      defaultInstance.initFields();
12947    }
12948
12949    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.DoUpgradeRequestProto)
12950  }
12951
12952  public interface DoUpgradeResponseProtoOrBuilder
12953      extends com.google.protobuf.MessageOrBuilder {
12954  }
12955  /**
12956   * Protobuf type {@code hadoop.hdfs.qjournal.DoUpgradeResponseProto}
12957   */
12958  public static final class DoUpgradeResponseProto extends
12959      com.google.protobuf.GeneratedMessage
12960      implements DoUpgradeResponseProtoOrBuilder {
12961    // Use DoUpgradeResponseProto.newBuilder() to construct.
12962    private DoUpgradeResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
12963      super(builder);
12964      this.unknownFields = builder.getUnknownFields();
12965    }
12966    private DoUpgradeResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
12967
12968    private static final DoUpgradeResponseProto defaultInstance;
12969    public static DoUpgradeResponseProto getDefaultInstance() {
12970      return defaultInstance;
12971    }
12972
12973    public DoUpgradeResponseProto getDefaultInstanceForType() {
12974      return defaultInstance;
12975    }
12976
12977    private final com.google.protobuf.UnknownFieldSet unknownFields;
12978    @java.lang.Override
12979    public final com.google.protobuf.UnknownFieldSet
12980        getUnknownFields() {
12981      return this.unknownFields;
12982    }
12983    private DoUpgradeResponseProto(
12984        com.google.protobuf.CodedInputStream input,
12985        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12986        throws com.google.protobuf.InvalidProtocolBufferException {
12987      initFields();
12988      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
12989          com.google.protobuf.UnknownFieldSet.newBuilder();
12990      try {
12991        boolean done = false;
12992        while (!done) {
12993          int tag = input.readTag();
12994          switch (tag) {
12995            case 0:
12996              done = true;
12997              break;
12998            default: {
12999              if (!parseUnknownField(input, unknownFields,
13000                                     extensionRegistry, tag)) {
13001                done = true;
13002              }
13003              break;
13004            }
13005          }
13006        }
13007      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
13008        throw e.setUnfinishedMessage(this);
13009      } catch (java.io.IOException e) {
13010        throw new com.google.protobuf.InvalidProtocolBufferException(
13011            e.getMessage()).setUnfinishedMessage(this);
13012      } finally {
13013        this.unknownFields = unknownFields.build();
13014        makeExtensionsImmutable();
13015      }
13016    }
13017    public static final com.google.protobuf.Descriptors.Descriptor
13018        getDescriptor() {
13019      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoUpgradeResponseProto_descriptor;
13020    }
13021
13022    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
13023        internalGetFieldAccessorTable() {
13024      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoUpgradeResponseProto_fieldAccessorTable
13025          .ensureFieldAccessorsInitialized(
13026              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.Builder.class);
13027    }
13028
13029    public static com.google.protobuf.Parser<DoUpgradeResponseProto> PARSER =
13030        new com.google.protobuf.AbstractParser<DoUpgradeResponseProto>() {
13031      public DoUpgradeResponseProto parsePartialFrom(
13032          com.google.protobuf.CodedInputStream input,
13033          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13034          throws com.google.protobuf.InvalidProtocolBufferException {
13035        return new DoUpgradeResponseProto(input, extensionRegistry);
13036      }
13037    };
13038
13039    @java.lang.Override
13040    public com.google.protobuf.Parser<DoUpgradeResponseProto> getParserForType() {
13041      return PARSER;
13042    }
13043
13044    private void initFields() {
13045    }
13046    private byte memoizedIsInitialized = -1;
13047    public final boolean isInitialized() {
13048      byte isInitialized = memoizedIsInitialized;
13049      if (isInitialized != -1) return isInitialized == 1;
13050
13051      memoizedIsInitialized = 1;
13052      return true;
13053    }
13054
13055    public void writeTo(com.google.protobuf.CodedOutputStream output)
13056                        throws java.io.IOException {
13057      getSerializedSize();
13058      getUnknownFields().writeTo(output);
13059    }
13060
13061    private int memoizedSerializedSize = -1;
13062    public int getSerializedSize() {
13063      int size = memoizedSerializedSize;
13064      if (size != -1) return size;
13065
13066      size = 0;
13067      size += getUnknownFields().getSerializedSize();
13068      memoizedSerializedSize = size;
13069      return size;
13070    }
13071
13072    private static final long serialVersionUID = 0L;
13073    @java.lang.Override
13074    protected java.lang.Object writeReplace()
13075        throws java.io.ObjectStreamException {
13076      return super.writeReplace();
13077    }
13078
13079    @java.lang.Override
13080    public boolean equals(final java.lang.Object obj) {
13081      if (obj == this) {
13082       return true;
13083      }
13084      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto)) {
13085        return super.equals(obj);
13086      }
13087      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto) obj;
13088
13089      boolean result = true;
13090      result = result &&
13091          getUnknownFields().equals(other.getUnknownFields());
13092      return result;
13093    }
13094
13095    private int memoizedHashCode = 0;
13096    @java.lang.Override
13097    public int hashCode() {
13098      if (memoizedHashCode != 0) {
13099        return memoizedHashCode;
13100      }
13101      int hash = 41;
13102      hash = (19 * hash) + getDescriptorForType().hashCode();
13103      hash = (29 * hash) + getUnknownFields().hashCode();
13104      memoizedHashCode = hash;
13105      return hash;
13106    }
13107
13108    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto parseFrom(
13109        com.google.protobuf.ByteString data)
13110        throws com.google.protobuf.InvalidProtocolBufferException {
13111      return PARSER.parseFrom(data);
13112    }
13113    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto parseFrom(
13114        com.google.protobuf.ByteString data,
13115        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13116        throws com.google.protobuf.InvalidProtocolBufferException {
13117      return PARSER.parseFrom(data, extensionRegistry);
13118    }
13119    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto parseFrom(byte[] data)
13120        throws com.google.protobuf.InvalidProtocolBufferException {
13121      return PARSER.parseFrom(data);
13122    }
13123    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto parseFrom(
13124        byte[] data,
13125        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13126        throws com.google.protobuf.InvalidProtocolBufferException {
13127      return PARSER.parseFrom(data, extensionRegistry);
13128    }
13129    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto parseFrom(java.io.InputStream input)
13130        throws java.io.IOException {
13131      return PARSER.parseFrom(input);
13132    }
13133    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto parseFrom(
13134        java.io.InputStream input,
13135        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13136        throws java.io.IOException {
13137      return PARSER.parseFrom(input, extensionRegistry);
13138    }
13139    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto parseDelimitedFrom(java.io.InputStream input)
13140        throws java.io.IOException {
13141      return PARSER.parseDelimitedFrom(input);
13142    }
13143    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto parseDelimitedFrom(
13144        java.io.InputStream input,
13145        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13146        throws java.io.IOException {
13147      return PARSER.parseDelimitedFrom(input, extensionRegistry);
13148    }
13149    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto parseFrom(
13150        com.google.protobuf.CodedInputStream input)
13151        throws java.io.IOException {
13152      return PARSER.parseFrom(input);
13153    }
13154    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto parseFrom(
13155        com.google.protobuf.CodedInputStream input,
13156        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13157        throws java.io.IOException {
13158      return PARSER.parseFrom(input, extensionRegistry);
13159    }
13160
13161    public static Builder newBuilder() { return Builder.create(); }
13162    public Builder newBuilderForType() { return newBuilder(); }
13163    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto prototype) {
13164      return newBuilder().mergeFrom(prototype);
13165    }
13166    public Builder toBuilder() { return newBuilder(this); }
13167
13168    @java.lang.Override
13169    protected Builder newBuilderForType(
13170        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
13171      Builder builder = new Builder(parent);
13172      return builder;
13173    }
13174    /**
13175     * Protobuf type {@code hadoop.hdfs.qjournal.DoUpgradeResponseProto}
13176     */
13177    public static final class Builder extends
13178        com.google.protobuf.GeneratedMessage.Builder<Builder>
13179       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProtoOrBuilder {
13180      public static final com.google.protobuf.Descriptors.Descriptor
13181          getDescriptor() {
13182        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoUpgradeResponseProto_descriptor;
13183      }
13184
13185      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
13186          internalGetFieldAccessorTable() {
13187        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoUpgradeResponseProto_fieldAccessorTable
13188            .ensureFieldAccessorsInitialized(
13189                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.Builder.class);
13190      }
13191
13192      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.newBuilder()
13193      private Builder() {
13194        maybeForceBuilderInitialization();
13195      }
13196
13197      private Builder(
13198          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
13199        super(parent);
13200        maybeForceBuilderInitialization();
13201      }
13202      private void maybeForceBuilderInitialization() {
13203        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
13204        }
13205      }
13206      private static Builder create() {
13207        return new Builder();
13208      }
13209
13210      public Builder clear() {
13211        super.clear();
13212        return this;
13213      }
13214
13215      public Builder clone() {
13216        return create().mergeFrom(buildPartial());
13217      }
13218
13219      public com.google.protobuf.Descriptors.Descriptor
13220          getDescriptorForType() {
13221        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoUpgradeResponseProto_descriptor;
13222      }
13223
13224      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto getDefaultInstanceForType() {
13225        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.getDefaultInstance();
13226      }
13227
13228      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto build() {
13229        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto result = buildPartial();
13230        if (!result.isInitialized()) {
13231          throw newUninitializedMessageException(result);
13232        }
13233        return result;
13234      }
13235
13236      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto buildPartial() {
13237        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto(this);
13238        onBuilt();
13239        return result;
13240      }
13241
13242      public Builder mergeFrom(com.google.protobuf.Message other) {
13243        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto) {
13244          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto)other);
13245        } else {
13246          super.mergeFrom(other);
13247          return this;
13248        }
13249      }
13250
13251      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto other) {
13252        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.getDefaultInstance()) return this;
13253        this.mergeUnknownFields(other.getUnknownFields());
13254        return this;
13255      }
13256
13257      public final boolean isInitialized() {
13258        return true;
13259      }
13260
13261      public Builder mergeFrom(
13262          com.google.protobuf.CodedInputStream input,
13263          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13264          throws java.io.IOException {
13265        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto parsedMessage = null;
13266        try {
13267          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
13268        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
13269          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto) e.getUnfinishedMessage();
13270          throw e;
13271        } finally {
13272          if (parsedMessage != null) {
13273            mergeFrom(parsedMessage);
13274          }
13275        }
13276        return this;
13277      }
13278
13279      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.DoUpgradeResponseProto)
13280    }
13281
13282    static {
13283      defaultInstance = new DoUpgradeResponseProto(true);
13284      defaultInstance.initFields();
13285    }
13286
13287    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.DoUpgradeResponseProto)
13288  }
13289
13290  public interface DoFinalizeRequestProtoOrBuilder
13291      extends com.google.protobuf.MessageOrBuilder {
13292
13293    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
13294    /**
13295     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13296     */
13297    boolean hasJid();
13298    /**
13299     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13300     */
13301    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
13302    /**
13303     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13304     */
13305    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
13306  }
13307  /**
13308   * Protobuf type {@code hadoop.hdfs.qjournal.DoFinalizeRequestProto}
13309   *
13310   * <pre>
13311   **
13312   * doFinalize()
13313   * </pre>
13314   */
13315  public static final class DoFinalizeRequestProto extends
13316      com.google.protobuf.GeneratedMessage
13317      implements DoFinalizeRequestProtoOrBuilder {
13318    // Use DoFinalizeRequestProto.newBuilder() to construct.
13319    private DoFinalizeRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
13320      super(builder);
13321      this.unknownFields = builder.getUnknownFields();
13322    }
13323    private DoFinalizeRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
13324
13325    private static final DoFinalizeRequestProto defaultInstance;
13326    public static DoFinalizeRequestProto getDefaultInstance() {
13327      return defaultInstance;
13328    }
13329
13330    public DoFinalizeRequestProto getDefaultInstanceForType() {
13331      return defaultInstance;
13332    }
13333
13334    private final com.google.protobuf.UnknownFieldSet unknownFields;
13335    @java.lang.Override
13336    public final com.google.protobuf.UnknownFieldSet
13337        getUnknownFields() {
13338      return this.unknownFields;
13339    }
13340    private DoFinalizeRequestProto(
13341        com.google.protobuf.CodedInputStream input,
13342        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13343        throws com.google.protobuf.InvalidProtocolBufferException {
13344      initFields();
13345      int mutable_bitField0_ = 0;
13346      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
13347          com.google.protobuf.UnknownFieldSet.newBuilder();
13348      try {
13349        boolean done = false;
13350        while (!done) {
13351          int tag = input.readTag();
13352          switch (tag) {
13353            case 0:
13354              done = true;
13355              break;
13356            default: {
13357              if (!parseUnknownField(input, unknownFields,
13358                                     extensionRegistry, tag)) {
13359                done = true;
13360              }
13361              break;
13362            }
13363            case 10: {
13364              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
13365              if (((bitField0_ & 0x00000001) == 0x00000001)) {
13366                subBuilder = jid_.toBuilder();
13367              }
13368              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
13369              if (subBuilder != null) {
13370                subBuilder.mergeFrom(jid_);
13371                jid_ = subBuilder.buildPartial();
13372              }
13373              bitField0_ |= 0x00000001;
13374              break;
13375            }
13376          }
13377        }
13378      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
13379        throw e.setUnfinishedMessage(this);
13380      } catch (java.io.IOException e) {
13381        throw new com.google.protobuf.InvalidProtocolBufferException(
13382            e.getMessage()).setUnfinishedMessage(this);
13383      } finally {
13384        this.unknownFields = unknownFields.build();
13385        makeExtensionsImmutable();
13386      }
13387    }
13388    public static final com.google.protobuf.Descriptors.Descriptor
13389        getDescriptor() {
13390      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoFinalizeRequestProto_descriptor;
13391    }
13392
13393    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
13394        internalGetFieldAccessorTable() {
13395      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoFinalizeRequestProto_fieldAccessorTable
13396          .ensureFieldAccessorsInitialized(
13397              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto.Builder.class);
13398    }
13399
13400    public static com.google.protobuf.Parser<DoFinalizeRequestProto> PARSER =
13401        new com.google.protobuf.AbstractParser<DoFinalizeRequestProto>() {
13402      public DoFinalizeRequestProto parsePartialFrom(
13403          com.google.protobuf.CodedInputStream input,
13404          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13405          throws com.google.protobuf.InvalidProtocolBufferException {
13406        return new DoFinalizeRequestProto(input, extensionRegistry);
13407      }
13408    };
13409
13410    @java.lang.Override
13411    public com.google.protobuf.Parser<DoFinalizeRequestProto> getParserForType() {
13412      return PARSER;
13413    }
13414
13415    private int bitField0_;
13416    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
13417    public static final int JID_FIELD_NUMBER = 1;
13418    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
13419    /**
13420     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13421     */
13422    public boolean hasJid() {
13423      return ((bitField0_ & 0x00000001) == 0x00000001);
13424    }
13425    /**
13426     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13427     */
13428    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
13429      return jid_;
13430    }
13431    /**
13432     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13433     */
13434    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
13435      return jid_;
13436    }
13437
13438    private void initFields() {
13439      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
13440    }
13441    private byte memoizedIsInitialized = -1;
13442    public final boolean isInitialized() {
13443      byte isInitialized = memoizedIsInitialized;
13444      if (isInitialized != -1) return isInitialized == 1;
13445
13446      if (!hasJid()) {
13447        memoizedIsInitialized = 0;
13448        return false;
13449      }
13450      if (!getJid().isInitialized()) {
13451        memoizedIsInitialized = 0;
13452        return false;
13453      }
13454      memoizedIsInitialized = 1;
13455      return true;
13456    }
13457
13458    public void writeTo(com.google.protobuf.CodedOutputStream output)
13459                        throws java.io.IOException {
13460      getSerializedSize();
13461      if (((bitField0_ & 0x00000001) == 0x00000001)) {
13462        output.writeMessage(1, jid_);
13463      }
13464      getUnknownFields().writeTo(output);
13465    }
13466
13467    private int memoizedSerializedSize = -1;
13468    public int getSerializedSize() {
13469      int size = memoizedSerializedSize;
13470      if (size != -1) return size;
13471
13472      size = 0;
13473      if (((bitField0_ & 0x00000001) == 0x00000001)) {
13474        size += com.google.protobuf.CodedOutputStream
13475          .computeMessageSize(1, jid_);
13476      }
13477      size += getUnknownFields().getSerializedSize();
13478      memoizedSerializedSize = size;
13479      return size;
13480    }
13481
13482    private static final long serialVersionUID = 0L;
13483    @java.lang.Override
13484    protected java.lang.Object writeReplace()
13485        throws java.io.ObjectStreamException {
13486      return super.writeReplace();
13487    }
13488
13489    @java.lang.Override
13490    public boolean equals(final java.lang.Object obj) {
13491      if (obj == this) {
13492       return true;
13493      }
13494      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto)) {
13495        return super.equals(obj);
13496      }
13497      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto) obj;
13498
13499      boolean result = true;
13500      result = result && (hasJid() == other.hasJid());
13501      if (hasJid()) {
13502        result = result && getJid()
13503            .equals(other.getJid());
13504      }
13505      result = result &&
13506          getUnknownFields().equals(other.getUnknownFields());
13507      return result;
13508    }
13509
13510    private int memoizedHashCode = 0;
13511    @java.lang.Override
13512    public int hashCode() {
13513      if (memoizedHashCode != 0) {
13514        return memoizedHashCode;
13515      }
13516      int hash = 41;
13517      hash = (19 * hash) + getDescriptorForType().hashCode();
13518      if (hasJid()) {
13519        hash = (37 * hash) + JID_FIELD_NUMBER;
13520        hash = (53 * hash) + getJid().hashCode();
13521      }
13522      hash = (29 * hash) + getUnknownFields().hashCode();
13523      memoizedHashCode = hash;
13524      return hash;
13525    }
13526
13527    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto parseFrom(
13528        com.google.protobuf.ByteString data)
13529        throws com.google.protobuf.InvalidProtocolBufferException {
13530      return PARSER.parseFrom(data);
13531    }
13532    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto parseFrom(
13533        com.google.protobuf.ByteString data,
13534        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13535        throws com.google.protobuf.InvalidProtocolBufferException {
13536      return PARSER.parseFrom(data, extensionRegistry);
13537    }
13538    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto parseFrom(byte[] data)
13539        throws com.google.protobuf.InvalidProtocolBufferException {
13540      return PARSER.parseFrom(data);
13541    }
13542    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto parseFrom(
13543        byte[] data,
13544        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13545        throws com.google.protobuf.InvalidProtocolBufferException {
13546      return PARSER.parseFrom(data, extensionRegistry);
13547    }
13548    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto parseFrom(java.io.InputStream input)
13549        throws java.io.IOException {
13550      return PARSER.parseFrom(input);
13551    }
13552    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto parseFrom(
13553        java.io.InputStream input,
13554        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13555        throws java.io.IOException {
13556      return PARSER.parseFrom(input, extensionRegistry);
13557    }
13558    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto parseDelimitedFrom(java.io.InputStream input)
13559        throws java.io.IOException {
13560      return PARSER.parseDelimitedFrom(input);
13561    }
13562    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto parseDelimitedFrom(
13563        java.io.InputStream input,
13564        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13565        throws java.io.IOException {
13566      return PARSER.parseDelimitedFrom(input, extensionRegistry);
13567    }
13568    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto parseFrom(
13569        com.google.protobuf.CodedInputStream input)
13570        throws java.io.IOException {
13571      return PARSER.parseFrom(input);
13572    }
13573    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto parseFrom(
13574        com.google.protobuf.CodedInputStream input,
13575        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13576        throws java.io.IOException {
13577      return PARSER.parseFrom(input, extensionRegistry);
13578    }
13579
13580    public static Builder newBuilder() { return Builder.create(); }
13581    public Builder newBuilderForType() { return newBuilder(); }
13582    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto prototype) {
13583      return newBuilder().mergeFrom(prototype);
13584    }
13585    public Builder toBuilder() { return newBuilder(this); }
13586
13587    @java.lang.Override
13588    protected Builder newBuilderForType(
13589        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
13590      Builder builder = new Builder(parent);
13591      return builder;
13592    }
13593    /**
13594     * Protobuf type {@code hadoop.hdfs.qjournal.DoFinalizeRequestProto}
13595     *
13596     * <pre>
13597     **
13598     * doFinalize()
13599     * </pre>
13600     */
13601    public static final class Builder extends
13602        com.google.protobuf.GeneratedMessage.Builder<Builder>
13603       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProtoOrBuilder {
13604      public static final com.google.protobuf.Descriptors.Descriptor
13605          getDescriptor() {
13606        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoFinalizeRequestProto_descriptor;
13607      }
13608
13609      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
13610          internalGetFieldAccessorTable() {
13611        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoFinalizeRequestProto_fieldAccessorTable
13612            .ensureFieldAccessorsInitialized(
13613                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto.Builder.class);
13614      }
13615
13616      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto.newBuilder()
13617      private Builder() {
13618        maybeForceBuilderInitialization();
13619      }
13620
13621      private Builder(
13622          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
13623        super(parent);
13624        maybeForceBuilderInitialization();
13625      }
13626      private void maybeForceBuilderInitialization() {
13627        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
13628          getJidFieldBuilder();
13629        }
13630      }
13631      private static Builder create() {
13632        return new Builder();
13633      }
13634
13635      public Builder clear() {
13636        super.clear();
13637        if (jidBuilder_ == null) {
13638          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
13639        } else {
13640          jidBuilder_.clear();
13641        }
13642        bitField0_ = (bitField0_ & ~0x00000001);
13643        return this;
13644      }
13645
13646      public Builder clone() {
13647        return create().mergeFrom(buildPartial());
13648      }
13649
13650      public com.google.protobuf.Descriptors.Descriptor
13651          getDescriptorForType() {
13652        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoFinalizeRequestProto_descriptor;
13653      }
13654
13655      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto getDefaultInstanceForType() {
13656        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto.getDefaultInstance();
13657      }
13658
13659      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto build() {
13660        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto result = buildPartial();
13661        if (!result.isInitialized()) {
13662          throw newUninitializedMessageException(result);
13663        }
13664        return result;
13665      }
13666
13667      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto buildPartial() {
13668        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto(this);
13669        int from_bitField0_ = bitField0_;
13670        int to_bitField0_ = 0;
13671        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
13672          to_bitField0_ |= 0x00000001;
13673        }
13674        if (jidBuilder_ == null) {
13675          result.jid_ = jid_;
13676        } else {
13677          result.jid_ = jidBuilder_.build();
13678        }
13679        result.bitField0_ = to_bitField0_;
13680        onBuilt();
13681        return result;
13682      }
13683
13684      public Builder mergeFrom(com.google.protobuf.Message other) {
13685        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto) {
13686          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto)other);
13687        } else {
13688          super.mergeFrom(other);
13689          return this;
13690        }
13691      }
13692
13693      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto other) {
13694        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto.getDefaultInstance()) return this;
13695        if (other.hasJid()) {
13696          mergeJid(other.getJid());
13697        }
13698        this.mergeUnknownFields(other.getUnknownFields());
13699        return this;
13700      }
13701
13702      public final boolean isInitialized() {
13703        if (!hasJid()) {
13704          
13705          return false;
13706        }
13707        if (!getJid().isInitialized()) {
13708          
13709          return false;
13710        }
13711        return true;
13712      }
13713
13714      public Builder mergeFrom(
13715          com.google.protobuf.CodedInputStream input,
13716          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13717          throws java.io.IOException {
13718        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto parsedMessage = null;
13719        try {
13720          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
13721        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
13722          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto) e.getUnfinishedMessage();
13723          throw e;
13724        } finally {
13725          if (parsedMessage != null) {
13726            mergeFrom(parsedMessage);
13727          }
13728        }
13729        return this;
13730      }
13731      private int bitField0_;
13732
13733      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
13734      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
13735      private com.google.protobuf.SingleFieldBuilder<
13736          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
13737      /**
13738       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13739       */
13740      public boolean hasJid() {
13741        return ((bitField0_ & 0x00000001) == 0x00000001);
13742      }
13743      /**
13744       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13745       */
13746      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
13747        if (jidBuilder_ == null) {
13748          return jid_;
13749        } else {
13750          return jidBuilder_.getMessage();
13751        }
13752      }
13753      /**
13754       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13755       */
13756      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
13757        if (jidBuilder_ == null) {
13758          if (value == null) {
13759            throw new NullPointerException();
13760          }
13761          jid_ = value;
13762          onChanged();
13763        } else {
13764          jidBuilder_.setMessage(value);
13765        }
13766        bitField0_ |= 0x00000001;
13767        return this;
13768      }
13769      /**
13770       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13771       */
13772      public Builder setJid(
13773          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
13774        if (jidBuilder_ == null) {
13775          jid_ = builderForValue.build();
13776          onChanged();
13777        } else {
13778          jidBuilder_.setMessage(builderForValue.build());
13779        }
13780        bitField0_ |= 0x00000001;
13781        return this;
13782      }
13783      /**
13784       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13785       */
13786      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
13787        if (jidBuilder_ == null) {
13788          if (((bitField0_ & 0x00000001) == 0x00000001) &&
13789              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
13790            jid_ =
13791              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
13792          } else {
13793            jid_ = value;
13794          }
13795          onChanged();
13796        } else {
13797          jidBuilder_.mergeFrom(value);
13798        }
13799        bitField0_ |= 0x00000001;
13800        return this;
13801      }
13802      /**
13803       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13804       */
13805      public Builder clearJid() {
13806        if (jidBuilder_ == null) {
13807          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
13808          onChanged();
13809        } else {
13810          jidBuilder_.clear();
13811        }
13812        bitField0_ = (bitField0_ & ~0x00000001);
13813        return this;
13814      }
13815      /**
13816       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13817       */
13818      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
13819        bitField0_ |= 0x00000001;
13820        onChanged();
13821        return getJidFieldBuilder().getBuilder();
13822      }
13823      /**
13824       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13825       */
13826      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
13827        if (jidBuilder_ != null) {
13828          return jidBuilder_.getMessageOrBuilder();
13829        } else {
13830          return jid_;
13831        }
13832      }
13833      /**
13834       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
13835       */
13836      private com.google.protobuf.SingleFieldBuilder<
13837          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
13838          getJidFieldBuilder() {
13839        if (jidBuilder_ == null) {
13840          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
13841              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
13842                  jid_,
13843                  getParentForChildren(),
13844                  isClean());
13845          jid_ = null;
13846        }
13847        return jidBuilder_;
13848      }
13849
13850      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.DoFinalizeRequestProto)
13851    }
13852
13853    static {
13854      defaultInstance = new DoFinalizeRequestProto(true);
13855      defaultInstance.initFields();
13856    }
13857
13858    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.DoFinalizeRequestProto)
13859  }
13860
13861  public interface DoFinalizeResponseProtoOrBuilder
13862      extends com.google.protobuf.MessageOrBuilder {
13863  }
13864  /**
13865   * Protobuf type {@code hadoop.hdfs.qjournal.DoFinalizeResponseProto}
13866   */
13867  public static final class DoFinalizeResponseProto extends
13868      com.google.protobuf.GeneratedMessage
13869      implements DoFinalizeResponseProtoOrBuilder {
13870    // Use DoFinalizeResponseProto.newBuilder() to construct.
13871    private DoFinalizeResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
13872      super(builder);
13873      this.unknownFields = builder.getUnknownFields();
13874    }
13875    private DoFinalizeResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
13876
13877    private static final DoFinalizeResponseProto defaultInstance;
13878    public static DoFinalizeResponseProto getDefaultInstance() {
13879      return defaultInstance;
13880    }
13881
13882    public DoFinalizeResponseProto getDefaultInstanceForType() {
13883      return defaultInstance;
13884    }
13885
13886    private final com.google.protobuf.UnknownFieldSet unknownFields;
13887    @java.lang.Override
13888    public final com.google.protobuf.UnknownFieldSet
13889        getUnknownFields() {
13890      return this.unknownFields;
13891    }
13892    private DoFinalizeResponseProto(
13893        com.google.protobuf.CodedInputStream input,
13894        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13895        throws com.google.protobuf.InvalidProtocolBufferException {
13896      initFields();
13897      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
13898          com.google.protobuf.UnknownFieldSet.newBuilder();
13899      try {
13900        boolean done = false;
13901        while (!done) {
13902          int tag = input.readTag();
13903          switch (tag) {
13904            case 0:
13905              done = true;
13906              break;
13907            default: {
13908              if (!parseUnknownField(input, unknownFields,
13909                                     extensionRegistry, tag)) {
13910                done = true;
13911              }
13912              break;
13913            }
13914          }
13915        }
13916      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
13917        throw e.setUnfinishedMessage(this);
13918      } catch (java.io.IOException e) {
13919        throw new com.google.protobuf.InvalidProtocolBufferException(
13920            e.getMessage()).setUnfinishedMessage(this);
13921      } finally {
13922        this.unknownFields = unknownFields.build();
13923        makeExtensionsImmutable();
13924      }
13925    }
13926    public static final com.google.protobuf.Descriptors.Descriptor
13927        getDescriptor() {
13928      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoFinalizeResponseProto_descriptor;
13929    }
13930
13931    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
13932        internalGetFieldAccessorTable() {
13933      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoFinalizeResponseProto_fieldAccessorTable
13934          .ensureFieldAccessorsInitialized(
13935              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.Builder.class);
13936    }
13937
13938    public static com.google.protobuf.Parser<DoFinalizeResponseProto> PARSER =
13939        new com.google.protobuf.AbstractParser<DoFinalizeResponseProto>() {
13940      public DoFinalizeResponseProto parsePartialFrom(
13941          com.google.protobuf.CodedInputStream input,
13942          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13943          throws com.google.protobuf.InvalidProtocolBufferException {
13944        return new DoFinalizeResponseProto(input, extensionRegistry);
13945      }
13946    };
13947
13948    @java.lang.Override
13949    public com.google.protobuf.Parser<DoFinalizeResponseProto> getParserForType() {
13950      return PARSER;
13951    }
13952
13953    private void initFields() {
13954    }
13955    private byte memoizedIsInitialized = -1;
13956    public final boolean isInitialized() {
13957      byte isInitialized = memoizedIsInitialized;
13958      if (isInitialized != -1) return isInitialized == 1;
13959
13960      memoizedIsInitialized = 1;
13961      return true;
13962    }
13963
13964    public void writeTo(com.google.protobuf.CodedOutputStream output)
13965                        throws java.io.IOException {
13966      getSerializedSize();
13967      getUnknownFields().writeTo(output);
13968    }
13969
13970    private int memoizedSerializedSize = -1;
13971    public int getSerializedSize() {
13972      int size = memoizedSerializedSize;
13973      if (size != -1) return size;
13974
13975      size = 0;
13976      size += getUnknownFields().getSerializedSize();
13977      memoizedSerializedSize = size;
13978      return size;
13979    }
13980
13981    private static final long serialVersionUID = 0L;
13982    @java.lang.Override
13983    protected java.lang.Object writeReplace()
13984        throws java.io.ObjectStreamException {
13985      return super.writeReplace();
13986    }
13987
13988    @java.lang.Override
13989    public boolean equals(final java.lang.Object obj) {
13990      if (obj == this) {
13991       return true;
13992      }
13993      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto)) {
13994        return super.equals(obj);
13995      }
13996      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto) obj;
13997
13998      boolean result = true;
13999      result = result &&
14000          getUnknownFields().equals(other.getUnknownFields());
14001      return result;
14002    }
14003
14004    private int memoizedHashCode = 0;
14005    @java.lang.Override
14006    public int hashCode() {
14007      if (memoizedHashCode != 0) {
14008        return memoizedHashCode;
14009      }
14010      int hash = 41;
14011      hash = (19 * hash) + getDescriptorForType().hashCode();
14012      hash = (29 * hash) + getUnknownFields().hashCode();
14013      memoizedHashCode = hash;
14014      return hash;
14015    }
14016
14017    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto parseFrom(
14018        com.google.protobuf.ByteString data)
14019        throws com.google.protobuf.InvalidProtocolBufferException {
14020      return PARSER.parseFrom(data);
14021    }
14022    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto parseFrom(
14023        com.google.protobuf.ByteString data,
14024        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14025        throws com.google.protobuf.InvalidProtocolBufferException {
14026      return PARSER.parseFrom(data, extensionRegistry);
14027    }
14028    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto parseFrom(byte[] data)
14029        throws com.google.protobuf.InvalidProtocolBufferException {
14030      return PARSER.parseFrom(data);
14031    }
14032    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto parseFrom(
14033        byte[] data,
14034        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14035        throws com.google.protobuf.InvalidProtocolBufferException {
14036      return PARSER.parseFrom(data, extensionRegistry);
14037    }
14038    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto parseFrom(java.io.InputStream input)
14039        throws java.io.IOException {
14040      return PARSER.parseFrom(input);
14041    }
14042    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto parseFrom(
14043        java.io.InputStream input,
14044        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14045        throws java.io.IOException {
14046      return PARSER.parseFrom(input, extensionRegistry);
14047    }
14048    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto parseDelimitedFrom(java.io.InputStream input)
14049        throws java.io.IOException {
14050      return PARSER.parseDelimitedFrom(input);
14051    }
14052    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto parseDelimitedFrom(
14053        java.io.InputStream input,
14054        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14055        throws java.io.IOException {
14056      return PARSER.parseDelimitedFrom(input, extensionRegistry);
14057    }
14058    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto parseFrom(
14059        com.google.protobuf.CodedInputStream input)
14060        throws java.io.IOException {
14061      return PARSER.parseFrom(input);
14062    }
14063    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto parseFrom(
14064        com.google.protobuf.CodedInputStream input,
14065        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14066        throws java.io.IOException {
14067      return PARSER.parseFrom(input, extensionRegistry);
14068    }
14069
14070    public static Builder newBuilder() { return Builder.create(); }
14071    public Builder newBuilderForType() { return newBuilder(); }
14072    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto prototype) {
14073      return newBuilder().mergeFrom(prototype);
14074    }
14075    public Builder toBuilder() { return newBuilder(this); }
14076
14077    @java.lang.Override
14078    protected Builder newBuilderForType(
14079        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
14080      Builder builder = new Builder(parent);
14081      return builder;
14082    }
14083    /**
14084     * Protobuf type {@code hadoop.hdfs.qjournal.DoFinalizeResponseProto}
14085     */
14086    public static final class Builder extends
14087        com.google.protobuf.GeneratedMessage.Builder<Builder>
14088       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProtoOrBuilder {
14089      public static final com.google.protobuf.Descriptors.Descriptor
14090          getDescriptor() {
14091        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoFinalizeResponseProto_descriptor;
14092      }
14093
14094      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
14095          internalGetFieldAccessorTable() {
14096        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoFinalizeResponseProto_fieldAccessorTable
14097            .ensureFieldAccessorsInitialized(
14098                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.Builder.class);
14099      }
14100
14101      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.newBuilder()
14102      private Builder() {
14103        maybeForceBuilderInitialization();
14104      }
14105
14106      private Builder(
14107          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
14108        super(parent);
14109        maybeForceBuilderInitialization();
14110      }
14111      private void maybeForceBuilderInitialization() {
14112        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
14113        }
14114      }
14115      private static Builder create() {
14116        return new Builder();
14117      }
14118
14119      public Builder clear() {
14120        super.clear();
14121        return this;
14122      }
14123
14124      public Builder clone() {
14125        return create().mergeFrom(buildPartial());
14126      }
14127
14128      public com.google.protobuf.Descriptors.Descriptor
14129          getDescriptorForType() {
14130        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoFinalizeResponseProto_descriptor;
14131      }
14132
14133      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto getDefaultInstanceForType() {
14134        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.getDefaultInstance();
14135      }
14136
14137      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto build() {
14138        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto result = buildPartial();
14139        if (!result.isInitialized()) {
14140          throw newUninitializedMessageException(result);
14141        }
14142        return result;
14143      }
14144
14145      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto buildPartial() {
14146        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto(this);
14147        onBuilt();
14148        return result;
14149      }
14150
14151      public Builder mergeFrom(com.google.protobuf.Message other) {
14152        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto) {
14153          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto)other);
14154        } else {
14155          super.mergeFrom(other);
14156          return this;
14157        }
14158      }
14159
14160      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto other) {
14161        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.getDefaultInstance()) return this;
14162        this.mergeUnknownFields(other.getUnknownFields());
14163        return this;
14164      }
14165
14166      public final boolean isInitialized() {
14167        return true;
14168      }
14169
14170      public Builder mergeFrom(
14171          com.google.protobuf.CodedInputStream input,
14172          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14173          throws java.io.IOException {
14174        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto parsedMessage = null;
14175        try {
14176          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
14177        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
14178          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto) e.getUnfinishedMessage();
14179          throw e;
14180        } finally {
14181          if (parsedMessage != null) {
14182            mergeFrom(parsedMessage);
14183          }
14184        }
14185        return this;
14186      }
14187
14188      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.DoFinalizeResponseProto)
14189    }
14190
14191    static {
14192      defaultInstance = new DoFinalizeResponseProto(true);
14193      defaultInstance.initFields();
14194    }
14195
14196    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.DoFinalizeResponseProto)
14197  }
14198
14199  public interface CanRollBackRequestProtoOrBuilder
14200      extends com.google.protobuf.MessageOrBuilder {
14201
14202    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
14203    /**
14204     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14205     */
14206    boolean hasJid();
14207    /**
14208     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14209     */
14210    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
14211    /**
14212     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14213     */
14214    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
14215
14216    // required .hadoop.hdfs.StorageInfoProto storage = 2;
14217    /**
14218     * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
14219     */
14220    boolean hasStorage();
14221    /**
14222     * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
14223     */
14224    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getStorage();
14225    /**
14226     * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
14227     */
14228    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getStorageOrBuilder();
14229
14230    // required .hadoop.hdfs.StorageInfoProto prevStorage = 3;
14231    /**
14232     * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
14233     */
14234    boolean hasPrevStorage();
14235    /**
14236     * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
14237     */
14238    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getPrevStorage();
14239    /**
14240     * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
14241     */
14242    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getPrevStorageOrBuilder();
14243
14244    // required int32 targetLayoutVersion = 4;
14245    /**
14246     * <code>required int32 targetLayoutVersion = 4;</code>
14247     */
14248    boolean hasTargetLayoutVersion();
14249    /**
14250     * <code>required int32 targetLayoutVersion = 4;</code>
14251     */
14252    int getTargetLayoutVersion();
14253  }
14254  /**
14255   * Protobuf type {@code hadoop.hdfs.qjournal.CanRollBackRequestProto}
14256   *
14257   * <pre>
14258   **
14259   * canRollBack()
14260   * </pre>
14261   */
14262  public static final class CanRollBackRequestProto extends
14263      com.google.protobuf.GeneratedMessage
14264      implements CanRollBackRequestProtoOrBuilder {
14265    // Use CanRollBackRequestProto.newBuilder() to construct.
14266    private CanRollBackRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
14267      super(builder);
14268      this.unknownFields = builder.getUnknownFields();
14269    }
14270    private CanRollBackRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
14271
14272    private static final CanRollBackRequestProto defaultInstance;
14273    public static CanRollBackRequestProto getDefaultInstance() {
14274      return defaultInstance;
14275    }
14276
14277    public CanRollBackRequestProto getDefaultInstanceForType() {
14278      return defaultInstance;
14279    }
14280
14281    private final com.google.protobuf.UnknownFieldSet unknownFields;
14282    @java.lang.Override
14283    public final com.google.protobuf.UnknownFieldSet
14284        getUnknownFields() {
14285      return this.unknownFields;
14286    }
14287    private CanRollBackRequestProto(
14288        com.google.protobuf.CodedInputStream input,
14289        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14290        throws com.google.protobuf.InvalidProtocolBufferException {
14291      initFields();
14292      int mutable_bitField0_ = 0;
14293      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
14294          com.google.protobuf.UnknownFieldSet.newBuilder();
14295      try {
14296        boolean done = false;
14297        while (!done) {
14298          int tag = input.readTag();
14299          switch (tag) {
14300            case 0:
14301              done = true;
14302              break;
14303            default: {
14304              if (!parseUnknownField(input, unknownFields,
14305                                     extensionRegistry, tag)) {
14306                done = true;
14307              }
14308              break;
14309            }
14310            case 10: {
14311              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
14312              if (((bitField0_ & 0x00000001) == 0x00000001)) {
14313                subBuilder = jid_.toBuilder();
14314              }
14315              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
14316              if (subBuilder != null) {
14317                subBuilder.mergeFrom(jid_);
14318                jid_ = subBuilder.buildPartial();
14319              }
14320              bitField0_ |= 0x00000001;
14321              break;
14322            }
14323            case 18: {
14324              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder subBuilder = null;
14325              if (((bitField0_ & 0x00000002) == 0x00000002)) {
14326                subBuilder = storage_.toBuilder();
14327              }
14328              storage_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.PARSER, extensionRegistry);
14329              if (subBuilder != null) {
14330                subBuilder.mergeFrom(storage_);
14331                storage_ = subBuilder.buildPartial();
14332              }
14333              bitField0_ |= 0x00000002;
14334              break;
14335            }
14336            case 26: {
14337              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder subBuilder = null;
14338              if (((bitField0_ & 0x00000004) == 0x00000004)) {
14339                subBuilder = prevStorage_.toBuilder();
14340              }
14341              prevStorage_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.PARSER, extensionRegistry);
14342              if (subBuilder != null) {
14343                subBuilder.mergeFrom(prevStorage_);
14344                prevStorage_ = subBuilder.buildPartial();
14345              }
14346              bitField0_ |= 0x00000004;
14347              break;
14348            }
14349            case 32: {
14350              bitField0_ |= 0x00000008;
14351              targetLayoutVersion_ = input.readInt32();
14352              break;
14353            }
14354          }
14355        }
14356      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
14357        throw e.setUnfinishedMessage(this);
14358      } catch (java.io.IOException e) {
14359        throw new com.google.protobuf.InvalidProtocolBufferException(
14360            e.getMessage()).setUnfinishedMessage(this);
14361      } finally {
14362        this.unknownFields = unknownFields.build();
14363        makeExtensionsImmutable();
14364      }
14365    }
14366    public static final com.google.protobuf.Descriptors.Descriptor
14367        getDescriptor() {
14368      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_CanRollBackRequestProto_descriptor;
14369    }
14370
14371    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
14372        internalGetFieldAccessorTable() {
14373      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_CanRollBackRequestProto_fieldAccessorTable
14374          .ensureFieldAccessorsInitialized(
14375              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto.Builder.class);
14376    }
14377
14378    public static com.google.protobuf.Parser<CanRollBackRequestProto> PARSER =
14379        new com.google.protobuf.AbstractParser<CanRollBackRequestProto>() {
14380      public CanRollBackRequestProto parsePartialFrom(
14381          com.google.protobuf.CodedInputStream input,
14382          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14383          throws com.google.protobuf.InvalidProtocolBufferException {
14384        return new CanRollBackRequestProto(input, extensionRegistry);
14385      }
14386    };
14387
14388    @java.lang.Override
14389    public com.google.protobuf.Parser<CanRollBackRequestProto> getParserForType() {
14390      return PARSER;
14391    }
14392
14393    private int bitField0_;
14394    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
14395    public static final int JID_FIELD_NUMBER = 1;
14396    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
14397    /**
14398     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14399     */
14400    public boolean hasJid() {
14401      return ((bitField0_ & 0x00000001) == 0x00000001);
14402    }
14403    /**
14404     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14405     */
14406    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
14407      return jid_;
14408    }
14409    /**
14410     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14411     */
14412    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
14413      return jid_;
14414    }
14415
14416    // required .hadoop.hdfs.StorageInfoProto storage = 2;
14417    public static final int STORAGE_FIELD_NUMBER = 2;
14418    private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto storage_;
14419    /**
14420     * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
14421     */
14422    public boolean hasStorage() {
14423      return ((bitField0_ & 0x00000002) == 0x00000002);
14424    }
14425    /**
14426     * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
14427     */
14428    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getStorage() {
14429      return storage_;
14430    }
14431    /**
14432     * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
14433     */
14434    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getStorageOrBuilder() {
14435      return storage_;
14436    }
14437
14438    // required .hadoop.hdfs.StorageInfoProto prevStorage = 3;
14439    public static final int PREVSTORAGE_FIELD_NUMBER = 3;
14440    private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto prevStorage_;
14441    /**
14442     * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
14443     */
14444    public boolean hasPrevStorage() {
14445      return ((bitField0_ & 0x00000004) == 0x00000004);
14446    }
14447    /**
14448     * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
14449     */
14450    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getPrevStorage() {
14451      return prevStorage_;
14452    }
14453    /**
14454     * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
14455     */
14456    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getPrevStorageOrBuilder() {
14457      return prevStorage_;
14458    }
14459
14460    // required int32 targetLayoutVersion = 4;
14461    public static final int TARGETLAYOUTVERSION_FIELD_NUMBER = 4;
14462    private int targetLayoutVersion_;
14463    /**
14464     * <code>required int32 targetLayoutVersion = 4;</code>
14465     */
14466    public boolean hasTargetLayoutVersion() {
14467      return ((bitField0_ & 0x00000008) == 0x00000008);
14468    }
14469    /**
14470     * <code>required int32 targetLayoutVersion = 4;</code>
14471     */
14472    public int getTargetLayoutVersion() {
14473      return targetLayoutVersion_;
14474    }
14475
14476    private void initFields() {
14477      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
14478      storage_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
14479      prevStorage_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
14480      targetLayoutVersion_ = 0;
14481    }
14482    private byte memoizedIsInitialized = -1;
14483    public final boolean isInitialized() {
14484      byte isInitialized = memoizedIsInitialized;
14485      if (isInitialized != -1) return isInitialized == 1;
14486
14487      if (!hasJid()) {
14488        memoizedIsInitialized = 0;
14489        return false;
14490      }
14491      if (!hasStorage()) {
14492        memoizedIsInitialized = 0;
14493        return false;
14494      }
14495      if (!hasPrevStorage()) {
14496        memoizedIsInitialized = 0;
14497        return false;
14498      }
14499      if (!hasTargetLayoutVersion()) {
14500        memoizedIsInitialized = 0;
14501        return false;
14502      }
14503      if (!getJid().isInitialized()) {
14504        memoizedIsInitialized = 0;
14505        return false;
14506      }
14507      if (!getStorage().isInitialized()) {
14508        memoizedIsInitialized = 0;
14509        return false;
14510      }
14511      if (!getPrevStorage().isInitialized()) {
14512        memoizedIsInitialized = 0;
14513        return false;
14514      }
14515      memoizedIsInitialized = 1;
14516      return true;
14517    }
14518
14519    public void writeTo(com.google.protobuf.CodedOutputStream output)
14520                        throws java.io.IOException {
14521      getSerializedSize();
14522      if (((bitField0_ & 0x00000001) == 0x00000001)) {
14523        output.writeMessage(1, jid_);
14524      }
14525      if (((bitField0_ & 0x00000002) == 0x00000002)) {
14526        output.writeMessage(2, storage_);
14527      }
14528      if (((bitField0_ & 0x00000004) == 0x00000004)) {
14529        output.writeMessage(3, prevStorage_);
14530      }
14531      if (((bitField0_ & 0x00000008) == 0x00000008)) {
14532        output.writeInt32(4, targetLayoutVersion_);
14533      }
14534      getUnknownFields().writeTo(output);
14535    }
14536
14537    private int memoizedSerializedSize = -1;
14538    public int getSerializedSize() {
14539      int size = memoizedSerializedSize;
14540      if (size != -1) return size;
14541
14542      size = 0;
14543      if (((bitField0_ & 0x00000001) == 0x00000001)) {
14544        size += com.google.protobuf.CodedOutputStream
14545          .computeMessageSize(1, jid_);
14546      }
14547      if (((bitField0_ & 0x00000002) == 0x00000002)) {
14548        size += com.google.protobuf.CodedOutputStream
14549          .computeMessageSize(2, storage_);
14550      }
14551      if (((bitField0_ & 0x00000004) == 0x00000004)) {
14552        size += com.google.protobuf.CodedOutputStream
14553          .computeMessageSize(3, prevStorage_);
14554      }
14555      if (((bitField0_ & 0x00000008) == 0x00000008)) {
14556        size += com.google.protobuf.CodedOutputStream
14557          .computeInt32Size(4, targetLayoutVersion_);
14558      }
14559      size += getUnknownFields().getSerializedSize();
14560      memoizedSerializedSize = size;
14561      return size;
14562    }
14563
14564    private static final long serialVersionUID = 0L;
14565    @java.lang.Override
14566    protected java.lang.Object writeReplace()
14567        throws java.io.ObjectStreamException {
14568      return super.writeReplace();
14569    }
14570
14571    @java.lang.Override
14572    public boolean equals(final java.lang.Object obj) {
14573      if (obj == this) {
14574       return true;
14575      }
14576      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto)) {
14577        return super.equals(obj);
14578      }
14579      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto) obj;
14580
14581      boolean result = true;
14582      result = result && (hasJid() == other.hasJid());
14583      if (hasJid()) {
14584        result = result && getJid()
14585            .equals(other.getJid());
14586      }
14587      result = result && (hasStorage() == other.hasStorage());
14588      if (hasStorage()) {
14589        result = result && getStorage()
14590            .equals(other.getStorage());
14591      }
14592      result = result && (hasPrevStorage() == other.hasPrevStorage());
14593      if (hasPrevStorage()) {
14594        result = result && getPrevStorage()
14595            .equals(other.getPrevStorage());
14596      }
14597      result = result && (hasTargetLayoutVersion() == other.hasTargetLayoutVersion());
14598      if (hasTargetLayoutVersion()) {
14599        result = result && (getTargetLayoutVersion()
14600            == other.getTargetLayoutVersion());
14601      }
14602      result = result &&
14603          getUnknownFields().equals(other.getUnknownFields());
14604      return result;
14605    }
14606
14607    private int memoizedHashCode = 0;
14608    @java.lang.Override
14609    public int hashCode() {
14610      if (memoizedHashCode != 0) {
14611        return memoizedHashCode;
14612      }
14613      int hash = 41;
14614      hash = (19 * hash) + getDescriptorForType().hashCode();
14615      if (hasJid()) {
14616        hash = (37 * hash) + JID_FIELD_NUMBER;
14617        hash = (53 * hash) + getJid().hashCode();
14618      }
14619      if (hasStorage()) {
14620        hash = (37 * hash) + STORAGE_FIELD_NUMBER;
14621        hash = (53 * hash) + getStorage().hashCode();
14622      }
14623      if (hasPrevStorage()) {
14624        hash = (37 * hash) + PREVSTORAGE_FIELD_NUMBER;
14625        hash = (53 * hash) + getPrevStorage().hashCode();
14626      }
14627      if (hasTargetLayoutVersion()) {
14628        hash = (37 * hash) + TARGETLAYOUTVERSION_FIELD_NUMBER;
14629        hash = (53 * hash) + getTargetLayoutVersion();
14630      }
14631      hash = (29 * hash) + getUnknownFields().hashCode();
14632      memoizedHashCode = hash;
14633      return hash;
14634    }
14635
14636    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto parseFrom(
14637        com.google.protobuf.ByteString data)
14638        throws com.google.protobuf.InvalidProtocolBufferException {
14639      return PARSER.parseFrom(data);
14640    }
14641    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto parseFrom(
14642        com.google.protobuf.ByteString data,
14643        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14644        throws com.google.protobuf.InvalidProtocolBufferException {
14645      return PARSER.parseFrom(data, extensionRegistry);
14646    }
14647    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto parseFrom(byte[] data)
14648        throws com.google.protobuf.InvalidProtocolBufferException {
14649      return PARSER.parseFrom(data);
14650    }
14651    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto parseFrom(
14652        byte[] data,
14653        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14654        throws com.google.protobuf.InvalidProtocolBufferException {
14655      return PARSER.parseFrom(data, extensionRegistry);
14656    }
14657    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto parseFrom(java.io.InputStream input)
14658        throws java.io.IOException {
14659      return PARSER.parseFrom(input);
14660    }
14661    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto parseFrom(
14662        java.io.InputStream input,
14663        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14664        throws java.io.IOException {
14665      return PARSER.parseFrom(input, extensionRegistry);
14666    }
14667    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto parseDelimitedFrom(java.io.InputStream input)
14668        throws java.io.IOException {
14669      return PARSER.parseDelimitedFrom(input);
14670    }
14671    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto parseDelimitedFrom(
14672        java.io.InputStream input,
14673        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14674        throws java.io.IOException {
14675      return PARSER.parseDelimitedFrom(input, extensionRegistry);
14676    }
14677    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto parseFrom(
14678        com.google.protobuf.CodedInputStream input)
14679        throws java.io.IOException {
14680      return PARSER.parseFrom(input);
14681    }
14682    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto parseFrom(
14683        com.google.protobuf.CodedInputStream input,
14684        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14685        throws java.io.IOException {
14686      return PARSER.parseFrom(input, extensionRegistry);
14687    }
14688
14689    public static Builder newBuilder() { return Builder.create(); }
14690    public Builder newBuilderForType() { return newBuilder(); }
14691    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto prototype) {
14692      return newBuilder().mergeFrom(prototype);
14693    }
14694    public Builder toBuilder() { return newBuilder(this); }
14695
14696    @java.lang.Override
14697    protected Builder newBuilderForType(
14698        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
14699      Builder builder = new Builder(parent);
14700      return builder;
14701    }
14702    /**
14703     * Protobuf type {@code hadoop.hdfs.qjournal.CanRollBackRequestProto}
14704     *
14705     * <pre>
14706     **
14707     * canRollBack()
14708     * </pre>
14709     */
14710    public static final class Builder extends
14711        com.google.protobuf.GeneratedMessage.Builder<Builder>
14712       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProtoOrBuilder {
14713      public static final com.google.protobuf.Descriptors.Descriptor
14714          getDescriptor() {
14715        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_CanRollBackRequestProto_descriptor;
14716      }
14717
14718      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
14719          internalGetFieldAccessorTable() {
14720        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_CanRollBackRequestProto_fieldAccessorTable
14721            .ensureFieldAccessorsInitialized(
14722                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto.Builder.class);
14723      }
14724
14725      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto.newBuilder()
14726      private Builder() {
14727        maybeForceBuilderInitialization();
14728      }
14729
14730      private Builder(
14731          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
14732        super(parent);
14733        maybeForceBuilderInitialization();
14734      }
14735      private void maybeForceBuilderInitialization() {
14736        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
14737          getJidFieldBuilder();
14738          getStorageFieldBuilder();
14739          getPrevStorageFieldBuilder();
14740        }
14741      }
14742      private static Builder create() {
14743        return new Builder();
14744      }
14745
14746      public Builder clear() {
14747        super.clear();
14748        if (jidBuilder_ == null) {
14749          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
14750        } else {
14751          jidBuilder_.clear();
14752        }
14753        bitField0_ = (bitField0_ & ~0x00000001);
14754        if (storageBuilder_ == null) {
14755          storage_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
14756        } else {
14757          storageBuilder_.clear();
14758        }
14759        bitField0_ = (bitField0_ & ~0x00000002);
14760        if (prevStorageBuilder_ == null) {
14761          prevStorage_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
14762        } else {
14763          prevStorageBuilder_.clear();
14764        }
14765        bitField0_ = (bitField0_ & ~0x00000004);
14766        targetLayoutVersion_ = 0;
14767        bitField0_ = (bitField0_ & ~0x00000008);
14768        return this;
14769      }
14770
14771      public Builder clone() {
14772        return create().mergeFrom(buildPartial());
14773      }
14774
14775      public com.google.protobuf.Descriptors.Descriptor
14776          getDescriptorForType() {
14777        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_CanRollBackRequestProto_descriptor;
14778      }
14779
14780      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto getDefaultInstanceForType() {
14781        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto.getDefaultInstance();
14782      }
14783
14784      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto build() {
14785        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto result = buildPartial();
14786        if (!result.isInitialized()) {
14787          throw newUninitializedMessageException(result);
14788        }
14789        return result;
14790      }
14791
14792      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto buildPartial() {
14793        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto(this);
14794        int from_bitField0_ = bitField0_;
14795        int to_bitField0_ = 0;
14796        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
14797          to_bitField0_ |= 0x00000001;
14798        }
14799        if (jidBuilder_ == null) {
14800          result.jid_ = jid_;
14801        } else {
14802          result.jid_ = jidBuilder_.build();
14803        }
14804        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
14805          to_bitField0_ |= 0x00000002;
14806        }
14807        if (storageBuilder_ == null) {
14808          result.storage_ = storage_;
14809        } else {
14810          result.storage_ = storageBuilder_.build();
14811        }
14812        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
14813          to_bitField0_ |= 0x00000004;
14814        }
14815        if (prevStorageBuilder_ == null) {
14816          result.prevStorage_ = prevStorage_;
14817        } else {
14818          result.prevStorage_ = prevStorageBuilder_.build();
14819        }
14820        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
14821          to_bitField0_ |= 0x00000008;
14822        }
14823        result.targetLayoutVersion_ = targetLayoutVersion_;
14824        result.bitField0_ = to_bitField0_;
14825        onBuilt();
14826        return result;
14827      }
14828
14829      public Builder mergeFrom(com.google.protobuf.Message other) {
14830        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto) {
14831          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto)other);
14832        } else {
14833          super.mergeFrom(other);
14834          return this;
14835        }
14836      }
14837
14838      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto other) {
14839        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto.getDefaultInstance()) return this;
14840        if (other.hasJid()) {
14841          mergeJid(other.getJid());
14842        }
14843        if (other.hasStorage()) {
14844          mergeStorage(other.getStorage());
14845        }
14846        if (other.hasPrevStorage()) {
14847          mergePrevStorage(other.getPrevStorage());
14848        }
14849        if (other.hasTargetLayoutVersion()) {
14850          setTargetLayoutVersion(other.getTargetLayoutVersion());
14851        }
14852        this.mergeUnknownFields(other.getUnknownFields());
14853        return this;
14854      }
14855
14856      public final boolean isInitialized() {
14857        if (!hasJid()) {
14858          
14859          return false;
14860        }
14861        if (!hasStorage()) {
14862          
14863          return false;
14864        }
14865        if (!hasPrevStorage()) {
14866          
14867          return false;
14868        }
14869        if (!hasTargetLayoutVersion()) {
14870          
14871          return false;
14872        }
14873        if (!getJid().isInitialized()) {
14874          
14875          return false;
14876        }
14877        if (!getStorage().isInitialized()) {
14878          
14879          return false;
14880        }
14881        if (!getPrevStorage().isInitialized()) {
14882          
14883          return false;
14884        }
14885        return true;
14886      }
14887
14888      public Builder mergeFrom(
14889          com.google.protobuf.CodedInputStream input,
14890          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14891          throws java.io.IOException {
14892        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto parsedMessage = null;
14893        try {
14894          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
14895        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
14896          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto) e.getUnfinishedMessage();
14897          throw e;
14898        } finally {
14899          if (parsedMessage != null) {
14900            mergeFrom(parsedMessage);
14901          }
14902        }
14903        return this;
14904      }
14905      private int bitField0_;
14906
14907      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
14908      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
14909      private com.google.protobuf.SingleFieldBuilder<
14910          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
14911      /**
14912       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14913       */
14914      public boolean hasJid() {
14915        return ((bitField0_ & 0x00000001) == 0x00000001);
14916      }
14917      /**
14918       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14919       */
14920      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
14921        if (jidBuilder_ == null) {
14922          return jid_;
14923        } else {
14924          return jidBuilder_.getMessage();
14925        }
14926      }
14927      /**
14928       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14929       */
14930      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
14931        if (jidBuilder_ == null) {
14932          if (value == null) {
14933            throw new NullPointerException();
14934          }
14935          jid_ = value;
14936          onChanged();
14937        } else {
14938          jidBuilder_.setMessage(value);
14939        }
14940        bitField0_ |= 0x00000001;
14941        return this;
14942      }
14943      /**
14944       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14945       */
14946      public Builder setJid(
14947          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
14948        if (jidBuilder_ == null) {
14949          jid_ = builderForValue.build();
14950          onChanged();
14951        } else {
14952          jidBuilder_.setMessage(builderForValue.build());
14953        }
14954        bitField0_ |= 0x00000001;
14955        return this;
14956      }
14957      /**
14958       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14959       */
14960      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
14961        if (jidBuilder_ == null) {
14962          if (((bitField0_ & 0x00000001) == 0x00000001) &&
14963              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
14964            jid_ =
14965              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
14966          } else {
14967            jid_ = value;
14968          }
14969          onChanged();
14970        } else {
14971          jidBuilder_.mergeFrom(value);
14972        }
14973        bitField0_ |= 0x00000001;
14974        return this;
14975      }
14976      /**
14977       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14978       */
14979      public Builder clearJid() {
14980        if (jidBuilder_ == null) {
14981          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
14982          onChanged();
14983        } else {
14984          jidBuilder_.clear();
14985        }
14986        bitField0_ = (bitField0_ & ~0x00000001);
14987        return this;
14988      }
14989      /**
14990       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14991       */
14992      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
14993        bitField0_ |= 0x00000001;
14994        onChanged();
14995        return getJidFieldBuilder().getBuilder();
14996      }
14997      /**
14998       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
14999       */
15000      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
15001        if (jidBuilder_ != null) {
15002          return jidBuilder_.getMessageOrBuilder();
15003        } else {
15004          return jid_;
15005        }
15006      }
15007      /**
15008       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
15009       */
15010      private com.google.protobuf.SingleFieldBuilder<
15011          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
15012          getJidFieldBuilder() {
15013        if (jidBuilder_ == null) {
15014          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
15015              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
15016                  jid_,
15017                  getParentForChildren(),
15018                  isClean());
15019          jid_ = null;
15020        }
15021        return jidBuilder_;
15022      }
15023
15024      // required .hadoop.hdfs.StorageInfoProto storage = 2;
15025      private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto storage_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
15026      private com.google.protobuf.SingleFieldBuilder<
15027          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder> storageBuilder_;
15028      /**
15029       * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
15030       */
15031      public boolean hasStorage() {
15032        return ((bitField0_ & 0x00000002) == 0x00000002);
15033      }
15034      /**
15035       * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
15036       */
15037      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getStorage() {
15038        if (storageBuilder_ == null) {
15039          return storage_;
15040        } else {
15041          return storageBuilder_.getMessage();
15042        }
15043      }
15044      /**
15045       * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
15046       */
15047      public Builder setStorage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto value) {
15048        if (storageBuilder_ == null) {
15049          if (value == null) {
15050            throw new NullPointerException();
15051          }
15052          storage_ = value;
15053          onChanged();
15054        } else {
15055          storageBuilder_.setMessage(value);
15056        }
15057        bitField0_ |= 0x00000002;
15058        return this;
15059      }
15060      /**
15061       * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
15062       */
15063      public Builder setStorage(
15064          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder builderForValue) {
15065        if (storageBuilder_ == null) {
15066          storage_ = builderForValue.build();
15067          onChanged();
15068        } else {
15069          storageBuilder_.setMessage(builderForValue.build());
15070        }
15071        bitField0_ |= 0x00000002;
15072        return this;
15073      }
15074      /**
15075       * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
15076       */
15077      public Builder mergeStorage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto value) {
15078        if (storageBuilder_ == null) {
15079          if (((bitField0_ & 0x00000002) == 0x00000002) &&
15080              storage_ != org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance()) {
15081            storage_ =
15082              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.newBuilder(storage_).mergeFrom(value).buildPartial();
15083          } else {
15084            storage_ = value;
15085          }
15086          onChanged();
15087        } else {
15088          storageBuilder_.mergeFrom(value);
15089        }
15090        bitField0_ |= 0x00000002;
15091        return this;
15092      }
15093      /**
15094       * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
15095       */
15096      public Builder clearStorage() {
15097        if (storageBuilder_ == null) {
15098          storage_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
15099          onChanged();
15100        } else {
15101          storageBuilder_.clear();
15102        }
15103        bitField0_ = (bitField0_ & ~0x00000002);
15104        return this;
15105      }
15106      /**
15107       * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
15108       */
15109      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder getStorageBuilder() {
15110        bitField0_ |= 0x00000002;
15111        onChanged();
15112        return getStorageFieldBuilder().getBuilder();
15113      }
15114      /**
15115       * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
15116       */
15117      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getStorageOrBuilder() {
15118        if (storageBuilder_ != null) {
15119          return storageBuilder_.getMessageOrBuilder();
15120        } else {
15121          return storage_;
15122        }
15123      }
15124      /**
15125       * <code>required .hadoop.hdfs.StorageInfoProto storage = 2;</code>
15126       */
15127      private com.google.protobuf.SingleFieldBuilder<
15128          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder> 
15129          getStorageFieldBuilder() {
15130        if (storageBuilder_ == null) {
15131          storageBuilder_ = new com.google.protobuf.SingleFieldBuilder<
15132              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder>(
15133                  storage_,
15134                  getParentForChildren(),
15135                  isClean());
15136          storage_ = null;
15137        }
15138        return storageBuilder_;
15139      }
15140
15141      // required .hadoop.hdfs.StorageInfoProto prevStorage = 3;
15142      private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto prevStorage_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
15143      private com.google.protobuf.SingleFieldBuilder<
15144          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder> prevStorageBuilder_;
15145      /**
15146       * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
15147       */
15148      public boolean hasPrevStorage() {
15149        return ((bitField0_ & 0x00000004) == 0x00000004);
15150      }
15151      /**
15152       * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
15153       */
15154      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getPrevStorage() {
15155        if (prevStorageBuilder_ == null) {
15156          return prevStorage_;
15157        } else {
15158          return prevStorageBuilder_.getMessage();
15159        }
15160      }
15161      /**
15162       * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
15163       */
15164      public Builder setPrevStorage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto value) {
15165        if (prevStorageBuilder_ == null) {
15166          if (value == null) {
15167            throw new NullPointerException();
15168          }
15169          prevStorage_ = value;
15170          onChanged();
15171        } else {
15172          prevStorageBuilder_.setMessage(value);
15173        }
15174        bitField0_ |= 0x00000004;
15175        return this;
15176      }
15177      /**
15178       * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
15179       */
15180      public Builder setPrevStorage(
15181          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder builderForValue) {
15182        if (prevStorageBuilder_ == null) {
15183          prevStorage_ = builderForValue.build();
15184          onChanged();
15185        } else {
15186          prevStorageBuilder_.setMessage(builderForValue.build());
15187        }
15188        bitField0_ |= 0x00000004;
15189        return this;
15190      }
15191      /**
15192       * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
15193       */
15194      public Builder mergePrevStorage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto value) {
15195        if (prevStorageBuilder_ == null) {
15196          if (((bitField0_ & 0x00000004) == 0x00000004) &&
15197              prevStorage_ != org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance()) {
15198            prevStorage_ =
15199              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.newBuilder(prevStorage_).mergeFrom(value).buildPartial();
15200          } else {
15201            prevStorage_ = value;
15202          }
15203          onChanged();
15204        } else {
15205          prevStorageBuilder_.mergeFrom(value);
15206        }
15207        bitField0_ |= 0x00000004;
15208        return this;
15209      }
15210      /**
15211       * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
15212       */
15213      public Builder clearPrevStorage() {
15214        if (prevStorageBuilder_ == null) {
15215          prevStorage_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance();
15216          onChanged();
15217        } else {
15218          prevStorageBuilder_.clear();
15219        }
15220        bitField0_ = (bitField0_ & ~0x00000004);
15221        return this;
15222      }
15223      /**
15224       * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
15225       */
15226      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder getPrevStorageBuilder() {
15227        bitField0_ |= 0x00000004;
15228        onChanged();
15229        return getPrevStorageFieldBuilder().getBuilder();
15230      }
15231      /**
15232       * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
15233       */
15234      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getPrevStorageOrBuilder() {
15235        if (prevStorageBuilder_ != null) {
15236          return prevStorageBuilder_.getMessageOrBuilder();
15237        } else {
15238          return prevStorage_;
15239        }
15240      }
15241      /**
15242       * <code>required .hadoop.hdfs.StorageInfoProto prevStorage = 3;</code>
15243       */
15244      private com.google.protobuf.SingleFieldBuilder<
15245          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder> 
15246          getPrevStorageFieldBuilder() {
15247        if (prevStorageBuilder_ == null) {
15248          prevStorageBuilder_ = new com.google.protobuf.SingleFieldBuilder<
15249              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder>(
15250                  prevStorage_,
15251                  getParentForChildren(),
15252                  isClean());
15253          prevStorage_ = null;
15254        }
15255        return prevStorageBuilder_;
15256      }
15257
15258      // required int32 targetLayoutVersion = 4;
15259      private int targetLayoutVersion_ ;
15260      /**
15261       * <code>required int32 targetLayoutVersion = 4;</code>
15262       */
15263      public boolean hasTargetLayoutVersion() {
15264        return ((bitField0_ & 0x00000008) == 0x00000008);
15265      }
15266      /**
15267       * <code>required int32 targetLayoutVersion = 4;</code>
15268       */
15269      public int getTargetLayoutVersion() {
15270        return targetLayoutVersion_;
15271      }
15272      /**
15273       * <code>required int32 targetLayoutVersion = 4;</code>
15274       */
15275      public Builder setTargetLayoutVersion(int value) {
15276        bitField0_ |= 0x00000008;
15277        targetLayoutVersion_ = value;
15278        onChanged();
15279        return this;
15280      }
15281      /**
15282       * <code>required int32 targetLayoutVersion = 4;</code>
15283       */
15284      public Builder clearTargetLayoutVersion() {
15285        bitField0_ = (bitField0_ & ~0x00000008);
15286        targetLayoutVersion_ = 0;
15287        onChanged();
15288        return this;
15289      }
15290
15291      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.CanRollBackRequestProto)
15292    }
15293
15294    static {
15295      defaultInstance = new CanRollBackRequestProto(true);
15296      defaultInstance.initFields();
15297    }
15298
15299    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.CanRollBackRequestProto)
15300  }
15301
15302  public interface CanRollBackResponseProtoOrBuilder
15303      extends com.google.protobuf.MessageOrBuilder {
15304
15305    // required bool canRollBack = 1;
15306    /**
15307     * <code>required bool canRollBack = 1;</code>
15308     */
15309    boolean hasCanRollBack();
15310    /**
15311     * <code>required bool canRollBack = 1;</code>
15312     */
15313    boolean getCanRollBack();
15314  }
15315  /**
15316   * Protobuf type {@code hadoop.hdfs.qjournal.CanRollBackResponseProto}
15317   */
15318  public static final class CanRollBackResponseProto extends
15319      com.google.protobuf.GeneratedMessage
15320      implements CanRollBackResponseProtoOrBuilder {
15321    // Use CanRollBackResponseProto.newBuilder() to construct.
15322    private CanRollBackResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
15323      super(builder);
15324      this.unknownFields = builder.getUnknownFields();
15325    }
15326    private CanRollBackResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
15327
15328    private static final CanRollBackResponseProto defaultInstance;
15329    public static CanRollBackResponseProto getDefaultInstance() {
15330      return defaultInstance;
15331    }
15332
15333    public CanRollBackResponseProto getDefaultInstanceForType() {
15334      return defaultInstance;
15335    }
15336
15337    private final com.google.protobuf.UnknownFieldSet unknownFields;
15338    @java.lang.Override
15339    public final com.google.protobuf.UnknownFieldSet
15340        getUnknownFields() {
15341      return this.unknownFields;
15342    }
15343    private CanRollBackResponseProto(
15344        com.google.protobuf.CodedInputStream input,
15345        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15346        throws com.google.protobuf.InvalidProtocolBufferException {
15347      initFields();
15348      int mutable_bitField0_ = 0;
15349      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
15350          com.google.protobuf.UnknownFieldSet.newBuilder();
15351      try {
15352        boolean done = false;
15353        while (!done) {
15354          int tag = input.readTag();
15355          switch (tag) {
15356            case 0:
15357              done = true;
15358              break;
15359            default: {
15360              if (!parseUnknownField(input, unknownFields,
15361                                     extensionRegistry, tag)) {
15362                done = true;
15363              }
15364              break;
15365            }
15366            case 8: {
15367              bitField0_ |= 0x00000001;
15368              canRollBack_ = input.readBool();
15369              break;
15370            }
15371          }
15372        }
15373      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
15374        throw e.setUnfinishedMessage(this);
15375      } catch (java.io.IOException e) {
15376        throw new com.google.protobuf.InvalidProtocolBufferException(
15377            e.getMessage()).setUnfinishedMessage(this);
15378      } finally {
15379        this.unknownFields = unknownFields.build();
15380        makeExtensionsImmutable();
15381      }
15382    }
15383    public static final com.google.protobuf.Descriptors.Descriptor
15384        getDescriptor() {
15385      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_CanRollBackResponseProto_descriptor;
15386    }
15387
15388    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
15389        internalGetFieldAccessorTable() {
15390      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_CanRollBackResponseProto_fieldAccessorTable
15391          .ensureFieldAccessorsInitialized(
15392              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.Builder.class);
15393    }
15394
15395    public static com.google.protobuf.Parser<CanRollBackResponseProto> PARSER =
15396        new com.google.protobuf.AbstractParser<CanRollBackResponseProto>() {
15397      public CanRollBackResponseProto parsePartialFrom(
15398          com.google.protobuf.CodedInputStream input,
15399          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15400          throws com.google.protobuf.InvalidProtocolBufferException {
15401        return new CanRollBackResponseProto(input, extensionRegistry);
15402      }
15403    };
15404
15405    @java.lang.Override
15406    public com.google.protobuf.Parser<CanRollBackResponseProto> getParserForType() {
15407      return PARSER;
15408    }
15409
15410    private int bitField0_;
15411    // required bool canRollBack = 1;
15412    public static final int CANROLLBACK_FIELD_NUMBER = 1;
15413    private boolean canRollBack_;
15414    /**
15415     * <code>required bool canRollBack = 1;</code>
15416     */
15417    public boolean hasCanRollBack() {
15418      return ((bitField0_ & 0x00000001) == 0x00000001);
15419    }
15420    /**
15421     * <code>required bool canRollBack = 1;</code>
15422     */
15423    public boolean getCanRollBack() {
15424      return canRollBack_;
15425    }
15426
15427    private void initFields() {
15428      canRollBack_ = false;
15429    }
15430    private byte memoizedIsInitialized = -1;
15431    public final boolean isInitialized() {
15432      byte isInitialized = memoizedIsInitialized;
15433      if (isInitialized != -1) return isInitialized == 1;
15434
15435      if (!hasCanRollBack()) {
15436        memoizedIsInitialized = 0;
15437        return false;
15438      }
15439      memoizedIsInitialized = 1;
15440      return true;
15441    }
15442
15443    public void writeTo(com.google.protobuf.CodedOutputStream output)
15444                        throws java.io.IOException {
15445      getSerializedSize();
15446      if (((bitField0_ & 0x00000001) == 0x00000001)) {
15447        output.writeBool(1, canRollBack_);
15448      }
15449      getUnknownFields().writeTo(output);
15450    }
15451
15452    private int memoizedSerializedSize = -1;
15453    public int getSerializedSize() {
15454      int size = memoizedSerializedSize;
15455      if (size != -1) return size;
15456
15457      size = 0;
15458      if (((bitField0_ & 0x00000001) == 0x00000001)) {
15459        size += com.google.protobuf.CodedOutputStream
15460          .computeBoolSize(1, canRollBack_);
15461      }
15462      size += getUnknownFields().getSerializedSize();
15463      memoizedSerializedSize = size;
15464      return size;
15465    }
15466
15467    private static final long serialVersionUID = 0L;
15468    @java.lang.Override
15469    protected java.lang.Object writeReplace()
15470        throws java.io.ObjectStreamException {
15471      return super.writeReplace();
15472    }
15473
15474    @java.lang.Override
15475    public boolean equals(final java.lang.Object obj) {
15476      if (obj == this) {
15477       return true;
15478      }
15479      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto)) {
15480        return super.equals(obj);
15481      }
15482      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto) obj;
15483
15484      boolean result = true;
15485      result = result && (hasCanRollBack() == other.hasCanRollBack());
15486      if (hasCanRollBack()) {
15487        result = result && (getCanRollBack()
15488            == other.getCanRollBack());
15489      }
15490      result = result &&
15491          getUnknownFields().equals(other.getUnknownFields());
15492      return result;
15493    }
15494
15495    private int memoizedHashCode = 0;
15496    @java.lang.Override
15497    public int hashCode() {
15498      if (memoizedHashCode != 0) {
15499        return memoizedHashCode;
15500      }
15501      int hash = 41;
15502      hash = (19 * hash) + getDescriptorForType().hashCode();
15503      if (hasCanRollBack()) {
15504        hash = (37 * hash) + CANROLLBACK_FIELD_NUMBER;
15505        hash = (53 * hash) + hashBoolean(getCanRollBack());
15506      }
15507      hash = (29 * hash) + getUnknownFields().hashCode();
15508      memoizedHashCode = hash;
15509      return hash;
15510    }
15511
15512    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto parseFrom(
15513        com.google.protobuf.ByteString data)
15514        throws com.google.protobuf.InvalidProtocolBufferException {
15515      return PARSER.parseFrom(data);
15516    }
15517    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto parseFrom(
15518        com.google.protobuf.ByteString data,
15519        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15520        throws com.google.protobuf.InvalidProtocolBufferException {
15521      return PARSER.parseFrom(data, extensionRegistry);
15522    }
15523    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto parseFrom(byte[] data)
15524        throws com.google.protobuf.InvalidProtocolBufferException {
15525      return PARSER.parseFrom(data);
15526    }
15527    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto parseFrom(
15528        byte[] data,
15529        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15530        throws com.google.protobuf.InvalidProtocolBufferException {
15531      return PARSER.parseFrom(data, extensionRegistry);
15532    }
15533    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto parseFrom(java.io.InputStream input)
15534        throws java.io.IOException {
15535      return PARSER.parseFrom(input);
15536    }
15537    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto parseFrom(
15538        java.io.InputStream input,
15539        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15540        throws java.io.IOException {
15541      return PARSER.parseFrom(input, extensionRegistry);
15542    }
15543    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto parseDelimitedFrom(java.io.InputStream input)
15544        throws java.io.IOException {
15545      return PARSER.parseDelimitedFrom(input);
15546    }
15547    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto parseDelimitedFrom(
15548        java.io.InputStream input,
15549        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15550        throws java.io.IOException {
15551      return PARSER.parseDelimitedFrom(input, extensionRegistry);
15552    }
15553    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto parseFrom(
15554        com.google.protobuf.CodedInputStream input)
15555        throws java.io.IOException {
15556      return PARSER.parseFrom(input);
15557    }
15558    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto parseFrom(
15559        com.google.protobuf.CodedInputStream input,
15560        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15561        throws java.io.IOException {
15562      return PARSER.parseFrom(input, extensionRegistry);
15563    }
15564
15565    public static Builder newBuilder() { return Builder.create(); }
15566    public Builder newBuilderForType() { return newBuilder(); }
15567    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto prototype) {
15568      return newBuilder().mergeFrom(prototype);
15569    }
15570    public Builder toBuilder() { return newBuilder(this); }
15571
15572    @java.lang.Override
15573    protected Builder newBuilderForType(
15574        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
15575      Builder builder = new Builder(parent);
15576      return builder;
15577    }
15578    /**
15579     * Protobuf type {@code hadoop.hdfs.qjournal.CanRollBackResponseProto}
15580     */
15581    public static final class Builder extends
15582        com.google.protobuf.GeneratedMessage.Builder<Builder>
15583       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProtoOrBuilder {
15584      public static final com.google.protobuf.Descriptors.Descriptor
15585          getDescriptor() {
15586        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_CanRollBackResponseProto_descriptor;
15587      }
15588
15589      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
15590          internalGetFieldAccessorTable() {
15591        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_CanRollBackResponseProto_fieldAccessorTable
15592            .ensureFieldAccessorsInitialized(
15593                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.Builder.class);
15594      }
15595
15596      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.newBuilder()
15597      private Builder() {
15598        maybeForceBuilderInitialization();
15599      }
15600
15601      private Builder(
15602          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
15603        super(parent);
15604        maybeForceBuilderInitialization();
15605      }
15606      private void maybeForceBuilderInitialization() {
15607        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
15608        }
15609      }
15610      private static Builder create() {
15611        return new Builder();
15612      }
15613
15614      public Builder clear() {
15615        super.clear();
15616        canRollBack_ = false;
15617        bitField0_ = (bitField0_ & ~0x00000001);
15618        return this;
15619      }
15620
15621      public Builder clone() {
15622        return create().mergeFrom(buildPartial());
15623      }
15624
15625      public com.google.protobuf.Descriptors.Descriptor
15626          getDescriptorForType() {
15627        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_CanRollBackResponseProto_descriptor;
15628      }
15629
15630      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto getDefaultInstanceForType() {
15631        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.getDefaultInstance();
15632      }
15633
15634      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto build() {
15635        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto result = buildPartial();
15636        if (!result.isInitialized()) {
15637          throw newUninitializedMessageException(result);
15638        }
15639        return result;
15640      }
15641
15642      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto buildPartial() {
15643        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto(this);
15644        int from_bitField0_ = bitField0_;
15645        int to_bitField0_ = 0;
15646        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
15647          to_bitField0_ |= 0x00000001;
15648        }
15649        result.canRollBack_ = canRollBack_;
15650        result.bitField0_ = to_bitField0_;
15651        onBuilt();
15652        return result;
15653      }
15654
15655      public Builder mergeFrom(com.google.protobuf.Message other) {
15656        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto) {
15657          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto)other);
15658        } else {
15659          super.mergeFrom(other);
15660          return this;
15661        }
15662      }
15663
15664      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto other) {
15665        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.getDefaultInstance()) return this;
15666        if (other.hasCanRollBack()) {
15667          setCanRollBack(other.getCanRollBack());
15668        }
15669        this.mergeUnknownFields(other.getUnknownFields());
15670        return this;
15671      }
15672
15673      public final boolean isInitialized() {
15674        if (!hasCanRollBack()) {
15675          
15676          return false;
15677        }
15678        return true;
15679      }
15680
15681      public Builder mergeFrom(
15682          com.google.protobuf.CodedInputStream input,
15683          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15684          throws java.io.IOException {
15685        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto parsedMessage = null;
15686        try {
15687          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
15688        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
15689          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto) e.getUnfinishedMessage();
15690          throw e;
15691        } finally {
15692          if (parsedMessage != null) {
15693            mergeFrom(parsedMessage);
15694          }
15695        }
15696        return this;
15697      }
15698      private int bitField0_;
15699
15700      // required bool canRollBack = 1;
15701      private boolean canRollBack_ ;
15702      /**
15703       * <code>required bool canRollBack = 1;</code>
15704       */
15705      public boolean hasCanRollBack() {
15706        return ((bitField0_ & 0x00000001) == 0x00000001);
15707      }
15708      /**
15709       * <code>required bool canRollBack = 1;</code>
15710       */
15711      public boolean getCanRollBack() {
15712        return canRollBack_;
15713      }
15714      /**
15715       * <code>required bool canRollBack = 1;</code>
15716       */
15717      public Builder setCanRollBack(boolean value) {
15718        bitField0_ |= 0x00000001;
15719        canRollBack_ = value;
15720        onChanged();
15721        return this;
15722      }
15723      /**
15724       * <code>required bool canRollBack = 1;</code>
15725       */
15726      public Builder clearCanRollBack() {
15727        bitField0_ = (bitField0_ & ~0x00000001);
15728        canRollBack_ = false;
15729        onChanged();
15730        return this;
15731      }
15732
15733      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.CanRollBackResponseProto)
15734    }
15735
15736    static {
15737      defaultInstance = new CanRollBackResponseProto(true);
15738      defaultInstance.initFields();
15739    }
15740
15741    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.CanRollBackResponseProto)
15742  }
15743
15744  public interface DoRollbackRequestProtoOrBuilder
15745      extends com.google.protobuf.MessageOrBuilder {
15746
15747    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
15748    /**
15749     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
15750     */
15751    boolean hasJid();
15752    /**
15753     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
15754     */
15755    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
15756    /**
15757     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
15758     */
15759    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
15760  }
15761  /**
15762   * Protobuf type {@code hadoop.hdfs.qjournal.DoRollbackRequestProto}
15763   *
15764   * <pre>
15765   **
15766   * doRollback()
15767   * </pre>
15768   */
15769  public static final class DoRollbackRequestProto extends
15770      com.google.protobuf.GeneratedMessage
15771      implements DoRollbackRequestProtoOrBuilder {
15772    // Use DoRollbackRequestProto.newBuilder() to construct.
15773    private DoRollbackRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
15774      super(builder);
15775      this.unknownFields = builder.getUnknownFields();
15776    }
15777    private DoRollbackRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
15778
15779    private static final DoRollbackRequestProto defaultInstance;
15780    public static DoRollbackRequestProto getDefaultInstance() {
15781      return defaultInstance;
15782    }
15783
15784    public DoRollbackRequestProto getDefaultInstanceForType() {
15785      return defaultInstance;
15786    }
15787
15788    private final com.google.protobuf.UnknownFieldSet unknownFields;
15789    @java.lang.Override
15790    public final com.google.protobuf.UnknownFieldSet
15791        getUnknownFields() {
15792      return this.unknownFields;
15793    }
15794    private DoRollbackRequestProto(
15795        com.google.protobuf.CodedInputStream input,
15796        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15797        throws com.google.protobuf.InvalidProtocolBufferException {
15798      initFields();
15799      int mutable_bitField0_ = 0;
15800      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
15801          com.google.protobuf.UnknownFieldSet.newBuilder();
15802      try {
15803        boolean done = false;
15804        while (!done) {
15805          int tag = input.readTag();
15806          switch (tag) {
15807            case 0:
15808              done = true;
15809              break;
15810            default: {
15811              if (!parseUnknownField(input, unknownFields,
15812                                     extensionRegistry, tag)) {
15813                done = true;
15814              }
15815              break;
15816            }
15817            case 10: {
15818              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
15819              if (((bitField0_ & 0x00000001) == 0x00000001)) {
15820                subBuilder = jid_.toBuilder();
15821              }
15822              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
15823              if (subBuilder != null) {
15824                subBuilder.mergeFrom(jid_);
15825                jid_ = subBuilder.buildPartial();
15826              }
15827              bitField0_ |= 0x00000001;
15828              break;
15829            }
15830          }
15831        }
15832      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
15833        throw e.setUnfinishedMessage(this);
15834      } catch (java.io.IOException e) {
15835        throw new com.google.protobuf.InvalidProtocolBufferException(
15836            e.getMessage()).setUnfinishedMessage(this);
15837      } finally {
15838        this.unknownFields = unknownFields.build();
15839        makeExtensionsImmutable();
15840      }
15841    }
15842    public static final com.google.protobuf.Descriptors.Descriptor
15843        getDescriptor() {
15844      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoRollbackRequestProto_descriptor;
15845    }
15846
15847    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
15848        internalGetFieldAccessorTable() {
15849      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoRollbackRequestProto_fieldAccessorTable
15850          .ensureFieldAccessorsInitialized(
15851              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto.Builder.class);
15852    }
15853
15854    public static com.google.protobuf.Parser<DoRollbackRequestProto> PARSER =
15855        new com.google.protobuf.AbstractParser<DoRollbackRequestProto>() {
15856      public DoRollbackRequestProto parsePartialFrom(
15857          com.google.protobuf.CodedInputStream input,
15858          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15859          throws com.google.protobuf.InvalidProtocolBufferException {
15860        return new DoRollbackRequestProto(input, extensionRegistry);
15861      }
15862    };
15863
15864    @java.lang.Override
15865    public com.google.protobuf.Parser<DoRollbackRequestProto> getParserForType() {
15866      return PARSER;
15867    }
15868
15869    private int bitField0_;
15870    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
15871    public static final int JID_FIELD_NUMBER = 1;
15872    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
15873    /**
15874     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
15875     */
15876    public boolean hasJid() {
15877      return ((bitField0_ & 0x00000001) == 0x00000001);
15878    }
15879    /**
15880     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
15881     */
15882    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
15883      return jid_;
15884    }
15885    /**
15886     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
15887     */
15888    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
15889      return jid_;
15890    }
15891
15892    private void initFields() {
15893      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
15894    }
15895    private byte memoizedIsInitialized = -1;
15896    public final boolean isInitialized() {
15897      byte isInitialized = memoizedIsInitialized;
15898      if (isInitialized != -1) return isInitialized == 1;
15899
15900      if (!hasJid()) {
15901        memoizedIsInitialized = 0;
15902        return false;
15903      }
15904      if (!getJid().isInitialized()) {
15905        memoizedIsInitialized = 0;
15906        return false;
15907      }
15908      memoizedIsInitialized = 1;
15909      return true;
15910    }
15911
15912    public void writeTo(com.google.protobuf.CodedOutputStream output)
15913                        throws java.io.IOException {
15914      getSerializedSize();
15915      if (((bitField0_ & 0x00000001) == 0x00000001)) {
15916        output.writeMessage(1, jid_);
15917      }
15918      getUnknownFields().writeTo(output);
15919    }
15920
15921    private int memoizedSerializedSize = -1;
15922    public int getSerializedSize() {
15923      int size = memoizedSerializedSize;
15924      if (size != -1) return size;
15925
15926      size = 0;
15927      if (((bitField0_ & 0x00000001) == 0x00000001)) {
15928        size += com.google.protobuf.CodedOutputStream
15929          .computeMessageSize(1, jid_);
15930      }
15931      size += getUnknownFields().getSerializedSize();
15932      memoizedSerializedSize = size;
15933      return size;
15934    }
15935
15936    private static final long serialVersionUID = 0L;
15937    @java.lang.Override
15938    protected java.lang.Object writeReplace()
15939        throws java.io.ObjectStreamException {
15940      return super.writeReplace();
15941    }
15942
15943    @java.lang.Override
15944    public boolean equals(final java.lang.Object obj) {
15945      if (obj == this) {
15946       return true;
15947      }
15948      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto)) {
15949        return super.equals(obj);
15950      }
15951      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto) obj;
15952
15953      boolean result = true;
15954      result = result && (hasJid() == other.hasJid());
15955      if (hasJid()) {
15956        result = result && getJid()
15957            .equals(other.getJid());
15958      }
15959      result = result &&
15960          getUnknownFields().equals(other.getUnknownFields());
15961      return result;
15962    }
15963
15964    private int memoizedHashCode = 0;
15965    @java.lang.Override
15966    public int hashCode() {
15967      if (memoizedHashCode != 0) {
15968        return memoizedHashCode;
15969      }
15970      int hash = 41;
15971      hash = (19 * hash) + getDescriptorForType().hashCode();
15972      if (hasJid()) {
15973        hash = (37 * hash) + JID_FIELD_NUMBER;
15974        hash = (53 * hash) + getJid().hashCode();
15975      }
15976      hash = (29 * hash) + getUnknownFields().hashCode();
15977      memoizedHashCode = hash;
15978      return hash;
15979    }
15980
15981    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto parseFrom(
15982        com.google.protobuf.ByteString data)
15983        throws com.google.protobuf.InvalidProtocolBufferException {
15984      return PARSER.parseFrom(data);
15985    }
15986    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto parseFrom(
15987        com.google.protobuf.ByteString data,
15988        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15989        throws com.google.protobuf.InvalidProtocolBufferException {
15990      return PARSER.parseFrom(data, extensionRegistry);
15991    }
15992    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto parseFrom(byte[] data)
15993        throws com.google.protobuf.InvalidProtocolBufferException {
15994      return PARSER.parseFrom(data);
15995    }
15996    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto parseFrom(
15997        byte[] data,
15998        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15999        throws com.google.protobuf.InvalidProtocolBufferException {
16000      return PARSER.parseFrom(data, extensionRegistry);
16001    }
16002    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto parseFrom(java.io.InputStream input)
16003        throws java.io.IOException {
16004      return PARSER.parseFrom(input);
16005    }
16006    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto parseFrom(
16007        java.io.InputStream input,
16008        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16009        throws java.io.IOException {
16010      return PARSER.parseFrom(input, extensionRegistry);
16011    }
16012    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto parseDelimitedFrom(java.io.InputStream input)
16013        throws java.io.IOException {
16014      return PARSER.parseDelimitedFrom(input);
16015    }
16016    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto parseDelimitedFrom(
16017        java.io.InputStream input,
16018        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16019        throws java.io.IOException {
16020      return PARSER.parseDelimitedFrom(input, extensionRegistry);
16021    }
16022    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto parseFrom(
16023        com.google.protobuf.CodedInputStream input)
16024        throws java.io.IOException {
16025      return PARSER.parseFrom(input);
16026    }
16027    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto parseFrom(
16028        com.google.protobuf.CodedInputStream input,
16029        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16030        throws java.io.IOException {
16031      return PARSER.parseFrom(input, extensionRegistry);
16032    }
16033
16034    public static Builder newBuilder() { return Builder.create(); }
16035    public Builder newBuilderForType() { return newBuilder(); }
16036    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto prototype) {
16037      return newBuilder().mergeFrom(prototype);
16038    }
16039    public Builder toBuilder() { return newBuilder(this); }
16040
16041    @java.lang.Override
16042    protected Builder newBuilderForType(
16043        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
16044      Builder builder = new Builder(parent);
16045      return builder;
16046    }
16047    /**
16048     * Protobuf type {@code hadoop.hdfs.qjournal.DoRollbackRequestProto}
16049     *
16050     * <pre>
16051     **
16052     * doRollback()
16053     * </pre>
16054     */
16055    public static final class Builder extends
16056        com.google.protobuf.GeneratedMessage.Builder<Builder>
16057       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProtoOrBuilder {
16058      public static final com.google.protobuf.Descriptors.Descriptor
16059          getDescriptor() {
16060        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoRollbackRequestProto_descriptor;
16061      }
16062
16063      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
16064          internalGetFieldAccessorTable() {
16065        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoRollbackRequestProto_fieldAccessorTable
16066            .ensureFieldAccessorsInitialized(
16067                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto.Builder.class);
16068      }
16069
16070      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto.newBuilder()
16071      private Builder() {
16072        maybeForceBuilderInitialization();
16073      }
16074
16075      private Builder(
16076          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
16077        super(parent);
16078        maybeForceBuilderInitialization();
16079      }
16080      private void maybeForceBuilderInitialization() {
16081        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
16082          getJidFieldBuilder();
16083        }
16084      }
16085      private static Builder create() {
16086        return new Builder();
16087      }
16088
16089      public Builder clear() {
16090        super.clear();
16091        if (jidBuilder_ == null) {
16092          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
16093        } else {
16094          jidBuilder_.clear();
16095        }
16096        bitField0_ = (bitField0_ & ~0x00000001);
16097        return this;
16098      }
16099
16100      public Builder clone() {
16101        return create().mergeFrom(buildPartial());
16102      }
16103
16104      public com.google.protobuf.Descriptors.Descriptor
16105          getDescriptorForType() {
16106        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoRollbackRequestProto_descriptor;
16107      }
16108
16109      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto getDefaultInstanceForType() {
16110        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto.getDefaultInstance();
16111      }
16112
16113      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto build() {
16114        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto result = buildPartial();
16115        if (!result.isInitialized()) {
16116          throw newUninitializedMessageException(result);
16117        }
16118        return result;
16119      }
16120
16121      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto buildPartial() {
16122        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto(this);
16123        int from_bitField0_ = bitField0_;
16124        int to_bitField0_ = 0;
16125        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
16126          to_bitField0_ |= 0x00000001;
16127        }
16128        if (jidBuilder_ == null) {
16129          result.jid_ = jid_;
16130        } else {
16131          result.jid_ = jidBuilder_.build();
16132        }
16133        result.bitField0_ = to_bitField0_;
16134        onBuilt();
16135        return result;
16136      }
16137
16138      public Builder mergeFrom(com.google.protobuf.Message other) {
16139        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto) {
16140          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto)other);
16141        } else {
16142          super.mergeFrom(other);
16143          return this;
16144        }
16145      }
16146
16147      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto other) {
16148        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto.getDefaultInstance()) return this;
16149        if (other.hasJid()) {
16150          mergeJid(other.getJid());
16151        }
16152        this.mergeUnknownFields(other.getUnknownFields());
16153        return this;
16154      }
16155
16156      public final boolean isInitialized() {
16157        if (!hasJid()) {
16158          
16159          return false;
16160        }
16161        if (!getJid().isInitialized()) {
16162          
16163          return false;
16164        }
16165        return true;
16166      }
16167
16168      public Builder mergeFrom(
16169          com.google.protobuf.CodedInputStream input,
16170          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16171          throws java.io.IOException {
16172        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto parsedMessage = null;
16173        try {
16174          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
16175        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
16176          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto) e.getUnfinishedMessage();
16177          throw e;
16178        } finally {
16179          if (parsedMessage != null) {
16180            mergeFrom(parsedMessage);
16181          }
16182        }
16183        return this;
16184      }
16185      private int bitField0_;
16186
16187      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
16188      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
16189      private com.google.protobuf.SingleFieldBuilder<
16190          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
16191      /**
16192       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16193       */
16194      public boolean hasJid() {
16195        return ((bitField0_ & 0x00000001) == 0x00000001);
16196      }
16197      /**
16198       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16199       */
16200      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
16201        if (jidBuilder_ == null) {
16202          return jid_;
16203        } else {
16204          return jidBuilder_.getMessage();
16205        }
16206      }
16207      /**
16208       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16209       */
16210      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
16211        if (jidBuilder_ == null) {
16212          if (value == null) {
16213            throw new NullPointerException();
16214          }
16215          jid_ = value;
16216          onChanged();
16217        } else {
16218          jidBuilder_.setMessage(value);
16219        }
16220        bitField0_ |= 0x00000001;
16221        return this;
16222      }
16223      /**
16224       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16225       */
16226      public Builder setJid(
16227          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
16228        if (jidBuilder_ == null) {
16229          jid_ = builderForValue.build();
16230          onChanged();
16231        } else {
16232          jidBuilder_.setMessage(builderForValue.build());
16233        }
16234        bitField0_ |= 0x00000001;
16235        return this;
16236      }
16237      /**
16238       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16239       */
16240      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
16241        if (jidBuilder_ == null) {
16242          if (((bitField0_ & 0x00000001) == 0x00000001) &&
16243              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
16244            jid_ =
16245              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
16246          } else {
16247            jid_ = value;
16248          }
16249          onChanged();
16250        } else {
16251          jidBuilder_.mergeFrom(value);
16252        }
16253        bitField0_ |= 0x00000001;
16254        return this;
16255      }
16256      /**
16257       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16258       */
16259      public Builder clearJid() {
16260        if (jidBuilder_ == null) {
16261          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
16262          onChanged();
16263        } else {
16264          jidBuilder_.clear();
16265        }
16266        bitField0_ = (bitField0_ & ~0x00000001);
16267        return this;
16268      }
16269      /**
16270       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16271       */
16272      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
16273        bitField0_ |= 0x00000001;
16274        onChanged();
16275        return getJidFieldBuilder().getBuilder();
16276      }
16277      /**
16278       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16279       */
16280      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
16281        if (jidBuilder_ != null) {
16282          return jidBuilder_.getMessageOrBuilder();
16283        } else {
16284          return jid_;
16285        }
16286      }
16287      /**
16288       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16289       */
16290      private com.google.protobuf.SingleFieldBuilder<
16291          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
16292          getJidFieldBuilder() {
16293        if (jidBuilder_ == null) {
16294          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
16295              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
16296                  jid_,
16297                  getParentForChildren(),
16298                  isClean());
16299          jid_ = null;
16300        }
16301        return jidBuilder_;
16302      }
16303
16304      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.DoRollbackRequestProto)
16305    }
16306
16307    static {
16308      defaultInstance = new DoRollbackRequestProto(true);
16309      defaultInstance.initFields();
16310    }
16311
16312    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.DoRollbackRequestProto)
16313  }
16314
16315  public interface DoRollbackResponseProtoOrBuilder
16316      extends com.google.protobuf.MessageOrBuilder {
16317  }
16318  /**
16319   * Protobuf type {@code hadoop.hdfs.qjournal.DoRollbackResponseProto}
16320   */
16321  public static final class DoRollbackResponseProto extends
16322      com.google.protobuf.GeneratedMessage
16323      implements DoRollbackResponseProtoOrBuilder {
16324    // Use DoRollbackResponseProto.newBuilder() to construct.
16325    private DoRollbackResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
16326      super(builder);
16327      this.unknownFields = builder.getUnknownFields();
16328    }
16329    private DoRollbackResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
16330
16331    private static final DoRollbackResponseProto defaultInstance;
16332    public static DoRollbackResponseProto getDefaultInstance() {
16333      return defaultInstance;
16334    }
16335
16336    public DoRollbackResponseProto getDefaultInstanceForType() {
16337      return defaultInstance;
16338    }
16339
16340    private final com.google.protobuf.UnknownFieldSet unknownFields;
16341    @java.lang.Override
16342    public final com.google.protobuf.UnknownFieldSet
16343        getUnknownFields() {
16344      return this.unknownFields;
16345    }
16346    private DoRollbackResponseProto(
16347        com.google.protobuf.CodedInputStream input,
16348        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16349        throws com.google.protobuf.InvalidProtocolBufferException {
16350      initFields();
16351      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
16352          com.google.protobuf.UnknownFieldSet.newBuilder();
16353      try {
16354        boolean done = false;
16355        while (!done) {
16356          int tag = input.readTag();
16357          switch (tag) {
16358            case 0:
16359              done = true;
16360              break;
16361            default: {
16362              if (!parseUnknownField(input, unknownFields,
16363                                     extensionRegistry, tag)) {
16364                done = true;
16365              }
16366              break;
16367            }
16368          }
16369        }
16370      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
16371        throw e.setUnfinishedMessage(this);
16372      } catch (java.io.IOException e) {
16373        throw new com.google.protobuf.InvalidProtocolBufferException(
16374            e.getMessage()).setUnfinishedMessage(this);
16375      } finally {
16376        this.unknownFields = unknownFields.build();
16377        makeExtensionsImmutable();
16378      }
16379    }
16380    public static final com.google.protobuf.Descriptors.Descriptor
16381        getDescriptor() {
16382      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoRollbackResponseProto_descriptor;
16383    }
16384
16385    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
16386        internalGetFieldAccessorTable() {
16387      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoRollbackResponseProto_fieldAccessorTable
16388          .ensureFieldAccessorsInitialized(
16389              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.Builder.class);
16390    }
16391
16392    public static com.google.protobuf.Parser<DoRollbackResponseProto> PARSER =
16393        new com.google.protobuf.AbstractParser<DoRollbackResponseProto>() {
16394      public DoRollbackResponseProto parsePartialFrom(
16395          com.google.protobuf.CodedInputStream input,
16396          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16397          throws com.google.protobuf.InvalidProtocolBufferException {
16398        return new DoRollbackResponseProto(input, extensionRegistry);
16399      }
16400    };
16401
16402    @java.lang.Override
16403    public com.google.protobuf.Parser<DoRollbackResponseProto> getParserForType() {
16404      return PARSER;
16405    }
16406
16407    private void initFields() {
16408    }
16409    private byte memoizedIsInitialized = -1;
16410    public final boolean isInitialized() {
16411      byte isInitialized = memoizedIsInitialized;
16412      if (isInitialized != -1) return isInitialized == 1;
16413
16414      memoizedIsInitialized = 1;
16415      return true;
16416    }
16417
16418    public void writeTo(com.google.protobuf.CodedOutputStream output)
16419                        throws java.io.IOException {
16420      getSerializedSize();
16421      getUnknownFields().writeTo(output);
16422    }
16423
16424    private int memoizedSerializedSize = -1;
16425    public int getSerializedSize() {
16426      int size = memoizedSerializedSize;
16427      if (size != -1) return size;
16428
16429      size = 0;
16430      size += getUnknownFields().getSerializedSize();
16431      memoizedSerializedSize = size;
16432      return size;
16433    }
16434
16435    private static final long serialVersionUID = 0L;
16436    @java.lang.Override
16437    protected java.lang.Object writeReplace()
16438        throws java.io.ObjectStreamException {
16439      return super.writeReplace();
16440    }
16441
16442    @java.lang.Override
16443    public boolean equals(final java.lang.Object obj) {
16444      if (obj == this) {
16445       return true;
16446      }
16447      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto)) {
16448        return super.equals(obj);
16449      }
16450      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto) obj;
16451
16452      boolean result = true;
16453      result = result &&
16454          getUnknownFields().equals(other.getUnknownFields());
16455      return result;
16456    }
16457
16458    private int memoizedHashCode = 0;
16459    @java.lang.Override
16460    public int hashCode() {
16461      if (memoizedHashCode != 0) {
16462        return memoizedHashCode;
16463      }
16464      int hash = 41;
16465      hash = (19 * hash) + getDescriptorForType().hashCode();
16466      hash = (29 * hash) + getUnknownFields().hashCode();
16467      memoizedHashCode = hash;
16468      return hash;
16469    }
16470
16471    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto parseFrom(
16472        com.google.protobuf.ByteString data)
16473        throws com.google.protobuf.InvalidProtocolBufferException {
16474      return PARSER.parseFrom(data);
16475    }
16476    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto parseFrom(
16477        com.google.protobuf.ByteString data,
16478        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16479        throws com.google.protobuf.InvalidProtocolBufferException {
16480      return PARSER.parseFrom(data, extensionRegistry);
16481    }
16482    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto parseFrom(byte[] data)
16483        throws com.google.protobuf.InvalidProtocolBufferException {
16484      return PARSER.parseFrom(data);
16485    }
16486    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto parseFrom(
16487        byte[] data,
16488        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16489        throws com.google.protobuf.InvalidProtocolBufferException {
16490      return PARSER.parseFrom(data, extensionRegistry);
16491    }
16492    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto parseFrom(java.io.InputStream input)
16493        throws java.io.IOException {
16494      return PARSER.parseFrom(input);
16495    }
16496    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto parseFrom(
16497        java.io.InputStream input,
16498        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16499        throws java.io.IOException {
16500      return PARSER.parseFrom(input, extensionRegistry);
16501    }
16502    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto parseDelimitedFrom(java.io.InputStream input)
16503        throws java.io.IOException {
16504      return PARSER.parseDelimitedFrom(input);
16505    }
16506    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto parseDelimitedFrom(
16507        java.io.InputStream input,
16508        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16509        throws java.io.IOException {
16510      return PARSER.parseDelimitedFrom(input, extensionRegistry);
16511    }
16512    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto parseFrom(
16513        com.google.protobuf.CodedInputStream input)
16514        throws java.io.IOException {
16515      return PARSER.parseFrom(input);
16516    }
16517    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto parseFrom(
16518        com.google.protobuf.CodedInputStream input,
16519        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16520        throws java.io.IOException {
16521      return PARSER.parseFrom(input, extensionRegistry);
16522    }
16523
16524    public static Builder newBuilder() { return Builder.create(); }
16525    public Builder newBuilderForType() { return newBuilder(); }
16526    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto prototype) {
16527      return newBuilder().mergeFrom(prototype);
16528    }
16529    public Builder toBuilder() { return newBuilder(this); }
16530
16531    @java.lang.Override
16532    protected Builder newBuilderForType(
16533        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
16534      Builder builder = new Builder(parent);
16535      return builder;
16536    }
16537    /**
16538     * Protobuf type {@code hadoop.hdfs.qjournal.DoRollbackResponseProto}
16539     */
16540    public static final class Builder extends
16541        com.google.protobuf.GeneratedMessage.Builder<Builder>
16542       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProtoOrBuilder {
16543      public static final com.google.protobuf.Descriptors.Descriptor
16544          getDescriptor() {
16545        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoRollbackResponseProto_descriptor;
16546      }
16547
16548      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
16549          internalGetFieldAccessorTable() {
16550        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoRollbackResponseProto_fieldAccessorTable
16551            .ensureFieldAccessorsInitialized(
16552                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.Builder.class);
16553      }
16554
16555      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.newBuilder()
16556      private Builder() {
16557        maybeForceBuilderInitialization();
16558      }
16559
16560      private Builder(
16561          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
16562        super(parent);
16563        maybeForceBuilderInitialization();
16564      }
16565      private void maybeForceBuilderInitialization() {
16566        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
16567        }
16568      }
16569      private static Builder create() {
16570        return new Builder();
16571      }
16572
16573      public Builder clear() {
16574        super.clear();
16575        return this;
16576      }
16577
16578      public Builder clone() {
16579        return create().mergeFrom(buildPartial());
16580      }
16581
16582      public com.google.protobuf.Descriptors.Descriptor
16583          getDescriptorForType() {
16584        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_DoRollbackResponseProto_descriptor;
16585      }
16586
16587      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto getDefaultInstanceForType() {
16588        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.getDefaultInstance();
16589      }
16590
16591      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto build() {
16592        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto result = buildPartial();
16593        if (!result.isInitialized()) {
16594          throw newUninitializedMessageException(result);
16595        }
16596        return result;
16597      }
16598
16599      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto buildPartial() {
16600        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto(this);
16601        onBuilt();
16602        return result;
16603      }
16604
16605      public Builder mergeFrom(com.google.protobuf.Message other) {
16606        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto) {
16607          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto)other);
16608        } else {
16609          super.mergeFrom(other);
16610          return this;
16611        }
16612      }
16613
16614      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto other) {
16615        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.getDefaultInstance()) return this;
16616        this.mergeUnknownFields(other.getUnknownFields());
16617        return this;
16618      }
16619
16620      public final boolean isInitialized() {
16621        return true;
16622      }
16623
16624      public Builder mergeFrom(
16625          com.google.protobuf.CodedInputStream input,
16626          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16627          throws java.io.IOException {
16628        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto parsedMessage = null;
16629        try {
16630          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
16631        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
16632          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto) e.getUnfinishedMessage();
16633          throw e;
16634        } finally {
16635          if (parsedMessage != null) {
16636            mergeFrom(parsedMessage);
16637          }
16638        }
16639        return this;
16640      }
16641
16642      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.DoRollbackResponseProto)
16643    }
16644
16645    static {
16646      defaultInstance = new DoRollbackResponseProto(true);
16647      defaultInstance.initFields();
16648    }
16649
16650    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.DoRollbackResponseProto)
16651  }
16652
16653  public interface GetJournalStateRequestProtoOrBuilder
16654      extends com.google.protobuf.MessageOrBuilder {
16655
16656    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
16657    /**
16658     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16659     */
16660    boolean hasJid();
16661    /**
16662     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16663     */
16664    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
16665    /**
16666     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16667     */
16668    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
16669  }
16670  /**
16671   * Protobuf type {@code hadoop.hdfs.qjournal.GetJournalStateRequestProto}
16672   *
16673   * <pre>
16674   **
16675   * getJournalState()
16676   * </pre>
16677   */
16678  public static final class GetJournalStateRequestProto extends
16679      com.google.protobuf.GeneratedMessage
16680      implements GetJournalStateRequestProtoOrBuilder {
16681    // Use GetJournalStateRequestProto.newBuilder() to construct.
16682    private GetJournalStateRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
16683      super(builder);
16684      this.unknownFields = builder.getUnknownFields();
16685    }
16686    private GetJournalStateRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
16687
16688    private static final GetJournalStateRequestProto defaultInstance;
16689    public static GetJournalStateRequestProto getDefaultInstance() {
16690      return defaultInstance;
16691    }
16692
16693    public GetJournalStateRequestProto getDefaultInstanceForType() {
16694      return defaultInstance;
16695    }
16696
16697    private final com.google.protobuf.UnknownFieldSet unknownFields;
16698    @java.lang.Override
16699    public final com.google.protobuf.UnknownFieldSet
16700        getUnknownFields() {
16701      return this.unknownFields;
16702    }
16703    private GetJournalStateRequestProto(
16704        com.google.protobuf.CodedInputStream input,
16705        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16706        throws com.google.protobuf.InvalidProtocolBufferException {
16707      initFields();
16708      int mutable_bitField0_ = 0;
16709      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
16710          com.google.protobuf.UnknownFieldSet.newBuilder();
16711      try {
16712        boolean done = false;
16713        while (!done) {
16714          int tag = input.readTag();
16715          switch (tag) {
16716            case 0:
16717              done = true;
16718              break;
16719            default: {
16720              if (!parseUnknownField(input, unknownFields,
16721                                     extensionRegistry, tag)) {
16722                done = true;
16723              }
16724              break;
16725            }
16726            case 10: {
16727              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
16728              if (((bitField0_ & 0x00000001) == 0x00000001)) {
16729                subBuilder = jid_.toBuilder();
16730              }
16731              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
16732              if (subBuilder != null) {
16733                subBuilder.mergeFrom(jid_);
16734                jid_ = subBuilder.buildPartial();
16735              }
16736              bitField0_ |= 0x00000001;
16737              break;
16738            }
16739          }
16740        }
16741      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
16742        throw e.setUnfinishedMessage(this);
16743      } catch (java.io.IOException e) {
16744        throw new com.google.protobuf.InvalidProtocolBufferException(
16745            e.getMessage()).setUnfinishedMessage(this);
16746      } finally {
16747        this.unknownFields = unknownFields.build();
16748        makeExtensionsImmutable();
16749      }
16750    }
16751    public static final com.google.protobuf.Descriptors.Descriptor
16752        getDescriptor() {
16753      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalStateRequestProto_descriptor;
16754    }
16755
16756    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
16757        internalGetFieldAccessorTable() {
16758      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalStateRequestProto_fieldAccessorTable
16759          .ensureFieldAccessorsInitialized(
16760              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.Builder.class);
16761    }
16762
16763    public static com.google.protobuf.Parser<GetJournalStateRequestProto> PARSER =
16764        new com.google.protobuf.AbstractParser<GetJournalStateRequestProto>() {
16765      public GetJournalStateRequestProto parsePartialFrom(
16766          com.google.protobuf.CodedInputStream input,
16767          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16768          throws com.google.protobuf.InvalidProtocolBufferException {
16769        return new GetJournalStateRequestProto(input, extensionRegistry);
16770      }
16771    };
16772
16773    @java.lang.Override
16774    public com.google.protobuf.Parser<GetJournalStateRequestProto> getParserForType() {
16775      return PARSER;
16776    }
16777
16778    private int bitField0_;
16779    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
16780    public static final int JID_FIELD_NUMBER = 1;
16781    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
16782    /**
16783     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16784     */
16785    public boolean hasJid() {
16786      return ((bitField0_ & 0x00000001) == 0x00000001);
16787    }
16788    /**
16789     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16790     */
16791    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
16792      return jid_;
16793    }
16794    /**
16795     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
16796     */
16797    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
16798      return jid_;
16799    }
16800
16801    private void initFields() {
16802      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
16803    }
16804    private byte memoizedIsInitialized = -1;
16805    public final boolean isInitialized() {
16806      byte isInitialized = memoizedIsInitialized;
16807      if (isInitialized != -1) return isInitialized == 1;
16808
16809      if (!hasJid()) {
16810        memoizedIsInitialized = 0;
16811        return false;
16812      }
16813      if (!getJid().isInitialized()) {
16814        memoizedIsInitialized = 0;
16815        return false;
16816      }
16817      memoizedIsInitialized = 1;
16818      return true;
16819    }
16820
16821    public void writeTo(com.google.protobuf.CodedOutputStream output)
16822                        throws java.io.IOException {
16823      getSerializedSize();
16824      if (((bitField0_ & 0x00000001) == 0x00000001)) {
16825        output.writeMessage(1, jid_);
16826      }
16827      getUnknownFields().writeTo(output);
16828    }
16829
16830    private int memoizedSerializedSize = -1;
16831    public int getSerializedSize() {
16832      int size = memoizedSerializedSize;
16833      if (size != -1) return size;
16834
16835      size = 0;
16836      if (((bitField0_ & 0x00000001) == 0x00000001)) {
16837        size += com.google.protobuf.CodedOutputStream
16838          .computeMessageSize(1, jid_);
16839      }
16840      size += getUnknownFields().getSerializedSize();
16841      memoizedSerializedSize = size;
16842      return size;
16843    }
16844
16845    private static final long serialVersionUID = 0L;
16846    @java.lang.Override
16847    protected java.lang.Object writeReplace()
16848        throws java.io.ObjectStreamException {
16849      return super.writeReplace();
16850    }
16851
16852    @java.lang.Override
16853    public boolean equals(final java.lang.Object obj) {
16854      if (obj == this) {
16855       return true;
16856      }
16857      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)) {
16858        return super.equals(obj);
16859      }
16860      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto) obj;
16861
16862      boolean result = true;
16863      result = result && (hasJid() == other.hasJid());
16864      if (hasJid()) {
16865        result = result && getJid()
16866            .equals(other.getJid());
16867      }
16868      result = result &&
16869          getUnknownFields().equals(other.getUnknownFields());
16870      return result;
16871    }
16872
16873    private int memoizedHashCode = 0;
16874    @java.lang.Override
16875    public int hashCode() {
16876      if (memoizedHashCode != 0) {
16877        return memoizedHashCode;
16878      }
16879      int hash = 41;
16880      hash = (19 * hash) + getDescriptorForType().hashCode();
16881      if (hasJid()) {
16882        hash = (37 * hash) + JID_FIELD_NUMBER;
16883        hash = (53 * hash) + getJid().hashCode();
16884      }
16885      hash = (29 * hash) + getUnknownFields().hashCode();
16886      memoizedHashCode = hash;
16887      return hash;
16888    }
16889
16890    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
16891        com.google.protobuf.ByteString data)
16892        throws com.google.protobuf.InvalidProtocolBufferException {
16893      return PARSER.parseFrom(data);
16894    }
16895    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
16896        com.google.protobuf.ByteString data,
16897        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16898        throws com.google.protobuf.InvalidProtocolBufferException {
16899      return PARSER.parseFrom(data, extensionRegistry);
16900    }
16901    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(byte[] data)
16902        throws com.google.protobuf.InvalidProtocolBufferException {
16903      return PARSER.parseFrom(data);
16904    }
16905    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
16906        byte[] data,
16907        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16908        throws com.google.protobuf.InvalidProtocolBufferException {
16909      return PARSER.parseFrom(data, extensionRegistry);
16910    }
16911    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(java.io.InputStream input)
16912        throws java.io.IOException {
16913      return PARSER.parseFrom(input);
16914    }
16915    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
16916        java.io.InputStream input,
16917        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16918        throws java.io.IOException {
16919      return PARSER.parseFrom(input, extensionRegistry);
16920    }
16921    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseDelimitedFrom(java.io.InputStream input)
16922        throws java.io.IOException {
16923      return PARSER.parseDelimitedFrom(input);
16924    }
16925    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseDelimitedFrom(
16926        java.io.InputStream input,
16927        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16928        throws java.io.IOException {
16929      return PARSER.parseDelimitedFrom(input, extensionRegistry);
16930    }
16931    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
16932        com.google.protobuf.CodedInputStream input)
16933        throws java.io.IOException {
16934      return PARSER.parseFrom(input);
16935    }
16936    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
16937        com.google.protobuf.CodedInputStream input,
16938        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16939        throws java.io.IOException {
16940      return PARSER.parseFrom(input, extensionRegistry);
16941    }
16942
16943    public static Builder newBuilder() { return Builder.create(); }
16944    public Builder newBuilderForType() { return newBuilder(); }
16945    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto prototype) {
16946      return newBuilder().mergeFrom(prototype);
16947    }
16948    public Builder toBuilder() { return newBuilder(this); }
16949
16950    @java.lang.Override
16951    protected Builder newBuilderForType(
16952        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
16953      Builder builder = new Builder(parent);
16954      return builder;
16955    }
16956    /**
16957     * Protobuf type {@code hadoop.hdfs.qjournal.GetJournalStateRequestProto}
16958     *
16959     * <pre>
16960     **
16961     * getJournalState()
16962     * </pre>
16963     */
16964    public static final class Builder extends
16965        com.google.protobuf.GeneratedMessage.Builder<Builder>
16966       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProtoOrBuilder {
16967      public static final com.google.protobuf.Descriptors.Descriptor
16968          getDescriptor() {
16969        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalStateRequestProto_descriptor;
16970      }
16971
16972      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
16973          internalGetFieldAccessorTable() {
16974        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalStateRequestProto_fieldAccessorTable
16975            .ensureFieldAccessorsInitialized(
16976                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.Builder.class);
16977      }
16978
16979      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.newBuilder()
16980      private Builder() {
16981        maybeForceBuilderInitialization();
16982      }
16983
16984      private Builder(
16985          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
16986        super(parent);
16987        maybeForceBuilderInitialization();
16988      }
16989      private void maybeForceBuilderInitialization() {
16990        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
16991          getJidFieldBuilder();
16992        }
16993      }
16994      private static Builder create() {
16995        return new Builder();
16996      }
16997
16998      public Builder clear() {
16999        super.clear();
17000        if (jidBuilder_ == null) {
17001          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
17002        } else {
17003          jidBuilder_.clear();
17004        }
17005        bitField0_ = (bitField0_ & ~0x00000001);
17006        return this;
17007      }
17008
17009      public Builder clone() {
17010        return create().mergeFrom(buildPartial());
17011      }
17012
17013      public com.google.protobuf.Descriptors.Descriptor
17014          getDescriptorForType() {
17015        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalStateRequestProto_descriptor;
17016      }
17017
17018      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto getDefaultInstanceForType() {
17019        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance();
17020      }
17021
17022      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto build() {
17023        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto result = buildPartial();
17024        if (!result.isInitialized()) {
17025          throw newUninitializedMessageException(result);
17026        }
17027        return result;
17028      }
17029
17030      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto buildPartial() {
17031        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto(this);
17032        int from_bitField0_ = bitField0_;
17033        int to_bitField0_ = 0;
17034        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
17035          to_bitField0_ |= 0x00000001;
17036        }
17037        if (jidBuilder_ == null) {
17038          result.jid_ = jid_;
17039        } else {
17040          result.jid_ = jidBuilder_.build();
17041        }
17042        result.bitField0_ = to_bitField0_;
17043        onBuilt();
17044        return result;
17045      }
17046
17047      public Builder mergeFrom(com.google.protobuf.Message other) {
17048        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto) {
17049          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)other);
17050        } else {
17051          super.mergeFrom(other);
17052          return this;
17053        }
17054      }
17055
17056      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto other) {
17057        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance()) return this;
17058        if (other.hasJid()) {
17059          mergeJid(other.getJid());
17060        }
17061        this.mergeUnknownFields(other.getUnknownFields());
17062        return this;
17063      }
17064
17065      public final boolean isInitialized() {
17066        if (!hasJid()) {
17067          
17068          return false;
17069        }
17070        if (!getJid().isInitialized()) {
17071          
17072          return false;
17073        }
17074        return true;
17075      }
17076
17077      public Builder mergeFrom(
17078          com.google.protobuf.CodedInputStream input,
17079          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17080          throws java.io.IOException {
17081        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parsedMessage = null;
17082        try {
17083          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
17084        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
17085          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto) e.getUnfinishedMessage();
17086          throw e;
17087        } finally {
17088          if (parsedMessage != null) {
17089            mergeFrom(parsedMessage);
17090          }
17091        }
17092        return this;
17093      }
17094      private int bitField0_;
17095
17096      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
17097      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
17098      private com.google.protobuf.SingleFieldBuilder<
17099          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
17100      /**
17101       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17102       */
17103      public boolean hasJid() {
17104        return ((bitField0_ & 0x00000001) == 0x00000001);
17105      }
17106      /**
17107       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17108       */
17109      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
17110        if (jidBuilder_ == null) {
17111          return jid_;
17112        } else {
17113          return jidBuilder_.getMessage();
17114        }
17115      }
17116      /**
17117       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17118       */
17119      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
17120        if (jidBuilder_ == null) {
17121          if (value == null) {
17122            throw new NullPointerException();
17123          }
17124          jid_ = value;
17125          onChanged();
17126        } else {
17127          jidBuilder_.setMessage(value);
17128        }
17129        bitField0_ |= 0x00000001;
17130        return this;
17131      }
17132      /**
17133       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17134       */
17135      public Builder setJid(
17136          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
17137        if (jidBuilder_ == null) {
17138          jid_ = builderForValue.build();
17139          onChanged();
17140        } else {
17141          jidBuilder_.setMessage(builderForValue.build());
17142        }
17143        bitField0_ |= 0x00000001;
17144        return this;
17145      }
17146      /**
17147       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17148       */
17149      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
17150        if (jidBuilder_ == null) {
17151          if (((bitField0_ & 0x00000001) == 0x00000001) &&
17152              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
17153            jid_ =
17154              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
17155          } else {
17156            jid_ = value;
17157          }
17158          onChanged();
17159        } else {
17160          jidBuilder_.mergeFrom(value);
17161        }
17162        bitField0_ |= 0x00000001;
17163        return this;
17164      }
17165      /**
17166       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17167       */
17168      public Builder clearJid() {
17169        if (jidBuilder_ == null) {
17170          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
17171          onChanged();
17172        } else {
17173          jidBuilder_.clear();
17174        }
17175        bitField0_ = (bitField0_ & ~0x00000001);
17176        return this;
17177      }
17178      /**
17179       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17180       */
17181      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
17182        bitField0_ |= 0x00000001;
17183        onChanged();
17184        return getJidFieldBuilder().getBuilder();
17185      }
17186      /**
17187       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17188       */
17189      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
17190        if (jidBuilder_ != null) {
17191          return jidBuilder_.getMessageOrBuilder();
17192        } else {
17193          return jid_;
17194        }
17195      }
17196      /**
17197       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17198       */
17199      private com.google.protobuf.SingleFieldBuilder<
17200          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
17201          getJidFieldBuilder() {
17202        if (jidBuilder_ == null) {
17203          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
17204              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
17205                  jid_,
17206                  getParentForChildren(),
17207                  isClean());
17208          jid_ = null;
17209        }
17210        return jidBuilder_;
17211      }
17212
17213      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.GetJournalStateRequestProto)
17214    }
17215
17216    static {
17217      defaultInstance = new GetJournalStateRequestProto(true);
17218      defaultInstance.initFields();
17219    }
17220
17221    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.GetJournalStateRequestProto)
17222  }
17223
17224  public interface GetJournalStateResponseProtoOrBuilder
17225      extends com.google.protobuf.MessageOrBuilder {
17226
17227    // required uint64 lastPromisedEpoch = 1;
17228    /**
17229     * <code>required uint64 lastPromisedEpoch = 1;</code>
17230     */
17231    boolean hasLastPromisedEpoch();
17232    /**
17233     * <code>required uint64 lastPromisedEpoch = 1;</code>
17234     */
17235    long getLastPromisedEpoch();
17236
17237    // required uint32 httpPort = 2;
17238    /**
17239     * <code>required uint32 httpPort = 2;</code>
17240     *
17241     * <pre>
17242     * Deprecated by fromURL
17243     * </pre>
17244     */
17245    boolean hasHttpPort();
17246    /**
17247     * <code>required uint32 httpPort = 2;</code>
17248     *
17249     * <pre>
17250     * Deprecated by fromURL
17251     * </pre>
17252     */
17253    int getHttpPort();
17254
17255    // optional string fromURL = 3;
17256    /**
17257     * <code>optional string fromURL = 3;</code>
17258     */
17259    boolean hasFromURL();
17260    /**
17261     * <code>optional string fromURL = 3;</code>
17262     */
17263    java.lang.String getFromURL();
17264    /**
17265     * <code>optional string fromURL = 3;</code>
17266     */
17267    com.google.protobuf.ByteString
17268        getFromURLBytes();
17269  }
17270  /**
17271   * Protobuf type {@code hadoop.hdfs.qjournal.GetJournalStateResponseProto}
17272   */
17273  public static final class GetJournalStateResponseProto extends
17274      com.google.protobuf.GeneratedMessage
17275      implements GetJournalStateResponseProtoOrBuilder {
17276    // Use GetJournalStateResponseProto.newBuilder() to construct.
17277    private GetJournalStateResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
17278      super(builder);
17279      this.unknownFields = builder.getUnknownFields();
17280    }
17281    private GetJournalStateResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
17282
17283    private static final GetJournalStateResponseProto defaultInstance;
17284    public static GetJournalStateResponseProto getDefaultInstance() {
17285      return defaultInstance;
17286    }
17287
17288    public GetJournalStateResponseProto getDefaultInstanceForType() {
17289      return defaultInstance;
17290    }
17291
17292    private final com.google.protobuf.UnknownFieldSet unknownFields;
17293    @java.lang.Override
17294    public final com.google.protobuf.UnknownFieldSet
17295        getUnknownFields() {
17296      return this.unknownFields;
17297    }
17298    private GetJournalStateResponseProto(
17299        com.google.protobuf.CodedInputStream input,
17300        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17301        throws com.google.protobuf.InvalidProtocolBufferException {
17302      initFields();
17303      int mutable_bitField0_ = 0;
17304      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
17305          com.google.protobuf.UnknownFieldSet.newBuilder();
17306      try {
17307        boolean done = false;
17308        while (!done) {
17309          int tag = input.readTag();
17310          switch (tag) {
17311            case 0:
17312              done = true;
17313              break;
17314            default: {
17315              if (!parseUnknownField(input, unknownFields,
17316                                     extensionRegistry, tag)) {
17317                done = true;
17318              }
17319              break;
17320            }
17321            case 8: {
17322              bitField0_ |= 0x00000001;
17323              lastPromisedEpoch_ = input.readUInt64();
17324              break;
17325            }
17326            case 16: {
17327              bitField0_ |= 0x00000002;
17328              httpPort_ = input.readUInt32();
17329              break;
17330            }
17331            case 26: {
17332              bitField0_ |= 0x00000004;
17333              fromURL_ = input.readBytes();
17334              break;
17335            }
17336          }
17337        }
17338      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
17339        throw e.setUnfinishedMessage(this);
17340      } catch (java.io.IOException e) {
17341        throw new com.google.protobuf.InvalidProtocolBufferException(
17342            e.getMessage()).setUnfinishedMessage(this);
17343      } finally {
17344        this.unknownFields = unknownFields.build();
17345        makeExtensionsImmutable();
17346      }
17347    }
17348    public static final com.google.protobuf.Descriptors.Descriptor
17349        getDescriptor() {
17350      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalStateResponseProto_descriptor;
17351    }
17352
17353    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
17354        internalGetFieldAccessorTable() {
17355      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalStateResponseProto_fieldAccessorTable
17356          .ensureFieldAccessorsInitialized(
17357              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.Builder.class);
17358    }
17359
17360    public static com.google.protobuf.Parser<GetJournalStateResponseProto> PARSER =
17361        new com.google.protobuf.AbstractParser<GetJournalStateResponseProto>() {
17362      public GetJournalStateResponseProto parsePartialFrom(
17363          com.google.protobuf.CodedInputStream input,
17364          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17365          throws com.google.protobuf.InvalidProtocolBufferException {
17366        return new GetJournalStateResponseProto(input, extensionRegistry);
17367      }
17368    };
17369
17370    @java.lang.Override
17371    public com.google.protobuf.Parser<GetJournalStateResponseProto> getParserForType() {
17372      return PARSER;
17373    }
17374
17375    private int bitField0_;
17376    // required uint64 lastPromisedEpoch = 1;
17377    public static final int LASTPROMISEDEPOCH_FIELD_NUMBER = 1;
17378    private long lastPromisedEpoch_;
17379    /**
17380     * <code>required uint64 lastPromisedEpoch = 1;</code>
17381     */
17382    public boolean hasLastPromisedEpoch() {
17383      return ((bitField0_ & 0x00000001) == 0x00000001);
17384    }
17385    /**
17386     * <code>required uint64 lastPromisedEpoch = 1;</code>
17387     */
17388    public long getLastPromisedEpoch() {
17389      return lastPromisedEpoch_;
17390    }
17391
17392    // required uint32 httpPort = 2;
17393    public static final int HTTPPORT_FIELD_NUMBER = 2;
17394    private int httpPort_;
17395    /**
17396     * <code>required uint32 httpPort = 2;</code>
17397     *
17398     * <pre>
17399     * Deprecated by fromURL
17400     * </pre>
17401     */
17402    public boolean hasHttpPort() {
17403      return ((bitField0_ & 0x00000002) == 0x00000002);
17404    }
17405    /**
17406     * <code>required uint32 httpPort = 2;</code>
17407     *
17408     * <pre>
17409     * Deprecated by fromURL
17410     * </pre>
17411     */
17412    public int getHttpPort() {
17413      return httpPort_;
17414    }
17415
17416    // optional string fromURL = 3;
17417    public static final int FROMURL_FIELD_NUMBER = 3;
17418    private java.lang.Object fromURL_;
17419    /**
17420     * <code>optional string fromURL = 3;</code>
17421     */
17422    public boolean hasFromURL() {
17423      return ((bitField0_ & 0x00000004) == 0x00000004);
17424    }
17425    /**
17426     * <code>optional string fromURL = 3;</code>
17427     */
17428    public java.lang.String getFromURL() {
17429      java.lang.Object ref = fromURL_;
17430      if (ref instanceof java.lang.String) {
17431        return (java.lang.String) ref;
17432      } else {
17433        com.google.protobuf.ByteString bs = 
17434            (com.google.protobuf.ByteString) ref;
17435        java.lang.String s = bs.toStringUtf8();
17436        if (bs.isValidUtf8()) {
17437          fromURL_ = s;
17438        }
17439        return s;
17440      }
17441    }
17442    /**
17443     * <code>optional string fromURL = 3;</code>
17444     */
17445    public com.google.protobuf.ByteString
17446        getFromURLBytes() {
17447      java.lang.Object ref = fromURL_;
17448      if (ref instanceof java.lang.String) {
17449        com.google.protobuf.ByteString b = 
17450            com.google.protobuf.ByteString.copyFromUtf8(
17451                (java.lang.String) ref);
17452        fromURL_ = b;
17453        return b;
17454      } else {
17455        return (com.google.protobuf.ByteString) ref;
17456      }
17457    }
17458
17459    private void initFields() {
17460      lastPromisedEpoch_ = 0L;
17461      httpPort_ = 0;
17462      fromURL_ = "";
17463    }
17464    private byte memoizedIsInitialized = -1;
17465    public final boolean isInitialized() {
17466      byte isInitialized = memoizedIsInitialized;
17467      if (isInitialized != -1) return isInitialized == 1;
17468
17469      if (!hasLastPromisedEpoch()) {
17470        memoizedIsInitialized = 0;
17471        return false;
17472      }
17473      if (!hasHttpPort()) {
17474        memoizedIsInitialized = 0;
17475        return false;
17476      }
17477      memoizedIsInitialized = 1;
17478      return true;
17479    }
17480
17481    public void writeTo(com.google.protobuf.CodedOutputStream output)
17482                        throws java.io.IOException {
17483      getSerializedSize();
17484      if (((bitField0_ & 0x00000001) == 0x00000001)) {
17485        output.writeUInt64(1, lastPromisedEpoch_);
17486      }
17487      if (((bitField0_ & 0x00000002) == 0x00000002)) {
17488        output.writeUInt32(2, httpPort_);
17489      }
17490      if (((bitField0_ & 0x00000004) == 0x00000004)) {
17491        output.writeBytes(3, getFromURLBytes());
17492      }
17493      getUnknownFields().writeTo(output);
17494    }
17495
17496    private int memoizedSerializedSize = -1;
17497    public int getSerializedSize() {
17498      int size = memoizedSerializedSize;
17499      if (size != -1) return size;
17500
17501      size = 0;
17502      if (((bitField0_ & 0x00000001) == 0x00000001)) {
17503        size += com.google.protobuf.CodedOutputStream
17504          .computeUInt64Size(1, lastPromisedEpoch_);
17505      }
17506      if (((bitField0_ & 0x00000002) == 0x00000002)) {
17507        size += com.google.protobuf.CodedOutputStream
17508          .computeUInt32Size(2, httpPort_);
17509      }
17510      if (((bitField0_ & 0x00000004) == 0x00000004)) {
17511        size += com.google.protobuf.CodedOutputStream
17512          .computeBytesSize(3, getFromURLBytes());
17513      }
17514      size += getUnknownFields().getSerializedSize();
17515      memoizedSerializedSize = size;
17516      return size;
17517    }
17518
17519    private static final long serialVersionUID = 0L;
17520    @java.lang.Override
17521    protected java.lang.Object writeReplace()
17522        throws java.io.ObjectStreamException {
17523      return super.writeReplace();
17524    }
17525
17526    @java.lang.Override
17527    public boolean equals(final java.lang.Object obj) {
17528      if (obj == this) {
17529       return true;
17530      }
17531      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto)) {
17532        return super.equals(obj);
17533      }
17534      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) obj;
17535
17536      boolean result = true;
17537      result = result && (hasLastPromisedEpoch() == other.hasLastPromisedEpoch());
17538      if (hasLastPromisedEpoch()) {
17539        result = result && (getLastPromisedEpoch()
17540            == other.getLastPromisedEpoch());
17541      }
17542      result = result && (hasHttpPort() == other.hasHttpPort());
17543      if (hasHttpPort()) {
17544        result = result && (getHttpPort()
17545            == other.getHttpPort());
17546      }
17547      result = result && (hasFromURL() == other.hasFromURL());
17548      if (hasFromURL()) {
17549        result = result && getFromURL()
17550            .equals(other.getFromURL());
17551      }
17552      result = result &&
17553          getUnknownFields().equals(other.getUnknownFields());
17554      return result;
17555    }
17556
17557    private int memoizedHashCode = 0;
17558    @java.lang.Override
17559    public int hashCode() {
17560      if (memoizedHashCode != 0) {
17561        return memoizedHashCode;
17562      }
17563      int hash = 41;
17564      hash = (19 * hash) + getDescriptorForType().hashCode();
17565      if (hasLastPromisedEpoch()) {
17566        hash = (37 * hash) + LASTPROMISEDEPOCH_FIELD_NUMBER;
17567        hash = (53 * hash) + hashLong(getLastPromisedEpoch());
17568      }
17569      if (hasHttpPort()) {
17570        hash = (37 * hash) + HTTPPORT_FIELD_NUMBER;
17571        hash = (53 * hash) + getHttpPort();
17572      }
17573      if (hasFromURL()) {
17574        hash = (37 * hash) + FROMURL_FIELD_NUMBER;
17575        hash = (53 * hash) + getFromURL().hashCode();
17576      }
17577      hash = (29 * hash) + getUnknownFields().hashCode();
17578      memoizedHashCode = hash;
17579      return hash;
17580    }
17581
17582    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
17583        com.google.protobuf.ByteString data)
17584        throws com.google.protobuf.InvalidProtocolBufferException {
17585      return PARSER.parseFrom(data);
17586    }
17587    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
17588        com.google.protobuf.ByteString data,
17589        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17590        throws com.google.protobuf.InvalidProtocolBufferException {
17591      return PARSER.parseFrom(data, extensionRegistry);
17592    }
17593    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(byte[] data)
17594        throws com.google.protobuf.InvalidProtocolBufferException {
17595      return PARSER.parseFrom(data);
17596    }
17597    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
17598        byte[] data,
17599        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17600        throws com.google.protobuf.InvalidProtocolBufferException {
17601      return PARSER.parseFrom(data, extensionRegistry);
17602    }
17603    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(java.io.InputStream input)
17604        throws java.io.IOException {
17605      return PARSER.parseFrom(input);
17606    }
17607    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
17608        java.io.InputStream input,
17609        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17610        throws java.io.IOException {
17611      return PARSER.parseFrom(input, extensionRegistry);
17612    }
17613    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseDelimitedFrom(java.io.InputStream input)
17614        throws java.io.IOException {
17615      return PARSER.parseDelimitedFrom(input);
17616    }
17617    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseDelimitedFrom(
17618        java.io.InputStream input,
17619        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17620        throws java.io.IOException {
17621      return PARSER.parseDelimitedFrom(input, extensionRegistry);
17622    }
17623    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
17624        com.google.protobuf.CodedInputStream input)
17625        throws java.io.IOException {
17626      return PARSER.parseFrom(input);
17627    }
17628    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
17629        com.google.protobuf.CodedInputStream input,
17630        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17631        throws java.io.IOException {
17632      return PARSER.parseFrom(input, extensionRegistry);
17633    }
17634
17635    public static Builder newBuilder() { return Builder.create(); }
17636    public Builder newBuilderForType() { return newBuilder(); }
17637    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto prototype) {
17638      return newBuilder().mergeFrom(prototype);
17639    }
17640    public Builder toBuilder() { return newBuilder(this); }
17641
17642    @java.lang.Override
17643    protected Builder newBuilderForType(
17644        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
17645      Builder builder = new Builder(parent);
17646      return builder;
17647    }
17648    /**
17649     * Protobuf type {@code hadoop.hdfs.qjournal.GetJournalStateResponseProto}
17650     */
17651    public static final class Builder extends
17652        com.google.protobuf.GeneratedMessage.Builder<Builder>
17653       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProtoOrBuilder {
17654      public static final com.google.protobuf.Descriptors.Descriptor
17655          getDescriptor() {
17656        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalStateResponseProto_descriptor;
17657      }
17658
17659      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
17660          internalGetFieldAccessorTable() {
17661        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalStateResponseProto_fieldAccessorTable
17662            .ensureFieldAccessorsInitialized(
17663                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.Builder.class);
17664      }
17665
17666      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.newBuilder()
17667      private Builder() {
17668        maybeForceBuilderInitialization();
17669      }
17670
17671      private Builder(
17672          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
17673        super(parent);
17674        maybeForceBuilderInitialization();
17675      }
17676      private void maybeForceBuilderInitialization() {
17677        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
17678        }
17679      }
17680      private static Builder create() {
17681        return new Builder();
17682      }
17683
17684      public Builder clear() {
17685        super.clear();
17686        lastPromisedEpoch_ = 0L;
17687        bitField0_ = (bitField0_ & ~0x00000001);
17688        httpPort_ = 0;
17689        bitField0_ = (bitField0_ & ~0x00000002);
17690        fromURL_ = "";
17691        bitField0_ = (bitField0_ & ~0x00000004);
17692        return this;
17693      }
17694
17695      public Builder clone() {
17696        return create().mergeFrom(buildPartial());
17697      }
17698
17699      public com.google.protobuf.Descriptors.Descriptor
17700          getDescriptorForType() {
17701        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetJournalStateResponseProto_descriptor;
17702      }
17703
17704      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getDefaultInstanceForType() {
17705        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance();
17706      }
17707
17708      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto build() {
17709        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto result = buildPartial();
17710        if (!result.isInitialized()) {
17711          throw newUninitializedMessageException(result);
17712        }
17713        return result;
17714      }
17715
17716      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto buildPartial() {
17717        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto(this);
17718        int from_bitField0_ = bitField0_;
17719        int to_bitField0_ = 0;
17720        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
17721          to_bitField0_ |= 0x00000001;
17722        }
17723        result.lastPromisedEpoch_ = lastPromisedEpoch_;
17724        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
17725          to_bitField0_ |= 0x00000002;
17726        }
17727        result.httpPort_ = httpPort_;
17728        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
17729          to_bitField0_ |= 0x00000004;
17730        }
17731        result.fromURL_ = fromURL_;
17732        result.bitField0_ = to_bitField0_;
17733        onBuilt();
17734        return result;
17735      }
17736
17737      public Builder mergeFrom(com.google.protobuf.Message other) {
17738        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) {
17739          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto)other);
17740        } else {
17741          super.mergeFrom(other);
17742          return this;
17743        }
17744      }
17745
17746      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto other) {
17747        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance()) return this;
17748        if (other.hasLastPromisedEpoch()) {
17749          setLastPromisedEpoch(other.getLastPromisedEpoch());
17750        }
17751        if (other.hasHttpPort()) {
17752          setHttpPort(other.getHttpPort());
17753        }
17754        if (other.hasFromURL()) {
17755          bitField0_ |= 0x00000004;
17756          fromURL_ = other.fromURL_;
17757          onChanged();
17758        }
17759        this.mergeUnknownFields(other.getUnknownFields());
17760        return this;
17761      }
17762
17763      public final boolean isInitialized() {
17764        if (!hasLastPromisedEpoch()) {
17765          
17766          return false;
17767        }
17768        if (!hasHttpPort()) {
17769          
17770          return false;
17771        }
17772        return true;
17773      }
17774
17775      public Builder mergeFrom(
17776          com.google.protobuf.CodedInputStream input,
17777          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17778          throws java.io.IOException {
17779        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parsedMessage = null;
17780        try {
17781          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
17782        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
17783          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) e.getUnfinishedMessage();
17784          throw e;
17785        } finally {
17786          if (parsedMessage != null) {
17787            mergeFrom(parsedMessage);
17788          }
17789        }
17790        return this;
17791      }
17792      private int bitField0_;
17793
17794      // required uint64 lastPromisedEpoch = 1;
17795      private long lastPromisedEpoch_ ;
17796      /**
17797       * <code>required uint64 lastPromisedEpoch = 1;</code>
17798       */
17799      public boolean hasLastPromisedEpoch() {
17800        return ((bitField0_ & 0x00000001) == 0x00000001);
17801      }
17802      /**
17803       * <code>required uint64 lastPromisedEpoch = 1;</code>
17804       */
17805      public long getLastPromisedEpoch() {
17806        return lastPromisedEpoch_;
17807      }
17808      /**
17809       * <code>required uint64 lastPromisedEpoch = 1;</code>
17810       */
17811      public Builder setLastPromisedEpoch(long value) {
17812        bitField0_ |= 0x00000001;
17813        lastPromisedEpoch_ = value;
17814        onChanged();
17815        return this;
17816      }
17817      /**
17818       * <code>required uint64 lastPromisedEpoch = 1;</code>
17819       */
17820      public Builder clearLastPromisedEpoch() {
17821        bitField0_ = (bitField0_ & ~0x00000001);
17822        lastPromisedEpoch_ = 0L;
17823        onChanged();
17824        return this;
17825      }
17826
17827      // required uint32 httpPort = 2;
17828      private int httpPort_ ;
17829      /**
17830       * <code>required uint32 httpPort = 2;</code>
17831       *
17832       * <pre>
17833       * Deprecated by fromURL
17834       * </pre>
17835       */
17836      public boolean hasHttpPort() {
17837        return ((bitField0_ & 0x00000002) == 0x00000002);
17838      }
17839      /**
17840       * <code>required uint32 httpPort = 2;</code>
17841       *
17842       * <pre>
17843       * Deprecated by fromURL
17844       * </pre>
17845       */
17846      public int getHttpPort() {
17847        return httpPort_;
17848      }
17849      /**
17850       * <code>required uint32 httpPort = 2;</code>
17851       *
17852       * <pre>
17853       * Deprecated by fromURL
17854       * </pre>
17855       */
17856      public Builder setHttpPort(int value) {
17857        bitField0_ |= 0x00000002;
17858        httpPort_ = value;
17859        onChanged();
17860        return this;
17861      }
17862      /**
17863       * <code>required uint32 httpPort = 2;</code>
17864       *
17865       * <pre>
17866       * Deprecated by fromURL
17867       * </pre>
17868       */
17869      public Builder clearHttpPort() {
17870        bitField0_ = (bitField0_ & ~0x00000002);
17871        httpPort_ = 0;
17872        onChanged();
17873        return this;
17874      }
17875
17876      // optional string fromURL = 3;
17877      private java.lang.Object fromURL_ = "";
17878      /**
17879       * <code>optional string fromURL = 3;</code>
17880       */
17881      public boolean hasFromURL() {
17882        return ((bitField0_ & 0x00000004) == 0x00000004);
17883      }
17884      /**
17885       * <code>optional string fromURL = 3;</code>
17886       */
17887      public java.lang.String getFromURL() {
17888        java.lang.Object ref = fromURL_;
17889        if (!(ref instanceof java.lang.String)) {
17890          java.lang.String s = ((com.google.protobuf.ByteString) ref)
17891              .toStringUtf8();
17892          fromURL_ = s;
17893          return s;
17894        } else {
17895          return (java.lang.String) ref;
17896        }
17897      }
17898      /**
17899       * <code>optional string fromURL = 3;</code>
17900       */
17901      public com.google.protobuf.ByteString
17902          getFromURLBytes() {
17903        java.lang.Object ref = fromURL_;
17904        if (ref instanceof String) {
17905          com.google.protobuf.ByteString b = 
17906              com.google.protobuf.ByteString.copyFromUtf8(
17907                  (java.lang.String) ref);
17908          fromURL_ = b;
17909          return b;
17910        } else {
17911          return (com.google.protobuf.ByteString) ref;
17912        }
17913      }
17914      /**
17915       * <code>optional string fromURL = 3;</code>
17916       */
17917      public Builder setFromURL(
17918          java.lang.String value) {
17919        if (value == null) {
17920    throw new NullPointerException();
17921  }
17922  bitField0_ |= 0x00000004;
17923        fromURL_ = value;
17924        onChanged();
17925        return this;
17926      }
17927      /**
17928       * <code>optional string fromURL = 3;</code>
17929       */
17930      public Builder clearFromURL() {
17931        bitField0_ = (bitField0_ & ~0x00000004);
17932        fromURL_ = getDefaultInstance().getFromURL();
17933        onChanged();
17934        return this;
17935      }
17936      /**
17937       * <code>optional string fromURL = 3;</code>
17938       */
17939      public Builder setFromURLBytes(
17940          com.google.protobuf.ByteString value) {
17941        if (value == null) {
17942    throw new NullPointerException();
17943  }
17944  bitField0_ |= 0x00000004;
17945        fromURL_ = value;
17946        onChanged();
17947        return this;
17948      }
17949
17950      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.GetJournalStateResponseProto)
17951    }
17952
17953    static {
17954      defaultInstance = new GetJournalStateResponseProto(true);
17955      defaultInstance.initFields();
17956    }
17957
17958    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.GetJournalStateResponseProto)
17959  }
17960
17961  public interface FormatRequestProtoOrBuilder
17962      extends com.google.protobuf.MessageOrBuilder {
17963
17964    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
17965    /**
17966     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17967     */
17968    boolean hasJid();
17969    /**
17970     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17971     */
17972    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
17973    /**
17974     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
17975     */
17976    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
17977
17978    // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
17979    /**
17980     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
17981     */
17982    boolean hasNsInfo();
17983    /**
17984     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
17985     */
17986    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto getNsInfo();
17987    /**
17988     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
17989     */
17990    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder();
17991  }
17992  /**
17993   * Protobuf type {@code hadoop.hdfs.qjournal.FormatRequestProto}
17994   *
17995   * <pre>
17996   **
17997   * format()
17998   * </pre>
17999   */
18000  public static final class FormatRequestProto extends
18001      com.google.protobuf.GeneratedMessage
18002      implements FormatRequestProtoOrBuilder {
18003    // Use FormatRequestProto.newBuilder() to construct.
18004    private FormatRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
18005      super(builder);
18006      this.unknownFields = builder.getUnknownFields();
18007    }
18008    private FormatRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
18009
18010    private static final FormatRequestProto defaultInstance;
18011    public static FormatRequestProto getDefaultInstance() {
18012      return defaultInstance;
18013    }
18014
18015    public FormatRequestProto getDefaultInstanceForType() {
18016      return defaultInstance;
18017    }
18018
18019    private final com.google.protobuf.UnknownFieldSet unknownFields;
18020    @java.lang.Override
18021    public final com.google.protobuf.UnknownFieldSet
18022        getUnknownFields() {
18023      return this.unknownFields;
18024    }
18025    private FormatRequestProto(
18026        com.google.protobuf.CodedInputStream input,
18027        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18028        throws com.google.protobuf.InvalidProtocolBufferException {
18029      initFields();
18030      int mutable_bitField0_ = 0;
18031      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
18032          com.google.protobuf.UnknownFieldSet.newBuilder();
18033      try {
18034        boolean done = false;
18035        while (!done) {
18036          int tag = input.readTag();
18037          switch (tag) {
18038            case 0:
18039              done = true;
18040              break;
18041            default: {
18042              if (!parseUnknownField(input, unknownFields,
18043                                     extensionRegistry, tag)) {
18044                done = true;
18045              }
18046              break;
18047            }
18048            case 10: {
18049              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
18050              if (((bitField0_ & 0x00000001) == 0x00000001)) {
18051                subBuilder = jid_.toBuilder();
18052              }
18053              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
18054              if (subBuilder != null) {
18055                subBuilder.mergeFrom(jid_);
18056                jid_ = subBuilder.buildPartial();
18057              }
18058              bitField0_ |= 0x00000001;
18059              break;
18060            }
18061            case 18: {
18062              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder subBuilder = null;
18063              if (((bitField0_ & 0x00000002) == 0x00000002)) {
18064                subBuilder = nsInfo_.toBuilder();
18065              }
18066              nsInfo_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.PARSER, extensionRegistry);
18067              if (subBuilder != null) {
18068                subBuilder.mergeFrom(nsInfo_);
18069                nsInfo_ = subBuilder.buildPartial();
18070              }
18071              bitField0_ |= 0x00000002;
18072              break;
18073            }
18074          }
18075        }
18076      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
18077        throw e.setUnfinishedMessage(this);
18078      } catch (java.io.IOException e) {
18079        throw new com.google.protobuf.InvalidProtocolBufferException(
18080            e.getMessage()).setUnfinishedMessage(this);
18081      } finally {
18082        this.unknownFields = unknownFields.build();
18083        makeExtensionsImmutable();
18084      }
18085    }
18086    public static final com.google.protobuf.Descriptors.Descriptor
18087        getDescriptor() {
18088      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FormatRequestProto_descriptor;
18089    }
18090
18091    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
18092        internalGetFieldAccessorTable() {
18093      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FormatRequestProto_fieldAccessorTable
18094          .ensureFieldAccessorsInitialized(
18095              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.Builder.class);
18096    }
18097
18098    public static com.google.protobuf.Parser<FormatRequestProto> PARSER =
18099        new com.google.protobuf.AbstractParser<FormatRequestProto>() {
18100      public FormatRequestProto parsePartialFrom(
18101          com.google.protobuf.CodedInputStream input,
18102          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18103          throws com.google.protobuf.InvalidProtocolBufferException {
18104        return new FormatRequestProto(input, extensionRegistry);
18105      }
18106    };
18107
18108    @java.lang.Override
18109    public com.google.protobuf.Parser<FormatRequestProto> getParserForType() {
18110      return PARSER;
18111    }
18112
18113    private int bitField0_;
18114    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
18115    public static final int JID_FIELD_NUMBER = 1;
18116    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
18117    /**
18118     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18119     */
18120    public boolean hasJid() {
18121      return ((bitField0_ & 0x00000001) == 0x00000001);
18122    }
18123    /**
18124     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18125     */
18126    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
18127      return jid_;
18128    }
18129    /**
18130     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18131     */
18132    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
18133      return jid_;
18134    }
18135
18136    // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
18137    public static final int NSINFO_FIELD_NUMBER = 2;
18138    private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto nsInfo_;
18139    /**
18140     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18141     */
18142    public boolean hasNsInfo() {
18143      return ((bitField0_ & 0x00000002) == 0x00000002);
18144    }
18145    /**
18146     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18147     */
18148    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto getNsInfo() {
18149      return nsInfo_;
18150    }
18151    /**
18152     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18153     */
18154    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
18155      return nsInfo_;
18156    }
18157
18158    private void initFields() {
18159      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
18160      nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.getDefaultInstance();
18161    }
18162    private byte memoizedIsInitialized = -1;
18163    public final boolean isInitialized() {
18164      byte isInitialized = memoizedIsInitialized;
18165      if (isInitialized != -1) return isInitialized == 1;
18166
18167      if (!hasJid()) {
18168        memoizedIsInitialized = 0;
18169        return false;
18170      }
18171      if (!hasNsInfo()) {
18172        memoizedIsInitialized = 0;
18173        return false;
18174      }
18175      if (!getJid().isInitialized()) {
18176        memoizedIsInitialized = 0;
18177        return false;
18178      }
18179      if (!getNsInfo().isInitialized()) {
18180        memoizedIsInitialized = 0;
18181        return false;
18182      }
18183      memoizedIsInitialized = 1;
18184      return true;
18185    }
18186
18187    public void writeTo(com.google.protobuf.CodedOutputStream output)
18188                        throws java.io.IOException {
18189      getSerializedSize();
18190      if (((bitField0_ & 0x00000001) == 0x00000001)) {
18191        output.writeMessage(1, jid_);
18192      }
18193      if (((bitField0_ & 0x00000002) == 0x00000002)) {
18194        output.writeMessage(2, nsInfo_);
18195      }
18196      getUnknownFields().writeTo(output);
18197    }
18198
18199    private int memoizedSerializedSize = -1;
18200    public int getSerializedSize() {
18201      int size = memoizedSerializedSize;
18202      if (size != -1) return size;
18203
18204      size = 0;
18205      if (((bitField0_ & 0x00000001) == 0x00000001)) {
18206        size += com.google.protobuf.CodedOutputStream
18207          .computeMessageSize(1, jid_);
18208      }
18209      if (((bitField0_ & 0x00000002) == 0x00000002)) {
18210        size += com.google.protobuf.CodedOutputStream
18211          .computeMessageSize(2, nsInfo_);
18212      }
18213      size += getUnknownFields().getSerializedSize();
18214      memoizedSerializedSize = size;
18215      return size;
18216    }
18217
18218    private static final long serialVersionUID = 0L;
18219    @java.lang.Override
18220    protected java.lang.Object writeReplace()
18221        throws java.io.ObjectStreamException {
18222      return super.writeReplace();
18223    }
18224
18225    @java.lang.Override
18226    public boolean equals(final java.lang.Object obj) {
18227      if (obj == this) {
18228       return true;
18229      }
18230      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)) {
18231        return super.equals(obj);
18232      }
18233      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto) obj;
18234
18235      boolean result = true;
18236      result = result && (hasJid() == other.hasJid());
18237      if (hasJid()) {
18238        result = result && getJid()
18239            .equals(other.getJid());
18240      }
18241      result = result && (hasNsInfo() == other.hasNsInfo());
18242      if (hasNsInfo()) {
18243        result = result && getNsInfo()
18244            .equals(other.getNsInfo());
18245      }
18246      result = result &&
18247          getUnknownFields().equals(other.getUnknownFields());
18248      return result;
18249    }
18250
18251    private int memoizedHashCode = 0;
18252    @java.lang.Override
18253    public int hashCode() {
18254      if (memoizedHashCode != 0) {
18255        return memoizedHashCode;
18256      }
18257      int hash = 41;
18258      hash = (19 * hash) + getDescriptorForType().hashCode();
18259      if (hasJid()) {
18260        hash = (37 * hash) + JID_FIELD_NUMBER;
18261        hash = (53 * hash) + getJid().hashCode();
18262      }
18263      if (hasNsInfo()) {
18264        hash = (37 * hash) + NSINFO_FIELD_NUMBER;
18265        hash = (53 * hash) + getNsInfo().hashCode();
18266      }
18267      hash = (29 * hash) + getUnknownFields().hashCode();
18268      memoizedHashCode = hash;
18269      return hash;
18270    }
18271
18272    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
18273        com.google.protobuf.ByteString data)
18274        throws com.google.protobuf.InvalidProtocolBufferException {
18275      return PARSER.parseFrom(data);
18276    }
18277    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
18278        com.google.protobuf.ByteString data,
18279        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18280        throws com.google.protobuf.InvalidProtocolBufferException {
18281      return PARSER.parseFrom(data, extensionRegistry);
18282    }
18283    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(byte[] data)
18284        throws com.google.protobuf.InvalidProtocolBufferException {
18285      return PARSER.parseFrom(data);
18286    }
18287    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
18288        byte[] data,
18289        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18290        throws com.google.protobuf.InvalidProtocolBufferException {
18291      return PARSER.parseFrom(data, extensionRegistry);
18292    }
18293    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(java.io.InputStream input)
18294        throws java.io.IOException {
18295      return PARSER.parseFrom(input);
18296    }
18297    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
18298        java.io.InputStream input,
18299        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18300        throws java.io.IOException {
18301      return PARSER.parseFrom(input, extensionRegistry);
18302    }
18303    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseDelimitedFrom(java.io.InputStream input)
18304        throws java.io.IOException {
18305      return PARSER.parseDelimitedFrom(input);
18306    }
18307    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseDelimitedFrom(
18308        java.io.InputStream input,
18309        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18310        throws java.io.IOException {
18311      return PARSER.parseDelimitedFrom(input, extensionRegistry);
18312    }
18313    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
18314        com.google.protobuf.CodedInputStream input)
18315        throws java.io.IOException {
18316      return PARSER.parseFrom(input);
18317    }
18318    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
18319        com.google.protobuf.CodedInputStream input,
18320        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18321        throws java.io.IOException {
18322      return PARSER.parseFrom(input, extensionRegistry);
18323    }
18324
18325    public static Builder newBuilder() { return Builder.create(); }
18326    public Builder newBuilderForType() { return newBuilder(); }
18327    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto prototype) {
18328      return newBuilder().mergeFrom(prototype);
18329    }
18330    public Builder toBuilder() { return newBuilder(this); }
18331
18332    @java.lang.Override
18333    protected Builder newBuilderForType(
18334        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
18335      Builder builder = new Builder(parent);
18336      return builder;
18337    }
18338    /**
18339     * Protobuf type {@code hadoop.hdfs.qjournal.FormatRequestProto}
18340     *
18341     * <pre>
18342     **
18343     * format()
18344     * </pre>
18345     */
18346    public static final class Builder extends
18347        com.google.protobuf.GeneratedMessage.Builder<Builder>
18348       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProtoOrBuilder {
18349      public static final com.google.protobuf.Descriptors.Descriptor
18350          getDescriptor() {
18351        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FormatRequestProto_descriptor;
18352      }
18353
18354      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
18355          internalGetFieldAccessorTable() {
18356        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FormatRequestProto_fieldAccessorTable
18357            .ensureFieldAccessorsInitialized(
18358                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.Builder.class);
18359      }
18360
18361      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.newBuilder()
18362      private Builder() {
18363        maybeForceBuilderInitialization();
18364      }
18365
18366      private Builder(
18367          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
18368        super(parent);
18369        maybeForceBuilderInitialization();
18370      }
18371      private void maybeForceBuilderInitialization() {
18372        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
18373          getJidFieldBuilder();
18374          getNsInfoFieldBuilder();
18375        }
18376      }
18377      private static Builder create() {
18378        return new Builder();
18379      }
18380
18381      public Builder clear() {
18382        super.clear();
18383        if (jidBuilder_ == null) {
18384          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
18385        } else {
18386          jidBuilder_.clear();
18387        }
18388        bitField0_ = (bitField0_ & ~0x00000001);
18389        if (nsInfoBuilder_ == null) {
18390          nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.getDefaultInstance();
18391        } else {
18392          nsInfoBuilder_.clear();
18393        }
18394        bitField0_ = (bitField0_ & ~0x00000002);
18395        return this;
18396      }
18397
18398      public Builder clone() {
18399        return create().mergeFrom(buildPartial());
18400      }
18401
18402      public com.google.protobuf.Descriptors.Descriptor
18403          getDescriptorForType() {
18404        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FormatRequestProto_descriptor;
18405      }
18406
18407      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto getDefaultInstanceForType() {
18408        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance();
18409      }
18410
18411      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto build() {
18412        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto result = buildPartial();
18413        if (!result.isInitialized()) {
18414          throw newUninitializedMessageException(result);
18415        }
18416        return result;
18417      }
18418
18419      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto buildPartial() {
18420        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto(this);
18421        int from_bitField0_ = bitField0_;
18422        int to_bitField0_ = 0;
18423        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
18424          to_bitField0_ |= 0x00000001;
18425        }
18426        if (jidBuilder_ == null) {
18427          result.jid_ = jid_;
18428        } else {
18429          result.jid_ = jidBuilder_.build();
18430        }
18431        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
18432          to_bitField0_ |= 0x00000002;
18433        }
18434        if (nsInfoBuilder_ == null) {
18435          result.nsInfo_ = nsInfo_;
18436        } else {
18437          result.nsInfo_ = nsInfoBuilder_.build();
18438        }
18439        result.bitField0_ = to_bitField0_;
18440        onBuilt();
18441        return result;
18442      }
18443
18444      public Builder mergeFrom(com.google.protobuf.Message other) {
18445        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto) {
18446          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)other);
18447        } else {
18448          super.mergeFrom(other);
18449          return this;
18450        }
18451      }
18452
18453      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto other) {
18454        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance()) return this;
18455        if (other.hasJid()) {
18456          mergeJid(other.getJid());
18457        }
18458        if (other.hasNsInfo()) {
18459          mergeNsInfo(other.getNsInfo());
18460        }
18461        this.mergeUnknownFields(other.getUnknownFields());
18462        return this;
18463      }
18464
18465      public final boolean isInitialized() {
18466        if (!hasJid()) {
18467          
18468          return false;
18469        }
18470        if (!hasNsInfo()) {
18471          
18472          return false;
18473        }
18474        if (!getJid().isInitialized()) {
18475          
18476          return false;
18477        }
18478        if (!getNsInfo().isInitialized()) {
18479          
18480          return false;
18481        }
18482        return true;
18483      }
18484
18485      public Builder mergeFrom(
18486          com.google.protobuf.CodedInputStream input,
18487          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18488          throws java.io.IOException {
18489        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parsedMessage = null;
18490        try {
18491          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
18492        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
18493          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto) e.getUnfinishedMessage();
18494          throw e;
18495        } finally {
18496          if (parsedMessage != null) {
18497            mergeFrom(parsedMessage);
18498          }
18499        }
18500        return this;
18501      }
18502      private int bitField0_;
18503
18504      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
18505      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
18506      private com.google.protobuf.SingleFieldBuilder<
18507          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
18508      /**
18509       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18510       */
18511      public boolean hasJid() {
18512        return ((bitField0_ & 0x00000001) == 0x00000001);
18513      }
18514      /**
18515       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18516       */
18517      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
18518        if (jidBuilder_ == null) {
18519          return jid_;
18520        } else {
18521          return jidBuilder_.getMessage();
18522        }
18523      }
18524      /**
18525       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18526       */
18527      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
18528        if (jidBuilder_ == null) {
18529          if (value == null) {
18530            throw new NullPointerException();
18531          }
18532          jid_ = value;
18533          onChanged();
18534        } else {
18535          jidBuilder_.setMessage(value);
18536        }
18537        bitField0_ |= 0x00000001;
18538        return this;
18539      }
18540      /**
18541       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18542       */
18543      public Builder setJid(
18544          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
18545        if (jidBuilder_ == null) {
18546          jid_ = builderForValue.build();
18547          onChanged();
18548        } else {
18549          jidBuilder_.setMessage(builderForValue.build());
18550        }
18551        bitField0_ |= 0x00000001;
18552        return this;
18553      }
18554      /**
18555       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18556       */
18557      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
18558        if (jidBuilder_ == null) {
18559          if (((bitField0_ & 0x00000001) == 0x00000001) &&
18560              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
18561            jid_ =
18562              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
18563          } else {
18564            jid_ = value;
18565          }
18566          onChanged();
18567        } else {
18568          jidBuilder_.mergeFrom(value);
18569        }
18570        bitField0_ |= 0x00000001;
18571        return this;
18572      }
18573      /**
18574       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18575       */
18576      public Builder clearJid() {
18577        if (jidBuilder_ == null) {
18578          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
18579          onChanged();
18580        } else {
18581          jidBuilder_.clear();
18582        }
18583        bitField0_ = (bitField0_ & ~0x00000001);
18584        return this;
18585      }
18586      /**
18587       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18588       */
18589      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
18590        bitField0_ |= 0x00000001;
18591        onChanged();
18592        return getJidFieldBuilder().getBuilder();
18593      }
18594      /**
18595       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18596       */
18597      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
18598        if (jidBuilder_ != null) {
18599          return jidBuilder_.getMessageOrBuilder();
18600        } else {
18601          return jid_;
18602        }
18603      }
18604      /**
18605       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
18606       */
18607      private com.google.protobuf.SingleFieldBuilder<
18608          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
18609          getJidFieldBuilder() {
18610        if (jidBuilder_ == null) {
18611          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
18612              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
18613                  jid_,
18614                  getParentForChildren(),
18615                  isClean());
18616          jid_ = null;
18617        }
18618        return jidBuilder_;
18619      }
18620
18621      // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
18622      private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.getDefaultInstance();
18623      private com.google.protobuf.SingleFieldBuilder<
18624          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder> nsInfoBuilder_;
18625      /**
18626       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18627       */
18628      public boolean hasNsInfo() {
18629        return ((bitField0_ & 0x00000002) == 0x00000002);
18630      }
18631      /**
18632       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18633       */
18634      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto getNsInfo() {
18635        if (nsInfoBuilder_ == null) {
18636          return nsInfo_;
18637        } else {
18638          return nsInfoBuilder_.getMessage();
18639        }
18640      }
18641      /**
18642       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18643       */
18644      public Builder setNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto value) {
18645        if (nsInfoBuilder_ == null) {
18646          if (value == null) {
18647            throw new NullPointerException();
18648          }
18649          nsInfo_ = value;
18650          onChanged();
18651        } else {
18652          nsInfoBuilder_.setMessage(value);
18653        }
18654        bitField0_ |= 0x00000002;
18655        return this;
18656      }
18657      /**
18658       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18659       */
18660      public Builder setNsInfo(
18661          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder builderForValue) {
18662        if (nsInfoBuilder_ == null) {
18663          nsInfo_ = builderForValue.build();
18664          onChanged();
18665        } else {
18666          nsInfoBuilder_.setMessage(builderForValue.build());
18667        }
18668        bitField0_ |= 0x00000002;
18669        return this;
18670      }
18671      /**
18672       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18673       */
18674      public Builder mergeNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto value) {
18675        if (nsInfoBuilder_ == null) {
18676          if (((bitField0_ & 0x00000002) == 0x00000002) &&
18677              nsInfo_ != org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.getDefaultInstance()) {
18678            nsInfo_ =
18679              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.newBuilder(nsInfo_).mergeFrom(value).buildPartial();
18680          } else {
18681            nsInfo_ = value;
18682          }
18683          onChanged();
18684        } else {
18685          nsInfoBuilder_.mergeFrom(value);
18686        }
18687        bitField0_ |= 0x00000002;
18688        return this;
18689      }
18690      /**
18691       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18692       */
18693      public Builder clearNsInfo() {
18694        if (nsInfoBuilder_ == null) {
18695          nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.getDefaultInstance();
18696          onChanged();
18697        } else {
18698          nsInfoBuilder_.clear();
18699        }
18700        bitField0_ = (bitField0_ & ~0x00000002);
18701        return this;
18702      }
18703      /**
18704       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18705       */
18706      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder getNsInfoBuilder() {
18707        bitField0_ |= 0x00000002;
18708        onChanged();
18709        return getNsInfoFieldBuilder().getBuilder();
18710      }
18711      /**
18712       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18713       */
18714      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
18715        if (nsInfoBuilder_ != null) {
18716          return nsInfoBuilder_.getMessageOrBuilder();
18717        } else {
18718          return nsInfo_;
18719        }
18720      }
18721      /**
18722       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
18723       */
18724      private com.google.protobuf.SingleFieldBuilder<
18725          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder> 
18726          getNsInfoFieldBuilder() {
18727        if (nsInfoBuilder_ == null) {
18728          nsInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
18729              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder>(
18730                  nsInfo_,
18731                  getParentForChildren(),
18732                  isClean());
18733          nsInfo_ = null;
18734        }
18735        return nsInfoBuilder_;
18736      }
18737
18738      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.FormatRequestProto)
18739    }
18740
18741    static {
18742      defaultInstance = new FormatRequestProto(true);
18743      defaultInstance.initFields();
18744    }
18745
18746    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.FormatRequestProto)
18747  }
18748
18749  public interface FormatResponseProtoOrBuilder
18750      extends com.google.protobuf.MessageOrBuilder {
18751  }
18752  /**
18753   * Protobuf type {@code hadoop.hdfs.qjournal.FormatResponseProto}
18754   */
18755  public static final class FormatResponseProto extends
18756      com.google.protobuf.GeneratedMessage
18757      implements FormatResponseProtoOrBuilder {
18758    // Use FormatResponseProto.newBuilder() to construct.
18759    private FormatResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
18760      super(builder);
18761      this.unknownFields = builder.getUnknownFields();
18762    }
18763    private FormatResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
18764
18765    private static final FormatResponseProto defaultInstance;
18766    public static FormatResponseProto getDefaultInstance() {
18767      return defaultInstance;
18768    }
18769
18770    public FormatResponseProto getDefaultInstanceForType() {
18771      return defaultInstance;
18772    }
18773
18774    private final com.google.protobuf.UnknownFieldSet unknownFields;
18775    @java.lang.Override
18776    public final com.google.protobuf.UnknownFieldSet
18777        getUnknownFields() {
18778      return this.unknownFields;
18779    }
18780    private FormatResponseProto(
18781        com.google.protobuf.CodedInputStream input,
18782        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18783        throws com.google.protobuf.InvalidProtocolBufferException {
18784      initFields();
18785      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
18786          com.google.protobuf.UnknownFieldSet.newBuilder();
18787      try {
18788        boolean done = false;
18789        while (!done) {
18790          int tag = input.readTag();
18791          switch (tag) {
18792            case 0:
18793              done = true;
18794              break;
18795            default: {
18796              if (!parseUnknownField(input, unknownFields,
18797                                     extensionRegistry, tag)) {
18798                done = true;
18799              }
18800              break;
18801            }
18802          }
18803        }
18804      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
18805        throw e.setUnfinishedMessage(this);
18806      } catch (java.io.IOException e) {
18807        throw new com.google.protobuf.InvalidProtocolBufferException(
18808            e.getMessage()).setUnfinishedMessage(this);
18809      } finally {
18810        this.unknownFields = unknownFields.build();
18811        makeExtensionsImmutable();
18812      }
18813    }
18814    public static final com.google.protobuf.Descriptors.Descriptor
18815        getDescriptor() {
18816      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FormatResponseProto_descriptor;
18817    }
18818
18819    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
18820        internalGetFieldAccessorTable() {
18821      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FormatResponseProto_fieldAccessorTable
18822          .ensureFieldAccessorsInitialized(
18823              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.Builder.class);
18824    }
18825
18826    public static com.google.protobuf.Parser<FormatResponseProto> PARSER =
18827        new com.google.protobuf.AbstractParser<FormatResponseProto>() {
18828      public FormatResponseProto parsePartialFrom(
18829          com.google.protobuf.CodedInputStream input,
18830          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18831          throws com.google.protobuf.InvalidProtocolBufferException {
18832        return new FormatResponseProto(input, extensionRegistry);
18833      }
18834    };
18835
18836    @java.lang.Override
18837    public com.google.protobuf.Parser<FormatResponseProto> getParserForType() {
18838      return PARSER;
18839    }
18840
18841    private void initFields() {
18842    }
18843    private byte memoizedIsInitialized = -1;
18844    public final boolean isInitialized() {
18845      byte isInitialized = memoizedIsInitialized;
18846      if (isInitialized != -1) return isInitialized == 1;
18847
18848      memoizedIsInitialized = 1;
18849      return true;
18850    }
18851
18852    public void writeTo(com.google.protobuf.CodedOutputStream output)
18853                        throws java.io.IOException {
18854      getSerializedSize();
18855      getUnknownFields().writeTo(output);
18856    }
18857
18858    private int memoizedSerializedSize = -1;
18859    public int getSerializedSize() {
18860      int size = memoizedSerializedSize;
18861      if (size != -1) return size;
18862
18863      size = 0;
18864      size += getUnknownFields().getSerializedSize();
18865      memoizedSerializedSize = size;
18866      return size;
18867    }
18868
18869    private static final long serialVersionUID = 0L;
18870    @java.lang.Override
18871    protected java.lang.Object writeReplace()
18872        throws java.io.ObjectStreamException {
18873      return super.writeReplace();
18874    }
18875
18876    @java.lang.Override
18877    public boolean equals(final java.lang.Object obj) {
18878      if (obj == this) {
18879       return true;
18880      }
18881      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto)) {
18882        return super.equals(obj);
18883      }
18884      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) obj;
18885
18886      boolean result = true;
18887      result = result &&
18888          getUnknownFields().equals(other.getUnknownFields());
18889      return result;
18890    }
18891
18892    private int memoizedHashCode = 0;
18893    @java.lang.Override
18894    public int hashCode() {
18895      if (memoizedHashCode != 0) {
18896        return memoizedHashCode;
18897      }
18898      int hash = 41;
18899      hash = (19 * hash) + getDescriptorForType().hashCode();
18900      hash = (29 * hash) + getUnknownFields().hashCode();
18901      memoizedHashCode = hash;
18902      return hash;
18903    }
18904
18905    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
18906        com.google.protobuf.ByteString data)
18907        throws com.google.protobuf.InvalidProtocolBufferException {
18908      return PARSER.parseFrom(data);
18909    }
18910    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
18911        com.google.protobuf.ByteString data,
18912        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18913        throws com.google.protobuf.InvalidProtocolBufferException {
18914      return PARSER.parseFrom(data, extensionRegistry);
18915    }
18916    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(byte[] data)
18917        throws com.google.protobuf.InvalidProtocolBufferException {
18918      return PARSER.parseFrom(data);
18919    }
18920    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
18921        byte[] data,
18922        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18923        throws com.google.protobuf.InvalidProtocolBufferException {
18924      return PARSER.parseFrom(data, extensionRegistry);
18925    }
18926    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(java.io.InputStream input)
18927        throws java.io.IOException {
18928      return PARSER.parseFrom(input);
18929    }
18930    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
18931        java.io.InputStream input,
18932        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18933        throws java.io.IOException {
18934      return PARSER.parseFrom(input, extensionRegistry);
18935    }
18936    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseDelimitedFrom(java.io.InputStream input)
18937        throws java.io.IOException {
18938      return PARSER.parseDelimitedFrom(input);
18939    }
18940    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseDelimitedFrom(
18941        java.io.InputStream input,
18942        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18943        throws java.io.IOException {
18944      return PARSER.parseDelimitedFrom(input, extensionRegistry);
18945    }
18946    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
18947        com.google.protobuf.CodedInputStream input)
18948        throws java.io.IOException {
18949      return PARSER.parseFrom(input);
18950    }
18951    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
18952        com.google.protobuf.CodedInputStream input,
18953        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
18954        throws java.io.IOException {
18955      return PARSER.parseFrom(input, extensionRegistry);
18956    }
18957
18958    public static Builder newBuilder() { return Builder.create(); }
18959    public Builder newBuilderForType() { return newBuilder(); }
18960    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto prototype) {
18961      return newBuilder().mergeFrom(prototype);
18962    }
18963    public Builder toBuilder() { return newBuilder(this); }
18964
18965    @java.lang.Override
18966    protected Builder newBuilderForType(
18967        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
18968      Builder builder = new Builder(parent);
18969      return builder;
18970    }
18971    /**
18972     * Protobuf type {@code hadoop.hdfs.qjournal.FormatResponseProto}
18973     */
18974    public static final class Builder extends
18975        com.google.protobuf.GeneratedMessage.Builder<Builder>
18976       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProtoOrBuilder {
18977      public static final com.google.protobuf.Descriptors.Descriptor
18978          getDescriptor() {
18979        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FormatResponseProto_descriptor;
18980      }
18981
18982      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
18983          internalGetFieldAccessorTable() {
18984        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FormatResponseProto_fieldAccessorTable
18985            .ensureFieldAccessorsInitialized(
18986                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.Builder.class);
18987      }
18988
18989      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.newBuilder()
18990      private Builder() {
18991        maybeForceBuilderInitialization();
18992      }
18993
18994      private Builder(
18995          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
18996        super(parent);
18997        maybeForceBuilderInitialization();
18998      }
18999      private void maybeForceBuilderInitialization() {
19000        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
19001        }
19002      }
19003      private static Builder create() {
19004        return new Builder();
19005      }
19006
19007      public Builder clear() {
19008        super.clear();
19009        return this;
19010      }
19011
19012      public Builder clone() {
19013        return create().mergeFrom(buildPartial());
19014      }
19015
19016      public com.google.protobuf.Descriptors.Descriptor
19017          getDescriptorForType() {
19018        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_FormatResponseProto_descriptor;
19019      }
19020
19021      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto getDefaultInstanceForType() {
19022        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance();
19023      }
19024
19025      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto build() {
19026        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto result = buildPartial();
19027        if (!result.isInitialized()) {
19028          throw newUninitializedMessageException(result);
19029        }
19030        return result;
19031      }
19032
19033      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto buildPartial() {
19034        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto(this);
19035        onBuilt();
19036        return result;
19037      }
19038
19039      public Builder mergeFrom(com.google.protobuf.Message other) {
19040        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) {
19041          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto)other);
19042        } else {
19043          super.mergeFrom(other);
19044          return this;
19045        }
19046      }
19047
19048      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto other) {
19049        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance()) return this;
19050        this.mergeUnknownFields(other.getUnknownFields());
19051        return this;
19052      }
19053
19054      public final boolean isInitialized() {
19055        return true;
19056      }
19057
19058      public Builder mergeFrom(
19059          com.google.protobuf.CodedInputStream input,
19060          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
19061          throws java.io.IOException {
19062        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parsedMessage = null;
19063        try {
19064          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
19065        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
19066          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) e.getUnfinishedMessage();
19067          throw e;
19068        } finally {
19069          if (parsedMessage != null) {
19070            mergeFrom(parsedMessage);
19071          }
19072        }
19073        return this;
19074      }
19075
19076      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.FormatResponseProto)
19077    }
19078
19079    static {
19080      defaultInstance = new FormatResponseProto(true);
19081      defaultInstance.initFields();
19082    }
19083
19084    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.FormatResponseProto)
19085  }
19086
19087  public interface NewEpochRequestProtoOrBuilder
19088      extends com.google.protobuf.MessageOrBuilder {
19089
19090    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
19091    /**
19092     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19093     */
19094    boolean hasJid();
19095    /**
19096     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19097     */
19098    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
19099    /**
19100     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19101     */
19102    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
19103
19104    // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
19105    /**
19106     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19107     */
19108    boolean hasNsInfo();
19109    /**
19110     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19111     */
19112    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto getNsInfo();
19113    /**
19114     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19115     */
19116    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder();
19117
19118    // required uint64 epoch = 3;
19119    /**
19120     * <code>required uint64 epoch = 3;</code>
19121     */
19122    boolean hasEpoch();
19123    /**
19124     * <code>required uint64 epoch = 3;</code>
19125     */
19126    long getEpoch();
19127  }
19128  /**
19129   * Protobuf type {@code hadoop.hdfs.qjournal.NewEpochRequestProto}
19130   *
19131   * <pre>
19132   **
19133   * newEpoch()
19134   * </pre>
19135   */
19136  public static final class NewEpochRequestProto extends
19137      com.google.protobuf.GeneratedMessage
19138      implements NewEpochRequestProtoOrBuilder {
19139    // Use NewEpochRequestProto.newBuilder() to construct.
19140    private NewEpochRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
19141      super(builder);
19142      this.unknownFields = builder.getUnknownFields();
19143    }
19144    private NewEpochRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
19145
19146    private static final NewEpochRequestProto defaultInstance;
19147    public static NewEpochRequestProto getDefaultInstance() {
19148      return defaultInstance;
19149    }
19150
19151    public NewEpochRequestProto getDefaultInstanceForType() {
19152      return defaultInstance;
19153    }
19154
19155    private final com.google.protobuf.UnknownFieldSet unknownFields;
19156    @java.lang.Override
19157    public final com.google.protobuf.UnknownFieldSet
19158        getUnknownFields() {
19159      return this.unknownFields;
19160    }
19161    private NewEpochRequestProto(
19162        com.google.protobuf.CodedInputStream input,
19163        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
19164        throws com.google.protobuf.InvalidProtocolBufferException {
19165      initFields();
19166      int mutable_bitField0_ = 0;
19167      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
19168          com.google.protobuf.UnknownFieldSet.newBuilder();
19169      try {
19170        boolean done = false;
19171        while (!done) {
19172          int tag = input.readTag();
19173          switch (tag) {
19174            case 0:
19175              done = true;
19176              break;
19177            default: {
19178              if (!parseUnknownField(input, unknownFields,
19179                                     extensionRegistry, tag)) {
19180                done = true;
19181              }
19182              break;
19183            }
19184            case 10: {
19185              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
19186              if (((bitField0_ & 0x00000001) == 0x00000001)) {
19187                subBuilder = jid_.toBuilder();
19188              }
19189              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
19190              if (subBuilder != null) {
19191                subBuilder.mergeFrom(jid_);
19192                jid_ = subBuilder.buildPartial();
19193              }
19194              bitField0_ |= 0x00000001;
19195              break;
19196            }
19197            case 18: {
19198              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder subBuilder = null;
19199              if (((bitField0_ & 0x00000002) == 0x00000002)) {
19200                subBuilder = nsInfo_.toBuilder();
19201              }
19202              nsInfo_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.PARSER, extensionRegistry);
19203              if (subBuilder != null) {
19204                subBuilder.mergeFrom(nsInfo_);
19205                nsInfo_ = subBuilder.buildPartial();
19206              }
19207              bitField0_ |= 0x00000002;
19208              break;
19209            }
19210            case 24: {
19211              bitField0_ |= 0x00000004;
19212              epoch_ = input.readUInt64();
19213              break;
19214            }
19215          }
19216        }
19217      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
19218        throw e.setUnfinishedMessage(this);
19219      } catch (java.io.IOException e) {
19220        throw new com.google.protobuf.InvalidProtocolBufferException(
19221            e.getMessage()).setUnfinishedMessage(this);
19222      } finally {
19223        this.unknownFields = unknownFields.build();
19224        makeExtensionsImmutable();
19225      }
19226    }
19227    public static final com.google.protobuf.Descriptors.Descriptor
19228        getDescriptor() {
19229      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_NewEpochRequestProto_descriptor;
19230    }
19231
19232    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
19233        internalGetFieldAccessorTable() {
19234      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_NewEpochRequestProto_fieldAccessorTable
19235          .ensureFieldAccessorsInitialized(
19236              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.Builder.class);
19237    }
19238
19239    public static com.google.protobuf.Parser<NewEpochRequestProto> PARSER =
19240        new com.google.protobuf.AbstractParser<NewEpochRequestProto>() {
19241      public NewEpochRequestProto parsePartialFrom(
19242          com.google.protobuf.CodedInputStream input,
19243          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
19244          throws com.google.protobuf.InvalidProtocolBufferException {
19245        return new NewEpochRequestProto(input, extensionRegistry);
19246      }
19247    };
19248
19249    @java.lang.Override
19250    public com.google.protobuf.Parser<NewEpochRequestProto> getParserForType() {
19251      return PARSER;
19252    }
19253
19254    private int bitField0_;
19255    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
19256    public static final int JID_FIELD_NUMBER = 1;
19257    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
19258    /**
19259     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19260     */
19261    public boolean hasJid() {
19262      return ((bitField0_ & 0x00000001) == 0x00000001);
19263    }
19264    /**
19265     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19266     */
19267    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
19268      return jid_;
19269    }
19270    /**
19271     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19272     */
19273    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
19274      return jid_;
19275    }
19276
19277    // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
19278    public static final int NSINFO_FIELD_NUMBER = 2;
19279    private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto nsInfo_;
19280    /**
19281     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19282     */
19283    public boolean hasNsInfo() {
19284      return ((bitField0_ & 0x00000002) == 0x00000002);
19285    }
19286    /**
19287     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19288     */
19289    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto getNsInfo() {
19290      return nsInfo_;
19291    }
19292    /**
19293     * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19294     */
19295    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
19296      return nsInfo_;
19297    }
19298
19299    // required uint64 epoch = 3;
19300    public static final int EPOCH_FIELD_NUMBER = 3;
19301    private long epoch_;
19302    /**
19303     * <code>required uint64 epoch = 3;</code>
19304     */
19305    public boolean hasEpoch() {
19306      return ((bitField0_ & 0x00000004) == 0x00000004);
19307    }
19308    /**
19309     * <code>required uint64 epoch = 3;</code>
19310     */
19311    public long getEpoch() {
19312      return epoch_;
19313    }
19314
19315    private void initFields() {
19316      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
19317      nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.getDefaultInstance();
19318      epoch_ = 0L;
19319    }
19320    private byte memoizedIsInitialized = -1;
19321    public final boolean isInitialized() {
19322      byte isInitialized = memoizedIsInitialized;
19323      if (isInitialized != -1) return isInitialized == 1;
19324
19325      if (!hasJid()) {
19326        memoizedIsInitialized = 0;
19327        return false;
19328      }
19329      if (!hasNsInfo()) {
19330        memoizedIsInitialized = 0;
19331        return false;
19332      }
19333      if (!hasEpoch()) {
19334        memoizedIsInitialized = 0;
19335        return false;
19336      }
19337      if (!getJid().isInitialized()) {
19338        memoizedIsInitialized = 0;
19339        return false;
19340      }
19341      if (!getNsInfo().isInitialized()) {
19342        memoizedIsInitialized = 0;
19343        return false;
19344      }
19345      memoizedIsInitialized = 1;
19346      return true;
19347    }
19348
19349    public void writeTo(com.google.protobuf.CodedOutputStream output)
19350                        throws java.io.IOException {
19351      getSerializedSize();
19352      if (((bitField0_ & 0x00000001) == 0x00000001)) {
19353        output.writeMessage(1, jid_);
19354      }
19355      if (((bitField0_ & 0x00000002) == 0x00000002)) {
19356        output.writeMessage(2, nsInfo_);
19357      }
19358      if (((bitField0_ & 0x00000004) == 0x00000004)) {
19359        output.writeUInt64(3, epoch_);
19360      }
19361      getUnknownFields().writeTo(output);
19362    }
19363
19364    private int memoizedSerializedSize = -1;
19365    public int getSerializedSize() {
19366      int size = memoizedSerializedSize;
19367      if (size != -1) return size;
19368
19369      size = 0;
19370      if (((bitField0_ & 0x00000001) == 0x00000001)) {
19371        size += com.google.protobuf.CodedOutputStream
19372          .computeMessageSize(1, jid_);
19373      }
19374      if (((bitField0_ & 0x00000002) == 0x00000002)) {
19375        size += com.google.protobuf.CodedOutputStream
19376          .computeMessageSize(2, nsInfo_);
19377      }
19378      if (((bitField0_ & 0x00000004) == 0x00000004)) {
19379        size += com.google.protobuf.CodedOutputStream
19380          .computeUInt64Size(3, epoch_);
19381      }
19382      size += getUnknownFields().getSerializedSize();
19383      memoizedSerializedSize = size;
19384      return size;
19385    }
19386
19387    private static final long serialVersionUID = 0L;
19388    @java.lang.Override
19389    protected java.lang.Object writeReplace()
19390        throws java.io.ObjectStreamException {
19391      return super.writeReplace();
19392    }
19393
19394    @java.lang.Override
19395    public boolean equals(final java.lang.Object obj) {
19396      if (obj == this) {
19397       return true;
19398      }
19399      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)) {
19400        return super.equals(obj);
19401      }
19402      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto) obj;
19403
19404      boolean result = true;
19405      result = result && (hasJid() == other.hasJid());
19406      if (hasJid()) {
19407        result = result && getJid()
19408            .equals(other.getJid());
19409      }
19410      result = result && (hasNsInfo() == other.hasNsInfo());
19411      if (hasNsInfo()) {
19412        result = result && getNsInfo()
19413            .equals(other.getNsInfo());
19414      }
19415      result = result && (hasEpoch() == other.hasEpoch());
19416      if (hasEpoch()) {
19417        result = result && (getEpoch()
19418            == other.getEpoch());
19419      }
19420      result = result &&
19421          getUnknownFields().equals(other.getUnknownFields());
19422      return result;
19423    }
19424
19425    private int memoizedHashCode = 0;
19426    @java.lang.Override
19427    public int hashCode() {
19428      if (memoizedHashCode != 0) {
19429        return memoizedHashCode;
19430      }
19431      int hash = 41;
19432      hash = (19 * hash) + getDescriptorForType().hashCode();
19433      if (hasJid()) {
19434        hash = (37 * hash) + JID_FIELD_NUMBER;
19435        hash = (53 * hash) + getJid().hashCode();
19436      }
19437      if (hasNsInfo()) {
19438        hash = (37 * hash) + NSINFO_FIELD_NUMBER;
19439        hash = (53 * hash) + getNsInfo().hashCode();
19440      }
19441      if (hasEpoch()) {
19442        hash = (37 * hash) + EPOCH_FIELD_NUMBER;
19443        hash = (53 * hash) + hashLong(getEpoch());
19444      }
19445      hash = (29 * hash) + getUnknownFields().hashCode();
19446      memoizedHashCode = hash;
19447      return hash;
19448    }
19449
19450    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
19451        com.google.protobuf.ByteString data)
19452        throws com.google.protobuf.InvalidProtocolBufferException {
19453      return PARSER.parseFrom(data);
19454    }
19455    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
19456        com.google.protobuf.ByteString data,
19457        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
19458        throws com.google.protobuf.InvalidProtocolBufferException {
19459      return PARSER.parseFrom(data, extensionRegistry);
19460    }
19461    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(byte[] data)
19462        throws com.google.protobuf.InvalidProtocolBufferException {
19463      return PARSER.parseFrom(data);
19464    }
19465    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
19466        byte[] data,
19467        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
19468        throws com.google.protobuf.InvalidProtocolBufferException {
19469      return PARSER.parseFrom(data, extensionRegistry);
19470    }
19471    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(java.io.InputStream input)
19472        throws java.io.IOException {
19473      return PARSER.parseFrom(input);
19474    }
19475    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
19476        java.io.InputStream input,
19477        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
19478        throws java.io.IOException {
19479      return PARSER.parseFrom(input, extensionRegistry);
19480    }
19481    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseDelimitedFrom(java.io.InputStream input)
19482        throws java.io.IOException {
19483      return PARSER.parseDelimitedFrom(input);
19484    }
19485    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseDelimitedFrom(
19486        java.io.InputStream input,
19487        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
19488        throws java.io.IOException {
19489      return PARSER.parseDelimitedFrom(input, extensionRegistry);
19490    }
19491    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
19492        com.google.protobuf.CodedInputStream input)
19493        throws java.io.IOException {
19494      return PARSER.parseFrom(input);
19495    }
19496    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
19497        com.google.protobuf.CodedInputStream input,
19498        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
19499        throws java.io.IOException {
19500      return PARSER.parseFrom(input, extensionRegistry);
19501    }
19502
19503    public static Builder newBuilder() { return Builder.create(); }
19504    public Builder newBuilderForType() { return newBuilder(); }
19505    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto prototype) {
19506      return newBuilder().mergeFrom(prototype);
19507    }
19508    public Builder toBuilder() { return newBuilder(this); }
19509
19510    @java.lang.Override
19511    protected Builder newBuilderForType(
19512        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
19513      Builder builder = new Builder(parent);
19514      return builder;
19515    }
19516    /**
19517     * Protobuf type {@code hadoop.hdfs.qjournal.NewEpochRequestProto}
19518     *
19519     * <pre>
19520     **
19521     * newEpoch()
19522     * </pre>
19523     */
19524    public static final class Builder extends
19525        com.google.protobuf.GeneratedMessage.Builder<Builder>
19526       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProtoOrBuilder {
19527      public static final com.google.protobuf.Descriptors.Descriptor
19528          getDescriptor() {
19529        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_NewEpochRequestProto_descriptor;
19530      }
19531
19532      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
19533          internalGetFieldAccessorTable() {
19534        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_NewEpochRequestProto_fieldAccessorTable
19535            .ensureFieldAccessorsInitialized(
19536                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.Builder.class);
19537      }
19538
19539      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.newBuilder()
19540      private Builder() {
19541        maybeForceBuilderInitialization();
19542      }
19543
19544      private Builder(
19545          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
19546        super(parent);
19547        maybeForceBuilderInitialization();
19548      }
19549      private void maybeForceBuilderInitialization() {
19550        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
19551          getJidFieldBuilder();
19552          getNsInfoFieldBuilder();
19553        }
19554      }
19555      private static Builder create() {
19556        return new Builder();
19557      }
19558
19559      public Builder clear() {
19560        super.clear();
19561        if (jidBuilder_ == null) {
19562          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
19563        } else {
19564          jidBuilder_.clear();
19565        }
19566        bitField0_ = (bitField0_ & ~0x00000001);
19567        if (nsInfoBuilder_ == null) {
19568          nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.getDefaultInstance();
19569        } else {
19570          nsInfoBuilder_.clear();
19571        }
19572        bitField0_ = (bitField0_ & ~0x00000002);
19573        epoch_ = 0L;
19574        bitField0_ = (bitField0_ & ~0x00000004);
19575        return this;
19576      }
19577
19578      public Builder clone() {
19579        return create().mergeFrom(buildPartial());
19580      }
19581
19582      public com.google.protobuf.Descriptors.Descriptor
19583          getDescriptorForType() {
19584        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_NewEpochRequestProto_descriptor;
19585      }
19586
19587      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto getDefaultInstanceForType() {
19588        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance();
19589      }
19590
19591      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto build() {
19592        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto result = buildPartial();
19593        if (!result.isInitialized()) {
19594          throw newUninitializedMessageException(result);
19595        }
19596        return result;
19597      }
19598
19599      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto buildPartial() {
19600        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto(this);
19601        int from_bitField0_ = bitField0_;
19602        int to_bitField0_ = 0;
19603        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
19604          to_bitField0_ |= 0x00000001;
19605        }
19606        if (jidBuilder_ == null) {
19607          result.jid_ = jid_;
19608        } else {
19609          result.jid_ = jidBuilder_.build();
19610        }
19611        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
19612          to_bitField0_ |= 0x00000002;
19613        }
19614        if (nsInfoBuilder_ == null) {
19615          result.nsInfo_ = nsInfo_;
19616        } else {
19617          result.nsInfo_ = nsInfoBuilder_.build();
19618        }
19619        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
19620          to_bitField0_ |= 0x00000004;
19621        }
19622        result.epoch_ = epoch_;
19623        result.bitField0_ = to_bitField0_;
19624        onBuilt();
19625        return result;
19626      }
19627
19628      public Builder mergeFrom(com.google.protobuf.Message other) {
19629        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto) {
19630          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)other);
19631        } else {
19632          super.mergeFrom(other);
19633          return this;
19634        }
19635      }
19636
19637      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto other) {
19638        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance()) return this;
19639        if (other.hasJid()) {
19640          mergeJid(other.getJid());
19641        }
19642        if (other.hasNsInfo()) {
19643          mergeNsInfo(other.getNsInfo());
19644        }
19645        if (other.hasEpoch()) {
19646          setEpoch(other.getEpoch());
19647        }
19648        this.mergeUnknownFields(other.getUnknownFields());
19649        return this;
19650      }
19651
19652      public final boolean isInitialized() {
19653        if (!hasJid()) {
19654          
19655          return false;
19656        }
19657        if (!hasNsInfo()) {
19658          
19659          return false;
19660        }
19661        if (!hasEpoch()) {
19662          
19663          return false;
19664        }
19665        if (!getJid().isInitialized()) {
19666          
19667          return false;
19668        }
19669        if (!getNsInfo().isInitialized()) {
19670          
19671          return false;
19672        }
19673        return true;
19674      }
19675
19676      public Builder mergeFrom(
19677          com.google.protobuf.CodedInputStream input,
19678          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
19679          throws java.io.IOException {
19680        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parsedMessage = null;
19681        try {
19682          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
19683        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
19684          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto) e.getUnfinishedMessage();
19685          throw e;
19686        } finally {
19687          if (parsedMessage != null) {
19688            mergeFrom(parsedMessage);
19689          }
19690        }
19691        return this;
19692      }
19693      private int bitField0_;
19694
19695      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
19696      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
19697      private com.google.protobuf.SingleFieldBuilder<
19698          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
19699      /**
19700       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19701       */
19702      public boolean hasJid() {
19703        return ((bitField0_ & 0x00000001) == 0x00000001);
19704      }
19705      /**
19706       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19707       */
19708      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
19709        if (jidBuilder_ == null) {
19710          return jid_;
19711        } else {
19712          return jidBuilder_.getMessage();
19713        }
19714      }
19715      /**
19716       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19717       */
19718      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
19719        if (jidBuilder_ == null) {
19720          if (value == null) {
19721            throw new NullPointerException();
19722          }
19723          jid_ = value;
19724          onChanged();
19725        } else {
19726          jidBuilder_.setMessage(value);
19727        }
19728        bitField0_ |= 0x00000001;
19729        return this;
19730      }
19731      /**
19732       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19733       */
19734      public Builder setJid(
19735          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
19736        if (jidBuilder_ == null) {
19737          jid_ = builderForValue.build();
19738          onChanged();
19739        } else {
19740          jidBuilder_.setMessage(builderForValue.build());
19741        }
19742        bitField0_ |= 0x00000001;
19743        return this;
19744      }
19745      /**
19746       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19747       */
19748      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
19749        if (jidBuilder_ == null) {
19750          if (((bitField0_ & 0x00000001) == 0x00000001) &&
19751              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
19752            jid_ =
19753              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
19754          } else {
19755            jid_ = value;
19756          }
19757          onChanged();
19758        } else {
19759          jidBuilder_.mergeFrom(value);
19760        }
19761        bitField0_ |= 0x00000001;
19762        return this;
19763      }
19764      /**
19765       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19766       */
19767      public Builder clearJid() {
19768        if (jidBuilder_ == null) {
19769          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
19770          onChanged();
19771        } else {
19772          jidBuilder_.clear();
19773        }
19774        bitField0_ = (bitField0_ & ~0x00000001);
19775        return this;
19776      }
19777      /**
19778       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19779       */
19780      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
19781        bitField0_ |= 0x00000001;
19782        onChanged();
19783        return getJidFieldBuilder().getBuilder();
19784      }
19785      /**
19786       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19787       */
19788      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
19789        if (jidBuilder_ != null) {
19790          return jidBuilder_.getMessageOrBuilder();
19791        } else {
19792          return jid_;
19793        }
19794      }
19795      /**
19796       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
19797       */
19798      private com.google.protobuf.SingleFieldBuilder<
19799          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
19800          getJidFieldBuilder() {
19801        if (jidBuilder_ == null) {
19802          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
19803              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
19804                  jid_,
19805                  getParentForChildren(),
19806                  isClean());
19807          jid_ = null;
19808        }
19809        return jidBuilder_;
19810      }
19811
19812      // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
19813      private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.getDefaultInstance();
19814      private com.google.protobuf.SingleFieldBuilder<
19815          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder> nsInfoBuilder_;
19816      /**
19817       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19818       */
19819      public boolean hasNsInfo() {
19820        return ((bitField0_ & 0x00000002) == 0x00000002);
19821      }
19822      /**
19823       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19824       */
19825      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto getNsInfo() {
19826        if (nsInfoBuilder_ == null) {
19827          return nsInfo_;
19828        } else {
19829          return nsInfoBuilder_.getMessage();
19830        }
19831      }
19832      /**
19833       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19834       */
19835      public Builder setNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto value) {
19836        if (nsInfoBuilder_ == null) {
19837          if (value == null) {
19838            throw new NullPointerException();
19839          }
19840          nsInfo_ = value;
19841          onChanged();
19842        } else {
19843          nsInfoBuilder_.setMessage(value);
19844        }
19845        bitField0_ |= 0x00000002;
19846        return this;
19847      }
19848      /**
19849       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19850       */
19851      public Builder setNsInfo(
19852          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder builderForValue) {
19853        if (nsInfoBuilder_ == null) {
19854          nsInfo_ = builderForValue.build();
19855          onChanged();
19856        } else {
19857          nsInfoBuilder_.setMessage(builderForValue.build());
19858        }
19859        bitField0_ |= 0x00000002;
19860        return this;
19861      }
19862      /**
19863       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19864       */
19865      public Builder mergeNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto value) {
19866        if (nsInfoBuilder_ == null) {
19867          if (((bitField0_ & 0x00000002) == 0x00000002) &&
19868              nsInfo_ != org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.getDefaultInstance()) {
19869            nsInfo_ =
19870              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.newBuilder(nsInfo_).mergeFrom(value).buildPartial();
19871          } else {
19872            nsInfo_ = value;
19873          }
19874          onChanged();
19875        } else {
19876          nsInfoBuilder_.mergeFrom(value);
19877        }
19878        bitField0_ |= 0x00000002;
19879        return this;
19880      }
19881      /**
19882       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19883       */
19884      public Builder clearNsInfo() {
19885        if (nsInfoBuilder_ == null) {
19886          nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.getDefaultInstance();
19887          onChanged();
19888        } else {
19889          nsInfoBuilder_.clear();
19890        }
19891        bitField0_ = (bitField0_ & ~0x00000002);
19892        return this;
19893      }
19894      /**
19895       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19896       */
19897      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder getNsInfoBuilder() {
19898        bitField0_ |= 0x00000002;
19899        onChanged();
19900        return getNsInfoFieldBuilder().getBuilder();
19901      }
19902      /**
19903       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19904       */
19905      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
19906        if (nsInfoBuilder_ != null) {
19907          return nsInfoBuilder_.getMessageOrBuilder();
19908        } else {
19909          return nsInfo_;
19910        }
19911      }
19912      /**
19913       * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
19914       */
19915      private com.google.protobuf.SingleFieldBuilder<
19916          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder> 
19917          getNsInfoFieldBuilder() {
19918        if (nsInfoBuilder_ == null) {
19919          nsInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
19920              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProtoOrBuilder>(
19921                  nsInfo_,
19922                  getParentForChildren(),
19923                  isClean());
19924          nsInfo_ = null;
19925        }
19926        return nsInfoBuilder_;
19927      }
19928
19929      // required uint64 epoch = 3;
19930      private long epoch_ ;
19931      /**
19932       * <code>required uint64 epoch = 3;</code>
19933       */
19934      public boolean hasEpoch() {
19935        return ((bitField0_ & 0x00000004) == 0x00000004);
19936      }
19937      /**
19938       * <code>required uint64 epoch = 3;</code>
19939       */
19940      public long getEpoch() {
19941        return epoch_;
19942      }
19943      /**
19944       * <code>required uint64 epoch = 3;</code>
19945       */
19946      public Builder setEpoch(long value) {
19947        bitField0_ |= 0x00000004;
19948        epoch_ = value;
19949        onChanged();
19950        return this;
19951      }
19952      /**
19953       * <code>required uint64 epoch = 3;</code>
19954       */
19955      public Builder clearEpoch() {
19956        bitField0_ = (bitField0_ & ~0x00000004);
19957        epoch_ = 0L;
19958        onChanged();
19959        return this;
19960      }
19961
19962      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.NewEpochRequestProto)
19963    }
19964
19965    static {
19966      defaultInstance = new NewEpochRequestProto(true);
19967      defaultInstance.initFields();
19968    }
19969
19970    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.NewEpochRequestProto)
19971  }
19972
19973  public interface NewEpochResponseProtoOrBuilder
19974      extends com.google.protobuf.MessageOrBuilder {
19975
19976    // optional uint64 lastSegmentTxId = 1;
19977    /**
19978     * <code>optional uint64 lastSegmentTxId = 1;</code>
19979     */
19980    boolean hasLastSegmentTxId();
19981    /**
19982     * <code>optional uint64 lastSegmentTxId = 1;</code>
19983     */
19984    long getLastSegmentTxId();
19985  }
19986  /**
19987   * Protobuf type {@code hadoop.hdfs.qjournal.NewEpochResponseProto}
19988   */
19989  public static final class NewEpochResponseProto extends
19990      com.google.protobuf.GeneratedMessage
19991      implements NewEpochResponseProtoOrBuilder {
19992    // Use NewEpochResponseProto.newBuilder() to construct.
19993    private NewEpochResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
19994      super(builder);
19995      this.unknownFields = builder.getUnknownFields();
19996    }
19997    private NewEpochResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
19998
19999    private static final NewEpochResponseProto defaultInstance;
20000    public static NewEpochResponseProto getDefaultInstance() {
20001      return defaultInstance;
20002    }
20003
20004    public NewEpochResponseProto getDefaultInstanceForType() {
20005      return defaultInstance;
20006    }
20007
20008    private final com.google.protobuf.UnknownFieldSet unknownFields;
20009    @java.lang.Override
20010    public final com.google.protobuf.UnknownFieldSet
20011        getUnknownFields() {
20012      return this.unknownFields;
20013    }
20014    private NewEpochResponseProto(
20015        com.google.protobuf.CodedInputStream input,
20016        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20017        throws com.google.protobuf.InvalidProtocolBufferException {
20018      initFields();
20019      int mutable_bitField0_ = 0;
20020      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
20021          com.google.protobuf.UnknownFieldSet.newBuilder();
20022      try {
20023        boolean done = false;
20024        while (!done) {
20025          int tag = input.readTag();
20026          switch (tag) {
20027            case 0:
20028              done = true;
20029              break;
20030            default: {
20031              if (!parseUnknownField(input, unknownFields,
20032                                     extensionRegistry, tag)) {
20033                done = true;
20034              }
20035              break;
20036            }
20037            case 8: {
20038              bitField0_ |= 0x00000001;
20039              lastSegmentTxId_ = input.readUInt64();
20040              break;
20041            }
20042          }
20043        }
20044      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
20045        throw e.setUnfinishedMessage(this);
20046      } catch (java.io.IOException e) {
20047        throw new com.google.protobuf.InvalidProtocolBufferException(
20048            e.getMessage()).setUnfinishedMessage(this);
20049      } finally {
20050        this.unknownFields = unknownFields.build();
20051        makeExtensionsImmutable();
20052      }
20053    }
20054    public static final com.google.protobuf.Descriptors.Descriptor
20055        getDescriptor() {
20056      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_NewEpochResponseProto_descriptor;
20057    }
20058
20059    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
20060        internalGetFieldAccessorTable() {
20061      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_NewEpochResponseProto_fieldAccessorTable
20062          .ensureFieldAccessorsInitialized(
20063              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.Builder.class);
20064    }
20065
20066    public static com.google.protobuf.Parser<NewEpochResponseProto> PARSER =
20067        new com.google.protobuf.AbstractParser<NewEpochResponseProto>() {
20068      public NewEpochResponseProto parsePartialFrom(
20069          com.google.protobuf.CodedInputStream input,
20070          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20071          throws com.google.protobuf.InvalidProtocolBufferException {
20072        return new NewEpochResponseProto(input, extensionRegistry);
20073      }
20074    };
20075
20076    @java.lang.Override
20077    public com.google.protobuf.Parser<NewEpochResponseProto> getParserForType() {
20078      return PARSER;
20079    }
20080
20081    private int bitField0_;
20082    // optional uint64 lastSegmentTxId = 1;
20083    public static final int LASTSEGMENTTXID_FIELD_NUMBER = 1;
20084    private long lastSegmentTxId_;
20085    /**
20086     * <code>optional uint64 lastSegmentTxId = 1;</code>
20087     */
20088    public boolean hasLastSegmentTxId() {
20089      return ((bitField0_ & 0x00000001) == 0x00000001);
20090    }
20091    /**
20092     * <code>optional uint64 lastSegmentTxId = 1;</code>
20093     */
20094    public long getLastSegmentTxId() {
20095      return lastSegmentTxId_;
20096    }
20097
20098    private void initFields() {
20099      lastSegmentTxId_ = 0L;
20100    }
20101    private byte memoizedIsInitialized = -1;
20102    public final boolean isInitialized() {
20103      byte isInitialized = memoizedIsInitialized;
20104      if (isInitialized != -1) return isInitialized == 1;
20105
20106      memoizedIsInitialized = 1;
20107      return true;
20108    }
20109
20110    public void writeTo(com.google.protobuf.CodedOutputStream output)
20111                        throws java.io.IOException {
20112      getSerializedSize();
20113      if (((bitField0_ & 0x00000001) == 0x00000001)) {
20114        output.writeUInt64(1, lastSegmentTxId_);
20115      }
20116      getUnknownFields().writeTo(output);
20117    }
20118
20119    private int memoizedSerializedSize = -1;
20120    public int getSerializedSize() {
20121      int size = memoizedSerializedSize;
20122      if (size != -1) return size;
20123
20124      size = 0;
20125      if (((bitField0_ & 0x00000001) == 0x00000001)) {
20126        size += com.google.protobuf.CodedOutputStream
20127          .computeUInt64Size(1, lastSegmentTxId_);
20128      }
20129      size += getUnknownFields().getSerializedSize();
20130      memoizedSerializedSize = size;
20131      return size;
20132    }
20133
20134    private static final long serialVersionUID = 0L;
20135    @java.lang.Override
20136    protected java.lang.Object writeReplace()
20137        throws java.io.ObjectStreamException {
20138      return super.writeReplace();
20139    }
20140
20141    @java.lang.Override
20142    public boolean equals(final java.lang.Object obj) {
20143      if (obj == this) {
20144       return true;
20145      }
20146      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto)) {
20147        return super.equals(obj);
20148      }
20149      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) obj;
20150
20151      boolean result = true;
20152      result = result && (hasLastSegmentTxId() == other.hasLastSegmentTxId());
20153      if (hasLastSegmentTxId()) {
20154        result = result && (getLastSegmentTxId()
20155            == other.getLastSegmentTxId());
20156      }
20157      result = result &&
20158          getUnknownFields().equals(other.getUnknownFields());
20159      return result;
20160    }
20161
20162    private int memoizedHashCode = 0;
20163    @java.lang.Override
20164    public int hashCode() {
20165      if (memoizedHashCode != 0) {
20166        return memoizedHashCode;
20167      }
20168      int hash = 41;
20169      hash = (19 * hash) + getDescriptorForType().hashCode();
20170      if (hasLastSegmentTxId()) {
20171        hash = (37 * hash) + LASTSEGMENTTXID_FIELD_NUMBER;
20172        hash = (53 * hash) + hashLong(getLastSegmentTxId());
20173      }
20174      hash = (29 * hash) + getUnknownFields().hashCode();
20175      memoizedHashCode = hash;
20176      return hash;
20177    }
20178
20179    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
20180        com.google.protobuf.ByteString data)
20181        throws com.google.protobuf.InvalidProtocolBufferException {
20182      return PARSER.parseFrom(data);
20183    }
20184    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
20185        com.google.protobuf.ByteString data,
20186        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20187        throws com.google.protobuf.InvalidProtocolBufferException {
20188      return PARSER.parseFrom(data, extensionRegistry);
20189    }
20190    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(byte[] data)
20191        throws com.google.protobuf.InvalidProtocolBufferException {
20192      return PARSER.parseFrom(data);
20193    }
20194    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
20195        byte[] data,
20196        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20197        throws com.google.protobuf.InvalidProtocolBufferException {
20198      return PARSER.parseFrom(data, extensionRegistry);
20199    }
20200    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(java.io.InputStream input)
20201        throws java.io.IOException {
20202      return PARSER.parseFrom(input);
20203    }
20204    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
20205        java.io.InputStream input,
20206        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20207        throws java.io.IOException {
20208      return PARSER.parseFrom(input, extensionRegistry);
20209    }
20210    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseDelimitedFrom(java.io.InputStream input)
20211        throws java.io.IOException {
20212      return PARSER.parseDelimitedFrom(input);
20213    }
20214    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseDelimitedFrom(
20215        java.io.InputStream input,
20216        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20217        throws java.io.IOException {
20218      return PARSER.parseDelimitedFrom(input, extensionRegistry);
20219    }
20220    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
20221        com.google.protobuf.CodedInputStream input)
20222        throws java.io.IOException {
20223      return PARSER.parseFrom(input);
20224    }
20225    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
20226        com.google.protobuf.CodedInputStream input,
20227        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20228        throws java.io.IOException {
20229      return PARSER.parseFrom(input, extensionRegistry);
20230    }
20231
20232    public static Builder newBuilder() { return Builder.create(); }
20233    public Builder newBuilderForType() { return newBuilder(); }
20234    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto prototype) {
20235      return newBuilder().mergeFrom(prototype);
20236    }
20237    public Builder toBuilder() { return newBuilder(this); }
20238
20239    @java.lang.Override
20240    protected Builder newBuilderForType(
20241        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
20242      Builder builder = new Builder(parent);
20243      return builder;
20244    }
20245    /**
20246     * Protobuf type {@code hadoop.hdfs.qjournal.NewEpochResponseProto}
20247     */
20248    public static final class Builder extends
20249        com.google.protobuf.GeneratedMessage.Builder<Builder>
20250       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProtoOrBuilder {
20251      public static final com.google.protobuf.Descriptors.Descriptor
20252          getDescriptor() {
20253        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_NewEpochResponseProto_descriptor;
20254      }
20255
20256      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
20257          internalGetFieldAccessorTable() {
20258        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_NewEpochResponseProto_fieldAccessorTable
20259            .ensureFieldAccessorsInitialized(
20260                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.Builder.class);
20261      }
20262
20263      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.newBuilder()
20264      private Builder() {
20265        maybeForceBuilderInitialization();
20266      }
20267
20268      private Builder(
20269          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
20270        super(parent);
20271        maybeForceBuilderInitialization();
20272      }
20273      private void maybeForceBuilderInitialization() {
20274        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
20275        }
20276      }
20277      private static Builder create() {
20278        return new Builder();
20279      }
20280
20281      public Builder clear() {
20282        super.clear();
20283        lastSegmentTxId_ = 0L;
20284        bitField0_ = (bitField0_ & ~0x00000001);
20285        return this;
20286      }
20287
20288      public Builder clone() {
20289        return create().mergeFrom(buildPartial());
20290      }
20291
20292      public com.google.protobuf.Descriptors.Descriptor
20293          getDescriptorForType() {
20294        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_NewEpochResponseProto_descriptor;
20295      }
20296
20297      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto getDefaultInstanceForType() {
20298        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance();
20299      }
20300
20301      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto build() {
20302        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto result = buildPartial();
20303        if (!result.isInitialized()) {
20304          throw newUninitializedMessageException(result);
20305        }
20306        return result;
20307      }
20308
20309      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto buildPartial() {
20310        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto(this);
20311        int from_bitField0_ = bitField0_;
20312        int to_bitField0_ = 0;
20313        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
20314          to_bitField0_ |= 0x00000001;
20315        }
20316        result.lastSegmentTxId_ = lastSegmentTxId_;
20317        result.bitField0_ = to_bitField0_;
20318        onBuilt();
20319        return result;
20320      }
20321
20322      public Builder mergeFrom(com.google.protobuf.Message other) {
20323        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) {
20324          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto)other);
20325        } else {
20326          super.mergeFrom(other);
20327          return this;
20328        }
20329      }
20330
20331      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto other) {
20332        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance()) return this;
20333        if (other.hasLastSegmentTxId()) {
20334          setLastSegmentTxId(other.getLastSegmentTxId());
20335        }
20336        this.mergeUnknownFields(other.getUnknownFields());
20337        return this;
20338      }
20339
20340      public final boolean isInitialized() {
20341        return true;
20342      }
20343
20344      public Builder mergeFrom(
20345          com.google.protobuf.CodedInputStream input,
20346          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20347          throws java.io.IOException {
20348        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parsedMessage = null;
20349        try {
20350          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
20351        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
20352          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) e.getUnfinishedMessage();
20353          throw e;
20354        } finally {
20355          if (parsedMessage != null) {
20356            mergeFrom(parsedMessage);
20357          }
20358        }
20359        return this;
20360      }
20361      private int bitField0_;
20362
20363      // optional uint64 lastSegmentTxId = 1;
20364      private long lastSegmentTxId_ ;
20365      /**
20366       * <code>optional uint64 lastSegmentTxId = 1;</code>
20367       */
20368      public boolean hasLastSegmentTxId() {
20369        return ((bitField0_ & 0x00000001) == 0x00000001);
20370      }
20371      /**
20372       * <code>optional uint64 lastSegmentTxId = 1;</code>
20373       */
20374      public long getLastSegmentTxId() {
20375        return lastSegmentTxId_;
20376      }
20377      /**
20378       * <code>optional uint64 lastSegmentTxId = 1;</code>
20379       */
20380      public Builder setLastSegmentTxId(long value) {
20381        bitField0_ |= 0x00000001;
20382        lastSegmentTxId_ = value;
20383        onChanged();
20384        return this;
20385      }
20386      /**
20387       * <code>optional uint64 lastSegmentTxId = 1;</code>
20388       */
20389      public Builder clearLastSegmentTxId() {
20390        bitField0_ = (bitField0_ & ~0x00000001);
20391        lastSegmentTxId_ = 0L;
20392        onChanged();
20393        return this;
20394      }
20395
20396      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.NewEpochResponseProto)
20397    }
20398
20399    static {
20400      defaultInstance = new NewEpochResponseProto(true);
20401      defaultInstance.initFields();
20402    }
20403
20404    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.NewEpochResponseProto)
20405  }
20406
20407  public interface GetEditLogManifestRequestProtoOrBuilder
20408      extends com.google.protobuf.MessageOrBuilder {
20409
20410    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
20411    /**
20412     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
20413     */
20414    boolean hasJid();
20415    /**
20416     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
20417     */
20418    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
20419    /**
20420     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
20421     */
20422    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
20423
20424    // required uint64 sinceTxId = 2;
20425    /**
20426     * <code>required uint64 sinceTxId = 2;</code>
20427     *
20428     * <pre>
20429     * Transaction ID
20430     * </pre>
20431     */
20432    boolean hasSinceTxId();
20433    /**
20434     * <code>required uint64 sinceTxId = 2;</code>
20435     *
20436     * <pre>
20437     * Transaction ID
20438     * </pre>
20439     */
20440    long getSinceTxId();
20441
20442    // optional bool inProgressOk = 4 [default = false];
20443    /**
20444     * <code>optional bool inProgressOk = 4 [default = false];</code>
20445     *
20446     * <pre>
20447     * Whether or not the client will be reading from the returned streams.
20448     * optional bool forReading = 3 [default = true]; &lt;obsolete, do not reuse&gt;
20449     * </pre>
20450     */
20451    boolean hasInProgressOk();
20452    /**
20453     * <code>optional bool inProgressOk = 4 [default = false];</code>
20454     *
20455     * <pre>
20456     * Whether or not the client will be reading from the returned streams.
20457     * optional bool forReading = 3 [default = true]; &lt;obsolete, do not reuse&gt;
20458     * </pre>
20459     */
20460    boolean getInProgressOk();
20461  }
20462  /**
20463   * Protobuf type {@code hadoop.hdfs.qjournal.GetEditLogManifestRequestProto}
20464   *
20465   * <pre>
20466   **
20467   * getEditLogManifest()
20468   * </pre>
20469   */
20470  public static final class GetEditLogManifestRequestProto extends
20471      com.google.protobuf.GeneratedMessage
20472      implements GetEditLogManifestRequestProtoOrBuilder {
20473    // Use GetEditLogManifestRequestProto.newBuilder() to construct.
20474    private GetEditLogManifestRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
20475      super(builder);
20476      this.unknownFields = builder.getUnknownFields();
20477    }
20478    private GetEditLogManifestRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
20479
20480    private static final GetEditLogManifestRequestProto defaultInstance;
20481    public static GetEditLogManifestRequestProto getDefaultInstance() {
20482      return defaultInstance;
20483    }
20484
20485    public GetEditLogManifestRequestProto getDefaultInstanceForType() {
20486      return defaultInstance;
20487    }
20488
20489    private final com.google.protobuf.UnknownFieldSet unknownFields;
20490    @java.lang.Override
20491    public final com.google.protobuf.UnknownFieldSet
20492        getUnknownFields() {
20493      return this.unknownFields;
20494    }
20495    private GetEditLogManifestRequestProto(
20496        com.google.protobuf.CodedInputStream input,
20497        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20498        throws com.google.protobuf.InvalidProtocolBufferException {
20499      initFields();
20500      int mutable_bitField0_ = 0;
20501      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
20502          com.google.protobuf.UnknownFieldSet.newBuilder();
20503      try {
20504        boolean done = false;
20505        while (!done) {
20506          int tag = input.readTag();
20507          switch (tag) {
20508            case 0:
20509              done = true;
20510              break;
20511            default: {
20512              if (!parseUnknownField(input, unknownFields,
20513                                     extensionRegistry, tag)) {
20514                done = true;
20515              }
20516              break;
20517            }
20518            case 10: {
20519              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
20520              if (((bitField0_ & 0x00000001) == 0x00000001)) {
20521                subBuilder = jid_.toBuilder();
20522              }
20523              jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
20524              if (subBuilder != null) {
20525                subBuilder.mergeFrom(jid_);
20526                jid_ = subBuilder.buildPartial();
20527              }
20528              bitField0_ |= 0x00000001;
20529              break;
20530            }
20531            case 16: {
20532              bitField0_ |= 0x00000002;
20533              sinceTxId_ = input.readUInt64();
20534              break;
20535            }
20536            case 32: {
20537              bitField0_ |= 0x00000004;
20538              inProgressOk_ = input.readBool();
20539              break;
20540            }
20541          }
20542        }
20543      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
20544        throw e.setUnfinishedMessage(this);
20545      } catch (java.io.IOException e) {
20546        throw new com.google.protobuf.InvalidProtocolBufferException(
20547            e.getMessage()).setUnfinishedMessage(this);
20548      } finally {
20549        this.unknownFields = unknownFields.build();
20550        makeExtensionsImmutable();
20551      }
20552    }
20553    public static final com.google.protobuf.Descriptors.Descriptor
20554        getDescriptor() {
20555      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetEditLogManifestRequestProto_descriptor;
20556    }
20557
20558    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
20559        internalGetFieldAccessorTable() {
20560      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetEditLogManifestRequestProto_fieldAccessorTable
20561          .ensureFieldAccessorsInitialized(
20562              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.Builder.class);
20563    }
20564
20565    public static com.google.protobuf.Parser<GetEditLogManifestRequestProto> PARSER =
20566        new com.google.protobuf.AbstractParser<GetEditLogManifestRequestProto>() {
20567      public GetEditLogManifestRequestProto parsePartialFrom(
20568          com.google.protobuf.CodedInputStream input,
20569          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20570          throws com.google.protobuf.InvalidProtocolBufferException {
20571        return new GetEditLogManifestRequestProto(input, extensionRegistry);
20572      }
20573    };
20574
20575    @java.lang.Override
20576    public com.google.protobuf.Parser<GetEditLogManifestRequestProto> getParserForType() {
20577      return PARSER;
20578    }
20579
20580    private int bitField0_;
20581    // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
20582    public static final int JID_FIELD_NUMBER = 1;
20583    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
20584    /**
20585     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
20586     */
20587    public boolean hasJid() {
20588      return ((bitField0_ & 0x00000001) == 0x00000001);
20589    }
20590    /**
20591     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
20592     */
20593    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
20594      return jid_;
20595    }
20596    /**
20597     * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
20598     */
20599    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
20600      return jid_;
20601    }
20602
20603    // required uint64 sinceTxId = 2;
20604    public static final int SINCETXID_FIELD_NUMBER = 2;
20605    private long sinceTxId_;
20606    /**
20607     * <code>required uint64 sinceTxId = 2;</code>
20608     *
20609     * <pre>
20610     * Transaction ID
20611     * </pre>
20612     */
20613    public boolean hasSinceTxId() {
20614      return ((bitField0_ & 0x00000002) == 0x00000002);
20615    }
20616    /**
20617     * <code>required uint64 sinceTxId = 2;</code>
20618     *
20619     * <pre>
20620     * Transaction ID
20621     * </pre>
20622     */
20623    public long getSinceTxId() {
20624      return sinceTxId_;
20625    }
20626
20627    // optional bool inProgressOk = 4 [default = false];
20628    public static final int INPROGRESSOK_FIELD_NUMBER = 4;
20629    private boolean inProgressOk_;
20630    /**
20631     * <code>optional bool inProgressOk = 4 [default = false];</code>
20632     *
20633     * <pre>
20634     * Whether or not the client will be reading from the returned streams.
20635     * optional bool forReading = 3 [default = true]; &lt;obsolete, do not reuse&gt;
20636     * </pre>
20637     */
20638    public boolean hasInProgressOk() {
20639      return ((bitField0_ & 0x00000004) == 0x00000004);
20640    }
20641    /**
20642     * <code>optional bool inProgressOk = 4 [default = false];</code>
20643     *
20644     * <pre>
20645     * Whether or not the client will be reading from the returned streams.
20646     * optional bool forReading = 3 [default = true]; &lt;obsolete, do not reuse&gt;
20647     * </pre>
20648     */
20649    public boolean getInProgressOk() {
20650      return inProgressOk_;
20651    }
20652
20653    private void initFields() {
20654      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
20655      sinceTxId_ = 0L;
20656      inProgressOk_ = false;
20657    }
20658    private byte memoizedIsInitialized = -1;
20659    public final boolean isInitialized() {
20660      byte isInitialized = memoizedIsInitialized;
20661      if (isInitialized != -1) return isInitialized == 1;
20662
20663      if (!hasJid()) {
20664        memoizedIsInitialized = 0;
20665        return false;
20666      }
20667      if (!hasSinceTxId()) {
20668        memoizedIsInitialized = 0;
20669        return false;
20670      }
20671      if (!getJid().isInitialized()) {
20672        memoizedIsInitialized = 0;
20673        return false;
20674      }
20675      memoizedIsInitialized = 1;
20676      return true;
20677    }
20678
20679    public void writeTo(com.google.protobuf.CodedOutputStream output)
20680                        throws java.io.IOException {
20681      getSerializedSize();
20682      if (((bitField0_ & 0x00000001) == 0x00000001)) {
20683        output.writeMessage(1, jid_);
20684      }
20685      if (((bitField0_ & 0x00000002) == 0x00000002)) {
20686        output.writeUInt64(2, sinceTxId_);
20687      }
20688      if (((bitField0_ & 0x00000004) == 0x00000004)) {
20689        output.writeBool(4, inProgressOk_);
20690      }
20691      getUnknownFields().writeTo(output);
20692    }
20693
20694    private int memoizedSerializedSize = -1;
20695    public int getSerializedSize() {
20696      int size = memoizedSerializedSize;
20697      if (size != -1) return size;
20698
20699      size = 0;
20700      if (((bitField0_ & 0x00000001) == 0x00000001)) {
20701        size += com.google.protobuf.CodedOutputStream
20702          .computeMessageSize(1, jid_);
20703      }
20704      if (((bitField0_ & 0x00000002) == 0x00000002)) {
20705        size += com.google.protobuf.CodedOutputStream
20706          .computeUInt64Size(2, sinceTxId_);
20707      }
20708      if (((bitField0_ & 0x00000004) == 0x00000004)) {
20709        size += com.google.protobuf.CodedOutputStream
20710          .computeBoolSize(4, inProgressOk_);
20711      }
20712      size += getUnknownFields().getSerializedSize();
20713      memoizedSerializedSize = size;
20714      return size;
20715    }
20716
20717    private static final long serialVersionUID = 0L;
20718    @java.lang.Override
20719    protected java.lang.Object writeReplace()
20720        throws java.io.ObjectStreamException {
20721      return super.writeReplace();
20722    }
20723
20724    @java.lang.Override
20725    public boolean equals(final java.lang.Object obj) {
20726      if (obj == this) {
20727       return true;
20728      }
20729      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)) {
20730        return super.equals(obj);
20731      }
20732      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto) obj;
20733
20734      boolean result = true;
20735      result = result && (hasJid() == other.hasJid());
20736      if (hasJid()) {
20737        result = result && getJid()
20738            .equals(other.getJid());
20739      }
20740      result = result && (hasSinceTxId() == other.hasSinceTxId());
20741      if (hasSinceTxId()) {
20742        result = result && (getSinceTxId()
20743            == other.getSinceTxId());
20744      }
20745      result = result && (hasInProgressOk() == other.hasInProgressOk());
20746      if (hasInProgressOk()) {
20747        result = result && (getInProgressOk()
20748            == other.getInProgressOk());
20749      }
20750      result = result &&
20751          getUnknownFields().equals(other.getUnknownFields());
20752      return result;
20753    }
20754
20755    private int memoizedHashCode = 0;
20756    @java.lang.Override
20757    public int hashCode() {
20758      if (memoizedHashCode != 0) {
20759        return memoizedHashCode;
20760      }
20761      int hash = 41;
20762      hash = (19 * hash) + getDescriptorForType().hashCode();
20763      if (hasJid()) {
20764        hash = (37 * hash) + JID_FIELD_NUMBER;
20765        hash = (53 * hash) + getJid().hashCode();
20766      }
20767      if (hasSinceTxId()) {
20768        hash = (37 * hash) + SINCETXID_FIELD_NUMBER;
20769        hash = (53 * hash) + hashLong(getSinceTxId());
20770      }
20771      if (hasInProgressOk()) {
20772        hash = (37 * hash) + INPROGRESSOK_FIELD_NUMBER;
20773        hash = (53 * hash) + hashBoolean(getInProgressOk());
20774      }
20775      hash = (29 * hash) + getUnknownFields().hashCode();
20776      memoizedHashCode = hash;
20777      return hash;
20778    }
20779
20780    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
20781        com.google.protobuf.ByteString data)
20782        throws com.google.protobuf.InvalidProtocolBufferException {
20783      return PARSER.parseFrom(data);
20784    }
20785    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
20786        com.google.protobuf.ByteString data,
20787        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20788        throws com.google.protobuf.InvalidProtocolBufferException {
20789      return PARSER.parseFrom(data, extensionRegistry);
20790    }
20791    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(byte[] data)
20792        throws com.google.protobuf.InvalidProtocolBufferException {
20793      return PARSER.parseFrom(data);
20794    }
20795    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
20796        byte[] data,
20797        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20798        throws com.google.protobuf.InvalidProtocolBufferException {
20799      return PARSER.parseFrom(data, extensionRegistry);
20800    }
20801    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(java.io.InputStream input)
20802        throws java.io.IOException {
20803      return PARSER.parseFrom(input);
20804    }
20805    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
20806        java.io.InputStream input,
20807        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20808        throws java.io.IOException {
20809      return PARSER.parseFrom(input, extensionRegistry);
20810    }
20811    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseDelimitedFrom(java.io.InputStream input)
20812        throws java.io.IOException {
20813      return PARSER.parseDelimitedFrom(input);
20814    }
20815    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseDelimitedFrom(
20816        java.io.InputStream input,
20817        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20818        throws java.io.IOException {
20819      return PARSER.parseDelimitedFrom(input, extensionRegistry);
20820    }
20821    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
20822        com.google.protobuf.CodedInputStream input)
20823        throws java.io.IOException {
20824      return PARSER.parseFrom(input);
20825    }
20826    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
20827        com.google.protobuf.CodedInputStream input,
20828        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20829        throws java.io.IOException {
20830      return PARSER.parseFrom(input, extensionRegistry);
20831    }
20832
20833    public static Builder newBuilder() { return Builder.create(); }
20834    public Builder newBuilderForType() { return newBuilder(); }
20835    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto prototype) {
20836      return newBuilder().mergeFrom(prototype);
20837    }
20838    public Builder toBuilder() { return newBuilder(this); }
20839
20840    @java.lang.Override
20841    protected Builder newBuilderForType(
20842        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
20843      Builder builder = new Builder(parent);
20844      return builder;
20845    }
20846    /**
20847     * Protobuf type {@code hadoop.hdfs.qjournal.GetEditLogManifestRequestProto}
20848     *
20849     * <pre>
20850     **
20851     * getEditLogManifest()
20852     * </pre>
20853     */
20854    public static final class Builder extends
20855        com.google.protobuf.GeneratedMessage.Builder<Builder>
20856       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProtoOrBuilder {
20857      public static final com.google.protobuf.Descriptors.Descriptor
20858          getDescriptor() {
20859        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetEditLogManifestRequestProto_descriptor;
20860      }
20861
20862      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
20863          internalGetFieldAccessorTable() {
20864        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetEditLogManifestRequestProto_fieldAccessorTable
20865            .ensureFieldAccessorsInitialized(
20866                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.Builder.class);
20867      }
20868
20869      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.newBuilder()
20870      private Builder() {
20871        maybeForceBuilderInitialization();
20872      }
20873
20874      private Builder(
20875          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
20876        super(parent);
20877        maybeForceBuilderInitialization();
20878      }
20879      private void maybeForceBuilderInitialization() {
20880        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
20881          getJidFieldBuilder();
20882        }
20883      }
20884      private static Builder create() {
20885        return new Builder();
20886      }
20887
20888      public Builder clear() {
20889        super.clear();
20890        if (jidBuilder_ == null) {
20891          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
20892        } else {
20893          jidBuilder_.clear();
20894        }
20895        bitField0_ = (bitField0_ & ~0x00000001);
20896        sinceTxId_ = 0L;
20897        bitField0_ = (bitField0_ & ~0x00000002);
20898        inProgressOk_ = false;
20899        bitField0_ = (bitField0_ & ~0x00000004);
20900        return this;
20901      }
20902
20903      public Builder clone() {
20904        return create().mergeFrom(buildPartial());
20905      }
20906
20907      public com.google.protobuf.Descriptors.Descriptor
20908          getDescriptorForType() {
20909        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetEditLogManifestRequestProto_descriptor;
20910      }
20911
20912      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto getDefaultInstanceForType() {
20913        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance();
20914      }
20915
20916      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto build() {
20917        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto result = buildPartial();
20918        if (!result.isInitialized()) {
20919          throw newUninitializedMessageException(result);
20920        }
20921        return result;
20922      }
20923
20924      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto buildPartial() {
20925        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto(this);
20926        int from_bitField0_ = bitField0_;
20927        int to_bitField0_ = 0;
20928        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
20929          to_bitField0_ |= 0x00000001;
20930        }
20931        if (jidBuilder_ == null) {
20932          result.jid_ = jid_;
20933        } else {
20934          result.jid_ = jidBuilder_.build();
20935        }
20936        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
20937          to_bitField0_ |= 0x00000002;
20938        }
20939        result.sinceTxId_ = sinceTxId_;
20940        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
20941          to_bitField0_ |= 0x00000004;
20942        }
20943        result.inProgressOk_ = inProgressOk_;
20944        result.bitField0_ = to_bitField0_;
20945        onBuilt();
20946        return result;
20947      }
20948
20949      public Builder mergeFrom(com.google.protobuf.Message other) {
20950        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto) {
20951          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)other);
20952        } else {
20953          super.mergeFrom(other);
20954          return this;
20955        }
20956      }
20957
20958      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto other) {
20959        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance()) return this;
20960        if (other.hasJid()) {
20961          mergeJid(other.getJid());
20962        }
20963        if (other.hasSinceTxId()) {
20964          setSinceTxId(other.getSinceTxId());
20965        }
20966        if (other.hasInProgressOk()) {
20967          setInProgressOk(other.getInProgressOk());
20968        }
20969        this.mergeUnknownFields(other.getUnknownFields());
20970        return this;
20971      }
20972
20973      public final boolean isInitialized() {
20974        if (!hasJid()) {
20975          
20976          return false;
20977        }
20978        if (!hasSinceTxId()) {
20979          
20980          return false;
20981        }
20982        if (!getJid().isInitialized()) {
20983          
20984          return false;
20985        }
20986        return true;
20987      }
20988
20989      public Builder mergeFrom(
20990          com.google.protobuf.CodedInputStream input,
20991          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
20992          throws java.io.IOException {
20993        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parsedMessage = null;
20994        try {
20995          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
20996        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
20997          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto) e.getUnfinishedMessage();
20998          throw e;
20999        } finally {
21000          if (parsedMessage != null) {
21001            mergeFrom(parsedMessage);
21002          }
21003        }
21004        return this;
21005      }
21006      private int bitField0_;
21007
21008      // required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;
21009      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
21010      private com.google.protobuf.SingleFieldBuilder<
21011          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
21012      /**
21013       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
21014       */
21015      public boolean hasJid() {
21016        return ((bitField0_ & 0x00000001) == 0x00000001);
21017      }
21018      /**
21019       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
21020       */
21021      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
21022        if (jidBuilder_ == null) {
21023          return jid_;
21024        } else {
21025          return jidBuilder_.getMessage();
21026        }
21027      }
21028      /**
21029       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
21030       */
21031      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
21032        if (jidBuilder_ == null) {
21033          if (value == null) {
21034            throw new NullPointerException();
21035          }
21036          jid_ = value;
21037          onChanged();
21038        } else {
21039          jidBuilder_.setMessage(value);
21040        }
21041        bitField0_ |= 0x00000001;
21042        return this;
21043      }
21044      /**
21045       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
21046       */
21047      public Builder setJid(
21048          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
21049        if (jidBuilder_ == null) {
21050          jid_ = builderForValue.build();
21051          onChanged();
21052        } else {
21053          jidBuilder_.setMessage(builderForValue.build());
21054        }
21055        bitField0_ |= 0x00000001;
21056        return this;
21057      }
21058      /**
21059       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
21060       */
21061      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
21062        if (jidBuilder_ == null) {
21063          if (((bitField0_ & 0x00000001) == 0x00000001) &&
21064              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
21065            jid_ =
21066              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
21067          } else {
21068            jid_ = value;
21069          }
21070          onChanged();
21071        } else {
21072          jidBuilder_.mergeFrom(value);
21073        }
21074        bitField0_ |= 0x00000001;
21075        return this;
21076      }
21077      /**
21078       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
21079       */
21080      public Builder clearJid() {
21081        if (jidBuilder_ == null) {
21082          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
21083          onChanged();
21084        } else {
21085          jidBuilder_.clear();
21086        }
21087        bitField0_ = (bitField0_ & ~0x00000001);
21088        return this;
21089      }
21090      /**
21091       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
21092       */
21093      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
21094        bitField0_ |= 0x00000001;
21095        onChanged();
21096        return getJidFieldBuilder().getBuilder();
21097      }
21098      /**
21099       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
21100       */
21101      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
21102        if (jidBuilder_ != null) {
21103          return jidBuilder_.getMessageOrBuilder();
21104        } else {
21105          return jid_;
21106        }
21107      }
21108      /**
21109       * <code>required .hadoop.hdfs.qjournal.JournalIdProto jid = 1;</code>
21110       */
21111      private com.google.protobuf.SingleFieldBuilder<
21112          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
21113          getJidFieldBuilder() {
21114        if (jidBuilder_ == null) {
21115          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
21116              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
21117                  jid_,
21118                  getParentForChildren(),
21119                  isClean());
21120          jid_ = null;
21121        }
21122        return jidBuilder_;
21123      }
21124
21125      // required uint64 sinceTxId = 2;
21126      private long sinceTxId_ ;
21127      /**
21128       * <code>required uint64 sinceTxId = 2;</code>
21129       *
21130       * <pre>
21131       * Transaction ID
21132       * </pre>
21133       */
21134      public boolean hasSinceTxId() {
21135        return ((bitField0_ & 0x00000002) == 0x00000002);
21136      }
21137      /**
21138       * <code>required uint64 sinceTxId = 2;</code>
21139       *
21140       * <pre>
21141       * Transaction ID
21142       * </pre>
21143       */
21144      public long getSinceTxId() {
21145        return sinceTxId_;
21146      }
21147      /**
21148       * <code>required uint64 sinceTxId = 2;</code>
21149       *
21150       * <pre>
21151       * Transaction ID
21152       * </pre>
21153       */
21154      public Builder setSinceTxId(long value) {
21155        bitField0_ |= 0x00000002;
21156        sinceTxId_ = value;
21157        onChanged();
21158        return this;
21159      }
21160      /**
21161       * <code>required uint64 sinceTxId = 2;</code>
21162       *
21163       * <pre>
21164       * Transaction ID
21165       * </pre>
21166       */
21167      public Builder clearSinceTxId() {
21168        bitField0_ = (bitField0_ & ~0x00000002);
21169        sinceTxId_ = 0L;
21170        onChanged();
21171        return this;
21172      }
21173
21174      // optional bool inProgressOk = 4 [default = false];
21175      private boolean inProgressOk_ ;
21176      /**
21177       * <code>optional bool inProgressOk = 4 [default = false];</code>
21178       *
21179       * <pre>
21180       * Whether or not the client will be reading from the returned streams.
21181       * optional bool forReading = 3 [default = true]; &lt;obsolete, do not reuse&gt;
21182       * </pre>
21183       */
21184      public boolean hasInProgressOk() {
21185        return ((bitField0_ & 0x00000004) == 0x00000004);
21186      }
21187      /**
21188       * <code>optional bool inProgressOk = 4 [default = false];</code>
21189       *
21190       * <pre>
21191       * Whether or not the client will be reading from the returned streams.
21192       * optional bool forReading = 3 [default = true]; &lt;obsolete, do not reuse&gt;
21193       * </pre>
21194       */
21195      public boolean getInProgressOk() {
21196        return inProgressOk_;
21197      }
21198      /**
21199       * <code>optional bool inProgressOk = 4 [default = false];</code>
21200       *
21201       * <pre>
21202       * Whether or not the client will be reading from the returned streams.
21203       * optional bool forReading = 3 [default = true]; &lt;obsolete, do not reuse&gt;
21204       * </pre>
21205       */
21206      public Builder setInProgressOk(boolean value) {
21207        bitField0_ |= 0x00000004;
21208        inProgressOk_ = value;
21209        onChanged();
21210        return this;
21211      }
21212      /**
21213       * <code>optional bool inProgressOk = 4 [default = false];</code>
21214       *
21215       * <pre>
21216       * Whether or not the client will be reading from the returned streams.
21217       * optional bool forReading = 3 [default = true]; &lt;obsolete, do not reuse&gt;
21218       * </pre>
21219       */
21220      public Builder clearInProgressOk() {
21221        bitField0_ = (bitField0_ & ~0x00000004);
21222        inProgressOk_ = false;
21223        onChanged();
21224        return this;
21225      }
21226
21227      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.GetEditLogManifestRequestProto)
21228    }
21229
21230    static {
21231      defaultInstance = new GetEditLogManifestRequestProto(true);
21232      defaultInstance.initFields();
21233    }
21234
21235    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.GetEditLogManifestRequestProto)
21236  }
21237
21238  public interface GetEditLogManifestResponseProtoOrBuilder
21239      extends com.google.protobuf.MessageOrBuilder {
21240
21241    // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;
21242    /**
21243     * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21244     */
21245    boolean hasManifest();
21246    /**
21247     * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21248     */
21249    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto getManifest();
21250    /**
21251     * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21252     */
21253    org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder();
21254
21255    // required uint32 httpPort = 2;
21256    /**
21257     * <code>required uint32 httpPort = 2;</code>
21258     *
21259     * <pre>
21260     * Deprecated by fromURL
21261     * </pre>
21262     */
21263    boolean hasHttpPort();
21264    /**
21265     * <code>required uint32 httpPort = 2;</code>
21266     *
21267     * <pre>
21268     * Deprecated by fromURL
21269     * </pre>
21270     */
21271    int getHttpPort();
21272
21273    // optional string fromURL = 3;
21274    /**
21275     * <code>optional string fromURL = 3;</code>
21276     */
21277    boolean hasFromURL();
21278    /**
21279     * <code>optional string fromURL = 3;</code>
21280     */
21281    java.lang.String getFromURL();
21282    /**
21283     * <code>optional string fromURL = 3;</code>
21284     */
21285    com.google.protobuf.ByteString
21286        getFromURLBytes();
21287  }
21288  /**
21289   * Protobuf type {@code hadoop.hdfs.qjournal.GetEditLogManifestResponseProto}
21290   */
21291  public static final class GetEditLogManifestResponseProto extends
21292      com.google.protobuf.GeneratedMessage
21293      implements GetEditLogManifestResponseProtoOrBuilder {
21294    // Use GetEditLogManifestResponseProto.newBuilder() to construct.
21295    private GetEditLogManifestResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
21296      super(builder);
21297      this.unknownFields = builder.getUnknownFields();
21298    }
21299    private GetEditLogManifestResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
21300
21301    private static final GetEditLogManifestResponseProto defaultInstance;
21302    public static GetEditLogManifestResponseProto getDefaultInstance() {
21303      return defaultInstance;
21304    }
21305
21306    public GetEditLogManifestResponseProto getDefaultInstanceForType() {
21307      return defaultInstance;
21308    }
21309
21310    private final com.google.protobuf.UnknownFieldSet unknownFields;
21311    @java.lang.Override
21312    public final com.google.protobuf.UnknownFieldSet
21313        getUnknownFields() {
21314      return this.unknownFields;
21315    }
21316    private GetEditLogManifestResponseProto(
21317        com.google.protobuf.CodedInputStream input,
21318        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
21319        throws com.google.protobuf.InvalidProtocolBufferException {
21320      initFields();
21321      int mutable_bitField0_ = 0;
21322      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
21323          com.google.protobuf.UnknownFieldSet.newBuilder();
21324      try {
21325        boolean done = false;
21326        while (!done) {
21327          int tag = input.readTag();
21328          switch (tag) {
21329            case 0:
21330              done = true;
21331              break;
21332            default: {
21333              if (!parseUnknownField(input, unknownFields,
21334                                     extensionRegistry, tag)) {
21335                done = true;
21336              }
21337              break;
21338            }
21339            case 10: {
21340              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.Builder subBuilder = null;
21341              if (((bitField0_ & 0x00000001) == 0x00000001)) {
21342                subBuilder = manifest_.toBuilder();
21343              }
21344              manifest_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.PARSER, extensionRegistry);
21345              if (subBuilder != null) {
21346                subBuilder.mergeFrom(manifest_);
21347                manifest_ = subBuilder.buildPartial();
21348              }
21349              bitField0_ |= 0x00000001;
21350              break;
21351            }
21352            case 16: {
21353              bitField0_ |= 0x00000002;
21354              httpPort_ = input.readUInt32();
21355              break;
21356            }
21357            case 26: {
21358              bitField0_ |= 0x00000004;
21359              fromURL_ = input.readBytes();
21360              break;
21361            }
21362          }
21363        }
21364      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
21365        throw e.setUnfinishedMessage(this);
21366      } catch (java.io.IOException e) {
21367        throw new com.google.protobuf.InvalidProtocolBufferException(
21368            e.getMessage()).setUnfinishedMessage(this);
21369      } finally {
21370        this.unknownFields = unknownFields.build();
21371        makeExtensionsImmutable();
21372      }
21373    }
21374    public static final com.google.protobuf.Descriptors.Descriptor
21375        getDescriptor() {
21376      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetEditLogManifestResponseProto_descriptor;
21377    }
21378
21379    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
21380        internalGetFieldAccessorTable() {
21381      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetEditLogManifestResponseProto_fieldAccessorTable
21382          .ensureFieldAccessorsInitialized(
21383              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.Builder.class);
21384    }
21385
21386    public static com.google.protobuf.Parser<GetEditLogManifestResponseProto> PARSER =
21387        new com.google.protobuf.AbstractParser<GetEditLogManifestResponseProto>() {
21388      public GetEditLogManifestResponseProto parsePartialFrom(
21389          com.google.protobuf.CodedInputStream input,
21390          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
21391          throws com.google.protobuf.InvalidProtocolBufferException {
21392        return new GetEditLogManifestResponseProto(input, extensionRegistry);
21393      }
21394    };
21395
21396    @java.lang.Override
21397    public com.google.protobuf.Parser<GetEditLogManifestResponseProto> getParserForType() {
21398      return PARSER;
21399    }
21400
21401    private int bitField0_;
21402    // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;
21403    public static final int MANIFEST_FIELD_NUMBER = 1;
21404    private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto manifest_;
21405    /**
21406     * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21407     */
21408    public boolean hasManifest() {
21409      return ((bitField0_ & 0x00000001) == 0x00000001);
21410    }
21411    /**
21412     * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21413     */
21414    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto getManifest() {
21415      return manifest_;
21416    }
21417    /**
21418     * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21419     */
21420    public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder() {
21421      return manifest_;
21422    }
21423
21424    // required uint32 httpPort = 2;
21425    public static final int HTTPPORT_FIELD_NUMBER = 2;
21426    private int httpPort_;
21427    /**
21428     * <code>required uint32 httpPort = 2;</code>
21429     *
21430     * <pre>
21431     * Deprecated by fromURL
21432     * </pre>
21433     */
21434    public boolean hasHttpPort() {
21435      return ((bitField0_ & 0x00000002) == 0x00000002);
21436    }
21437    /**
21438     * <code>required uint32 httpPort = 2;</code>
21439     *
21440     * <pre>
21441     * Deprecated by fromURL
21442     * </pre>
21443     */
21444    public int getHttpPort() {
21445      return httpPort_;
21446    }
21447
21448    // optional string fromURL = 3;
21449    public static final int FROMURL_FIELD_NUMBER = 3;
21450    private java.lang.Object fromURL_;
21451    /**
21452     * <code>optional string fromURL = 3;</code>
21453     */
21454    public boolean hasFromURL() {
21455      return ((bitField0_ & 0x00000004) == 0x00000004);
21456    }
21457    /**
21458     * <code>optional string fromURL = 3;</code>
21459     */
21460    public java.lang.String getFromURL() {
21461      java.lang.Object ref = fromURL_;
21462      if (ref instanceof java.lang.String) {
21463        return (java.lang.String) ref;
21464      } else {
21465        com.google.protobuf.ByteString bs = 
21466            (com.google.protobuf.ByteString) ref;
21467        java.lang.String s = bs.toStringUtf8();
21468        if (bs.isValidUtf8()) {
21469          fromURL_ = s;
21470        }
21471        return s;
21472      }
21473    }
21474    /**
21475     * <code>optional string fromURL = 3;</code>
21476     */
21477    public com.google.protobuf.ByteString
21478        getFromURLBytes() {
21479      java.lang.Object ref = fromURL_;
21480      if (ref instanceof java.lang.String) {
21481        com.google.protobuf.ByteString b = 
21482            com.google.protobuf.ByteString.copyFromUtf8(
21483                (java.lang.String) ref);
21484        fromURL_ = b;
21485        return b;
21486      } else {
21487        return (com.google.protobuf.ByteString) ref;
21488      }
21489    }
21490
21491    private void initFields() {
21492      manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.getDefaultInstance();
21493      httpPort_ = 0;
21494      fromURL_ = "";
21495    }
21496    private byte memoizedIsInitialized = -1;
21497    public final boolean isInitialized() {
21498      byte isInitialized = memoizedIsInitialized;
21499      if (isInitialized != -1) return isInitialized == 1;
21500
21501      if (!hasManifest()) {
21502        memoizedIsInitialized = 0;
21503        return false;
21504      }
21505      if (!hasHttpPort()) {
21506        memoizedIsInitialized = 0;
21507        return false;
21508      }
21509      if (!getManifest().isInitialized()) {
21510        memoizedIsInitialized = 0;
21511        return false;
21512      }
21513      memoizedIsInitialized = 1;
21514      return true;
21515    }
21516
21517    public void writeTo(com.google.protobuf.CodedOutputStream output)
21518                        throws java.io.IOException {
21519      getSerializedSize();
21520      if (((bitField0_ & 0x00000001) == 0x00000001)) {
21521        output.writeMessage(1, manifest_);
21522      }
21523      if (((bitField0_ & 0x00000002) == 0x00000002)) {
21524        output.writeUInt32(2, httpPort_);
21525      }
21526      if (((bitField0_ & 0x00000004) == 0x00000004)) {
21527        output.writeBytes(3, getFromURLBytes());
21528      }
21529      getUnknownFields().writeTo(output);
21530    }
21531
21532    private int memoizedSerializedSize = -1;
21533    public int getSerializedSize() {
21534      int size = memoizedSerializedSize;
21535      if (size != -1) return size;
21536
21537      size = 0;
21538      if (((bitField0_ & 0x00000001) == 0x00000001)) {
21539        size += com.google.protobuf.CodedOutputStream
21540          .computeMessageSize(1, manifest_);
21541      }
21542      if (((bitField0_ & 0x00000002) == 0x00000002)) {
21543        size += com.google.protobuf.CodedOutputStream
21544          .computeUInt32Size(2, httpPort_);
21545      }
21546      if (((bitField0_ & 0x00000004) == 0x00000004)) {
21547        size += com.google.protobuf.CodedOutputStream
21548          .computeBytesSize(3, getFromURLBytes());
21549      }
21550      size += getUnknownFields().getSerializedSize();
21551      memoizedSerializedSize = size;
21552      return size;
21553    }
21554
21555    private static final long serialVersionUID = 0L;
21556    @java.lang.Override
21557    protected java.lang.Object writeReplace()
21558        throws java.io.ObjectStreamException {
21559      return super.writeReplace();
21560    }
21561
21562    @java.lang.Override
21563    public boolean equals(final java.lang.Object obj) {
21564      if (obj == this) {
21565       return true;
21566      }
21567      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto)) {
21568        return super.equals(obj);
21569      }
21570      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) obj;
21571
21572      boolean result = true;
21573      result = result && (hasManifest() == other.hasManifest());
21574      if (hasManifest()) {
21575        result = result && getManifest()
21576            .equals(other.getManifest());
21577      }
21578      result = result && (hasHttpPort() == other.hasHttpPort());
21579      if (hasHttpPort()) {
21580        result = result && (getHttpPort()
21581            == other.getHttpPort());
21582      }
21583      result = result && (hasFromURL() == other.hasFromURL());
21584      if (hasFromURL()) {
21585        result = result && getFromURL()
21586            .equals(other.getFromURL());
21587      }
21588      result = result &&
21589          getUnknownFields().equals(other.getUnknownFields());
21590      return result;
21591    }
21592
21593    private int memoizedHashCode = 0;
21594    @java.lang.Override
21595    public int hashCode() {
21596      if (memoizedHashCode != 0) {
21597        return memoizedHashCode;
21598      }
21599      int hash = 41;
21600      hash = (19 * hash) + getDescriptorForType().hashCode();
21601      if (hasManifest()) {
21602        hash = (37 * hash) + MANIFEST_FIELD_NUMBER;
21603        hash = (53 * hash) + getManifest().hashCode();
21604      }
21605      if (hasHttpPort()) {
21606        hash = (37 * hash) + HTTPPORT_FIELD_NUMBER;
21607        hash = (53 * hash) + getHttpPort();
21608      }
21609      if (hasFromURL()) {
21610        hash = (37 * hash) + FROMURL_FIELD_NUMBER;
21611        hash = (53 * hash) + getFromURL().hashCode();
21612      }
21613      hash = (29 * hash) + getUnknownFields().hashCode();
21614      memoizedHashCode = hash;
21615      return hash;
21616    }
21617
21618    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
21619        com.google.protobuf.ByteString data)
21620        throws com.google.protobuf.InvalidProtocolBufferException {
21621      return PARSER.parseFrom(data);
21622    }
21623    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
21624        com.google.protobuf.ByteString data,
21625        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
21626        throws com.google.protobuf.InvalidProtocolBufferException {
21627      return PARSER.parseFrom(data, extensionRegistry);
21628    }
21629    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(byte[] data)
21630        throws com.google.protobuf.InvalidProtocolBufferException {
21631      return PARSER.parseFrom(data);
21632    }
21633    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
21634        byte[] data,
21635        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
21636        throws com.google.protobuf.InvalidProtocolBufferException {
21637      return PARSER.parseFrom(data, extensionRegistry);
21638    }
21639    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(java.io.InputStream input)
21640        throws java.io.IOException {
21641      return PARSER.parseFrom(input);
21642    }
21643    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
21644        java.io.InputStream input,
21645        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
21646        throws java.io.IOException {
21647      return PARSER.parseFrom(input, extensionRegistry);
21648    }
21649    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseDelimitedFrom(java.io.InputStream input)
21650        throws java.io.IOException {
21651      return PARSER.parseDelimitedFrom(input);
21652    }
21653    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseDelimitedFrom(
21654        java.io.InputStream input,
21655        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
21656        throws java.io.IOException {
21657      return PARSER.parseDelimitedFrom(input, extensionRegistry);
21658    }
21659    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
21660        com.google.protobuf.CodedInputStream input)
21661        throws java.io.IOException {
21662      return PARSER.parseFrom(input);
21663    }
21664    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
21665        com.google.protobuf.CodedInputStream input,
21666        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
21667        throws java.io.IOException {
21668      return PARSER.parseFrom(input, extensionRegistry);
21669    }
21670
21671    public static Builder newBuilder() { return Builder.create(); }
21672    public Builder newBuilderForType() { return newBuilder(); }
21673    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto prototype) {
21674      return newBuilder().mergeFrom(prototype);
21675    }
21676    public Builder toBuilder() { return newBuilder(this); }
21677
21678    @java.lang.Override
21679    protected Builder newBuilderForType(
21680        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
21681      Builder builder = new Builder(parent);
21682      return builder;
21683    }
21684    /**
21685     * Protobuf type {@code hadoop.hdfs.qjournal.GetEditLogManifestResponseProto}
21686     */
21687    public static final class Builder extends
21688        com.google.protobuf.GeneratedMessage.Builder<Builder>
21689       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProtoOrBuilder {
21690      public static final com.google.protobuf.Descriptors.Descriptor
21691          getDescriptor() {
21692        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetEditLogManifestResponseProto_descriptor;
21693      }
21694
21695      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
21696          internalGetFieldAccessorTable() {
21697        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetEditLogManifestResponseProto_fieldAccessorTable
21698            .ensureFieldAccessorsInitialized(
21699                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.Builder.class);
21700      }
21701
21702      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.newBuilder()
21703      private Builder() {
21704        maybeForceBuilderInitialization();
21705      }
21706
21707      private Builder(
21708          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
21709        super(parent);
21710        maybeForceBuilderInitialization();
21711      }
21712      private void maybeForceBuilderInitialization() {
21713        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
21714          getManifestFieldBuilder();
21715        }
21716      }
21717      private static Builder create() {
21718        return new Builder();
21719      }
21720
21721      public Builder clear() {
21722        super.clear();
21723        if (manifestBuilder_ == null) {
21724          manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.getDefaultInstance();
21725        } else {
21726          manifestBuilder_.clear();
21727        }
21728        bitField0_ = (bitField0_ & ~0x00000001);
21729        httpPort_ = 0;
21730        bitField0_ = (bitField0_ & ~0x00000002);
21731        fromURL_ = "";
21732        bitField0_ = (bitField0_ & ~0x00000004);
21733        return this;
21734      }
21735
21736      public Builder clone() {
21737        return create().mergeFrom(buildPartial());
21738      }
21739
21740      public com.google.protobuf.Descriptors.Descriptor
21741          getDescriptorForType() {
21742        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_GetEditLogManifestResponseProto_descriptor;
21743      }
21744
21745      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getDefaultInstanceForType() {
21746        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance();
21747      }
21748
21749      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto build() {
21750        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto result = buildPartial();
21751        if (!result.isInitialized()) {
21752          throw newUninitializedMessageException(result);
21753        }
21754        return result;
21755      }
21756
21757      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto buildPartial() {
21758        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto(this);
21759        int from_bitField0_ = bitField0_;
21760        int to_bitField0_ = 0;
21761        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
21762          to_bitField0_ |= 0x00000001;
21763        }
21764        if (manifestBuilder_ == null) {
21765          result.manifest_ = manifest_;
21766        } else {
21767          result.manifest_ = manifestBuilder_.build();
21768        }
21769        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
21770          to_bitField0_ |= 0x00000002;
21771        }
21772        result.httpPort_ = httpPort_;
21773        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
21774          to_bitField0_ |= 0x00000004;
21775        }
21776        result.fromURL_ = fromURL_;
21777        result.bitField0_ = to_bitField0_;
21778        onBuilt();
21779        return result;
21780      }
21781
21782      public Builder mergeFrom(com.google.protobuf.Message other) {
21783        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) {
21784          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto)other);
21785        } else {
21786          super.mergeFrom(other);
21787          return this;
21788        }
21789      }
21790
21791      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto other) {
21792        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance()) return this;
21793        if (other.hasManifest()) {
21794          mergeManifest(other.getManifest());
21795        }
21796        if (other.hasHttpPort()) {
21797          setHttpPort(other.getHttpPort());
21798        }
21799        if (other.hasFromURL()) {
21800          bitField0_ |= 0x00000004;
21801          fromURL_ = other.fromURL_;
21802          onChanged();
21803        }
21804        this.mergeUnknownFields(other.getUnknownFields());
21805        return this;
21806      }
21807
21808      public final boolean isInitialized() {
21809        if (!hasManifest()) {
21810          
21811          return false;
21812        }
21813        if (!hasHttpPort()) {
21814          
21815          return false;
21816        }
21817        if (!getManifest().isInitialized()) {
21818          
21819          return false;
21820        }
21821        return true;
21822      }
21823
21824      public Builder mergeFrom(
21825          com.google.protobuf.CodedInputStream input,
21826          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
21827          throws java.io.IOException {
21828        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parsedMessage = null;
21829        try {
21830          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
21831        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
21832          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) e.getUnfinishedMessage();
21833          throw e;
21834        } finally {
21835          if (parsedMessage != null) {
21836            mergeFrom(parsedMessage);
21837          }
21838        }
21839        return this;
21840      }
21841      private int bitField0_;
21842
21843      // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;
21844      private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.getDefaultInstance();
21845      private com.google.protobuf.SingleFieldBuilder<
21846          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProtoOrBuilder> manifestBuilder_;
21847      /**
21848       * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21849       */
21850      public boolean hasManifest() {
21851        return ((bitField0_ & 0x00000001) == 0x00000001);
21852      }
21853      /**
21854       * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21855       */
21856      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto getManifest() {
21857        if (manifestBuilder_ == null) {
21858          return manifest_;
21859        } else {
21860          return manifestBuilder_.getMessage();
21861        }
21862      }
21863      /**
21864       * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21865       */
21866      public Builder setManifest(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto value) {
21867        if (manifestBuilder_ == null) {
21868          if (value == null) {
21869            throw new NullPointerException();
21870          }
21871          manifest_ = value;
21872          onChanged();
21873        } else {
21874          manifestBuilder_.setMessage(value);
21875        }
21876        bitField0_ |= 0x00000001;
21877        return this;
21878      }
21879      /**
21880       * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21881       */
21882      public Builder setManifest(
21883          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.Builder builderForValue) {
21884        if (manifestBuilder_ == null) {
21885          manifest_ = builderForValue.build();
21886          onChanged();
21887        } else {
21888          manifestBuilder_.setMessage(builderForValue.build());
21889        }
21890        bitField0_ |= 0x00000001;
21891        return this;
21892      }
21893      /**
21894       * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21895       */
21896      public Builder mergeManifest(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto value) {
21897        if (manifestBuilder_ == null) {
21898          if (((bitField0_ & 0x00000001) == 0x00000001) &&
21899              manifest_ != org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.getDefaultInstance()) {
21900            manifest_ =
21901              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.newBuilder(manifest_).mergeFrom(value).buildPartial();
21902          } else {
21903            manifest_ = value;
21904          }
21905          onChanged();
21906        } else {
21907          manifestBuilder_.mergeFrom(value);
21908        }
21909        bitField0_ |= 0x00000001;
21910        return this;
21911      }
21912      /**
21913       * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21914       */
21915      public Builder clearManifest() {
21916        if (manifestBuilder_ == null) {
21917          manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.getDefaultInstance();
21918          onChanged();
21919        } else {
21920          manifestBuilder_.clear();
21921        }
21922        bitField0_ = (bitField0_ & ~0x00000001);
21923        return this;
21924      }
21925      /**
21926       * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21927       */
21928      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.Builder getManifestBuilder() {
21929        bitField0_ |= 0x00000001;
21930        onChanged();
21931        return getManifestFieldBuilder().getBuilder();
21932      }
21933      /**
21934       * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21935       */
21936      public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder() {
21937        if (manifestBuilder_ != null) {
21938          return manifestBuilder_.getMessageOrBuilder();
21939        } else {
21940          return manifest_;
21941        }
21942      }
21943      /**
21944       * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
21945       */
21946      private com.google.protobuf.SingleFieldBuilder<
21947          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProtoOrBuilder> 
21948          getManifestFieldBuilder() {
21949        if (manifestBuilder_ == null) {
21950          manifestBuilder_ = new com.google.protobuf.SingleFieldBuilder<
21951              org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RemoteEditLogManifestProtoOrBuilder>(
21952                  manifest_,
21953                  getParentForChildren(),
21954                  isClean());
21955          manifest_ = null;
21956        }
21957        return manifestBuilder_;
21958      }
21959
21960      // required uint32 httpPort = 2;
21961      private int httpPort_ ;
21962      /**
21963       * <code>required uint32 httpPort = 2;</code>
21964       *
21965       * <pre>
21966       * Deprecated by fromURL
21967       * </pre>
21968       */
21969      public boolean hasHttpPort() {
21970        return ((bitField0_ & 0x00000002) == 0x00000002);
21971      }
21972      /**
21973       * <code>required uint32 httpPort = 2;</code>
21974       *
21975       * <pre>
21976       * Deprecated by fromURL
21977       * </pre>
21978       */
21979      public int getHttpPort() {
21980        return httpPort_;
21981      }
21982      /**
21983       * <code>required uint32 httpPort = 2;</code>
21984       *
21985       * <pre>
21986       * Deprecated by fromURL
21987       * </pre>
21988       */
21989      public Builder setHttpPort(int value) {
21990        bitField0_ |= 0x00000002;
21991        httpPort_ = value;
21992        onChanged();
21993        return this;
21994      }
21995      /**
21996       * <code>required uint32 httpPort = 2;</code>
21997       *
21998       * <pre>
21999       * Deprecated by fromURL
22000       * </pre>
22001       */
22002      public Builder clearHttpPort() {
22003        bitField0_ = (bitField0_ & ~0x00000002);
22004        httpPort_ = 0;
22005        onChanged();
22006        return this;
22007      }
22008
22009      // optional string fromURL = 3;
22010      private java.lang.Object fromURL_ = "";
22011      /**
22012       * <code>optional string fromURL = 3;</code>
22013       */
22014      public boolean hasFromURL() {
22015        return ((bitField0_ & 0x00000004) == 0x00000004);
22016      }
22017      /**
22018       * <code>optional string fromURL = 3;</code>
22019       */
22020      public java.lang.String getFromURL() {
22021        java.lang.Object ref = fromURL_;
22022        if (!(ref instanceof java.lang.String)) {
22023          java.lang.String s = ((com.google.protobuf.ByteString) ref)
22024              .toStringUtf8();
22025          fromURL_ = s;
22026          return s;
22027        } else {
22028          return (java.lang.String) ref;
22029        }
22030      }
22031      /**
22032       * <code>optional string fromURL = 3;</code>
22033       */
22034      public com.google.protobuf.ByteString
22035          getFromURLBytes() {
22036        java.lang.Object ref = fromURL_;
22037        if (ref instanceof String) {
22038          com.google.protobuf.ByteString b = 
22039              com.google.protobuf.ByteString.copyFromUtf8(
22040                  (java.lang.String) ref);
22041          fromURL_ = b;
22042          return b;
22043        } else {
22044          return (com.google.protobuf.ByteString) ref;
22045        }
22046      }
22047      /**
22048       * <code>optional string fromURL = 3;</code>
22049       */
22050      public Builder setFromURL(
22051          java.lang.String value) {
22052        if (value == null) {
22053    throw new NullPointerException();
22054  }
22055  bitField0_ |= 0x00000004;
22056        fromURL_ = value;
22057        onChanged();
22058        return this;
22059      }
22060      /**
22061       * <code>optional string fromURL = 3;</code>
22062       */
22063      public Builder clearFromURL() {
22064        bitField0_ = (bitField0_ & ~0x00000004);
22065        fromURL_ = getDefaultInstance().getFromURL();
22066        onChanged();
22067        return this;
22068      }
22069      /**
22070       * <code>optional string fromURL = 3;</code>
22071       */
22072      public Builder setFromURLBytes(
22073          com.google.protobuf.ByteString value) {
22074        if (value == null) {
22075    throw new NullPointerException();
22076  }
22077  bitField0_ |= 0x00000004;
22078        fromURL_ = value;
22079        onChanged();
22080        return this;
22081      }
22082
22083      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.GetEditLogManifestResponseProto)
22084    }
22085
22086    static {
22087      defaultInstance = new GetEditLogManifestResponseProto(true);
22088      defaultInstance.initFields();
22089    }
22090
22091    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.GetEditLogManifestResponseProto)
22092  }
22093
22094  public interface PrepareRecoveryRequestProtoOrBuilder
22095      extends com.google.protobuf.MessageOrBuilder {
22096
22097    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
22098    /**
22099     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22100     */
22101    boolean hasReqInfo();
22102    /**
22103     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22104     */
22105    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
22106    /**
22107     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22108     */
22109    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
22110
22111    // required uint64 segmentTxId = 2;
22112    /**
22113     * <code>required uint64 segmentTxId = 2;</code>
22114     */
22115    boolean hasSegmentTxId();
22116    /**
22117     * <code>required uint64 segmentTxId = 2;</code>
22118     */
22119    long getSegmentTxId();
22120  }
22121  /**
22122   * Protobuf type {@code hadoop.hdfs.qjournal.PrepareRecoveryRequestProto}
22123   *
22124   * <pre>
22125   **
22126   * prepareRecovery()
22127   * </pre>
22128   */
22129  public static final class PrepareRecoveryRequestProto extends
22130      com.google.protobuf.GeneratedMessage
22131      implements PrepareRecoveryRequestProtoOrBuilder {
22132    // Use PrepareRecoveryRequestProto.newBuilder() to construct.
22133    private PrepareRecoveryRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
22134      super(builder);
22135      this.unknownFields = builder.getUnknownFields();
22136    }
22137    private PrepareRecoveryRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
22138
22139    private static final PrepareRecoveryRequestProto defaultInstance;
22140    public static PrepareRecoveryRequestProto getDefaultInstance() {
22141      return defaultInstance;
22142    }
22143
22144    public PrepareRecoveryRequestProto getDefaultInstanceForType() {
22145      return defaultInstance;
22146    }
22147
22148    private final com.google.protobuf.UnknownFieldSet unknownFields;
22149    @java.lang.Override
22150    public final com.google.protobuf.UnknownFieldSet
22151        getUnknownFields() {
22152      return this.unknownFields;
22153    }
22154    private PrepareRecoveryRequestProto(
22155        com.google.protobuf.CodedInputStream input,
22156        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
22157        throws com.google.protobuf.InvalidProtocolBufferException {
22158      initFields();
22159      int mutable_bitField0_ = 0;
22160      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
22161          com.google.protobuf.UnknownFieldSet.newBuilder();
22162      try {
22163        boolean done = false;
22164        while (!done) {
22165          int tag = input.readTag();
22166          switch (tag) {
22167            case 0:
22168              done = true;
22169              break;
22170            default: {
22171              if (!parseUnknownField(input, unknownFields,
22172                                     extensionRegistry, tag)) {
22173                done = true;
22174              }
22175              break;
22176            }
22177            case 10: {
22178              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
22179              if (((bitField0_ & 0x00000001) == 0x00000001)) {
22180                subBuilder = reqInfo_.toBuilder();
22181              }
22182              reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
22183              if (subBuilder != null) {
22184                subBuilder.mergeFrom(reqInfo_);
22185                reqInfo_ = subBuilder.buildPartial();
22186              }
22187              bitField0_ |= 0x00000001;
22188              break;
22189            }
22190            case 16: {
22191              bitField0_ |= 0x00000002;
22192              segmentTxId_ = input.readUInt64();
22193              break;
22194            }
22195          }
22196        }
22197      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
22198        throw e.setUnfinishedMessage(this);
22199      } catch (java.io.IOException e) {
22200        throw new com.google.protobuf.InvalidProtocolBufferException(
22201            e.getMessage()).setUnfinishedMessage(this);
22202      } finally {
22203        this.unknownFields = unknownFields.build();
22204        makeExtensionsImmutable();
22205      }
22206    }
22207    public static final com.google.protobuf.Descriptors.Descriptor
22208        getDescriptor() {
22209      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PrepareRecoveryRequestProto_descriptor;
22210    }
22211
22212    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
22213        internalGetFieldAccessorTable() {
22214      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PrepareRecoveryRequestProto_fieldAccessorTable
22215          .ensureFieldAccessorsInitialized(
22216              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.Builder.class);
22217    }
22218
22219    public static com.google.protobuf.Parser<PrepareRecoveryRequestProto> PARSER =
22220        new com.google.protobuf.AbstractParser<PrepareRecoveryRequestProto>() {
22221      public PrepareRecoveryRequestProto parsePartialFrom(
22222          com.google.protobuf.CodedInputStream input,
22223          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
22224          throws com.google.protobuf.InvalidProtocolBufferException {
22225        return new PrepareRecoveryRequestProto(input, extensionRegistry);
22226      }
22227    };
22228
22229    @java.lang.Override
22230    public com.google.protobuf.Parser<PrepareRecoveryRequestProto> getParserForType() {
22231      return PARSER;
22232    }
22233
22234    private int bitField0_;
22235    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
22236    public static final int REQINFO_FIELD_NUMBER = 1;
22237    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
22238    /**
22239     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22240     */
22241    public boolean hasReqInfo() {
22242      return ((bitField0_ & 0x00000001) == 0x00000001);
22243    }
22244    /**
22245     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22246     */
22247    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
22248      return reqInfo_;
22249    }
22250    /**
22251     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22252     */
22253    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
22254      return reqInfo_;
22255    }
22256
22257    // required uint64 segmentTxId = 2;
22258    public static final int SEGMENTTXID_FIELD_NUMBER = 2;
22259    private long segmentTxId_;
22260    /**
22261     * <code>required uint64 segmentTxId = 2;</code>
22262     */
22263    public boolean hasSegmentTxId() {
22264      return ((bitField0_ & 0x00000002) == 0x00000002);
22265    }
22266    /**
22267     * <code>required uint64 segmentTxId = 2;</code>
22268     */
22269    public long getSegmentTxId() {
22270      return segmentTxId_;
22271    }
22272
22273    private void initFields() {
22274      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
22275      segmentTxId_ = 0L;
22276    }
22277    private byte memoizedIsInitialized = -1;
22278    public final boolean isInitialized() {
22279      byte isInitialized = memoizedIsInitialized;
22280      if (isInitialized != -1) return isInitialized == 1;
22281
22282      if (!hasReqInfo()) {
22283        memoizedIsInitialized = 0;
22284        return false;
22285      }
22286      if (!hasSegmentTxId()) {
22287        memoizedIsInitialized = 0;
22288        return false;
22289      }
22290      if (!getReqInfo().isInitialized()) {
22291        memoizedIsInitialized = 0;
22292        return false;
22293      }
22294      memoizedIsInitialized = 1;
22295      return true;
22296    }
22297
22298    public void writeTo(com.google.protobuf.CodedOutputStream output)
22299                        throws java.io.IOException {
22300      getSerializedSize();
22301      if (((bitField0_ & 0x00000001) == 0x00000001)) {
22302        output.writeMessage(1, reqInfo_);
22303      }
22304      if (((bitField0_ & 0x00000002) == 0x00000002)) {
22305        output.writeUInt64(2, segmentTxId_);
22306      }
22307      getUnknownFields().writeTo(output);
22308    }
22309
22310    private int memoizedSerializedSize = -1;
22311    public int getSerializedSize() {
22312      int size = memoizedSerializedSize;
22313      if (size != -1) return size;
22314
22315      size = 0;
22316      if (((bitField0_ & 0x00000001) == 0x00000001)) {
22317        size += com.google.protobuf.CodedOutputStream
22318          .computeMessageSize(1, reqInfo_);
22319      }
22320      if (((bitField0_ & 0x00000002) == 0x00000002)) {
22321        size += com.google.protobuf.CodedOutputStream
22322          .computeUInt64Size(2, segmentTxId_);
22323      }
22324      size += getUnknownFields().getSerializedSize();
22325      memoizedSerializedSize = size;
22326      return size;
22327    }
22328
22329    private static final long serialVersionUID = 0L;
22330    @java.lang.Override
22331    protected java.lang.Object writeReplace()
22332        throws java.io.ObjectStreamException {
22333      return super.writeReplace();
22334    }
22335
22336    @java.lang.Override
22337    public boolean equals(final java.lang.Object obj) {
22338      if (obj == this) {
22339       return true;
22340      }
22341      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)) {
22342        return super.equals(obj);
22343      }
22344      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto) obj;
22345
22346      boolean result = true;
22347      result = result && (hasReqInfo() == other.hasReqInfo());
22348      if (hasReqInfo()) {
22349        result = result && getReqInfo()
22350            .equals(other.getReqInfo());
22351      }
22352      result = result && (hasSegmentTxId() == other.hasSegmentTxId());
22353      if (hasSegmentTxId()) {
22354        result = result && (getSegmentTxId()
22355            == other.getSegmentTxId());
22356      }
22357      result = result &&
22358          getUnknownFields().equals(other.getUnknownFields());
22359      return result;
22360    }
22361
22362    private int memoizedHashCode = 0;
22363    @java.lang.Override
22364    public int hashCode() {
22365      if (memoizedHashCode != 0) {
22366        return memoizedHashCode;
22367      }
22368      int hash = 41;
22369      hash = (19 * hash) + getDescriptorForType().hashCode();
22370      if (hasReqInfo()) {
22371        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
22372        hash = (53 * hash) + getReqInfo().hashCode();
22373      }
22374      if (hasSegmentTxId()) {
22375        hash = (37 * hash) + SEGMENTTXID_FIELD_NUMBER;
22376        hash = (53 * hash) + hashLong(getSegmentTxId());
22377      }
22378      hash = (29 * hash) + getUnknownFields().hashCode();
22379      memoizedHashCode = hash;
22380      return hash;
22381    }
22382
22383    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
22384        com.google.protobuf.ByteString data)
22385        throws com.google.protobuf.InvalidProtocolBufferException {
22386      return PARSER.parseFrom(data);
22387    }
22388    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
22389        com.google.protobuf.ByteString data,
22390        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
22391        throws com.google.protobuf.InvalidProtocolBufferException {
22392      return PARSER.parseFrom(data, extensionRegistry);
22393    }
22394    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(byte[] data)
22395        throws com.google.protobuf.InvalidProtocolBufferException {
22396      return PARSER.parseFrom(data);
22397    }
22398    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
22399        byte[] data,
22400        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
22401        throws com.google.protobuf.InvalidProtocolBufferException {
22402      return PARSER.parseFrom(data, extensionRegistry);
22403    }
22404    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(java.io.InputStream input)
22405        throws java.io.IOException {
22406      return PARSER.parseFrom(input);
22407    }
22408    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
22409        java.io.InputStream input,
22410        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
22411        throws java.io.IOException {
22412      return PARSER.parseFrom(input, extensionRegistry);
22413    }
22414    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseDelimitedFrom(java.io.InputStream input)
22415        throws java.io.IOException {
22416      return PARSER.parseDelimitedFrom(input);
22417    }
22418    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseDelimitedFrom(
22419        java.io.InputStream input,
22420        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
22421        throws java.io.IOException {
22422      return PARSER.parseDelimitedFrom(input, extensionRegistry);
22423    }
22424    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
22425        com.google.protobuf.CodedInputStream input)
22426        throws java.io.IOException {
22427      return PARSER.parseFrom(input);
22428    }
22429    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
22430        com.google.protobuf.CodedInputStream input,
22431        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
22432        throws java.io.IOException {
22433      return PARSER.parseFrom(input, extensionRegistry);
22434    }
22435
22436    public static Builder newBuilder() { return Builder.create(); }
22437    public Builder newBuilderForType() { return newBuilder(); }
22438    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto prototype) {
22439      return newBuilder().mergeFrom(prototype);
22440    }
22441    public Builder toBuilder() { return newBuilder(this); }
22442
22443    @java.lang.Override
22444    protected Builder newBuilderForType(
22445        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
22446      Builder builder = new Builder(parent);
22447      return builder;
22448    }
22449    /**
22450     * Protobuf type {@code hadoop.hdfs.qjournal.PrepareRecoveryRequestProto}
22451     *
22452     * <pre>
22453     **
22454     * prepareRecovery()
22455     * </pre>
22456     */
22457    public static final class Builder extends
22458        com.google.protobuf.GeneratedMessage.Builder<Builder>
22459       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProtoOrBuilder {
22460      public static final com.google.protobuf.Descriptors.Descriptor
22461          getDescriptor() {
22462        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PrepareRecoveryRequestProto_descriptor;
22463      }
22464
22465      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
22466          internalGetFieldAccessorTable() {
22467        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PrepareRecoveryRequestProto_fieldAccessorTable
22468            .ensureFieldAccessorsInitialized(
22469                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.Builder.class);
22470      }
22471
22472      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.newBuilder()
22473      private Builder() {
22474        maybeForceBuilderInitialization();
22475      }
22476
22477      private Builder(
22478          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
22479        super(parent);
22480        maybeForceBuilderInitialization();
22481      }
22482      private void maybeForceBuilderInitialization() {
22483        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
22484          getReqInfoFieldBuilder();
22485        }
22486      }
22487      private static Builder create() {
22488        return new Builder();
22489      }
22490
22491      public Builder clear() {
22492        super.clear();
22493        if (reqInfoBuilder_ == null) {
22494          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
22495        } else {
22496          reqInfoBuilder_.clear();
22497        }
22498        bitField0_ = (bitField0_ & ~0x00000001);
22499        segmentTxId_ = 0L;
22500        bitField0_ = (bitField0_ & ~0x00000002);
22501        return this;
22502      }
22503
22504      public Builder clone() {
22505        return create().mergeFrom(buildPartial());
22506      }
22507
22508      public com.google.protobuf.Descriptors.Descriptor
22509          getDescriptorForType() {
22510        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PrepareRecoveryRequestProto_descriptor;
22511      }
22512
22513      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto getDefaultInstanceForType() {
22514        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance();
22515      }
22516
22517      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto build() {
22518        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto result = buildPartial();
22519        if (!result.isInitialized()) {
22520          throw newUninitializedMessageException(result);
22521        }
22522        return result;
22523      }
22524
22525      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto buildPartial() {
22526        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto(this);
22527        int from_bitField0_ = bitField0_;
22528        int to_bitField0_ = 0;
22529        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
22530          to_bitField0_ |= 0x00000001;
22531        }
22532        if (reqInfoBuilder_ == null) {
22533          result.reqInfo_ = reqInfo_;
22534        } else {
22535          result.reqInfo_ = reqInfoBuilder_.build();
22536        }
22537        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
22538          to_bitField0_ |= 0x00000002;
22539        }
22540        result.segmentTxId_ = segmentTxId_;
22541        result.bitField0_ = to_bitField0_;
22542        onBuilt();
22543        return result;
22544      }
22545
22546      public Builder mergeFrom(com.google.protobuf.Message other) {
22547        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto) {
22548          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)other);
22549        } else {
22550          super.mergeFrom(other);
22551          return this;
22552        }
22553      }
22554
22555      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto other) {
22556        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance()) return this;
22557        if (other.hasReqInfo()) {
22558          mergeReqInfo(other.getReqInfo());
22559        }
22560        if (other.hasSegmentTxId()) {
22561          setSegmentTxId(other.getSegmentTxId());
22562        }
22563        this.mergeUnknownFields(other.getUnknownFields());
22564        return this;
22565      }
22566
22567      public final boolean isInitialized() {
22568        if (!hasReqInfo()) {
22569          
22570          return false;
22571        }
22572        if (!hasSegmentTxId()) {
22573          
22574          return false;
22575        }
22576        if (!getReqInfo().isInitialized()) {
22577          
22578          return false;
22579        }
22580        return true;
22581      }
22582
22583      public Builder mergeFrom(
22584          com.google.protobuf.CodedInputStream input,
22585          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
22586          throws java.io.IOException {
22587        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parsedMessage = null;
22588        try {
22589          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
22590        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
22591          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto) e.getUnfinishedMessage();
22592          throw e;
22593        } finally {
22594          if (parsedMessage != null) {
22595            mergeFrom(parsedMessage);
22596          }
22597        }
22598        return this;
22599      }
22600      private int bitField0_;
22601
22602      // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
22603      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
22604      private com.google.protobuf.SingleFieldBuilder<
22605          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
22606      /**
22607       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22608       */
22609      public boolean hasReqInfo() {
22610        return ((bitField0_ & 0x00000001) == 0x00000001);
22611      }
22612      /**
22613       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22614       */
22615      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
22616        if (reqInfoBuilder_ == null) {
22617          return reqInfo_;
22618        } else {
22619          return reqInfoBuilder_.getMessage();
22620        }
22621      }
22622      /**
22623       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22624       */
22625      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
22626        if (reqInfoBuilder_ == null) {
22627          if (value == null) {
22628            throw new NullPointerException();
22629          }
22630          reqInfo_ = value;
22631          onChanged();
22632        } else {
22633          reqInfoBuilder_.setMessage(value);
22634        }
22635        bitField0_ |= 0x00000001;
22636        return this;
22637      }
22638      /**
22639       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22640       */
22641      public Builder setReqInfo(
22642          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
22643        if (reqInfoBuilder_ == null) {
22644          reqInfo_ = builderForValue.build();
22645          onChanged();
22646        } else {
22647          reqInfoBuilder_.setMessage(builderForValue.build());
22648        }
22649        bitField0_ |= 0x00000001;
22650        return this;
22651      }
22652      /**
22653       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22654       */
22655      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
22656        if (reqInfoBuilder_ == null) {
22657          if (((bitField0_ & 0x00000001) == 0x00000001) &&
22658              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
22659            reqInfo_ =
22660              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
22661          } else {
22662            reqInfo_ = value;
22663          }
22664          onChanged();
22665        } else {
22666          reqInfoBuilder_.mergeFrom(value);
22667        }
22668        bitField0_ |= 0x00000001;
22669        return this;
22670      }
22671      /**
22672       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22673       */
22674      public Builder clearReqInfo() {
22675        if (reqInfoBuilder_ == null) {
22676          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
22677          onChanged();
22678        } else {
22679          reqInfoBuilder_.clear();
22680        }
22681        bitField0_ = (bitField0_ & ~0x00000001);
22682        return this;
22683      }
22684      /**
22685       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22686       */
22687      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
22688        bitField0_ |= 0x00000001;
22689        onChanged();
22690        return getReqInfoFieldBuilder().getBuilder();
22691      }
22692      /**
22693       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22694       */
22695      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
22696        if (reqInfoBuilder_ != null) {
22697          return reqInfoBuilder_.getMessageOrBuilder();
22698        } else {
22699          return reqInfo_;
22700        }
22701      }
22702      /**
22703       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
22704       */
22705      private com.google.protobuf.SingleFieldBuilder<
22706          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
22707          getReqInfoFieldBuilder() {
22708        if (reqInfoBuilder_ == null) {
22709          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
22710              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
22711                  reqInfo_,
22712                  getParentForChildren(),
22713                  isClean());
22714          reqInfo_ = null;
22715        }
22716        return reqInfoBuilder_;
22717      }
22718
22719      // required uint64 segmentTxId = 2;
22720      private long segmentTxId_ ;
22721      /**
22722       * <code>required uint64 segmentTxId = 2;</code>
22723       */
22724      public boolean hasSegmentTxId() {
22725        return ((bitField0_ & 0x00000002) == 0x00000002);
22726      }
22727      /**
22728       * <code>required uint64 segmentTxId = 2;</code>
22729       */
22730      public long getSegmentTxId() {
22731        return segmentTxId_;
22732      }
22733      /**
22734       * <code>required uint64 segmentTxId = 2;</code>
22735       */
22736      public Builder setSegmentTxId(long value) {
22737        bitField0_ |= 0x00000002;
22738        segmentTxId_ = value;
22739        onChanged();
22740        return this;
22741      }
22742      /**
22743       * <code>required uint64 segmentTxId = 2;</code>
22744       */
22745      public Builder clearSegmentTxId() {
22746        bitField0_ = (bitField0_ & ~0x00000002);
22747        segmentTxId_ = 0L;
22748        onChanged();
22749        return this;
22750      }
22751
22752      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.PrepareRecoveryRequestProto)
22753    }
22754
22755    static {
22756      defaultInstance = new PrepareRecoveryRequestProto(true);
22757      defaultInstance.initFields();
22758    }
22759
22760    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.PrepareRecoveryRequestProto)
22761  }
22762
22763  public interface PrepareRecoveryResponseProtoOrBuilder
22764      extends com.google.protobuf.MessageOrBuilder {
22765
22766    // optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;
22767    /**
22768     * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
22769     */
22770    boolean hasSegmentState();
22771    /**
22772     * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
22773     */
22774    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState();
22775    /**
22776     * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
22777     */
22778    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder();
22779
22780    // optional uint64 acceptedInEpoch = 2;
22781    /**
22782     * <code>optional uint64 acceptedInEpoch = 2;</code>
22783     */
22784    boolean hasAcceptedInEpoch();
22785    /**
22786     * <code>optional uint64 acceptedInEpoch = 2;</code>
22787     */
22788    long getAcceptedInEpoch();
22789
22790    // required uint64 lastWriterEpoch = 3;
22791    /**
22792     * <code>required uint64 lastWriterEpoch = 3;</code>
22793     */
22794    boolean hasLastWriterEpoch();
22795    /**
22796     * <code>required uint64 lastWriterEpoch = 3;</code>
22797     */
22798    long getLastWriterEpoch();
22799
22800    // optional uint64 lastCommittedTxId = 4;
22801    /**
22802     * <code>optional uint64 lastCommittedTxId = 4;</code>
22803     *
22804     * <pre>
22805     * The highest committed txid that this logger has ever seen.
22806     * This may be higher than the data it actually has, in the case
22807     * that it was lagging before the old writer crashed.
22808     * </pre>
22809     */
22810    boolean hasLastCommittedTxId();
22811    /**
22812     * <code>optional uint64 lastCommittedTxId = 4;</code>
22813     *
22814     * <pre>
22815     * The highest committed txid that this logger has ever seen.
22816     * This may be higher than the data it actually has, in the case
22817     * that it was lagging before the old writer crashed.
22818     * </pre>
22819     */
22820    long getLastCommittedTxId();
22821  }
22822  /**
22823   * Protobuf type {@code hadoop.hdfs.qjournal.PrepareRecoveryResponseProto}
22824   */
22825  public static final class PrepareRecoveryResponseProto extends
22826      com.google.protobuf.GeneratedMessage
22827      implements PrepareRecoveryResponseProtoOrBuilder {
22828    // Use PrepareRecoveryResponseProto.newBuilder() to construct.
22829    private PrepareRecoveryResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
22830      super(builder);
22831      this.unknownFields = builder.getUnknownFields();
22832    }
22833    private PrepareRecoveryResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
22834
22835    private static final PrepareRecoveryResponseProto defaultInstance;
22836    public static PrepareRecoveryResponseProto getDefaultInstance() {
22837      return defaultInstance;
22838    }
22839
22840    public PrepareRecoveryResponseProto getDefaultInstanceForType() {
22841      return defaultInstance;
22842    }
22843
22844    private final com.google.protobuf.UnknownFieldSet unknownFields;
22845    @java.lang.Override
22846    public final com.google.protobuf.UnknownFieldSet
22847        getUnknownFields() {
22848      return this.unknownFields;
22849    }
22850    private PrepareRecoveryResponseProto(
22851        com.google.protobuf.CodedInputStream input,
22852        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
22853        throws com.google.protobuf.InvalidProtocolBufferException {
22854      initFields();
22855      int mutable_bitField0_ = 0;
22856      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
22857          com.google.protobuf.UnknownFieldSet.newBuilder();
22858      try {
22859        boolean done = false;
22860        while (!done) {
22861          int tag = input.readTag();
22862          switch (tag) {
22863            case 0:
22864              done = true;
22865              break;
22866            default: {
22867              if (!parseUnknownField(input, unknownFields,
22868                                     extensionRegistry, tag)) {
22869                done = true;
22870              }
22871              break;
22872            }
22873            case 10: {
22874              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = null;
22875              if (((bitField0_ & 0x00000001) == 0x00000001)) {
22876                subBuilder = segmentState_.toBuilder();
22877              }
22878              segmentState_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.PARSER, extensionRegistry);
22879              if (subBuilder != null) {
22880                subBuilder.mergeFrom(segmentState_);
22881                segmentState_ = subBuilder.buildPartial();
22882              }
22883              bitField0_ |= 0x00000001;
22884              break;
22885            }
22886            case 16: {
22887              bitField0_ |= 0x00000002;
22888              acceptedInEpoch_ = input.readUInt64();
22889              break;
22890            }
22891            case 24: {
22892              bitField0_ |= 0x00000004;
22893              lastWriterEpoch_ = input.readUInt64();
22894              break;
22895            }
22896            case 32: {
22897              bitField0_ |= 0x00000008;
22898              lastCommittedTxId_ = input.readUInt64();
22899              break;
22900            }
22901          }
22902        }
22903      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
22904        throw e.setUnfinishedMessage(this);
22905      } catch (java.io.IOException e) {
22906        throw new com.google.protobuf.InvalidProtocolBufferException(
22907            e.getMessage()).setUnfinishedMessage(this);
22908      } finally {
22909        this.unknownFields = unknownFields.build();
22910        makeExtensionsImmutable();
22911      }
22912    }
22913    public static final com.google.protobuf.Descriptors.Descriptor
22914        getDescriptor() {
22915      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PrepareRecoveryResponseProto_descriptor;
22916    }
22917
22918    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
22919        internalGetFieldAccessorTable() {
22920      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PrepareRecoveryResponseProto_fieldAccessorTable
22921          .ensureFieldAccessorsInitialized(
22922              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.Builder.class);
22923    }
22924
22925    public static com.google.protobuf.Parser<PrepareRecoveryResponseProto> PARSER =
22926        new com.google.protobuf.AbstractParser<PrepareRecoveryResponseProto>() {
22927      public PrepareRecoveryResponseProto parsePartialFrom(
22928          com.google.protobuf.CodedInputStream input,
22929          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
22930          throws com.google.protobuf.InvalidProtocolBufferException {
22931        return new PrepareRecoveryResponseProto(input, extensionRegistry);
22932      }
22933    };
22934
22935    @java.lang.Override
22936    public com.google.protobuf.Parser<PrepareRecoveryResponseProto> getParserForType() {
22937      return PARSER;
22938    }
22939
22940    private int bitField0_;
22941    // optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;
22942    public static final int SEGMENTSTATE_FIELD_NUMBER = 1;
22943    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_;
22944    /**
22945     * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
22946     */
22947    public boolean hasSegmentState() {
22948      return ((bitField0_ & 0x00000001) == 0x00000001);
22949    }
22950    /**
22951     * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
22952     */
22953    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
22954      return segmentState_;
22955    }
22956    /**
22957     * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
22958     */
22959    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
22960      return segmentState_;
22961    }
22962
22963    // optional uint64 acceptedInEpoch = 2;
22964    public static final int ACCEPTEDINEPOCH_FIELD_NUMBER = 2;
22965    private long acceptedInEpoch_;
22966    /**
22967     * <code>optional uint64 acceptedInEpoch = 2;</code>
22968     */
22969    public boolean hasAcceptedInEpoch() {
22970      return ((bitField0_ & 0x00000002) == 0x00000002);
22971    }
22972    /**
22973     * <code>optional uint64 acceptedInEpoch = 2;</code>
22974     */
22975    public long getAcceptedInEpoch() {
22976      return acceptedInEpoch_;
22977    }
22978
22979    // required uint64 lastWriterEpoch = 3;
22980    public static final int LASTWRITEREPOCH_FIELD_NUMBER = 3;
22981    private long lastWriterEpoch_;
22982    /**
22983     * <code>required uint64 lastWriterEpoch = 3;</code>
22984     */
22985    public boolean hasLastWriterEpoch() {
22986      return ((bitField0_ & 0x00000004) == 0x00000004);
22987    }
22988    /**
22989     * <code>required uint64 lastWriterEpoch = 3;</code>
22990     */
22991    public long getLastWriterEpoch() {
22992      return lastWriterEpoch_;
22993    }
22994
22995    // optional uint64 lastCommittedTxId = 4;
22996    public static final int LASTCOMMITTEDTXID_FIELD_NUMBER = 4;
22997    private long lastCommittedTxId_;
22998    /**
22999     * <code>optional uint64 lastCommittedTxId = 4;</code>
23000     *
23001     * <pre>
23002     * The highest committed txid that this logger has ever seen.
23003     * This may be higher than the data it actually has, in the case
23004     * that it was lagging before the old writer crashed.
23005     * </pre>
23006     */
23007    public boolean hasLastCommittedTxId() {
23008      return ((bitField0_ & 0x00000008) == 0x00000008);
23009    }
23010    /**
23011     * <code>optional uint64 lastCommittedTxId = 4;</code>
23012     *
23013     * <pre>
23014     * The highest committed txid that this logger has ever seen.
23015     * This may be higher than the data it actually has, in the case
23016     * that it was lagging before the old writer crashed.
23017     * </pre>
23018     */
23019    public long getLastCommittedTxId() {
23020      return lastCommittedTxId_;
23021    }
23022
23023    private void initFields() {
23024      segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
23025      acceptedInEpoch_ = 0L;
23026      lastWriterEpoch_ = 0L;
23027      lastCommittedTxId_ = 0L;
23028    }
23029    private byte memoizedIsInitialized = -1;
23030    public final boolean isInitialized() {
23031      byte isInitialized = memoizedIsInitialized;
23032      if (isInitialized != -1) return isInitialized == 1;
23033
23034      if (!hasLastWriterEpoch()) {
23035        memoizedIsInitialized = 0;
23036        return false;
23037      }
23038      if (hasSegmentState()) {
23039        if (!getSegmentState().isInitialized()) {
23040          memoizedIsInitialized = 0;
23041          return false;
23042        }
23043      }
23044      memoizedIsInitialized = 1;
23045      return true;
23046    }
23047
23048    public void writeTo(com.google.protobuf.CodedOutputStream output)
23049                        throws java.io.IOException {
23050      getSerializedSize();
23051      if (((bitField0_ & 0x00000001) == 0x00000001)) {
23052        output.writeMessage(1, segmentState_);
23053      }
23054      if (((bitField0_ & 0x00000002) == 0x00000002)) {
23055        output.writeUInt64(2, acceptedInEpoch_);
23056      }
23057      if (((bitField0_ & 0x00000004) == 0x00000004)) {
23058        output.writeUInt64(3, lastWriterEpoch_);
23059      }
23060      if (((bitField0_ & 0x00000008) == 0x00000008)) {
23061        output.writeUInt64(4, lastCommittedTxId_);
23062      }
23063      getUnknownFields().writeTo(output);
23064    }
23065
23066    private int memoizedSerializedSize = -1;
23067    public int getSerializedSize() {
23068      int size = memoizedSerializedSize;
23069      if (size != -1) return size;
23070
23071      size = 0;
23072      if (((bitField0_ & 0x00000001) == 0x00000001)) {
23073        size += com.google.protobuf.CodedOutputStream
23074          .computeMessageSize(1, segmentState_);
23075      }
23076      if (((bitField0_ & 0x00000002) == 0x00000002)) {
23077        size += com.google.protobuf.CodedOutputStream
23078          .computeUInt64Size(2, acceptedInEpoch_);
23079      }
23080      if (((bitField0_ & 0x00000004) == 0x00000004)) {
23081        size += com.google.protobuf.CodedOutputStream
23082          .computeUInt64Size(3, lastWriterEpoch_);
23083      }
23084      if (((bitField0_ & 0x00000008) == 0x00000008)) {
23085        size += com.google.protobuf.CodedOutputStream
23086          .computeUInt64Size(4, lastCommittedTxId_);
23087      }
23088      size += getUnknownFields().getSerializedSize();
23089      memoizedSerializedSize = size;
23090      return size;
23091    }
23092
23093    private static final long serialVersionUID = 0L;
23094    @java.lang.Override
23095    protected java.lang.Object writeReplace()
23096        throws java.io.ObjectStreamException {
23097      return super.writeReplace();
23098    }
23099
23100    @java.lang.Override
23101    public boolean equals(final java.lang.Object obj) {
23102      if (obj == this) {
23103       return true;
23104      }
23105      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto)) {
23106        return super.equals(obj);
23107      }
23108      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) obj;
23109
23110      boolean result = true;
23111      result = result && (hasSegmentState() == other.hasSegmentState());
23112      if (hasSegmentState()) {
23113        result = result && getSegmentState()
23114            .equals(other.getSegmentState());
23115      }
23116      result = result && (hasAcceptedInEpoch() == other.hasAcceptedInEpoch());
23117      if (hasAcceptedInEpoch()) {
23118        result = result && (getAcceptedInEpoch()
23119            == other.getAcceptedInEpoch());
23120      }
23121      result = result && (hasLastWriterEpoch() == other.hasLastWriterEpoch());
23122      if (hasLastWriterEpoch()) {
23123        result = result && (getLastWriterEpoch()
23124            == other.getLastWriterEpoch());
23125      }
23126      result = result && (hasLastCommittedTxId() == other.hasLastCommittedTxId());
23127      if (hasLastCommittedTxId()) {
23128        result = result && (getLastCommittedTxId()
23129            == other.getLastCommittedTxId());
23130      }
23131      result = result &&
23132          getUnknownFields().equals(other.getUnknownFields());
23133      return result;
23134    }
23135
23136    private int memoizedHashCode = 0;
23137    @java.lang.Override
23138    public int hashCode() {
23139      if (memoizedHashCode != 0) {
23140        return memoizedHashCode;
23141      }
23142      int hash = 41;
23143      hash = (19 * hash) + getDescriptorForType().hashCode();
23144      if (hasSegmentState()) {
23145        hash = (37 * hash) + SEGMENTSTATE_FIELD_NUMBER;
23146        hash = (53 * hash) + getSegmentState().hashCode();
23147      }
23148      if (hasAcceptedInEpoch()) {
23149        hash = (37 * hash) + ACCEPTEDINEPOCH_FIELD_NUMBER;
23150        hash = (53 * hash) + hashLong(getAcceptedInEpoch());
23151      }
23152      if (hasLastWriterEpoch()) {
23153        hash = (37 * hash) + LASTWRITEREPOCH_FIELD_NUMBER;
23154        hash = (53 * hash) + hashLong(getLastWriterEpoch());
23155      }
23156      if (hasLastCommittedTxId()) {
23157        hash = (37 * hash) + LASTCOMMITTEDTXID_FIELD_NUMBER;
23158        hash = (53 * hash) + hashLong(getLastCommittedTxId());
23159      }
23160      hash = (29 * hash) + getUnknownFields().hashCode();
23161      memoizedHashCode = hash;
23162      return hash;
23163    }
23164
23165    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
23166        com.google.protobuf.ByteString data)
23167        throws com.google.protobuf.InvalidProtocolBufferException {
23168      return PARSER.parseFrom(data);
23169    }
23170    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
23171        com.google.protobuf.ByteString data,
23172        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
23173        throws com.google.protobuf.InvalidProtocolBufferException {
23174      return PARSER.parseFrom(data, extensionRegistry);
23175    }
23176    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(byte[] data)
23177        throws com.google.protobuf.InvalidProtocolBufferException {
23178      return PARSER.parseFrom(data);
23179    }
23180    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
23181        byte[] data,
23182        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
23183        throws com.google.protobuf.InvalidProtocolBufferException {
23184      return PARSER.parseFrom(data, extensionRegistry);
23185    }
23186    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(java.io.InputStream input)
23187        throws java.io.IOException {
23188      return PARSER.parseFrom(input);
23189    }
23190    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
23191        java.io.InputStream input,
23192        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
23193        throws java.io.IOException {
23194      return PARSER.parseFrom(input, extensionRegistry);
23195    }
23196    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseDelimitedFrom(java.io.InputStream input)
23197        throws java.io.IOException {
23198      return PARSER.parseDelimitedFrom(input);
23199    }
23200    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseDelimitedFrom(
23201        java.io.InputStream input,
23202        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
23203        throws java.io.IOException {
23204      return PARSER.parseDelimitedFrom(input, extensionRegistry);
23205    }
23206    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
23207        com.google.protobuf.CodedInputStream input)
23208        throws java.io.IOException {
23209      return PARSER.parseFrom(input);
23210    }
23211    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
23212        com.google.protobuf.CodedInputStream input,
23213        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
23214        throws java.io.IOException {
23215      return PARSER.parseFrom(input, extensionRegistry);
23216    }
23217
23218    public static Builder newBuilder() { return Builder.create(); }
23219    public Builder newBuilderForType() { return newBuilder(); }
23220    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prototype) {
23221      return newBuilder().mergeFrom(prototype);
23222    }
23223    public Builder toBuilder() { return newBuilder(this); }
23224
23225    @java.lang.Override
23226    protected Builder newBuilderForType(
23227        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
23228      Builder builder = new Builder(parent);
23229      return builder;
23230    }
23231    /**
23232     * Protobuf type {@code hadoop.hdfs.qjournal.PrepareRecoveryResponseProto}
23233     */
23234    public static final class Builder extends
23235        com.google.protobuf.GeneratedMessage.Builder<Builder>
23236       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProtoOrBuilder {
23237      public static final com.google.protobuf.Descriptors.Descriptor
23238          getDescriptor() {
23239        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PrepareRecoveryResponseProto_descriptor;
23240      }
23241
23242      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
23243          internalGetFieldAccessorTable() {
23244        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PrepareRecoveryResponseProto_fieldAccessorTable
23245            .ensureFieldAccessorsInitialized(
23246                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.Builder.class);
23247      }
23248
23249      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.newBuilder()
23250      private Builder() {
23251        maybeForceBuilderInitialization();
23252      }
23253
23254      private Builder(
23255          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
23256        super(parent);
23257        maybeForceBuilderInitialization();
23258      }
23259      private void maybeForceBuilderInitialization() {
23260        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
23261          getSegmentStateFieldBuilder();
23262        }
23263      }
23264      private static Builder create() {
23265        return new Builder();
23266      }
23267
23268      public Builder clear() {
23269        super.clear();
23270        if (segmentStateBuilder_ == null) {
23271          segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
23272        } else {
23273          segmentStateBuilder_.clear();
23274        }
23275        bitField0_ = (bitField0_ & ~0x00000001);
23276        acceptedInEpoch_ = 0L;
23277        bitField0_ = (bitField0_ & ~0x00000002);
23278        lastWriterEpoch_ = 0L;
23279        bitField0_ = (bitField0_ & ~0x00000004);
23280        lastCommittedTxId_ = 0L;
23281        bitField0_ = (bitField0_ & ~0x00000008);
23282        return this;
23283      }
23284
23285      public Builder clone() {
23286        return create().mergeFrom(buildPartial());
23287      }
23288
23289      public com.google.protobuf.Descriptors.Descriptor
23290          getDescriptorForType() {
23291        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_PrepareRecoveryResponseProto_descriptor;
23292      }
23293
23294      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto getDefaultInstanceForType() {
23295        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance();
23296      }
23297
23298      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto build() {
23299        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto result = buildPartial();
23300        if (!result.isInitialized()) {
23301          throw newUninitializedMessageException(result);
23302        }
23303        return result;
23304      }
23305
23306      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto buildPartial() {
23307        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto(this);
23308        int from_bitField0_ = bitField0_;
23309        int to_bitField0_ = 0;
23310        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
23311          to_bitField0_ |= 0x00000001;
23312        }
23313        if (segmentStateBuilder_ == null) {
23314          result.segmentState_ = segmentState_;
23315        } else {
23316          result.segmentState_ = segmentStateBuilder_.build();
23317        }
23318        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
23319          to_bitField0_ |= 0x00000002;
23320        }
23321        result.acceptedInEpoch_ = acceptedInEpoch_;
23322        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
23323          to_bitField0_ |= 0x00000004;
23324        }
23325        result.lastWriterEpoch_ = lastWriterEpoch_;
23326        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
23327          to_bitField0_ |= 0x00000008;
23328        }
23329        result.lastCommittedTxId_ = lastCommittedTxId_;
23330        result.bitField0_ = to_bitField0_;
23331        onBuilt();
23332        return result;
23333      }
23334
23335      public Builder mergeFrom(com.google.protobuf.Message other) {
23336        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) {
23337          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto)other);
23338        } else {
23339          super.mergeFrom(other);
23340          return this;
23341        }
23342      }
23343
23344      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto other) {
23345        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance()) return this;
23346        if (other.hasSegmentState()) {
23347          mergeSegmentState(other.getSegmentState());
23348        }
23349        if (other.hasAcceptedInEpoch()) {
23350          setAcceptedInEpoch(other.getAcceptedInEpoch());
23351        }
23352        if (other.hasLastWriterEpoch()) {
23353          setLastWriterEpoch(other.getLastWriterEpoch());
23354        }
23355        if (other.hasLastCommittedTxId()) {
23356          setLastCommittedTxId(other.getLastCommittedTxId());
23357        }
23358        this.mergeUnknownFields(other.getUnknownFields());
23359        return this;
23360      }
23361
23362      public final boolean isInitialized() {
23363        if (!hasLastWriterEpoch()) {
23364          
23365          return false;
23366        }
23367        if (hasSegmentState()) {
23368          if (!getSegmentState().isInitialized()) {
23369            
23370            return false;
23371          }
23372        }
23373        return true;
23374      }
23375
23376      public Builder mergeFrom(
23377          com.google.protobuf.CodedInputStream input,
23378          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
23379          throws java.io.IOException {
23380        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parsedMessage = null;
23381        try {
23382          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
23383        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
23384          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) e.getUnfinishedMessage();
23385          throw e;
23386        } finally {
23387          if (parsedMessage != null) {
23388            mergeFrom(parsedMessage);
23389          }
23390        }
23391        return this;
23392      }
23393      private int bitField0_;
23394
23395      // optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;
23396      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
23397      private com.google.protobuf.SingleFieldBuilder<
23398          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> segmentStateBuilder_;
23399      /**
23400       * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
23401       */
23402      public boolean hasSegmentState() {
23403        return ((bitField0_ & 0x00000001) == 0x00000001);
23404      }
23405      /**
23406       * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
23407       */
23408      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
23409        if (segmentStateBuilder_ == null) {
23410          return segmentState_;
23411        } else {
23412          return segmentStateBuilder_.getMessage();
23413        }
23414      }
23415      /**
23416       * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
23417       */
23418      public Builder setSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
23419        if (segmentStateBuilder_ == null) {
23420          if (value == null) {
23421            throw new NullPointerException();
23422          }
23423          segmentState_ = value;
23424          onChanged();
23425        } else {
23426          segmentStateBuilder_.setMessage(value);
23427        }
23428        bitField0_ |= 0x00000001;
23429        return this;
23430      }
23431      /**
23432       * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
23433       */
23434      public Builder setSegmentState(
23435          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) {
23436        if (segmentStateBuilder_ == null) {
23437          segmentState_ = builderForValue.build();
23438          onChanged();
23439        } else {
23440          segmentStateBuilder_.setMessage(builderForValue.build());
23441        }
23442        bitField0_ |= 0x00000001;
23443        return this;
23444      }
23445      /**
23446       * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
23447       */
23448      public Builder mergeSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
23449        if (segmentStateBuilder_ == null) {
23450          if (((bitField0_ & 0x00000001) == 0x00000001) &&
23451              segmentState_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) {
23452            segmentState_ =
23453              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(segmentState_).mergeFrom(value).buildPartial();
23454          } else {
23455            segmentState_ = value;
23456          }
23457          onChanged();
23458        } else {
23459          segmentStateBuilder_.mergeFrom(value);
23460        }
23461        bitField0_ |= 0x00000001;
23462        return this;
23463      }
23464      /**
23465       * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
23466       */
23467      public Builder clearSegmentState() {
23468        if (segmentStateBuilder_ == null) {
23469          segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
23470          onChanged();
23471        } else {
23472          segmentStateBuilder_.clear();
23473        }
23474        bitField0_ = (bitField0_ & ~0x00000001);
23475        return this;
23476      }
23477      /**
23478       * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
23479       */
23480      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getSegmentStateBuilder() {
23481        bitField0_ |= 0x00000001;
23482        onChanged();
23483        return getSegmentStateFieldBuilder().getBuilder();
23484      }
23485      /**
23486       * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
23487       */
23488      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
23489        if (segmentStateBuilder_ != null) {
23490          return segmentStateBuilder_.getMessageOrBuilder();
23491        } else {
23492          return segmentState_;
23493        }
23494      }
23495      /**
23496       * <code>optional .hadoop.hdfs.qjournal.SegmentStateProto segmentState = 1;</code>
23497       */
23498      private com.google.protobuf.SingleFieldBuilder<
23499          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 
23500          getSegmentStateFieldBuilder() {
23501        if (segmentStateBuilder_ == null) {
23502          segmentStateBuilder_ = new com.google.protobuf.SingleFieldBuilder<
23503              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>(
23504                  segmentState_,
23505                  getParentForChildren(),
23506                  isClean());
23507          segmentState_ = null;
23508        }
23509        return segmentStateBuilder_;
23510      }
23511
23512      // optional uint64 acceptedInEpoch = 2;
23513      private long acceptedInEpoch_ ;
23514      /**
23515       * <code>optional uint64 acceptedInEpoch = 2;</code>
23516       */
23517      public boolean hasAcceptedInEpoch() {
23518        return ((bitField0_ & 0x00000002) == 0x00000002);
23519      }
23520      /**
23521       * <code>optional uint64 acceptedInEpoch = 2;</code>
23522       */
23523      public long getAcceptedInEpoch() {
23524        return acceptedInEpoch_;
23525      }
23526      /**
23527       * <code>optional uint64 acceptedInEpoch = 2;</code>
23528       */
23529      public Builder setAcceptedInEpoch(long value) {
23530        bitField0_ |= 0x00000002;
23531        acceptedInEpoch_ = value;
23532        onChanged();
23533        return this;
23534      }
23535      /**
23536       * <code>optional uint64 acceptedInEpoch = 2;</code>
23537       */
23538      public Builder clearAcceptedInEpoch() {
23539        bitField0_ = (bitField0_ & ~0x00000002);
23540        acceptedInEpoch_ = 0L;
23541        onChanged();
23542        return this;
23543      }
23544
23545      // required uint64 lastWriterEpoch = 3;
23546      private long lastWriterEpoch_ ;
23547      /**
23548       * <code>required uint64 lastWriterEpoch = 3;</code>
23549       */
23550      public boolean hasLastWriterEpoch() {
23551        return ((bitField0_ & 0x00000004) == 0x00000004);
23552      }
23553      /**
23554       * <code>required uint64 lastWriterEpoch = 3;</code>
23555       */
23556      public long getLastWriterEpoch() {
23557        return lastWriterEpoch_;
23558      }
23559      /**
23560       * <code>required uint64 lastWriterEpoch = 3;</code>
23561       */
23562      public Builder setLastWriterEpoch(long value) {
23563        bitField0_ |= 0x00000004;
23564        lastWriterEpoch_ = value;
23565        onChanged();
23566        return this;
23567      }
23568      /**
23569       * <code>required uint64 lastWriterEpoch = 3;</code>
23570       */
23571      public Builder clearLastWriterEpoch() {
23572        bitField0_ = (bitField0_ & ~0x00000004);
23573        lastWriterEpoch_ = 0L;
23574        onChanged();
23575        return this;
23576      }
23577
23578      // optional uint64 lastCommittedTxId = 4;
23579      private long lastCommittedTxId_ ;
23580      /**
23581       * <code>optional uint64 lastCommittedTxId = 4;</code>
23582       *
23583       * <pre>
23584       * The highest committed txid that this logger has ever seen.
23585       * This may be higher than the data it actually has, in the case
23586       * that it was lagging before the old writer crashed.
23587       * </pre>
23588       */
23589      public boolean hasLastCommittedTxId() {
23590        return ((bitField0_ & 0x00000008) == 0x00000008);
23591      }
23592      /**
23593       * <code>optional uint64 lastCommittedTxId = 4;</code>
23594       *
23595       * <pre>
23596       * The highest committed txid that this logger has ever seen.
23597       * This may be higher than the data it actually has, in the case
23598       * that it was lagging before the old writer crashed.
23599       * </pre>
23600       */
23601      public long getLastCommittedTxId() {
23602        return lastCommittedTxId_;
23603      }
23604      /**
23605       * <code>optional uint64 lastCommittedTxId = 4;</code>
23606       *
23607       * <pre>
23608       * The highest committed txid that this logger has ever seen.
23609       * This may be higher than the data it actually has, in the case
23610       * that it was lagging before the old writer crashed.
23611       * </pre>
23612       */
23613      public Builder setLastCommittedTxId(long value) {
23614        bitField0_ |= 0x00000008;
23615        lastCommittedTxId_ = value;
23616        onChanged();
23617        return this;
23618      }
23619      /**
23620       * <code>optional uint64 lastCommittedTxId = 4;</code>
23621       *
23622       * <pre>
23623       * The highest committed txid that this logger has ever seen.
23624       * This may be higher than the data it actually has, in the case
23625       * that it was lagging before the old writer crashed.
23626       * </pre>
23627       */
23628      public Builder clearLastCommittedTxId() {
23629        bitField0_ = (bitField0_ & ~0x00000008);
23630        lastCommittedTxId_ = 0L;
23631        onChanged();
23632        return this;
23633      }
23634
23635      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.PrepareRecoveryResponseProto)
23636    }
23637
23638    static {
23639      defaultInstance = new PrepareRecoveryResponseProto(true);
23640      defaultInstance.initFields();
23641    }
23642
23643    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.PrepareRecoveryResponseProto)
23644  }
23645
23646  public interface AcceptRecoveryRequestProtoOrBuilder
23647      extends com.google.protobuf.MessageOrBuilder {
23648
23649    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
23650    /**
23651     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
23652     */
23653    boolean hasReqInfo();
23654    /**
23655     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
23656     */
23657    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
23658    /**
23659     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
23660     */
23661    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
23662
23663    // required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;
23664    /**
23665     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
23666     *
23667     * <pre>
23668     ** Details on the segment to recover 
23669     * </pre>
23670     */
23671    boolean hasStateToAccept();
23672    /**
23673     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
23674     *
23675     * <pre>
23676     ** Details on the segment to recover 
23677     * </pre>
23678     */
23679    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept();
23680    /**
23681     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
23682     *
23683     * <pre>
23684     ** Details on the segment to recover 
23685     * </pre>
23686     */
23687    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder();
23688
23689    // required string fromURL = 3;
23690    /**
23691     * <code>required string fromURL = 3;</code>
23692     *
23693     * <pre>
23694     ** The URL from which the log may be copied 
23695     * </pre>
23696     */
23697    boolean hasFromURL();
23698    /**
23699     * <code>required string fromURL = 3;</code>
23700     *
23701     * <pre>
23702     ** The URL from which the log may be copied 
23703     * </pre>
23704     */
23705    java.lang.String getFromURL();
23706    /**
23707     * <code>required string fromURL = 3;</code>
23708     *
23709     * <pre>
23710     ** The URL from which the log may be copied 
23711     * </pre>
23712     */
23713    com.google.protobuf.ByteString
23714        getFromURLBytes();
23715  }
23716  /**
23717   * Protobuf type {@code hadoop.hdfs.qjournal.AcceptRecoveryRequestProto}
23718   *
23719   * <pre>
23720   **
23721   * acceptRecovery()
23722   * </pre>
23723   */
23724  public static final class AcceptRecoveryRequestProto extends
23725      com.google.protobuf.GeneratedMessage
23726      implements AcceptRecoveryRequestProtoOrBuilder {
23727    // Use AcceptRecoveryRequestProto.newBuilder() to construct.
23728    private AcceptRecoveryRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
23729      super(builder);
23730      this.unknownFields = builder.getUnknownFields();
23731    }
23732    private AcceptRecoveryRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
23733
23734    private static final AcceptRecoveryRequestProto defaultInstance;
23735    public static AcceptRecoveryRequestProto getDefaultInstance() {
23736      return defaultInstance;
23737    }
23738
23739    public AcceptRecoveryRequestProto getDefaultInstanceForType() {
23740      return defaultInstance;
23741    }
23742
23743    private final com.google.protobuf.UnknownFieldSet unknownFields;
23744    @java.lang.Override
23745    public final com.google.protobuf.UnknownFieldSet
23746        getUnknownFields() {
23747      return this.unknownFields;
23748    }
23749    private AcceptRecoveryRequestProto(
23750        com.google.protobuf.CodedInputStream input,
23751        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
23752        throws com.google.protobuf.InvalidProtocolBufferException {
23753      initFields();
23754      int mutable_bitField0_ = 0;
23755      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
23756          com.google.protobuf.UnknownFieldSet.newBuilder();
23757      try {
23758        boolean done = false;
23759        while (!done) {
23760          int tag = input.readTag();
23761          switch (tag) {
23762            case 0:
23763              done = true;
23764              break;
23765            default: {
23766              if (!parseUnknownField(input, unknownFields,
23767                                     extensionRegistry, tag)) {
23768                done = true;
23769              }
23770              break;
23771            }
23772            case 10: {
23773              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
23774              if (((bitField0_ & 0x00000001) == 0x00000001)) {
23775                subBuilder = reqInfo_.toBuilder();
23776              }
23777              reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
23778              if (subBuilder != null) {
23779                subBuilder.mergeFrom(reqInfo_);
23780                reqInfo_ = subBuilder.buildPartial();
23781              }
23782              bitField0_ |= 0x00000001;
23783              break;
23784            }
23785            case 18: {
23786              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = null;
23787              if (((bitField0_ & 0x00000002) == 0x00000002)) {
23788                subBuilder = stateToAccept_.toBuilder();
23789              }
23790              stateToAccept_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.PARSER, extensionRegistry);
23791              if (subBuilder != null) {
23792                subBuilder.mergeFrom(stateToAccept_);
23793                stateToAccept_ = subBuilder.buildPartial();
23794              }
23795              bitField0_ |= 0x00000002;
23796              break;
23797            }
23798            case 26: {
23799              bitField0_ |= 0x00000004;
23800              fromURL_ = input.readBytes();
23801              break;
23802            }
23803          }
23804        }
23805      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
23806        throw e.setUnfinishedMessage(this);
23807      } catch (java.io.IOException e) {
23808        throw new com.google.protobuf.InvalidProtocolBufferException(
23809            e.getMessage()).setUnfinishedMessage(this);
23810      } finally {
23811        this.unknownFields = unknownFields.build();
23812        makeExtensionsImmutable();
23813      }
23814    }
23815    public static final com.google.protobuf.Descriptors.Descriptor
23816        getDescriptor() {
23817      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_AcceptRecoveryRequestProto_descriptor;
23818    }
23819
23820    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
23821        internalGetFieldAccessorTable() {
23822      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_AcceptRecoveryRequestProto_fieldAccessorTable
23823          .ensureFieldAccessorsInitialized(
23824              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.Builder.class);
23825    }
23826
23827    public static com.google.protobuf.Parser<AcceptRecoveryRequestProto> PARSER =
23828        new com.google.protobuf.AbstractParser<AcceptRecoveryRequestProto>() {
23829      public AcceptRecoveryRequestProto parsePartialFrom(
23830          com.google.protobuf.CodedInputStream input,
23831          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
23832          throws com.google.protobuf.InvalidProtocolBufferException {
23833        return new AcceptRecoveryRequestProto(input, extensionRegistry);
23834      }
23835    };
23836
23837    @java.lang.Override
23838    public com.google.protobuf.Parser<AcceptRecoveryRequestProto> getParserForType() {
23839      return PARSER;
23840    }
23841
23842    private int bitField0_;
23843    // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
23844    public static final int REQINFO_FIELD_NUMBER = 1;
23845    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
23846    /**
23847     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
23848     */
23849    public boolean hasReqInfo() {
23850      return ((bitField0_ & 0x00000001) == 0x00000001);
23851    }
23852    /**
23853     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
23854     */
23855    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
23856      return reqInfo_;
23857    }
23858    /**
23859     * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
23860     */
23861    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
23862      return reqInfo_;
23863    }
23864
23865    // required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;
23866    public static final int STATETOACCEPT_FIELD_NUMBER = 2;
23867    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto stateToAccept_;
23868    /**
23869     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
23870     *
23871     * <pre>
23872     ** Details on the segment to recover 
23873     * </pre>
23874     */
23875    public boolean hasStateToAccept() {
23876      return ((bitField0_ & 0x00000002) == 0x00000002);
23877    }
23878    /**
23879     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
23880     *
23881     * <pre>
23882     ** Details on the segment to recover 
23883     * </pre>
23884     */
23885    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept() {
23886      return stateToAccept_;
23887    }
23888    /**
23889     * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
23890     *
23891     * <pre>
23892     ** Details on the segment to recover 
23893     * </pre>
23894     */
23895    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder() {
23896      return stateToAccept_;
23897    }
23898
23899    // required string fromURL = 3;
23900    public static final int FROMURL_FIELD_NUMBER = 3;
23901    private java.lang.Object fromURL_;
23902    /**
23903     * <code>required string fromURL = 3;</code>
23904     *
23905     * <pre>
23906     ** The URL from which the log may be copied 
23907     * </pre>
23908     */
23909    public boolean hasFromURL() {
23910      return ((bitField0_ & 0x00000004) == 0x00000004);
23911    }
23912    /**
23913     * <code>required string fromURL = 3;</code>
23914     *
23915     * <pre>
23916     ** The URL from which the log may be copied 
23917     * </pre>
23918     */
23919    public java.lang.String getFromURL() {
23920      java.lang.Object ref = fromURL_;
23921      if (ref instanceof java.lang.String) {
23922        return (java.lang.String) ref;
23923      } else {
23924        com.google.protobuf.ByteString bs = 
23925            (com.google.protobuf.ByteString) ref;
23926        java.lang.String s = bs.toStringUtf8();
23927        if (bs.isValidUtf8()) {
23928          fromURL_ = s;
23929        }
23930        return s;
23931      }
23932    }
23933    /**
23934     * <code>required string fromURL = 3;</code>
23935     *
23936     * <pre>
23937     ** The URL from which the log may be copied 
23938     * </pre>
23939     */
23940    public com.google.protobuf.ByteString
23941        getFromURLBytes() {
23942      java.lang.Object ref = fromURL_;
23943      if (ref instanceof java.lang.String) {
23944        com.google.protobuf.ByteString b = 
23945            com.google.protobuf.ByteString.copyFromUtf8(
23946                (java.lang.String) ref);
23947        fromURL_ = b;
23948        return b;
23949      } else {
23950        return (com.google.protobuf.ByteString) ref;
23951      }
23952    }
23953
23954    private void initFields() {
23955      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
23956      stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
23957      fromURL_ = "";
23958    }
23959    private byte memoizedIsInitialized = -1;
23960    public final boolean isInitialized() {
23961      byte isInitialized = memoizedIsInitialized;
23962      if (isInitialized != -1) return isInitialized == 1;
23963
23964      if (!hasReqInfo()) {
23965        memoizedIsInitialized = 0;
23966        return false;
23967      }
23968      if (!hasStateToAccept()) {
23969        memoizedIsInitialized = 0;
23970        return false;
23971      }
23972      if (!hasFromURL()) {
23973        memoizedIsInitialized = 0;
23974        return false;
23975      }
23976      if (!getReqInfo().isInitialized()) {
23977        memoizedIsInitialized = 0;
23978        return false;
23979      }
23980      if (!getStateToAccept().isInitialized()) {
23981        memoizedIsInitialized = 0;
23982        return false;
23983      }
23984      memoizedIsInitialized = 1;
23985      return true;
23986    }
23987
23988    public void writeTo(com.google.protobuf.CodedOutputStream output)
23989                        throws java.io.IOException {
23990      getSerializedSize();
23991      if (((bitField0_ & 0x00000001) == 0x00000001)) {
23992        output.writeMessage(1, reqInfo_);
23993      }
23994      if (((bitField0_ & 0x00000002) == 0x00000002)) {
23995        output.writeMessage(2, stateToAccept_);
23996      }
23997      if (((bitField0_ & 0x00000004) == 0x00000004)) {
23998        output.writeBytes(3, getFromURLBytes());
23999      }
24000      getUnknownFields().writeTo(output);
24001    }
24002
24003    private int memoizedSerializedSize = -1;
24004    public int getSerializedSize() {
24005      int size = memoizedSerializedSize;
24006      if (size != -1) return size;
24007
24008      size = 0;
24009      if (((bitField0_ & 0x00000001) == 0x00000001)) {
24010        size += com.google.protobuf.CodedOutputStream
24011          .computeMessageSize(1, reqInfo_);
24012      }
24013      if (((bitField0_ & 0x00000002) == 0x00000002)) {
24014        size += com.google.protobuf.CodedOutputStream
24015          .computeMessageSize(2, stateToAccept_);
24016      }
24017      if (((bitField0_ & 0x00000004) == 0x00000004)) {
24018        size += com.google.protobuf.CodedOutputStream
24019          .computeBytesSize(3, getFromURLBytes());
24020      }
24021      size += getUnknownFields().getSerializedSize();
24022      memoizedSerializedSize = size;
24023      return size;
24024    }
24025
24026    private static final long serialVersionUID = 0L;
24027    @java.lang.Override
24028    protected java.lang.Object writeReplace()
24029        throws java.io.ObjectStreamException {
24030      return super.writeReplace();
24031    }
24032
24033    @java.lang.Override
24034    public boolean equals(final java.lang.Object obj) {
24035      if (obj == this) {
24036       return true;
24037      }
24038      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)) {
24039        return super.equals(obj);
24040      }
24041      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto) obj;
24042
24043      boolean result = true;
24044      result = result && (hasReqInfo() == other.hasReqInfo());
24045      if (hasReqInfo()) {
24046        result = result && getReqInfo()
24047            .equals(other.getReqInfo());
24048      }
24049      result = result && (hasStateToAccept() == other.hasStateToAccept());
24050      if (hasStateToAccept()) {
24051        result = result && getStateToAccept()
24052            .equals(other.getStateToAccept());
24053      }
24054      result = result && (hasFromURL() == other.hasFromURL());
24055      if (hasFromURL()) {
24056        result = result && getFromURL()
24057            .equals(other.getFromURL());
24058      }
24059      result = result &&
24060          getUnknownFields().equals(other.getUnknownFields());
24061      return result;
24062    }
24063
24064    private int memoizedHashCode = 0;
24065    @java.lang.Override
24066    public int hashCode() {
24067      if (memoizedHashCode != 0) {
24068        return memoizedHashCode;
24069      }
24070      int hash = 41;
24071      hash = (19 * hash) + getDescriptorForType().hashCode();
24072      if (hasReqInfo()) {
24073        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
24074        hash = (53 * hash) + getReqInfo().hashCode();
24075      }
24076      if (hasStateToAccept()) {
24077        hash = (37 * hash) + STATETOACCEPT_FIELD_NUMBER;
24078        hash = (53 * hash) + getStateToAccept().hashCode();
24079      }
24080      if (hasFromURL()) {
24081        hash = (37 * hash) + FROMURL_FIELD_NUMBER;
24082        hash = (53 * hash) + getFromURL().hashCode();
24083      }
24084      hash = (29 * hash) + getUnknownFields().hashCode();
24085      memoizedHashCode = hash;
24086      return hash;
24087    }
24088
24089    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
24090        com.google.protobuf.ByteString data)
24091        throws com.google.protobuf.InvalidProtocolBufferException {
24092      return PARSER.parseFrom(data);
24093    }
24094    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
24095        com.google.protobuf.ByteString data,
24096        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24097        throws com.google.protobuf.InvalidProtocolBufferException {
24098      return PARSER.parseFrom(data, extensionRegistry);
24099    }
24100    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(byte[] data)
24101        throws com.google.protobuf.InvalidProtocolBufferException {
24102      return PARSER.parseFrom(data);
24103    }
24104    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
24105        byte[] data,
24106        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24107        throws com.google.protobuf.InvalidProtocolBufferException {
24108      return PARSER.parseFrom(data, extensionRegistry);
24109    }
24110    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(java.io.InputStream input)
24111        throws java.io.IOException {
24112      return PARSER.parseFrom(input);
24113    }
24114    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
24115        java.io.InputStream input,
24116        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24117        throws java.io.IOException {
24118      return PARSER.parseFrom(input, extensionRegistry);
24119    }
24120    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseDelimitedFrom(java.io.InputStream input)
24121        throws java.io.IOException {
24122      return PARSER.parseDelimitedFrom(input);
24123    }
24124    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseDelimitedFrom(
24125        java.io.InputStream input,
24126        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24127        throws java.io.IOException {
24128      return PARSER.parseDelimitedFrom(input, extensionRegistry);
24129    }
24130    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
24131        com.google.protobuf.CodedInputStream input)
24132        throws java.io.IOException {
24133      return PARSER.parseFrom(input);
24134    }
24135    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
24136        com.google.protobuf.CodedInputStream input,
24137        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24138        throws java.io.IOException {
24139      return PARSER.parseFrom(input, extensionRegistry);
24140    }
24141
24142    public static Builder newBuilder() { return Builder.create(); }
24143    public Builder newBuilderForType() { return newBuilder(); }
24144    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto prototype) {
24145      return newBuilder().mergeFrom(prototype);
24146    }
24147    public Builder toBuilder() { return newBuilder(this); }
24148
24149    @java.lang.Override
24150    protected Builder newBuilderForType(
24151        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
24152      Builder builder = new Builder(parent);
24153      return builder;
24154    }
24155    /**
24156     * Protobuf type {@code hadoop.hdfs.qjournal.AcceptRecoveryRequestProto}
24157     *
24158     * <pre>
24159     **
24160     * acceptRecovery()
24161     * </pre>
24162     */
24163    public static final class Builder extends
24164        com.google.protobuf.GeneratedMessage.Builder<Builder>
24165       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProtoOrBuilder {
24166      public static final com.google.protobuf.Descriptors.Descriptor
24167          getDescriptor() {
24168        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_AcceptRecoveryRequestProto_descriptor;
24169      }
24170
24171      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
24172          internalGetFieldAccessorTable() {
24173        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_AcceptRecoveryRequestProto_fieldAccessorTable
24174            .ensureFieldAccessorsInitialized(
24175                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.Builder.class);
24176      }
24177
24178      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.newBuilder()
24179      private Builder() {
24180        maybeForceBuilderInitialization();
24181      }
24182
24183      private Builder(
24184          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
24185        super(parent);
24186        maybeForceBuilderInitialization();
24187      }
24188      private void maybeForceBuilderInitialization() {
24189        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
24190          getReqInfoFieldBuilder();
24191          getStateToAcceptFieldBuilder();
24192        }
24193      }
24194      private static Builder create() {
24195        return new Builder();
24196      }
24197
24198      public Builder clear() {
24199        super.clear();
24200        if (reqInfoBuilder_ == null) {
24201          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
24202        } else {
24203          reqInfoBuilder_.clear();
24204        }
24205        bitField0_ = (bitField0_ & ~0x00000001);
24206        if (stateToAcceptBuilder_ == null) {
24207          stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
24208        } else {
24209          stateToAcceptBuilder_.clear();
24210        }
24211        bitField0_ = (bitField0_ & ~0x00000002);
24212        fromURL_ = "";
24213        bitField0_ = (bitField0_ & ~0x00000004);
24214        return this;
24215      }
24216
24217      public Builder clone() {
24218        return create().mergeFrom(buildPartial());
24219      }
24220
24221      public com.google.protobuf.Descriptors.Descriptor
24222          getDescriptorForType() {
24223        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_AcceptRecoveryRequestProto_descriptor;
24224      }
24225
24226      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto getDefaultInstanceForType() {
24227        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance();
24228      }
24229
24230      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto build() {
24231        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto result = buildPartial();
24232        if (!result.isInitialized()) {
24233          throw newUninitializedMessageException(result);
24234        }
24235        return result;
24236      }
24237
24238      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto buildPartial() {
24239        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto(this);
24240        int from_bitField0_ = bitField0_;
24241        int to_bitField0_ = 0;
24242        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
24243          to_bitField0_ |= 0x00000001;
24244        }
24245        if (reqInfoBuilder_ == null) {
24246          result.reqInfo_ = reqInfo_;
24247        } else {
24248          result.reqInfo_ = reqInfoBuilder_.build();
24249        }
24250        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
24251          to_bitField0_ |= 0x00000002;
24252        }
24253        if (stateToAcceptBuilder_ == null) {
24254          result.stateToAccept_ = stateToAccept_;
24255        } else {
24256          result.stateToAccept_ = stateToAcceptBuilder_.build();
24257        }
24258        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
24259          to_bitField0_ |= 0x00000004;
24260        }
24261        result.fromURL_ = fromURL_;
24262        result.bitField0_ = to_bitField0_;
24263        onBuilt();
24264        return result;
24265      }
24266
24267      public Builder mergeFrom(com.google.protobuf.Message other) {
24268        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto) {
24269          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)other);
24270        } else {
24271          super.mergeFrom(other);
24272          return this;
24273        }
24274      }
24275
24276      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto other) {
24277        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance()) return this;
24278        if (other.hasReqInfo()) {
24279          mergeReqInfo(other.getReqInfo());
24280        }
24281        if (other.hasStateToAccept()) {
24282          mergeStateToAccept(other.getStateToAccept());
24283        }
24284        if (other.hasFromURL()) {
24285          bitField0_ |= 0x00000004;
24286          fromURL_ = other.fromURL_;
24287          onChanged();
24288        }
24289        this.mergeUnknownFields(other.getUnknownFields());
24290        return this;
24291      }
24292
24293      public final boolean isInitialized() {
24294        if (!hasReqInfo()) {
24295          
24296          return false;
24297        }
24298        if (!hasStateToAccept()) {
24299          
24300          return false;
24301        }
24302        if (!hasFromURL()) {
24303          
24304          return false;
24305        }
24306        if (!getReqInfo().isInitialized()) {
24307          
24308          return false;
24309        }
24310        if (!getStateToAccept().isInitialized()) {
24311          
24312          return false;
24313        }
24314        return true;
24315      }
24316
24317      public Builder mergeFrom(
24318          com.google.protobuf.CodedInputStream input,
24319          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24320          throws java.io.IOException {
24321        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parsedMessage = null;
24322        try {
24323          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
24324        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
24325          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto) e.getUnfinishedMessage();
24326          throw e;
24327        } finally {
24328          if (parsedMessage != null) {
24329            mergeFrom(parsedMessage);
24330          }
24331        }
24332        return this;
24333      }
24334      private int bitField0_;
24335
24336      // required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;
24337      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
24338      private com.google.protobuf.SingleFieldBuilder<
24339          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
24340      /**
24341       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
24342       */
24343      public boolean hasReqInfo() {
24344        return ((bitField0_ & 0x00000001) == 0x00000001);
24345      }
24346      /**
24347       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
24348       */
24349      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
24350        if (reqInfoBuilder_ == null) {
24351          return reqInfo_;
24352        } else {
24353          return reqInfoBuilder_.getMessage();
24354        }
24355      }
24356      /**
24357       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
24358       */
24359      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
24360        if (reqInfoBuilder_ == null) {
24361          if (value == null) {
24362            throw new NullPointerException();
24363          }
24364          reqInfo_ = value;
24365          onChanged();
24366        } else {
24367          reqInfoBuilder_.setMessage(value);
24368        }
24369        bitField0_ |= 0x00000001;
24370        return this;
24371      }
24372      /**
24373       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
24374       */
24375      public Builder setReqInfo(
24376          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
24377        if (reqInfoBuilder_ == null) {
24378          reqInfo_ = builderForValue.build();
24379          onChanged();
24380        } else {
24381          reqInfoBuilder_.setMessage(builderForValue.build());
24382        }
24383        bitField0_ |= 0x00000001;
24384        return this;
24385      }
24386      /**
24387       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
24388       */
24389      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
24390        if (reqInfoBuilder_ == null) {
24391          if (((bitField0_ & 0x00000001) == 0x00000001) &&
24392              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
24393            reqInfo_ =
24394              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
24395          } else {
24396            reqInfo_ = value;
24397          }
24398          onChanged();
24399        } else {
24400          reqInfoBuilder_.mergeFrom(value);
24401        }
24402        bitField0_ |= 0x00000001;
24403        return this;
24404      }
24405      /**
24406       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
24407       */
24408      public Builder clearReqInfo() {
24409        if (reqInfoBuilder_ == null) {
24410          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
24411          onChanged();
24412        } else {
24413          reqInfoBuilder_.clear();
24414        }
24415        bitField0_ = (bitField0_ & ~0x00000001);
24416        return this;
24417      }
24418      /**
24419       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
24420       */
24421      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
24422        bitField0_ |= 0x00000001;
24423        onChanged();
24424        return getReqInfoFieldBuilder().getBuilder();
24425      }
24426      /**
24427       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
24428       */
24429      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
24430        if (reqInfoBuilder_ != null) {
24431          return reqInfoBuilder_.getMessageOrBuilder();
24432        } else {
24433          return reqInfo_;
24434        }
24435      }
24436      /**
24437       * <code>required .hadoop.hdfs.qjournal.RequestInfoProto reqInfo = 1;</code>
24438       */
24439      private com.google.protobuf.SingleFieldBuilder<
24440          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
24441          getReqInfoFieldBuilder() {
24442        if (reqInfoBuilder_ == null) {
24443          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
24444              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
24445                  reqInfo_,
24446                  getParentForChildren(),
24447                  isClean());
24448          reqInfo_ = null;
24449        }
24450        return reqInfoBuilder_;
24451      }
24452
24453      // required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;
24454      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
24455      private com.google.protobuf.SingleFieldBuilder<
24456          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> stateToAcceptBuilder_;
24457      /**
24458       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
24459       *
24460       * <pre>
24461       ** Details on the segment to recover 
24462       * </pre>
24463       */
24464      public boolean hasStateToAccept() {
24465        return ((bitField0_ & 0x00000002) == 0x00000002);
24466      }
24467      /**
24468       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
24469       *
24470       * <pre>
24471       ** Details on the segment to recover 
24472       * </pre>
24473       */
24474      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept() {
24475        if (stateToAcceptBuilder_ == null) {
24476          return stateToAccept_;
24477        } else {
24478          return stateToAcceptBuilder_.getMessage();
24479        }
24480      }
24481      /**
24482       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
24483       *
24484       * <pre>
24485       ** Details on the segment to recover 
24486       * </pre>
24487       */
24488      public Builder setStateToAccept(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
24489        if (stateToAcceptBuilder_ == null) {
24490          if (value == null) {
24491            throw new NullPointerException();
24492          }
24493          stateToAccept_ = value;
24494          onChanged();
24495        } else {
24496          stateToAcceptBuilder_.setMessage(value);
24497        }
24498        bitField0_ |= 0x00000002;
24499        return this;
24500      }
24501      /**
24502       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
24503       *
24504       * <pre>
24505       ** Details on the segment to recover 
24506       * </pre>
24507       */
24508      public Builder setStateToAccept(
24509          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) {
24510        if (stateToAcceptBuilder_ == null) {
24511          stateToAccept_ = builderForValue.build();
24512          onChanged();
24513        } else {
24514          stateToAcceptBuilder_.setMessage(builderForValue.build());
24515        }
24516        bitField0_ |= 0x00000002;
24517        return this;
24518      }
24519      /**
24520       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
24521       *
24522       * <pre>
24523       ** Details on the segment to recover 
24524       * </pre>
24525       */
24526      public Builder mergeStateToAccept(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
24527        if (stateToAcceptBuilder_ == null) {
24528          if (((bitField0_ & 0x00000002) == 0x00000002) &&
24529              stateToAccept_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) {
24530            stateToAccept_ =
24531              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(stateToAccept_).mergeFrom(value).buildPartial();
24532          } else {
24533            stateToAccept_ = value;
24534          }
24535          onChanged();
24536        } else {
24537          stateToAcceptBuilder_.mergeFrom(value);
24538        }
24539        bitField0_ |= 0x00000002;
24540        return this;
24541      }
24542      /**
24543       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
24544       *
24545       * <pre>
24546       ** Details on the segment to recover 
24547       * </pre>
24548       */
24549      public Builder clearStateToAccept() {
24550        if (stateToAcceptBuilder_ == null) {
24551          stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
24552          onChanged();
24553        } else {
24554          stateToAcceptBuilder_.clear();
24555        }
24556        bitField0_ = (bitField0_ & ~0x00000002);
24557        return this;
24558      }
24559      /**
24560       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
24561       *
24562       * <pre>
24563       ** Details on the segment to recover 
24564       * </pre>
24565       */
24566      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getStateToAcceptBuilder() {
24567        bitField0_ |= 0x00000002;
24568        onChanged();
24569        return getStateToAcceptFieldBuilder().getBuilder();
24570      }
24571      /**
24572       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
24573       *
24574       * <pre>
24575       ** Details on the segment to recover 
24576       * </pre>
24577       */
24578      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder() {
24579        if (stateToAcceptBuilder_ != null) {
24580          return stateToAcceptBuilder_.getMessageOrBuilder();
24581        } else {
24582          return stateToAccept_;
24583        }
24584      }
24585      /**
24586       * <code>required .hadoop.hdfs.qjournal.SegmentStateProto stateToAccept = 2;</code>
24587       *
24588       * <pre>
24589       ** Details on the segment to recover 
24590       * </pre>
24591       */
24592      private com.google.protobuf.SingleFieldBuilder<
24593          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 
24594          getStateToAcceptFieldBuilder() {
24595        if (stateToAcceptBuilder_ == null) {
24596          stateToAcceptBuilder_ = new com.google.protobuf.SingleFieldBuilder<
24597              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>(
24598                  stateToAccept_,
24599                  getParentForChildren(),
24600                  isClean());
24601          stateToAccept_ = null;
24602        }
24603        return stateToAcceptBuilder_;
24604      }
24605
24606      // required string fromURL = 3;
24607      private java.lang.Object fromURL_ = "";
24608      /**
24609       * <code>required string fromURL = 3;</code>
24610       *
24611       * <pre>
24612       ** The URL from which the log may be copied 
24613       * </pre>
24614       */
24615      public boolean hasFromURL() {
24616        return ((bitField0_ & 0x00000004) == 0x00000004);
24617      }
24618      /**
24619       * <code>required string fromURL = 3;</code>
24620       *
24621       * <pre>
24622       ** The URL from which the log may be copied 
24623       * </pre>
24624       */
24625      public java.lang.String getFromURL() {
24626        java.lang.Object ref = fromURL_;
24627        if (!(ref instanceof java.lang.String)) {
24628          java.lang.String s = ((com.google.protobuf.ByteString) ref)
24629              .toStringUtf8();
24630          fromURL_ = s;
24631          return s;
24632        } else {
24633          return (java.lang.String) ref;
24634        }
24635      }
24636      /**
24637       * <code>required string fromURL = 3;</code>
24638       *
24639       * <pre>
24640       ** The URL from which the log may be copied 
24641       * </pre>
24642       */
24643      public com.google.protobuf.ByteString
24644          getFromURLBytes() {
24645        java.lang.Object ref = fromURL_;
24646        if (ref instanceof String) {
24647          com.google.protobuf.ByteString b = 
24648              com.google.protobuf.ByteString.copyFromUtf8(
24649                  (java.lang.String) ref);
24650          fromURL_ = b;
24651          return b;
24652        } else {
24653          return (com.google.protobuf.ByteString) ref;
24654        }
24655      }
24656      /**
24657       * <code>required string fromURL = 3;</code>
24658       *
24659       * <pre>
24660       ** The URL from which the log may be copied 
24661       * </pre>
24662       */
24663      public Builder setFromURL(
24664          java.lang.String value) {
24665        if (value == null) {
24666    throw new NullPointerException();
24667  }
24668  bitField0_ |= 0x00000004;
24669        fromURL_ = value;
24670        onChanged();
24671        return this;
24672      }
24673      /**
24674       * <code>required string fromURL = 3;</code>
24675       *
24676       * <pre>
24677       ** The URL from which the log may be copied 
24678       * </pre>
24679       */
24680      public Builder clearFromURL() {
24681        bitField0_ = (bitField0_ & ~0x00000004);
24682        fromURL_ = getDefaultInstance().getFromURL();
24683        onChanged();
24684        return this;
24685      }
24686      /**
24687       * <code>required string fromURL = 3;</code>
24688       *
24689       * <pre>
24690       ** The URL from which the log may be copied 
24691       * </pre>
24692       */
24693      public Builder setFromURLBytes(
24694          com.google.protobuf.ByteString value) {
24695        if (value == null) {
24696    throw new NullPointerException();
24697  }
24698  bitField0_ |= 0x00000004;
24699        fromURL_ = value;
24700        onChanged();
24701        return this;
24702      }
24703
24704      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.AcceptRecoveryRequestProto)
24705    }
24706
24707    static {
24708      defaultInstance = new AcceptRecoveryRequestProto(true);
24709      defaultInstance.initFields();
24710    }
24711
24712    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.AcceptRecoveryRequestProto)
24713  }
24714
24715  public interface AcceptRecoveryResponseProtoOrBuilder
24716      extends com.google.protobuf.MessageOrBuilder {
24717  }
24718  /**
24719   * Protobuf type {@code hadoop.hdfs.qjournal.AcceptRecoveryResponseProto}
24720   */
24721  public static final class AcceptRecoveryResponseProto extends
24722      com.google.protobuf.GeneratedMessage
24723      implements AcceptRecoveryResponseProtoOrBuilder {
24724    // Use AcceptRecoveryResponseProto.newBuilder() to construct.
24725    private AcceptRecoveryResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
24726      super(builder);
24727      this.unknownFields = builder.getUnknownFields();
24728    }
24729    private AcceptRecoveryResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
24730
24731    private static final AcceptRecoveryResponseProto defaultInstance;
24732    public static AcceptRecoveryResponseProto getDefaultInstance() {
24733      return defaultInstance;
24734    }
24735
24736    public AcceptRecoveryResponseProto getDefaultInstanceForType() {
24737      return defaultInstance;
24738    }
24739
24740    private final com.google.protobuf.UnknownFieldSet unknownFields;
24741    @java.lang.Override
24742    public final com.google.protobuf.UnknownFieldSet
24743        getUnknownFields() {
24744      return this.unknownFields;
24745    }
24746    private AcceptRecoveryResponseProto(
24747        com.google.protobuf.CodedInputStream input,
24748        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24749        throws com.google.protobuf.InvalidProtocolBufferException {
24750      initFields();
24751      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
24752          com.google.protobuf.UnknownFieldSet.newBuilder();
24753      try {
24754        boolean done = false;
24755        while (!done) {
24756          int tag = input.readTag();
24757          switch (tag) {
24758            case 0:
24759              done = true;
24760              break;
24761            default: {
24762              if (!parseUnknownField(input, unknownFields,
24763                                     extensionRegistry, tag)) {
24764                done = true;
24765              }
24766              break;
24767            }
24768          }
24769        }
24770      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
24771        throw e.setUnfinishedMessage(this);
24772      } catch (java.io.IOException e) {
24773        throw new com.google.protobuf.InvalidProtocolBufferException(
24774            e.getMessage()).setUnfinishedMessage(this);
24775      } finally {
24776        this.unknownFields = unknownFields.build();
24777        makeExtensionsImmutable();
24778      }
24779    }
24780    public static final com.google.protobuf.Descriptors.Descriptor
24781        getDescriptor() {
24782      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_AcceptRecoveryResponseProto_descriptor;
24783    }
24784
24785    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
24786        internalGetFieldAccessorTable() {
24787      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_AcceptRecoveryResponseProto_fieldAccessorTable
24788          .ensureFieldAccessorsInitialized(
24789              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.Builder.class);
24790    }
24791
24792    public static com.google.protobuf.Parser<AcceptRecoveryResponseProto> PARSER =
24793        new com.google.protobuf.AbstractParser<AcceptRecoveryResponseProto>() {
24794      public AcceptRecoveryResponseProto parsePartialFrom(
24795          com.google.protobuf.CodedInputStream input,
24796          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24797          throws com.google.protobuf.InvalidProtocolBufferException {
24798        return new AcceptRecoveryResponseProto(input, extensionRegistry);
24799      }
24800    };
24801
24802    @java.lang.Override
24803    public com.google.protobuf.Parser<AcceptRecoveryResponseProto> getParserForType() {
24804      return PARSER;
24805    }
24806
24807    private void initFields() {
24808    }
24809    private byte memoizedIsInitialized = -1;
24810    public final boolean isInitialized() {
24811      byte isInitialized = memoizedIsInitialized;
24812      if (isInitialized != -1) return isInitialized == 1;
24813
24814      memoizedIsInitialized = 1;
24815      return true;
24816    }
24817
24818    public void writeTo(com.google.protobuf.CodedOutputStream output)
24819                        throws java.io.IOException {
24820      getSerializedSize();
24821      getUnknownFields().writeTo(output);
24822    }
24823
24824    private int memoizedSerializedSize = -1;
24825    public int getSerializedSize() {
24826      int size = memoizedSerializedSize;
24827      if (size != -1) return size;
24828
24829      size = 0;
24830      size += getUnknownFields().getSerializedSize();
24831      memoizedSerializedSize = size;
24832      return size;
24833    }
24834
24835    private static final long serialVersionUID = 0L;
24836    @java.lang.Override
24837    protected java.lang.Object writeReplace()
24838        throws java.io.ObjectStreamException {
24839      return super.writeReplace();
24840    }
24841
24842    @java.lang.Override
24843    public boolean equals(final java.lang.Object obj) {
24844      if (obj == this) {
24845       return true;
24846      }
24847      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto)) {
24848        return super.equals(obj);
24849      }
24850      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) obj;
24851
24852      boolean result = true;
24853      result = result &&
24854          getUnknownFields().equals(other.getUnknownFields());
24855      return result;
24856    }
24857
24858    private int memoizedHashCode = 0;
24859    @java.lang.Override
24860    public int hashCode() {
24861      if (memoizedHashCode != 0) {
24862        return memoizedHashCode;
24863      }
24864      int hash = 41;
24865      hash = (19 * hash) + getDescriptorForType().hashCode();
24866      hash = (29 * hash) + getUnknownFields().hashCode();
24867      memoizedHashCode = hash;
24868      return hash;
24869    }
24870
24871    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
24872        com.google.protobuf.ByteString data)
24873        throws com.google.protobuf.InvalidProtocolBufferException {
24874      return PARSER.parseFrom(data);
24875    }
24876    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
24877        com.google.protobuf.ByteString data,
24878        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24879        throws com.google.protobuf.InvalidProtocolBufferException {
24880      return PARSER.parseFrom(data, extensionRegistry);
24881    }
24882    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(byte[] data)
24883        throws com.google.protobuf.InvalidProtocolBufferException {
24884      return PARSER.parseFrom(data);
24885    }
24886    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
24887        byte[] data,
24888        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24889        throws com.google.protobuf.InvalidProtocolBufferException {
24890      return PARSER.parseFrom(data, extensionRegistry);
24891    }
24892    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(java.io.InputStream input)
24893        throws java.io.IOException {
24894      return PARSER.parseFrom(input);
24895    }
24896    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
24897        java.io.InputStream input,
24898        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24899        throws java.io.IOException {
24900      return PARSER.parseFrom(input, extensionRegistry);
24901    }
24902    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseDelimitedFrom(java.io.InputStream input)
24903        throws java.io.IOException {
24904      return PARSER.parseDelimitedFrom(input);
24905    }
24906    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseDelimitedFrom(
24907        java.io.InputStream input,
24908        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24909        throws java.io.IOException {
24910      return PARSER.parseDelimitedFrom(input, extensionRegistry);
24911    }
24912    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
24913        com.google.protobuf.CodedInputStream input)
24914        throws java.io.IOException {
24915      return PARSER.parseFrom(input);
24916    }
24917    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
24918        com.google.protobuf.CodedInputStream input,
24919        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
24920        throws java.io.IOException {
24921      return PARSER.parseFrom(input, extensionRegistry);
24922    }
24923
24924    public static Builder newBuilder() { return Builder.create(); }
24925    public Builder newBuilderForType() { return newBuilder(); }
24926    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto prototype) {
24927      return newBuilder().mergeFrom(prototype);
24928    }
24929    public Builder toBuilder() { return newBuilder(this); }
24930
24931    @java.lang.Override
24932    protected Builder newBuilderForType(
24933        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
24934      Builder builder = new Builder(parent);
24935      return builder;
24936    }
24937    /**
24938     * Protobuf type {@code hadoop.hdfs.qjournal.AcceptRecoveryResponseProto}
24939     */
24940    public static final class Builder extends
24941        com.google.protobuf.GeneratedMessage.Builder<Builder>
24942       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProtoOrBuilder {
24943      public static final com.google.protobuf.Descriptors.Descriptor
24944          getDescriptor() {
24945        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_AcceptRecoveryResponseProto_descriptor;
24946      }
24947
24948      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
24949          internalGetFieldAccessorTable() {
24950        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_AcceptRecoveryResponseProto_fieldAccessorTable
24951            .ensureFieldAccessorsInitialized(
24952                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.Builder.class);
24953      }
24954
24955      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.newBuilder()
24956      private Builder() {
24957        maybeForceBuilderInitialization();
24958      }
24959
24960      private Builder(
24961          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
24962        super(parent);
24963        maybeForceBuilderInitialization();
24964      }
24965      private void maybeForceBuilderInitialization() {
24966        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
24967        }
24968      }
24969      private static Builder create() {
24970        return new Builder();
24971      }
24972
24973      public Builder clear() {
24974        super.clear();
24975        return this;
24976      }
24977
24978      public Builder clone() {
24979        return create().mergeFrom(buildPartial());
24980      }
24981
24982      public com.google.protobuf.Descriptors.Descriptor
24983          getDescriptorForType() {
24984        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_qjournal_AcceptRecoveryResponseProto_descriptor;
24985      }
24986
24987      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto getDefaultInstanceForType() {
24988        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance();
24989      }
24990
24991      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto build() {
24992        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto result = buildPartial();
24993        if (!result.isInitialized()) {
24994          throw newUninitializedMessageException(result);
24995        }
24996        return result;
24997      }
24998
24999      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto buildPartial() {
25000        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto(this);
25001        onBuilt();
25002        return result;
25003      }
25004
25005      public Builder mergeFrom(com.google.protobuf.Message other) {
25006        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) {
25007          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto)other);
25008        } else {
25009          super.mergeFrom(other);
25010          return this;
25011        }
25012      }
25013
25014      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto other) {
25015        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance()) return this;
25016        this.mergeUnknownFields(other.getUnknownFields());
25017        return this;
25018      }
25019
25020      public final boolean isInitialized() {
25021        return true;
25022      }
25023
25024      public Builder mergeFrom(
25025          com.google.protobuf.CodedInputStream input,
25026          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
25027          throws java.io.IOException {
25028        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parsedMessage = null;
25029        try {
25030          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
25031        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
25032          parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) e.getUnfinishedMessage();
25033          throw e;
25034        } finally {
25035          if (parsedMessage != null) {
25036            mergeFrom(parsedMessage);
25037          }
25038        }
25039        return this;
25040      }
25041
25042      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.qjournal.AcceptRecoveryResponseProto)
25043    }
25044
25045    static {
25046      defaultInstance = new AcceptRecoveryResponseProto(true);
25047      defaultInstance.initFields();
25048    }
25049
25050    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.AcceptRecoveryResponseProto)
25051  }
25052
25053  /**
25054   * Protobuf service {@code hadoop.hdfs.qjournal.QJournalProtocolService}
25055   *
25056   * <pre>
25057   **
25058   * Protocol used to journal edits to a JournalNode.
25059   * See the request and response for details of rpc call.
25060   * </pre>
25061   */
25062  public static abstract class QJournalProtocolService
25063      implements com.google.protobuf.Service {
25064    protected QJournalProtocolService() {}
25065
25066    public interface Interface {
25067      /**
25068       * <code>rpc isFormatted(.hadoop.hdfs.qjournal.IsFormattedRequestProto) returns (.hadoop.hdfs.qjournal.IsFormattedResponseProto);</code>
25069       */
25070      public abstract void isFormatted(
25071          com.google.protobuf.RpcController controller,
25072          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
25073          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done);
25074
25075      /**
25076       * <code>rpc discardSegments(.hadoop.hdfs.qjournal.DiscardSegmentsRequestProto) returns (.hadoop.hdfs.qjournal.DiscardSegmentsResponseProto);</code>
25077       */
25078      public abstract void discardSegments(
25079          com.google.protobuf.RpcController controller,
25080          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto request,
25081          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto> done);
25082
25083      /**
25084       * <code>rpc getJournalCTime(.hadoop.hdfs.qjournal.GetJournalCTimeRequestProto) returns (.hadoop.hdfs.qjournal.GetJournalCTimeResponseProto);</code>
25085       */
25086      public abstract void getJournalCTime(
25087          com.google.protobuf.RpcController controller,
25088          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto request,
25089          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto> done);
25090
25091      /**
25092       * <code>rpc doPreUpgrade(.hadoop.hdfs.qjournal.DoPreUpgradeRequestProto) returns (.hadoop.hdfs.qjournal.DoPreUpgradeResponseProto);</code>
25093       */
25094      public abstract void doPreUpgrade(
25095          com.google.protobuf.RpcController controller,
25096          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto request,
25097          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto> done);
25098
25099      /**
25100       * <code>rpc doUpgrade(.hadoop.hdfs.qjournal.DoUpgradeRequestProto) returns (.hadoop.hdfs.qjournal.DoUpgradeResponseProto);</code>
25101       */
25102      public abstract void doUpgrade(
25103          com.google.protobuf.RpcController controller,
25104          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto request,
25105          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto> done);
25106
25107      /**
25108       * <code>rpc doFinalize(.hadoop.hdfs.qjournal.DoFinalizeRequestProto) returns (.hadoop.hdfs.qjournal.DoFinalizeResponseProto);</code>
25109       */
25110      public abstract void doFinalize(
25111          com.google.protobuf.RpcController controller,
25112          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto request,
25113          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto> done);
25114
25115      /**
25116       * <code>rpc canRollBack(.hadoop.hdfs.qjournal.CanRollBackRequestProto) returns (.hadoop.hdfs.qjournal.CanRollBackResponseProto);</code>
25117       */
25118      public abstract void canRollBack(
25119          com.google.protobuf.RpcController controller,
25120          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto request,
25121          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto> done);
25122
25123      /**
25124       * <code>rpc doRollback(.hadoop.hdfs.qjournal.DoRollbackRequestProto) returns (.hadoop.hdfs.qjournal.DoRollbackResponseProto);</code>
25125       */
25126      public abstract void doRollback(
25127          com.google.protobuf.RpcController controller,
25128          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto request,
25129          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto> done);
25130
25131      /**
25132       * <code>rpc getJournalState(.hadoop.hdfs.qjournal.GetJournalStateRequestProto) returns (.hadoop.hdfs.qjournal.GetJournalStateResponseProto);</code>
25133       */
25134      public abstract void getJournalState(
25135          com.google.protobuf.RpcController controller,
25136          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
25137          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done);
25138
25139      /**
25140       * <code>rpc newEpoch(.hadoop.hdfs.qjournal.NewEpochRequestProto) returns (.hadoop.hdfs.qjournal.NewEpochResponseProto);</code>
25141       */
25142      public abstract void newEpoch(
25143          com.google.protobuf.RpcController controller,
25144          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
25145          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done);
25146
25147      /**
25148       * <code>rpc format(.hadoop.hdfs.qjournal.FormatRequestProto) returns (.hadoop.hdfs.qjournal.FormatResponseProto);</code>
25149       */
25150      public abstract void format(
25151          com.google.protobuf.RpcController controller,
25152          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
25153          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done);
25154
25155      /**
25156       * <code>rpc journal(.hadoop.hdfs.qjournal.JournalRequestProto) returns (.hadoop.hdfs.qjournal.JournalResponseProto);</code>
25157       */
25158      public abstract void journal(
25159          com.google.protobuf.RpcController controller,
25160          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
25161          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done);
25162
25163      /**
25164       * <code>rpc heartbeat(.hadoop.hdfs.qjournal.HeartbeatRequestProto) returns (.hadoop.hdfs.qjournal.HeartbeatResponseProto);</code>
25165       */
25166      public abstract void heartbeat(
25167          com.google.protobuf.RpcController controller,
25168          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
25169          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done);
25170
25171      /**
25172       * <code>rpc startLogSegment(.hadoop.hdfs.qjournal.StartLogSegmentRequestProto) returns (.hadoop.hdfs.qjournal.StartLogSegmentResponseProto);</code>
25173       */
25174      public abstract void startLogSegment(
25175          com.google.protobuf.RpcController controller,
25176          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
25177          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done);
25178
25179      /**
25180       * <code>rpc finalizeLogSegment(.hadoop.hdfs.qjournal.FinalizeLogSegmentRequestProto) returns (.hadoop.hdfs.qjournal.FinalizeLogSegmentResponseProto);</code>
25181       */
25182      public abstract void finalizeLogSegment(
25183          com.google.protobuf.RpcController controller,
25184          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
25185          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done);
25186
25187      /**
25188       * <code>rpc purgeLogs(.hadoop.hdfs.qjournal.PurgeLogsRequestProto) returns (.hadoop.hdfs.qjournal.PurgeLogsResponseProto);</code>
25189       */
25190      public abstract void purgeLogs(
25191          com.google.protobuf.RpcController controller,
25192          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
25193          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done);
25194
25195      /**
25196       * <code>rpc getEditLogManifest(.hadoop.hdfs.qjournal.GetEditLogManifestRequestProto) returns (.hadoop.hdfs.qjournal.GetEditLogManifestResponseProto);</code>
25197       */
25198      public abstract void getEditLogManifest(
25199          com.google.protobuf.RpcController controller,
25200          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
25201          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done);
25202
25203      /**
25204       * <code>rpc prepareRecovery(.hadoop.hdfs.qjournal.PrepareRecoveryRequestProto) returns (.hadoop.hdfs.qjournal.PrepareRecoveryResponseProto);</code>
25205       */
25206      public abstract void prepareRecovery(
25207          com.google.protobuf.RpcController controller,
25208          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
25209          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done);
25210
25211      /**
25212       * <code>rpc acceptRecovery(.hadoop.hdfs.qjournal.AcceptRecoveryRequestProto) returns (.hadoop.hdfs.qjournal.AcceptRecoveryResponseProto);</code>
25213       */
25214      public abstract void acceptRecovery(
25215          com.google.protobuf.RpcController controller,
25216          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
25217          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done);
25218
25219    }
25220
25221    public static com.google.protobuf.Service newReflectiveService(
25222        final Interface impl) {
25223      return new QJournalProtocolService() {
25224        @java.lang.Override
25225        public  void isFormatted(
25226            com.google.protobuf.RpcController controller,
25227            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
25228            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done) {
25229          impl.isFormatted(controller, request, done);
25230        }
25231
25232        @java.lang.Override
25233        public  void discardSegments(
25234            com.google.protobuf.RpcController controller,
25235            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto request,
25236            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto> done) {
25237          impl.discardSegments(controller, request, done);
25238        }
25239
25240        @java.lang.Override
25241        public  void getJournalCTime(
25242            com.google.protobuf.RpcController controller,
25243            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto request,
25244            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto> done) {
25245          impl.getJournalCTime(controller, request, done);
25246        }
25247
25248        @java.lang.Override
25249        public  void doPreUpgrade(
25250            com.google.protobuf.RpcController controller,
25251            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto request,
25252            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto> done) {
25253          impl.doPreUpgrade(controller, request, done);
25254        }
25255
25256        @java.lang.Override
25257        public  void doUpgrade(
25258            com.google.protobuf.RpcController controller,
25259            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto request,
25260            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto> done) {
25261          impl.doUpgrade(controller, request, done);
25262        }
25263
25264        @java.lang.Override
25265        public  void doFinalize(
25266            com.google.protobuf.RpcController controller,
25267            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto request,
25268            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto> done) {
25269          impl.doFinalize(controller, request, done);
25270        }
25271
25272        @java.lang.Override
25273        public  void canRollBack(
25274            com.google.protobuf.RpcController controller,
25275            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto request,
25276            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto> done) {
25277          impl.canRollBack(controller, request, done);
25278        }
25279
25280        @java.lang.Override
25281        public  void doRollback(
25282            com.google.protobuf.RpcController controller,
25283            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto request,
25284            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto> done) {
25285          impl.doRollback(controller, request, done);
25286        }
25287
25288        @java.lang.Override
25289        public  void getJournalState(
25290            com.google.protobuf.RpcController controller,
25291            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
25292            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done) {
25293          impl.getJournalState(controller, request, done);
25294        }
25295
25296        @java.lang.Override
25297        public  void newEpoch(
25298            com.google.protobuf.RpcController controller,
25299            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
25300            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done) {
25301          impl.newEpoch(controller, request, done);
25302        }
25303
25304        @java.lang.Override
25305        public  void format(
25306            com.google.protobuf.RpcController controller,
25307            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
25308            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done) {
25309          impl.format(controller, request, done);
25310        }
25311
25312        @java.lang.Override
25313        public  void journal(
25314            com.google.protobuf.RpcController controller,
25315            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
25316            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done) {
25317          impl.journal(controller, request, done);
25318        }
25319
25320        @java.lang.Override
25321        public  void heartbeat(
25322            com.google.protobuf.RpcController controller,
25323            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
25324            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done) {
25325          impl.heartbeat(controller, request, done);
25326        }
25327
25328        @java.lang.Override
25329        public  void startLogSegment(
25330            com.google.protobuf.RpcController controller,
25331            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
25332            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done) {
25333          impl.startLogSegment(controller, request, done);
25334        }
25335
25336        @java.lang.Override
25337        public  void finalizeLogSegment(
25338            com.google.protobuf.RpcController controller,
25339            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
25340            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done) {
25341          impl.finalizeLogSegment(controller, request, done);
25342        }
25343
25344        @java.lang.Override
25345        public  void purgeLogs(
25346            com.google.protobuf.RpcController controller,
25347            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
25348            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done) {
25349          impl.purgeLogs(controller, request, done);
25350        }
25351
25352        @java.lang.Override
25353        public  void getEditLogManifest(
25354            com.google.protobuf.RpcController controller,
25355            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
25356            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done) {
25357          impl.getEditLogManifest(controller, request, done);
25358        }
25359
25360        @java.lang.Override
25361        public  void prepareRecovery(
25362            com.google.protobuf.RpcController controller,
25363            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
25364            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done) {
25365          impl.prepareRecovery(controller, request, done);
25366        }
25367
25368        @java.lang.Override
25369        public  void acceptRecovery(
25370            com.google.protobuf.RpcController controller,
25371            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
25372            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done) {
25373          impl.acceptRecovery(controller, request, done);
25374        }
25375
25376      };
25377    }
25378
25379    public static com.google.protobuf.BlockingService
25380        newReflectiveBlockingService(final BlockingInterface impl) {
25381      return new com.google.protobuf.BlockingService() {
25382        public final com.google.protobuf.Descriptors.ServiceDescriptor
25383            getDescriptorForType() {
25384          return getDescriptor();
25385        }
25386
25387        public final com.google.protobuf.Message callBlockingMethod(
25388            com.google.protobuf.Descriptors.MethodDescriptor method,
25389            com.google.protobuf.RpcController controller,
25390            com.google.protobuf.Message request)
25391            throws com.google.protobuf.ServiceException {
25392          if (method.getService() != getDescriptor()) {
25393            throw new java.lang.IllegalArgumentException(
25394              "Service.callBlockingMethod() given method descriptor for " +
25395              "wrong service type.");
25396          }
25397          switch(method.getIndex()) {
25398            case 0:
25399              return impl.isFormatted(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)request);
25400            case 1:
25401              return impl.discardSegments(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto)request);
25402            case 2:
25403              return impl.getJournalCTime(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto)request);
25404            case 3:
25405              return impl.doPreUpgrade(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto)request);
25406            case 4:
25407              return impl.doUpgrade(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto)request);
25408            case 5:
25409              return impl.doFinalize(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto)request);
25410            case 6:
25411              return impl.canRollBack(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto)request);
25412            case 7:
25413              return impl.doRollback(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto)request);
25414            case 8:
25415              return impl.getJournalState(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)request);
25416            case 9:
25417              return impl.newEpoch(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)request);
25418            case 10:
25419              return impl.format(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)request);
25420            case 11:
25421              return impl.journal(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)request);
25422            case 12:
25423              return impl.heartbeat(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)request);
25424            case 13:
25425              return impl.startLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)request);
25426            case 14:
25427              return impl.finalizeLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)request);
25428            case 15:
25429              return impl.purgeLogs(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)request);
25430            case 16:
25431              return impl.getEditLogManifest(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)request);
25432            case 17:
25433              return impl.prepareRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)request);
25434            case 18:
25435              return impl.acceptRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)request);
25436            default:
25437              throw new java.lang.AssertionError("Can't get here.");
25438          }
25439        }
25440
25441        public final com.google.protobuf.Message
25442            getRequestPrototype(
25443            com.google.protobuf.Descriptors.MethodDescriptor method) {
25444          if (method.getService() != getDescriptor()) {
25445            throw new java.lang.IllegalArgumentException(
25446              "Service.getRequestPrototype() given method " +
25447              "descriptor for wrong service type.");
25448          }
25449          switch(method.getIndex()) {
25450            case 0:
25451              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance();
25452            case 1:
25453              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto.getDefaultInstance();
25454            case 2:
25455              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto.getDefaultInstance();
25456            case 3:
25457              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto.getDefaultInstance();
25458            case 4:
25459              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto.getDefaultInstance();
25460            case 5:
25461              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto.getDefaultInstance();
25462            case 6:
25463              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto.getDefaultInstance();
25464            case 7:
25465              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto.getDefaultInstance();
25466            case 8:
25467              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance();
25468            case 9:
25469              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance();
25470            case 10:
25471              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance();
25472            case 11:
25473              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance();
25474            case 12:
25475              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance();
25476            case 13:
25477              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance();
25478            case 14:
25479              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance();
25480            case 15:
25481              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance();
25482            case 16:
25483              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance();
25484            case 17:
25485              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance();
25486            case 18:
25487              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance();
25488            default:
25489              throw new java.lang.AssertionError("Can't get here.");
25490          }
25491        }
25492
25493        public final com.google.protobuf.Message
25494            getResponsePrototype(
25495            com.google.protobuf.Descriptors.MethodDescriptor method) {
25496          if (method.getService() != getDescriptor()) {
25497            throw new java.lang.IllegalArgumentException(
25498              "Service.getResponsePrototype() given method " +
25499              "descriptor for wrong service type.");
25500          }
25501          switch(method.getIndex()) {
25502            case 0:
25503              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance();
25504            case 1:
25505              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.getDefaultInstance();
25506            case 2:
25507              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.getDefaultInstance();
25508            case 3:
25509              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.getDefaultInstance();
25510            case 4:
25511              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.getDefaultInstance();
25512            case 5:
25513              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.getDefaultInstance();
25514            case 6:
25515              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.getDefaultInstance();
25516            case 7:
25517              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.getDefaultInstance();
25518            case 8:
25519              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance();
25520            case 9:
25521              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance();
25522            case 10:
25523              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance();
25524            case 11:
25525              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance();
25526            case 12:
25527              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance();
25528            case 13:
25529              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance();
25530            case 14:
25531              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance();
25532            case 15:
25533              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance();
25534            case 16:
25535              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance();
25536            case 17:
25537              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance();
25538            case 18:
25539              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance();
25540            default:
25541              throw new java.lang.AssertionError("Can't get here.");
25542          }
25543        }
25544
25545      };
25546    }
25547
25548    /**
25549     * <code>rpc isFormatted(.hadoop.hdfs.qjournal.IsFormattedRequestProto) returns (.hadoop.hdfs.qjournal.IsFormattedResponseProto);</code>
25550     */
25551    public abstract void isFormatted(
25552        com.google.protobuf.RpcController controller,
25553        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
25554        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done);
25555
25556    /**
25557     * <code>rpc discardSegments(.hadoop.hdfs.qjournal.DiscardSegmentsRequestProto) returns (.hadoop.hdfs.qjournal.DiscardSegmentsResponseProto);</code>
25558     */
25559    public abstract void discardSegments(
25560        com.google.protobuf.RpcController controller,
25561        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto request,
25562        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto> done);
25563
25564    /**
25565     * <code>rpc getJournalCTime(.hadoop.hdfs.qjournal.GetJournalCTimeRequestProto) returns (.hadoop.hdfs.qjournal.GetJournalCTimeResponseProto);</code>
25566     */
25567    public abstract void getJournalCTime(
25568        com.google.protobuf.RpcController controller,
25569        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto request,
25570        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto> done);
25571
25572    /**
25573     * <code>rpc doPreUpgrade(.hadoop.hdfs.qjournal.DoPreUpgradeRequestProto) returns (.hadoop.hdfs.qjournal.DoPreUpgradeResponseProto);</code>
25574     */
25575    public abstract void doPreUpgrade(
25576        com.google.protobuf.RpcController controller,
25577        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto request,
25578        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto> done);
25579
25580    /**
25581     * <code>rpc doUpgrade(.hadoop.hdfs.qjournal.DoUpgradeRequestProto) returns (.hadoop.hdfs.qjournal.DoUpgradeResponseProto);</code>
25582     */
25583    public abstract void doUpgrade(
25584        com.google.protobuf.RpcController controller,
25585        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto request,
25586        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto> done);
25587
25588    /**
25589     * <code>rpc doFinalize(.hadoop.hdfs.qjournal.DoFinalizeRequestProto) returns (.hadoop.hdfs.qjournal.DoFinalizeResponseProto);</code>
25590     */
25591    public abstract void doFinalize(
25592        com.google.protobuf.RpcController controller,
25593        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto request,
25594        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto> done);
25595
25596    /**
25597     * <code>rpc canRollBack(.hadoop.hdfs.qjournal.CanRollBackRequestProto) returns (.hadoop.hdfs.qjournal.CanRollBackResponseProto);</code>
25598     */
25599    public abstract void canRollBack(
25600        com.google.protobuf.RpcController controller,
25601        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto request,
25602        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto> done);
25603
25604    /**
25605     * <code>rpc doRollback(.hadoop.hdfs.qjournal.DoRollbackRequestProto) returns (.hadoop.hdfs.qjournal.DoRollbackResponseProto);</code>
25606     */
25607    public abstract void doRollback(
25608        com.google.protobuf.RpcController controller,
25609        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto request,
25610        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto> done);
25611
25612    /**
25613     * <code>rpc getJournalState(.hadoop.hdfs.qjournal.GetJournalStateRequestProto) returns (.hadoop.hdfs.qjournal.GetJournalStateResponseProto);</code>
25614     */
25615    public abstract void getJournalState(
25616        com.google.protobuf.RpcController controller,
25617        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
25618        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done);
25619
25620    /**
25621     * <code>rpc newEpoch(.hadoop.hdfs.qjournal.NewEpochRequestProto) returns (.hadoop.hdfs.qjournal.NewEpochResponseProto);</code>
25622     */
25623    public abstract void newEpoch(
25624        com.google.protobuf.RpcController controller,
25625        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
25626        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done);
25627
25628    /**
25629     * <code>rpc format(.hadoop.hdfs.qjournal.FormatRequestProto) returns (.hadoop.hdfs.qjournal.FormatResponseProto);</code>
25630     */
25631    public abstract void format(
25632        com.google.protobuf.RpcController controller,
25633        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
25634        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done);
25635
25636    /**
25637     * <code>rpc journal(.hadoop.hdfs.qjournal.JournalRequestProto) returns (.hadoop.hdfs.qjournal.JournalResponseProto);</code>
25638     */
25639    public abstract void journal(
25640        com.google.protobuf.RpcController controller,
25641        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
25642        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done);
25643
25644    /**
25645     * <code>rpc heartbeat(.hadoop.hdfs.qjournal.HeartbeatRequestProto) returns (.hadoop.hdfs.qjournal.HeartbeatResponseProto);</code>
25646     */
25647    public abstract void heartbeat(
25648        com.google.protobuf.RpcController controller,
25649        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
25650        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done);
25651
25652    /**
25653     * <code>rpc startLogSegment(.hadoop.hdfs.qjournal.StartLogSegmentRequestProto) returns (.hadoop.hdfs.qjournal.StartLogSegmentResponseProto);</code>
25654     */
25655    public abstract void startLogSegment(
25656        com.google.protobuf.RpcController controller,
25657        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
25658        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done);
25659
25660    /**
25661     * <code>rpc finalizeLogSegment(.hadoop.hdfs.qjournal.FinalizeLogSegmentRequestProto) returns (.hadoop.hdfs.qjournal.FinalizeLogSegmentResponseProto);</code>
25662     */
25663    public abstract void finalizeLogSegment(
25664        com.google.protobuf.RpcController controller,
25665        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
25666        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done);
25667
25668    /**
25669     * <code>rpc purgeLogs(.hadoop.hdfs.qjournal.PurgeLogsRequestProto) returns (.hadoop.hdfs.qjournal.PurgeLogsResponseProto);</code>
25670     */
25671    public abstract void purgeLogs(
25672        com.google.protobuf.RpcController controller,
25673        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
25674        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done);
25675
25676    /**
25677     * <code>rpc getEditLogManifest(.hadoop.hdfs.qjournal.GetEditLogManifestRequestProto) returns (.hadoop.hdfs.qjournal.GetEditLogManifestResponseProto);</code>
25678     */
25679    public abstract void getEditLogManifest(
25680        com.google.protobuf.RpcController controller,
25681        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
25682        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done);
25683
25684    /**
25685     * <code>rpc prepareRecovery(.hadoop.hdfs.qjournal.PrepareRecoveryRequestProto) returns (.hadoop.hdfs.qjournal.PrepareRecoveryResponseProto);</code>
25686     */
25687    public abstract void prepareRecovery(
25688        com.google.protobuf.RpcController controller,
25689        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
25690        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done);
25691
25692    /**
25693     * <code>rpc acceptRecovery(.hadoop.hdfs.qjournal.AcceptRecoveryRequestProto) returns (.hadoop.hdfs.qjournal.AcceptRecoveryResponseProto);</code>
25694     */
25695    public abstract void acceptRecovery(
25696        com.google.protobuf.RpcController controller,
25697        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
25698        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done);
25699
25700    public static final
25701        com.google.protobuf.Descriptors.ServiceDescriptor
25702        getDescriptor() {
25703      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.getDescriptor().getServices().get(0);
25704    }
25705    public final com.google.protobuf.Descriptors.ServiceDescriptor
25706        getDescriptorForType() {
25707      return getDescriptor();
25708    }
25709
25710    public final void callMethod(
25711        com.google.protobuf.Descriptors.MethodDescriptor method,
25712        com.google.protobuf.RpcController controller,
25713        com.google.protobuf.Message request,
25714        com.google.protobuf.RpcCallback<
25715          com.google.protobuf.Message> done) {
25716      if (method.getService() != getDescriptor()) {
25717        throw new java.lang.IllegalArgumentException(
25718          "Service.callMethod() given method descriptor for wrong " +
25719          "service type.");
25720      }
25721      switch(method.getIndex()) {
25722        case 0:
25723          this.isFormatted(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)request,
25724            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto>specializeCallback(
25725              done));
25726          return;
25727        case 1:
25728          this.discardSegments(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto)request,
25729            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto>specializeCallback(
25730              done));
25731          return;
25732        case 2:
25733          this.getJournalCTime(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto)request,
25734            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto>specializeCallback(
25735              done));
25736          return;
25737        case 3:
25738          this.doPreUpgrade(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto)request,
25739            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto>specializeCallback(
25740              done));
25741          return;
25742        case 4:
25743          this.doUpgrade(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto)request,
25744            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto>specializeCallback(
25745              done));
25746          return;
25747        case 5:
25748          this.doFinalize(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto)request,
25749            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto>specializeCallback(
25750              done));
25751          return;
25752        case 6:
25753          this.canRollBack(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto)request,
25754            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto>specializeCallback(
25755              done));
25756          return;
25757        case 7:
25758          this.doRollback(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto)request,
25759            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto>specializeCallback(
25760              done));
25761          return;
25762        case 8:
25763          this.getJournalState(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)request,
25764            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto>specializeCallback(
25765              done));
25766          return;
25767        case 9:
25768          this.newEpoch(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)request,
25769            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto>specializeCallback(
25770              done));
25771          return;
25772        case 10:
25773          this.format(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)request,
25774            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto>specializeCallback(
25775              done));
25776          return;
25777        case 11:
25778          this.journal(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)request,
25779            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto>specializeCallback(
25780              done));
25781          return;
25782        case 12:
25783          this.heartbeat(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)request,
25784            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto>specializeCallback(
25785              done));
25786          return;
25787        case 13:
25788          this.startLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)request,
25789            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto>specializeCallback(
25790              done));
25791          return;
25792        case 14:
25793          this.finalizeLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)request,
25794            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto>specializeCallback(
25795              done));
25796          return;
25797        case 15:
25798          this.purgeLogs(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)request,
25799            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto>specializeCallback(
25800              done));
25801          return;
25802        case 16:
25803          this.getEditLogManifest(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)request,
25804            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto>specializeCallback(
25805              done));
25806          return;
25807        case 17:
25808          this.prepareRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)request,
25809            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto>specializeCallback(
25810              done));
25811          return;
25812        case 18:
25813          this.acceptRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)request,
25814            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto>specializeCallback(
25815              done));
25816          return;
25817        default:
25818          throw new java.lang.AssertionError("Can't get here.");
25819      }
25820    }
25821
25822    public final com.google.protobuf.Message
25823        getRequestPrototype(
25824        com.google.protobuf.Descriptors.MethodDescriptor method) {
25825      if (method.getService() != getDescriptor()) {
25826        throw new java.lang.IllegalArgumentException(
25827          "Service.getRequestPrototype() given method " +
25828          "descriptor for wrong service type.");
25829      }
25830      switch(method.getIndex()) {
25831        case 0:
25832          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance();
25833        case 1:
25834          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto.getDefaultInstance();
25835        case 2:
25836          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto.getDefaultInstance();
25837        case 3:
25838          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto.getDefaultInstance();
25839        case 4:
25840          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto.getDefaultInstance();
25841        case 5:
25842          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto.getDefaultInstance();
25843        case 6:
25844          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto.getDefaultInstance();
25845        case 7:
25846          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto.getDefaultInstance();
25847        case 8:
25848          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance();
25849        case 9:
25850          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance();
25851        case 10:
25852          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance();
25853        case 11:
25854          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance();
25855        case 12:
25856          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance();
25857        case 13:
25858          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance();
25859        case 14:
25860          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance();
25861        case 15:
25862          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance();
25863        case 16:
25864          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance();
25865        case 17:
25866          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance();
25867        case 18:
25868          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance();
25869        default:
25870          throw new java.lang.AssertionError("Can't get here.");
25871      }
25872    }
25873
25874    public final com.google.protobuf.Message
25875        getResponsePrototype(
25876        com.google.protobuf.Descriptors.MethodDescriptor method) {
25877      if (method.getService() != getDescriptor()) {
25878        throw new java.lang.IllegalArgumentException(
25879          "Service.getResponsePrototype() given method " +
25880          "descriptor for wrong service type.");
25881      }
25882      switch(method.getIndex()) {
25883        case 0:
25884          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance();
25885        case 1:
25886          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.getDefaultInstance();
25887        case 2:
25888          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.getDefaultInstance();
25889        case 3:
25890          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.getDefaultInstance();
25891        case 4:
25892          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.getDefaultInstance();
25893        case 5:
25894          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.getDefaultInstance();
25895        case 6:
25896          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.getDefaultInstance();
25897        case 7:
25898          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.getDefaultInstance();
25899        case 8:
25900          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance();
25901        case 9:
25902          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance();
25903        case 10:
25904          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance();
25905        case 11:
25906          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance();
25907        case 12:
25908          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance();
25909        case 13:
25910          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance();
25911        case 14:
25912          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance();
25913        case 15:
25914          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance();
25915        case 16:
25916          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance();
25917        case 17:
25918          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance();
25919        case 18:
25920          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance();
25921        default:
25922          throw new java.lang.AssertionError("Can't get here.");
25923      }
25924    }
25925
25926    public static Stub newStub(
25927        com.google.protobuf.RpcChannel channel) {
25928      return new Stub(channel);
25929    }
25930
25931    public static final class Stub extends org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.QJournalProtocolService implements Interface {
25932      private Stub(com.google.protobuf.RpcChannel channel) {
25933        this.channel = channel;
25934      }
25935
25936      private final com.google.protobuf.RpcChannel channel;
25937
25938      public com.google.protobuf.RpcChannel getChannel() {
25939        return channel;
25940      }
25941
25942      public  void isFormatted(
25943          com.google.protobuf.RpcController controller,
25944          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
25945          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done) {
25946        channel.callMethod(
25947          getDescriptor().getMethods().get(0),
25948          controller,
25949          request,
25950          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance(),
25951          com.google.protobuf.RpcUtil.generalizeCallback(
25952            done,
25953            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.class,
25954            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance()));
25955      }
25956
25957      public  void discardSegments(
25958          com.google.protobuf.RpcController controller,
25959          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto request,
25960          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto> done) {
25961        channel.callMethod(
25962          getDescriptor().getMethods().get(1),
25963          controller,
25964          request,
25965          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.getDefaultInstance(),
25966          com.google.protobuf.RpcUtil.generalizeCallback(
25967            done,
25968            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.class,
25969            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.getDefaultInstance()));
25970      }
25971
25972      public  void getJournalCTime(
25973          com.google.protobuf.RpcController controller,
25974          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto request,
25975          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto> done) {
25976        channel.callMethod(
25977          getDescriptor().getMethods().get(2),
25978          controller,
25979          request,
25980          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.getDefaultInstance(),
25981          com.google.protobuf.RpcUtil.generalizeCallback(
25982            done,
25983            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.class,
25984            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.getDefaultInstance()));
25985      }
25986
25987      public  void doPreUpgrade(
25988          com.google.protobuf.RpcController controller,
25989          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto request,
25990          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto> done) {
25991        channel.callMethod(
25992          getDescriptor().getMethods().get(3),
25993          controller,
25994          request,
25995          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.getDefaultInstance(),
25996          com.google.protobuf.RpcUtil.generalizeCallback(
25997            done,
25998            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.class,
25999            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.getDefaultInstance()));
26000      }
26001
26002      public  void doUpgrade(
26003          com.google.protobuf.RpcController controller,
26004          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto request,
26005          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto> done) {
26006        channel.callMethod(
26007          getDescriptor().getMethods().get(4),
26008          controller,
26009          request,
26010          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.getDefaultInstance(),
26011          com.google.protobuf.RpcUtil.generalizeCallback(
26012            done,
26013            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.class,
26014            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.getDefaultInstance()));
26015      }
26016
26017      public  void doFinalize(
26018          com.google.protobuf.RpcController controller,
26019          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto request,
26020          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto> done) {
26021        channel.callMethod(
26022          getDescriptor().getMethods().get(5),
26023          controller,
26024          request,
26025          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.getDefaultInstance(),
26026          com.google.protobuf.RpcUtil.generalizeCallback(
26027            done,
26028            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.class,
26029            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.getDefaultInstance()));
26030      }
26031
26032      public  void canRollBack(
26033          com.google.protobuf.RpcController controller,
26034          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto request,
26035          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto> done) {
26036        channel.callMethod(
26037          getDescriptor().getMethods().get(6),
26038          controller,
26039          request,
26040          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.getDefaultInstance(),
26041          com.google.protobuf.RpcUtil.generalizeCallback(
26042            done,
26043            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.class,
26044            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.getDefaultInstance()));
26045      }
26046
26047      public  void doRollback(
26048          com.google.protobuf.RpcController controller,
26049          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto request,
26050          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto> done) {
26051        channel.callMethod(
26052          getDescriptor().getMethods().get(7),
26053          controller,
26054          request,
26055          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.getDefaultInstance(),
26056          com.google.protobuf.RpcUtil.generalizeCallback(
26057            done,
26058            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.class,
26059            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.getDefaultInstance()));
26060      }
26061
26062      public  void getJournalState(
26063          com.google.protobuf.RpcController controller,
26064          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
26065          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done) {
26066        channel.callMethod(
26067          getDescriptor().getMethods().get(8),
26068          controller,
26069          request,
26070          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance(),
26071          com.google.protobuf.RpcUtil.generalizeCallback(
26072            done,
26073            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.class,
26074            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance()));
26075      }
26076
26077      public  void newEpoch(
26078          com.google.protobuf.RpcController controller,
26079          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
26080          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done) {
26081        channel.callMethod(
26082          getDescriptor().getMethods().get(9),
26083          controller,
26084          request,
26085          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance(),
26086          com.google.protobuf.RpcUtil.generalizeCallback(
26087            done,
26088            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.class,
26089            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance()));
26090      }
26091
26092      public  void format(
26093          com.google.protobuf.RpcController controller,
26094          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
26095          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done) {
26096        channel.callMethod(
26097          getDescriptor().getMethods().get(10),
26098          controller,
26099          request,
26100          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance(),
26101          com.google.protobuf.RpcUtil.generalizeCallback(
26102            done,
26103            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.class,
26104            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance()));
26105      }
26106
26107      public  void journal(
26108          com.google.protobuf.RpcController controller,
26109          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
26110          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done) {
26111        channel.callMethod(
26112          getDescriptor().getMethods().get(11),
26113          controller,
26114          request,
26115          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance(),
26116          com.google.protobuf.RpcUtil.generalizeCallback(
26117            done,
26118            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.class,
26119            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance()));
26120      }
26121
26122      public  void heartbeat(
26123          com.google.protobuf.RpcController controller,
26124          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
26125          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done) {
26126        channel.callMethod(
26127          getDescriptor().getMethods().get(12),
26128          controller,
26129          request,
26130          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance(),
26131          com.google.protobuf.RpcUtil.generalizeCallback(
26132            done,
26133            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.class,
26134            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance()));
26135      }
26136
26137      public  void startLogSegment(
26138          com.google.protobuf.RpcController controller,
26139          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
26140          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done) {
26141        channel.callMethod(
26142          getDescriptor().getMethods().get(13),
26143          controller,
26144          request,
26145          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance(),
26146          com.google.protobuf.RpcUtil.generalizeCallback(
26147            done,
26148            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.class,
26149            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance()));
26150      }
26151
26152      public  void finalizeLogSegment(
26153          com.google.protobuf.RpcController controller,
26154          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
26155          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done) {
26156        channel.callMethod(
26157          getDescriptor().getMethods().get(14),
26158          controller,
26159          request,
26160          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance(),
26161          com.google.protobuf.RpcUtil.generalizeCallback(
26162            done,
26163            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.class,
26164            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance()));
26165      }
26166
26167      public  void purgeLogs(
26168          com.google.protobuf.RpcController controller,
26169          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
26170          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done) {
26171        channel.callMethod(
26172          getDescriptor().getMethods().get(15),
26173          controller,
26174          request,
26175          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance(),
26176          com.google.protobuf.RpcUtil.generalizeCallback(
26177            done,
26178            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.class,
26179            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance()));
26180      }
26181
26182      public  void getEditLogManifest(
26183          com.google.protobuf.RpcController controller,
26184          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
26185          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done) {
26186        channel.callMethod(
26187          getDescriptor().getMethods().get(16),
26188          controller,
26189          request,
26190          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance(),
26191          com.google.protobuf.RpcUtil.generalizeCallback(
26192            done,
26193            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.class,
26194            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance()));
26195      }
26196
26197      public  void prepareRecovery(
26198          com.google.protobuf.RpcController controller,
26199          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
26200          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done) {
26201        channel.callMethod(
26202          getDescriptor().getMethods().get(17),
26203          controller,
26204          request,
26205          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance(),
26206          com.google.protobuf.RpcUtil.generalizeCallback(
26207            done,
26208            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.class,
26209            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance()));
26210      }
26211
26212      public  void acceptRecovery(
26213          com.google.protobuf.RpcController controller,
26214          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
26215          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done) {
26216        channel.callMethod(
26217          getDescriptor().getMethods().get(18),
26218          controller,
26219          request,
26220          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance(),
26221          com.google.protobuf.RpcUtil.generalizeCallback(
26222            done,
26223            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.class,
26224            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance()));
26225      }
26226    }
26227
26228    public static BlockingInterface newBlockingStub(
26229        com.google.protobuf.BlockingRpcChannel channel) {
26230      return new BlockingStub(channel);
26231    }
26232
26233    public interface BlockingInterface {
26234      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto isFormatted(
26235          com.google.protobuf.RpcController controller,
26236          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request)
26237          throws com.google.protobuf.ServiceException;
26238
26239      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto discardSegments(
26240          com.google.protobuf.RpcController controller,
26241          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto request)
26242          throws com.google.protobuf.ServiceException;
26243
26244      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto getJournalCTime(
26245          com.google.protobuf.RpcController controller,
26246          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto request)
26247          throws com.google.protobuf.ServiceException;
26248
26249      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto doPreUpgrade(
26250          com.google.protobuf.RpcController controller,
26251          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto request)
26252          throws com.google.protobuf.ServiceException;
26253
26254      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto doUpgrade(
26255          com.google.protobuf.RpcController controller,
26256          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto request)
26257          throws com.google.protobuf.ServiceException;
26258
26259      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto doFinalize(
26260          com.google.protobuf.RpcController controller,
26261          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto request)
26262          throws com.google.protobuf.ServiceException;
26263
26264      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto canRollBack(
26265          com.google.protobuf.RpcController controller,
26266          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto request)
26267          throws com.google.protobuf.ServiceException;
26268
26269      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto doRollback(
26270          com.google.protobuf.RpcController controller,
26271          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto request)
26272          throws com.google.protobuf.ServiceException;
26273
26274      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getJournalState(
26275          com.google.protobuf.RpcController controller,
26276          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request)
26277          throws com.google.protobuf.ServiceException;
26278
26279      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto newEpoch(
26280          com.google.protobuf.RpcController controller,
26281          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request)
26282          throws com.google.protobuf.ServiceException;
26283
26284      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto format(
26285          com.google.protobuf.RpcController controller,
26286          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request)
26287          throws com.google.protobuf.ServiceException;
26288
26289      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto journal(
26290          com.google.protobuf.RpcController controller,
26291          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request)
26292          throws com.google.protobuf.ServiceException;
26293
26294      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto heartbeat(
26295          com.google.protobuf.RpcController controller,
26296          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request)
26297          throws com.google.protobuf.ServiceException;
26298
26299      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto startLogSegment(
26300          com.google.protobuf.RpcController controller,
26301          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request)
26302          throws com.google.protobuf.ServiceException;
26303
26304      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto finalizeLogSegment(
26305          com.google.protobuf.RpcController controller,
26306          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request)
26307          throws com.google.protobuf.ServiceException;
26308
26309      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto purgeLogs(
26310          com.google.protobuf.RpcController controller,
26311          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request)
26312          throws com.google.protobuf.ServiceException;
26313
26314      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getEditLogManifest(
26315          com.google.protobuf.RpcController controller,
26316          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request)
26317          throws com.google.protobuf.ServiceException;
26318
26319      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prepareRecovery(
26320          com.google.protobuf.RpcController controller,
26321          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request)
26322          throws com.google.protobuf.ServiceException;
26323
26324      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto acceptRecovery(
26325          com.google.protobuf.RpcController controller,
26326          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request)
26327          throws com.google.protobuf.ServiceException;
26328    }
26329
26330    private static final class BlockingStub implements BlockingInterface {
26331      private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
26332        this.channel = channel;
26333      }
26334
26335      private final com.google.protobuf.BlockingRpcChannel channel;
26336
26337      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto isFormatted(
26338          com.google.protobuf.RpcController controller,
26339          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request)
26340          throws com.google.protobuf.ServiceException {
26341        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) channel.callBlockingMethod(
26342          getDescriptor().getMethods().get(0),
26343          controller,
26344          request,
26345          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance());
26346      }
26347
26348
26349      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto discardSegments(
26350          com.google.protobuf.RpcController controller,
26351          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsRequestProto request)
26352          throws com.google.protobuf.ServiceException {
26353        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto) channel.callBlockingMethod(
26354          getDescriptor().getMethods().get(1),
26355          controller,
26356          request,
26357          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DiscardSegmentsResponseProto.getDefaultInstance());
26358      }
26359
26360
26361      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto getJournalCTime(
26362          com.google.protobuf.RpcController controller,
26363          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeRequestProto request)
26364          throws com.google.protobuf.ServiceException {
26365        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto) channel.callBlockingMethod(
26366          getDescriptor().getMethods().get(2),
26367          controller,
26368          request,
26369          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalCTimeResponseProto.getDefaultInstance());
26370      }
26371
26372
26373      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto doPreUpgrade(
26374          com.google.protobuf.RpcController controller,
26375          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeRequestProto request)
26376          throws com.google.protobuf.ServiceException {
26377        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto) channel.callBlockingMethod(
26378          getDescriptor().getMethods().get(3),
26379          controller,
26380          request,
26381          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoPreUpgradeResponseProto.getDefaultInstance());
26382      }
26383
26384
26385      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto doUpgrade(
26386          com.google.protobuf.RpcController controller,
26387          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeRequestProto request)
26388          throws com.google.protobuf.ServiceException {
26389        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto) channel.callBlockingMethod(
26390          getDescriptor().getMethods().get(4),
26391          controller,
26392          request,
26393          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoUpgradeResponseProto.getDefaultInstance());
26394      }
26395
26396
26397      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto doFinalize(
26398          com.google.protobuf.RpcController controller,
26399          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeRequestProto request)
26400          throws com.google.protobuf.ServiceException {
26401        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto) channel.callBlockingMethod(
26402          getDescriptor().getMethods().get(5),
26403          controller,
26404          request,
26405          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoFinalizeResponseProto.getDefaultInstance());
26406      }
26407
26408
26409      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto canRollBack(
26410          com.google.protobuf.RpcController controller,
26411          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackRequestProto request)
26412          throws com.google.protobuf.ServiceException {
26413        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto) channel.callBlockingMethod(
26414          getDescriptor().getMethods().get(6),
26415          controller,
26416          request,
26417          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.CanRollBackResponseProto.getDefaultInstance());
26418      }
26419
26420
26421      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto doRollback(
26422          com.google.protobuf.RpcController controller,
26423          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackRequestProto request)
26424          throws com.google.protobuf.ServiceException {
26425        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto) channel.callBlockingMethod(
26426          getDescriptor().getMethods().get(7),
26427          controller,
26428          request,
26429          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.DoRollbackResponseProto.getDefaultInstance());
26430      }
26431
26432
26433      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getJournalState(
26434          com.google.protobuf.RpcController controller,
26435          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request)
26436          throws com.google.protobuf.ServiceException {
26437        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) channel.callBlockingMethod(
26438          getDescriptor().getMethods().get(8),
26439          controller,
26440          request,
26441          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance());
26442      }
26443
26444
26445      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto newEpoch(
26446          com.google.protobuf.RpcController controller,
26447          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request)
26448          throws com.google.protobuf.ServiceException {
26449        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) channel.callBlockingMethod(
26450          getDescriptor().getMethods().get(9),
26451          controller,
26452          request,
26453          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance());
26454      }
26455
26456
26457      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto format(
26458          com.google.protobuf.RpcController controller,
26459          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request)
26460          throws com.google.protobuf.ServiceException {
26461        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) channel.callBlockingMethod(
26462          getDescriptor().getMethods().get(10),
26463          controller,
26464          request,
26465          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance());
26466      }
26467
26468
26469      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto journal(
26470          com.google.protobuf.RpcController controller,
26471          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request)
26472          throws com.google.protobuf.ServiceException {
26473        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) channel.callBlockingMethod(
26474          getDescriptor().getMethods().get(11),
26475          controller,
26476          request,
26477          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance());
26478      }
26479
26480
26481      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto heartbeat(
26482          com.google.protobuf.RpcController controller,
26483          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request)
26484          throws com.google.protobuf.ServiceException {
26485        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) channel.callBlockingMethod(
26486          getDescriptor().getMethods().get(12),
26487          controller,
26488          request,
26489          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance());
26490      }
26491
26492
26493      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto startLogSegment(
26494          com.google.protobuf.RpcController controller,
26495          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request)
26496          throws com.google.protobuf.ServiceException {
26497        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) channel.callBlockingMethod(
26498          getDescriptor().getMethods().get(13),
26499          controller,
26500          request,
26501          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance());
26502      }
26503
26504
26505      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto finalizeLogSegment(
26506          com.google.protobuf.RpcController controller,
26507          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request)
26508          throws com.google.protobuf.ServiceException {
26509        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) channel.callBlockingMethod(
26510          getDescriptor().getMethods().get(14),
26511          controller,
26512          request,
26513          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance());
26514      }
26515
26516
26517      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto purgeLogs(
26518          com.google.protobuf.RpcController controller,
26519          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request)
26520          throws com.google.protobuf.ServiceException {
26521        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) channel.callBlockingMethod(
26522          getDescriptor().getMethods().get(15),
26523          controller,
26524          request,
26525          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance());
26526      }
26527
26528
26529      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getEditLogManifest(
26530          com.google.protobuf.RpcController controller,
26531          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request)
26532          throws com.google.protobuf.ServiceException {
26533        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) channel.callBlockingMethod(
26534          getDescriptor().getMethods().get(16),
26535          controller,
26536          request,
26537          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance());
26538      }
26539
26540
26541      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prepareRecovery(
26542          com.google.protobuf.RpcController controller,
26543          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request)
26544          throws com.google.protobuf.ServiceException {
26545        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) channel.callBlockingMethod(
26546          getDescriptor().getMethods().get(17),
26547          controller,
26548          request,
26549          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance());
26550      }
26551
26552
26553      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto acceptRecovery(
26554          com.google.protobuf.RpcController controller,
26555          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request)
26556          throws com.google.protobuf.ServiceException {
26557        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) channel.callBlockingMethod(
26558          getDescriptor().getMethods().get(18),
26559          controller,
26560          request,
26561          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance());
26562      }
26563
26564    }
26565
26566    // @@protoc_insertion_point(class_scope:hadoop.hdfs.qjournal.QJournalProtocolService)
26567  }
26568
26569  private static com.google.protobuf.Descriptors.Descriptor
26570    internal_static_hadoop_hdfs_qjournal_JournalIdProto_descriptor;
26571  private static
26572    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26573      internal_static_hadoop_hdfs_qjournal_JournalIdProto_fieldAccessorTable;
26574  private static com.google.protobuf.Descriptors.Descriptor
26575    internal_static_hadoop_hdfs_qjournal_RequestInfoProto_descriptor;
26576  private static
26577    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26578      internal_static_hadoop_hdfs_qjournal_RequestInfoProto_fieldAccessorTable;
26579  private static com.google.protobuf.Descriptors.Descriptor
26580    internal_static_hadoop_hdfs_qjournal_SegmentStateProto_descriptor;
26581  private static
26582    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26583      internal_static_hadoop_hdfs_qjournal_SegmentStateProto_fieldAccessorTable;
26584  private static com.google.protobuf.Descriptors.Descriptor
26585    internal_static_hadoop_hdfs_qjournal_PersistedRecoveryPaxosData_descriptor;
26586  private static
26587    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26588      internal_static_hadoop_hdfs_qjournal_PersistedRecoveryPaxosData_fieldAccessorTable;
26589  private static com.google.protobuf.Descriptors.Descriptor
26590    internal_static_hadoop_hdfs_qjournal_JournalRequestProto_descriptor;
26591  private static
26592    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26593      internal_static_hadoop_hdfs_qjournal_JournalRequestProto_fieldAccessorTable;
26594  private static com.google.protobuf.Descriptors.Descriptor
26595    internal_static_hadoop_hdfs_qjournal_JournalResponseProto_descriptor;
26596  private static
26597    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26598      internal_static_hadoop_hdfs_qjournal_JournalResponseProto_fieldAccessorTable;
26599  private static com.google.protobuf.Descriptors.Descriptor
26600    internal_static_hadoop_hdfs_qjournal_HeartbeatRequestProto_descriptor;
26601  private static
26602    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26603      internal_static_hadoop_hdfs_qjournal_HeartbeatRequestProto_fieldAccessorTable;
26604  private static com.google.protobuf.Descriptors.Descriptor
26605    internal_static_hadoop_hdfs_qjournal_HeartbeatResponseProto_descriptor;
26606  private static
26607    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26608      internal_static_hadoop_hdfs_qjournal_HeartbeatResponseProto_fieldAccessorTable;
26609  private static com.google.protobuf.Descriptors.Descriptor
26610    internal_static_hadoop_hdfs_qjournal_StartLogSegmentRequestProto_descriptor;
26611  private static
26612    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26613      internal_static_hadoop_hdfs_qjournal_StartLogSegmentRequestProto_fieldAccessorTable;
26614  private static com.google.protobuf.Descriptors.Descriptor
26615    internal_static_hadoop_hdfs_qjournal_StartLogSegmentResponseProto_descriptor;
26616  private static
26617    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26618      internal_static_hadoop_hdfs_qjournal_StartLogSegmentResponseProto_fieldAccessorTable;
26619  private static com.google.protobuf.Descriptors.Descriptor
26620    internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentRequestProto_descriptor;
26621  private static
26622    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26623      internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentRequestProto_fieldAccessorTable;
26624  private static com.google.protobuf.Descriptors.Descriptor
26625    internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentResponseProto_descriptor;
26626  private static
26627    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26628      internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentResponseProto_fieldAccessorTable;
26629  private static com.google.protobuf.Descriptors.Descriptor
26630    internal_static_hadoop_hdfs_qjournal_PurgeLogsRequestProto_descriptor;
26631  private static
26632    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26633      internal_static_hadoop_hdfs_qjournal_PurgeLogsRequestProto_fieldAccessorTable;
26634  private static com.google.protobuf.Descriptors.Descriptor
26635    internal_static_hadoop_hdfs_qjournal_PurgeLogsResponseProto_descriptor;
26636  private static
26637    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26638      internal_static_hadoop_hdfs_qjournal_PurgeLogsResponseProto_fieldAccessorTable;
26639  private static com.google.protobuf.Descriptors.Descriptor
26640    internal_static_hadoop_hdfs_qjournal_IsFormattedRequestProto_descriptor;
26641  private static
26642    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26643      internal_static_hadoop_hdfs_qjournal_IsFormattedRequestProto_fieldAccessorTable;
26644  private static com.google.protobuf.Descriptors.Descriptor
26645    internal_static_hadoop_hdfs_qjournal_IsFormattedResponseProto_descriptor;
26646  private static
26647    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26648      internal_static_hadoop_hdfs_qjournal_IsFormattedResponseProto_fieldAccessorTable;
26649  private static com.google.protobuf.Descriptors.Descriptor
26650    internal_static_hadoop_hdfs_qjournal_DiscardSegmentsRequestProto_descriptor;
26651  private static
26652    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26653      internal_static_hadoop_hdfs_qjournal_DiscardSegmentsRequestProto_fieldAccessorTable;
26654  private static com.google.protobuf.Descriptors.Descriptor
26655    internal_static_hadoop_hdfs_qjournal_DiscardSegmentsResponseProto_descriptor;
26656  private static
26657    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26658      internal_static_hadoop_hdfs_qjournal_DiscardSegmentsResponseProto_fieldAccessorTable;
26659  private static com.google.protobuf.Descriptors.Descriptor
26660    internal_static_hadoop_hdfs_qjournal_GetJournalCTimeRequestProto_descriptor;
26661  private static
26662    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26663      internal_static_hadoop_hdfs_qjournal_GetJournalCTimeRequestProto_fieldAccessorTable;
26664  private static com.google.protobuf.Descriptors.Descriptor
26665    internal_static_hadoop_hdfs_qjournal_GetJournalCTimeResponseProto_descriptor;
26666  private static
26667    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26668      internal_static_hadoop_hdfs_qjournal_GetJournalCTimeResponseProto_fieldAccessorTable;
26669  private static com.google.protobuf.Descriptors.Descriptor
26670    internal_static_hadoop_hdfs_qjournal_DoPreUpgradeRequestProto_descriptor;
26671  private static
26672    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26673      internal_static_hadoop_hdfs_qjournal_DoPreUpgradeRequestProto_fieldAccessorTable;
26674  private static com.google.protobuf.Descriptors.Descriptor
26675    internal_static_hadoop_hdfs_qjournal_DoPreUpgradeResponseProto_descriptor;
26676  private static
26677    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26678      internal_static_hadoop_hdfs_qjournal_DoPreUpgradeResponseProto_fieldAccessorTable;
26679  private static com.google.protobuf.Descriptors.Descriptor
26680    internal_static_hadoop_hdfs_qjournal_DoUpgradeRequestProto_descriptor;
26681  private static
26682    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26683      internal_static_hadoop_hdfs_qjournal_DoUpgradeRequestProto_fieldAccessorTable;
26684  private static com.google.protobuf.Descriptors.Descriptor
26685    internal_static_hadoop_hdfs_qjournal_DoUpgradeResponseProto_descriptor;
26686  private static
26687    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26688      internal_static_hadoop_hdfs_qjournal_DoUpgradeResponseProto_fieldAccessorTable;
26689  private static com.google.protobuf.Descriptors.Descriptor
26690    internal_static_hadoop_hdfs_qjournal_DoFinalizeRequestProto_descriptor;
26691  private static
26692    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26693      internal_static_hadoop_hdfs_qjournal_DoFinalizeRequestProto_fieldAccessorTable;
26694  private static com.google.protobuf.Descriptors.Descriptor
26695    internal_static_hadoop_hdfs_qjournal_DoFinalizeResponseProto_descriptor;
26696  private static
26697    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26698      internal_static_hadoop_hdfs_qjournal_DoFinalizeResponseProto_fieldAccessorTable;
26699  private static com.google.protobuf.Descriptors.Descriptor
26700    internal_static_hadoop_hdfs_qjournal_CanRollBackRequestProto_descriptor;
26701  private static
26702    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26703      internal_static_hadoop_hdfs_qjournal_CanRollBackRequestProto_fieldAccessorTable;
26704  private static com.google.protobuf.Descriptors.Descriptor
26705    internal_static_hadoop_hdfs_qjournal_CanRollBackResponseProto_descriptor;
26706  private static
26707    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26708      internal_static_hadoop_hdfs_qjournal_CanRollBackResponseProto_fieldAccessorTable;
26709  private static com.google.protobuf.Descriptors.Descriptor
26710    internal_static_hadoop_hdfs_qjournal_DoRollbackRequestProto_descriptor;
26711  private static
26712    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26713      internal_static_hadoop_hdfs_qjournal_DoRollbackRequestProto_fieldAccessorTable;
26714  private static com.google.protobuf.Descriptors.Descriptor
26715    internal_static_hadoop_hdfs_qjournal_DoRollbackResponseProto_descriptor;
26716  private static
26717    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26718      internal_static_hadoop_hdfs_qjournal_DoRollbackResponseProto_fieldAccessorTable;
26719  private static com.google.protobuf.Descriptors.Descriptor
26720    internal_static_hadoop_hdfs_qjournal_GetJournalStateRequestProto_descriptor;
26721  private static
26722    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26723      internal_static_hadoop_hdfs_qjournal_GetJournalStateRequestProto_fieldAccessorTable;
26724  private static com.google.protobuf.Descriptors.Descriptor
26725    internal_static_hadoop_hdfs_qjournal_GetJournalStateResponseProto_descriptor;
26726  private static
26727    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26728      internal_static_hadoop_hdfs_qjournal_GetJournalStateResponseProto_fieldAccessorTable;
26729  private static com.google.protobuf.Descriptors.Descriptor
26730    internal_static_hadoop_hdfs_qjournal_FormatRequestProto_descriptor;
26731  private static
26732    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26733      internal_static_hadoop_hdfs_qjournal_FormatRequestProto_fieldAccessorTable;
26734  private static com.google.protobuf.Descriptors.Descriptor
26735    internal_static_hadoop_hdfs_qjournal_FormatResponseProto_descriptor;
26736  private static
26737    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26738      internal_static_hadoop_hdfs_qjournal_FormatResponseProto_fieldAccessorTable;
26739  private static com.google.protobuf.Descriptors.Descriptor
26740    internal_static_hadoop_hdfs_qjournal_NewEpochRequestProto_descriptor;
26741  private static
26742    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26743      internal_static_hadoop_hdfs_qjournal_NewEpochRequestProto_fieldAccessorTable;
26744  private static com.google.protobuf.Descriptors.Descriptor
26745    internal_static_hadoop_hdfs_qjournal_NewEpochResponseProto_descriptor;
26746  private static
26747    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26748      internal_static_hadoop_hdfs_qjournal_NewEpochResponseProto_fieldAccessorTable;
26749  private static com.google.protobuf.Descriptors.Descriptor
26750    internal_static_hadoop_hdfs_qjournal_GetEditLogManifestRequestProto_descriptor;
26751  private static
26752    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26753      internal_static_hadoop_hdfs_qjournal_GetEditLogManifestRequestProto_fieldAccessorTable;
26754  private static com.google.protobuf.Descriptors.Descriptor
26755    internal_static_hadoop_hdfs_qjournal_GetEditLogManifestResponseProto_descriptor;
26756  private static
26757    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26758      internal_static_hadoop_hdfs_qjournal_GetEditLogManifestResponseProto_fieldAccessorTable;
26759  private static com.google.protobuf.Descriptors.Descriptor
26760    internal_static_hadoop_hdfs_qjournal_PrepareRecoveryRequestProto_descriptor;
26761  private static
26762    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26763      internal_static_hadoop_hdfs_qjournal_PrepareRecoveryRequestProto_fieldAccessorTable;
26764  private static com.google.protobuf.Descriptors.Descriptor
26765    internal_static_hadoop_hdfs_qjournal_PrepareRecoveryResponseProto_descriptor;
26766  private static
26767    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26768      internal_static_hadoop_hdfs_qjournal_PrepareRecoveryResponseProto_fieldAccessorTable;
26769  private static com.google.protobuf.Descriptors.Descriptor
26770    internal_static_hadoop_hdfs_qjournal_AcceptRecoveryRequestProto_descriptor;
26771  private static
26772    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26773      internal_static_hadoop_hdfs_qjournal_AcceptRecoveryRequestProto_fieldAccessorTable;
26774  private static com.google.protobuf.Descriptors.Descriptor
26775    internal_static_hadoop_hdfs_qjournal_AcceptRecoveryResponseProto_descriptor;
26776  private static
26777    com.google.protobuf.GeneratedMessage.FieldAccessorTable
26778      internal_static_hadoop_hdfs_qjournal_AcceptRecoveryResponseProto_fieldAccessorTable;
26779
26780  public static com.google.protobuf.Descriptors.FileDescriptor
26781      getDescriptor() {
26782    return descriptor;
26783  }
26784  private static com.google.protobuf.Descriptors.FileDescriptor
26785      descriptor;
26786  static {
26787    java.lang.String[] descriptorData = {
26788      "\n\026QJournalProtocol.proto\022\024hadoop.hdfs.qj" +
26789      "ournal\032\nhdfs.proto\032\020HdfsServer.proto\"$\n\016" +
26790      "JournalIdProto\022\022\n\nidentifier\030\001 \002(\t\"\212\001\n\020R" +
26791      "equestInfoProto\0227\n\tjournalId\030\001 \002(\0132$.had" +
26792      "oop.hdfs.qjournal.JournalIdProto\022\r\n\005epoc" +
26793      "h\030\002 \002(\004\022\027\n\017ipcSerialNumber\030\003 \002(\004\022\025\n\rcomm" +
26794      "ittedTxId\030\004 \001(\004\"M\n\021SegmentStateProto\022\021\n\t" +
26795      "startTxId\030\001 \002(\004\022\017\n\007endTxId\030\002 \002(\004\022\024\n\014isIn" +
26796      "Progress\030\003 \002(\010\"t\n\032PersistedRecoveryPaxos" +
26797      "Data\022=\n\014segmentState\030\001 \002(\0132\'.hadoop.hdfs",
26798      ".qjournal.SegmentStateProto\022\027\n\017acceptedI" +
26799      "nEpoch\030\002 \002(\004\"\232\001\n\023JournalRequestProto\0227\n\007" +
26800      "reqInfo\030\001 \002(\0132&.hadoop.hdfs.qjournal.Req" +
26801      "uestInfoProto\022\022\n\nfirstTxnId\030\002 \002(\004\022\017\n\007num" +
26802      "Txns\030\003 \002(\r\022\017\n\007records\030\004 \002(\014\022\024\n\014segmentTx" +
26803      "nId\030\005 \002(\004\"\026\n\024JournalResponseProto\"P\n\025Hea" +
26804      "rtbeatRequestProto\0227\n\007reqInfo\030\001 \002(\0132&.ha" +
26805      "doop.hdfs.qjournal.RequestInfoProto\"\030\n\026H" +
26806      "eartbeatResponseProto\"{\n\033StartLogSegment" +
26807      "RequestProto\0227\n\007reqInfo\030\001 \002(\0132&.hadoop.h",
26808      "dfs.qjournal.RequestInfoProto\022\014\n\004txid\030\002 " +
26809      "\002(\004\022\025\n\rlayoutVersion\030\003 \001(\021\"\036\n\034StartLogSe" +
26810      "gmentResponseProto\"}\n\036FinalizeLogSegment" +
26811      "RequestProto\0227\n\007reqInfo\030\001 \002(\0132&.hadoop.h" +
26812      "dfs.qjournal.RequestInfoProto\022\021\n\tstartTx" +
26813      "Id\030\002 \002(\004\022\017\n\007endTxId\030\003 \002(\004\"!\n\037FinalizeLog" +
26814      "SegmentResponseProto\"g\n\025PurgeLogsRequest" +
26815      "Proto\0227\n\007reqInfo\030\001 \002(\0132&.hadoop.hdfs.qjo" +
26816      "urnal.RequestInfoProto\022\025\n\rminTxIdToKeep\030" +
26817      "\002 \002(\004\"\030\n\026PurgeLogsResponseProto\"L\n\027IsFor",
26818      "mattedRequestProto\0221\n\003jid\030\001 \002(\0132$.hadoop" +
26819      ".hdfs.qjournal.JournalIdProto\"/\n\030IsForma" +
26820      "ttedResponseProto\022\023\n\013isFormatted\030\001 \002(\010\"c" +
26821      "\n\033DiscardSegmentsRequestProto\0221\n\003jid\030\001 \002" +
26822      "(\0132$.hadoop.hdfs.qjournal.JournalIdProto" +
26823      "\022\021\n\tstartTxId\030\002 \002(\004\"\036\n\034DiscardSegmentsRe" +
26824      "sponseProto\"P\n\033GetJournalCTimeRequestPro" +
26825      "to\0221\n\003jid\030\001 \002(\0132$.hadoop.hdfs.qjournal.J" +
26826      "ournalIdProto\"3\n\034GetJournalCTimeResponse" +
26827      "Proto\022\023\n\013resultCTime\030\001 \002(\003\"M\n\030DoPreUpgra",
26828      "deRequestProto\0221\n\003jid\030\001 \002(\0132$.hadoop.hdf" +
26829      "s.qjournal.JournalIdProto\"\033\n\031DoPreUpgrad" +
26830      "eResponseProto\"x\n\025DoUpgradeRequestProto\022" +
26831      "1\n\003jid\030\001 \002(\0132$.hadoop.hdfs.qjournal.Jour" +
26832      "nalIdProto\022,\n\005sInfo\030\002 \002(\0132\035.hadoop.hdfs." +
26833      "StorageInfoProto\"\030\n\026DoUpgradeResponsePro" +
26834      "to\"K\n\026DoFinalizeRequestProto\0221\n\003jid\030\001 \002(" +
26835      "\0132$.hadoop.hdfs.qjournal.JournalIdProto\"" +
26836      "\031\n\027DoFinalizeResponseProto\"\315\001\n\027CanRollBa" +
26837      "ckRequestProto\0221\n\003jid\030\001 \002(\0132$.hadoop.hdf",
26838      "s.qjournal.JournalIdProto\022.\n\007storage\030\002 \002" +
26839      "(\0132\035.hadoop.hdfs.StorageInfoProto\0222\n\013pre" +
26840      "vStorage\030\003 \002(\0132\035.hadoop.hdfs.StorageInfo" +
26841      "Proto\022\033\n\023targetLayoutVersion\030\004 \002(\005\"/\n\030Ca" +
26842      "nRollBackResponseProto\022\023\n\013canRollBack\030\001 " +
26843      "\002(\010\"K\n\026DoRollbackRequestProto\0221\n\003jid\030\001 \002" +
26844      "(\0132$.hadoop.hdfs.qjournal.JournalIdProto" +
26845      "\"\031\n\027DoRollbackResponseProto\"P\n\033GetJourna" +
26846      "lStateRequestProto\0221\n\003jid\030\001 \002(\0132$.hadoop" +
26847      ".hdfs.qjournal.JournalIdProto\"\\\n\034GetJour",
26848      "nalStateResponseProto\022\031\n\021lastPromisedEpo" +
26849      "ch\030\001 \002(\004\022\020\n\010httpPort\030\002 \002(\r\022\017\n\007fromURL\030\003 " +
26850      "\001(\t\"x\n\022FormatRequestProto\0221\n\003jid\030\001 \002(\0132$" +
26851      ".hadoop.hdfs.qjournal.JournalIdProto\022/\n\006" +
26852      "nsInfo\030\002 \002(\0132\037.hadoop.hdfs.NamespaceInfo" +
26853      "Proto\"\025\n\023FormatResponseProto\"\211\001\n\024NewEpoc" +
26854      "hRequestProto\0221\n\003jid\030\001 \002(\0132$.hadoop.hdfs" +
26855      ".qjournal.JournalIdProto\022/\n\006nsInfo\030\002 \002(\013" +
26856      "2\037.hadoop.hdfs.NamespaceInfoProto\022\r\n\005epo" +
26857      "ch\030\003 \002(\004\"0\n\025NewEpochResponseProto\022\027\n\017las",
26858      "tSegmentTxId\030\001 \001(\004\"\203\001\n\036GetEditLogManifes" +
26859      "tRequestProto\0221\n\003jid\030\001 \002(\0132$.hadoop.hdfs" +
26860      ".qjournal.JournalIdProto\022\021\n\tsinceTxId\030\002 " +
26861      "\002(\004\022\033\n\014inProgressOk\030\004 \001(\010:\005false\"\177\n\037GetE" +
26862      "ditLogManifestResponseProto\0229\n\010manifest\030" +
26863      "\001 \002(\0132\'.hadoop.hdfs.RemoteEditLogManifes" +
26864      "tProto\022\020\n\010httpPort\030\002 \002(\r\022\017\n\007fromURL\030\003 \001(" +
26865      "\t\"k\n\033PrepareRecoveryRequestProto\0227\n\007reqI" +
26866      "nfo\030\001 \002(\0132&.hadoop.hdfs.qjournal.Request" +
26867      "InfoProto\022\023\n\013segmentTxId\030\002 \002(\004\"\252\001\n\034Prepa",
26868      "reRecoveryResponseProto\022=\n\014segmentState\030" +
26869      "\001 \001(\0132\'.hadoop.hdfs.qjournal.SegmentStat" +
26870      "eProto\022\027\n\017acceptedInEpoch\030\002 \001(\004\022\027\n\017lastW" +
26871      "riterEpoch\030\003 \002(\004\022\031\n\021lastCommittedTxId\030\004 " +
26872      "\001(\004\"\246\001\n\032AcceptRecoveryRequestProto\0227\n\007re" +
26873      "qInfo\030\001 \002(\0132&.hadoop.hdfs.qjournal.Reque" +
26874      "stInfoProto\022>\n\rstateToAccept\030\002 \002(\0132\'.had" +
26875      "oop.hdfs.qjournal.SegmentStateProto\022\017\n\007f" +
26876      "romURL\030\003 \002(\t\"\035\n\033AcceptRecoveryResponsePr" +
26877      "oto2\373\020\n\027QJournalProtocolService\022l\n\013isFor",
26878      "matted\022-.hadoop.hdfs.qjournal.IsFormatte" +
26879      "dRequestProto\032..hadoop.hdfs.qjournal.IsF" +
26880      "ormattedResponseProto\022x\n\017discardSegments" +
26881      "\0221.hadoop.hdfs.qjournal.DiscardSegmentsR" +
26882      "equestProto\0322.hadoop.hdfs.qjournal.Disca" +
26883      "rdSegmentsResponseProto\022x\n\017getJournalCTi" +
26884      "me\0221.hadoop.hdfs.qjournal.GetJournalCTim" +
26885      "eRequestProto\0322.hadoop.hdfs.qjournal.Get" +
26886      "JournalCTimeResponseProto\022o\n\014doPreUpgrad" +
26887      "e\022..hadoop.hdfs.qjournal.DoPreUpgradeReq",
26888      "uestProto\032/.hadoop.hdfs.qjournal.DoPreUp" +
26889      "gradeResponseProto\022f\n\tdoUpgrade\022+.hadoop" +
26890      ".hdfs.qjournal.DoUpgradeRequestProto\032,.h" +
26891      "adoop.hdfs.qjournal.DoUpgradeResponsePro" +
26892      "to\022i\n\ndoFinalize\022,.hadoop.hdfs.qjournal." +
26893      "DoFinalizeRequestProto\032-.hadoop.hdfs.qjo" +
26894      "urnal.DoFinalizeResponseProto\022l\n\013canRoll" +
26895      "Back\022-.hadoop.hdfs.qjournal.CanRollBackR" +
26896      "equestProto\032..hadoop.hdfs.qjournal.CanRo" +
26897      "llBackResponseProto\022i\n\ndoRollback\022,.hado",
26898      "op.hdfs.qjournal.DoRollbackRequestProto\032" +
26899      "-.hadoop.hdfs.qjournal.DoRollbackRespons" +
26900      "eProto\022x\n\017getJournalState\0221.hadoop.hdfs." +
26901      "qjournal.GetJournalStateRequestProto\0322.h" +
26902      "adoop.hdfs.qjournal.GetJournalStateRespo" +
26903      "nseProto\022c\n\010newEpoch\022*.hadoop.hdfs.qjour" +
26904      "nal.NewEpochRequestProto\032+.hadoop.hdfs.q" +
26905      "journal.NewEpochResponseProto\022]\n\006format\022" +
26906      "(.hadoop.hdfs.qjournal.FormatRequestProt" +
26907      "o\032).hadoop.hdfs.qjournal.FormatResponseP",
26908      "roto\022`\n\007journal\022).hadoop.hdfs.qjournal.J" +
26909      "ournalRequestProto\032*.hadoop.hdfs.qjourna" +
26910      "l.JournalResponseProto\022f\n\theartbeat\022+.ha" +
26911      "doop.hdfs.qjournal.HeartbeatRequestProto" +
26912      "\032,.hadoop.hdfs.qjournal.HeartbeatRespons" +
26913      "eProto\022x\n\017startLogSegment\0221.hadoop.hdfs." +
26914      "qjournal.StartLogSegmentRequestProto\0322.h" +
26915      "adoop.hdfs.qjournal.StartLogSegmentRespo" +
26916      "nseProto\022\201\001\n\022finalizeLogSegment\0224.hadoop" +
26917      ".hdfs.qjournal.FinalizeLogSegmentRequest",
26918      "Proto\0325.hadoop.hdfs.qjournal.FinalizeLog" +
26919      "SegmentResponseProto\022f\n\tpurgeLogs\022+.hado" +
26920      "op.hdfs.qjournal.PurgeLogsRequestProto\032," +
26921      ".hadoop.hdfs.qjournal.PurgeLogsResponseP" +
26922      "roto\022\201\001\n\022getEditLogManifest\0224.hadoop.hdf" +
26923      "s.qjournal.GetEditLogManifestRequestProt" +
26924      "o\0325.hadoop.hdfs.qjournal.GetEditLogManif" +
26925      "estResponseProto\022x\n\017prepareRecovery\0221.ha" +
26926      "doop.hdfs.qjournal.PrepareRecoveryReques" +
26927      "tProto\0322.hadoop.hdfs.qjournal.PrepareRec",
26928      "overyResponseProto\022u\n\016acceptRecovery\0220.h" +
26929      "adoop.hdfs.qjournal.AcceptRecoveryReques" +
26930      "tProto\0321.hadoop.hdfs.qjournal.AcceptReco" +
26931      "veryResponseProtoBH\n(org.apache.hadoop.h" +
26932      "dfs.qjournal.protocolB\026QJournalProtocolP" +
26933      "rotos\210\001\001\240\001\001"
26934    };
26935    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
26936      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
26937        public com.google.protobuf.ExtensionRegistry assignDescriptors(
26938            com.google.protobuf.Descriptors.FileDescriptor root) {
26939          descriptor = root;
26940          internal_static_hadoop_hdfs_qjournal_JournalIdProto_descriptor =
26941            getDescriptor().getMessageTypes().get(0);
26942          internal_static_hadoop_hdfs_qjournal_JournalIdProto_fieldAccessorTable = new
26943            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
26944              internal_static_hadoop_hdfs_qjournal_JournalIdProto_descriptor,
26945              new java.lang.String[] { "Identifier", });
26946          internal_static_hadoop_hdfs_qjournal_RequestInfoProto_descriptor =
26947            getDescriptor().getMessageTypes().get(1);
26948          internal_static_hadoop_hdfs_qjournal_RequestInfoProto_fieldAccessorTable = new
26949            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
26950              internal_static_hadoop_hdfs_qjournal_RequestInfoProto_descriptor,
26951              new java.lang.String[] { "JournalId", "Epoch", "IpcSerialNumber", "CommittedTxId", });
26952          internal_static_hadoop_hdfs_qjournal_SegmentStateProto_descriptor =
26953            getDescriptor().getMessageTypes().get(2);
26954          internal_static_hadoop_hdfs_qjournal_SegmentStateProto_fieldAccessorTable = new
26955            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
26956              internal_static_hadoop_hdfs_qjournal_SegmentStateProto_descriptor,
26957              new java.lang.String[] { "StartTxId", "EndTxId", "IsInProgress", });
26958          internal_static_hadoop_hdfs_qjournal_PersistedRecoveryPaxosData_descriptor =
26959            getDescriptor().getMessageTypes().get(3);
26960          internal_static_hadoop_hdfs_qjournal_PersistedRecoveryPaxosData_fieldAccessorTable = new
26961            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
26962              internal_static_hadoop_hdfs_qjournal_PersistedRecoveryPaxosData_descriptor,
26963              new java.lang.String[] { "SegmentState", "AcceptedInEpoch", });
26964          internal_static_hadoop_hdfs_qjournal_JournalRequestProto_descriptor =
26965            getDescriptor().getMessageTypes().get(4);
26966          internal_static_hadoop_hdfs_qjournal_JournalRequestProto_fieldAccessorTable = new
26967            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
26968              internal_static_hadoop_hdfs_qjournal_JournalRequestProto_descriptor,
26969              new java.lang.String[] { "ReqInfo", "FirstTxnId", "NumTxns", "Records", "SegmentTxnId", });
26970          internal_static_hadoop_hdfs_qjournal_JournalResponseProto_descriptor =
26971            getDescriptor().getMessageTypes().get(5);
26972          internal_static_hadoop_hdfs_qjournal_JournalResponseProto_fieldAccessorTable = new
26973            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
26974              internal_static_hadoop_hdfs_qjournal_JournalResponseProto_descriptor,
26975              new java.lang.String[] { });
26976          internal_static_hadoop_hdfs_qjournal_HeartbeatRequestProto_descriptor =
26977            getDescriptor().getMessageTypes().get(6);
26978          internal_static_hadoop_hdfs_qjournal_HeartbeatRequestProto_fieldAccessorTable = new
26979            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
26980              internal_static_hadoop_hdfs_qjournal_HeartbeatRequestProto_descriptor,
26981              new java.lang.String[] { "ReqInfo", });
26982          internal_static_hadoop_hdfs_qjournal_HeartbeatResponseProto_descriptor =
26983            getDescriptor().getMessageTypes().get(7);
26984          internal_static_hadoop_hdfs_qjournal_HeartbeatResponseProto_fieldAccessorTable = new
26985            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
26986              internal_static_hadoop_hdfs_qjournal_HeartbeatResponseProto_descriptor,
26987              new java.lang.String[] { });
26988          internal_static_hadoop_hdfs_qjournal_StartLogSegmentRequestProto_descriptor =
26989            getDescriptor().getMessageTypes().get(8);
26990          internal_static_hadoop_hdfs_qjournal_StartLogSegmentRequestProto_fieldAccessorTable = new
26991            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
26992              internal_static_hadoop_hdfs_qjournal_StartLogSegmentRequestProto_descriptor,
26993              new java.lang.String[] { "ReqInfo", "Txid", "LayoutVersion", });
26994          internal_static_hadoop_hdfs_qjournal_StartLogSegmentResponseProto_descriptor =
26995            getDescriptor().getMessageTypes().get(9);
26996          internal_static_hadoop_hdfs_qjournal_StartLogSegmentResponseProto_fieldAccessorTable = new
26997            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
26998              internal_static_hadoop_hdfs_qjournal_StartLogSegmentResponseProto_descriptor,
26999              new java.lang.String[] { });
27000          internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentRequestProto_descriptor =
27001            getDescriptor().getMessageTypes().get(10);
27002          internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentRequestProto_fieldAccessorTable = new
27003            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27004              internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentRequestProto_descriptor,
27005              new java.lang.String[] { "ReqInfo", "StartTxId", "EndTxId", });
27006          internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentResponseProto_descriptor =
27007            getDescriptor().getMessageTypes().get(11);
27008          internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentResponseProto_fieldAccessorTable = new
27009            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27010              internal_static_hadoop_hdfs_qjournal_FinalizeLogSegmentResponseProto_descriptor,
27011              new java.lang.String[] { });
27012          internal_static_hadoop_hdfs_qjournal_PurgeLogsRequestProto_descriptor =
27013            getDescriptor().getMessageTypes().get(12);
27014          internal_static_hadoop_hdfs_qjournal_PurgeLogsRequestProto_fieldAccessorTable = new
27015            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27016              internal_static_hadoop_hdfs_qjournal_PurgeLogsRequestProto_descriptor,
27017              new java.lang.String[] { "ReqInfo", "MinTxIdToKeep", });
27018          internal_static_hadoop_hdfs_qjournal_PurgeLogsResponseProto_descriptor =
27019            getDescriptor().getMessageTypes().get(13);
27020          internal_static_hadoop_hdfs_qjournal_PurgeLogsResponseProto_fieldAccessorTable = new
27021            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27022              internal_static_hadoop_hdfs_qjournal_PurgeLogsResponseProto_descriptor,
27023              new java.lang.String[] { });
27024          internal_static_hadoop_hdfs_qjournal_IsFormattedRequestProto_descriptor =
27025            getDescriptor().getMessageTypes().get(14);
27026          internal_static_hadoop_hdfs_qjournal_IsFormattedRequestProto_fieldAccessorTable = new
27027            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27028              internal_static_hadoop_hdfs_qjournal_IsFormattedRequestProto_descriptor,
27029              new java.lang.String[] { "Jid", });
27030          internal_static_hadoop_hdfs_qjournal_IsFormattedResponseProto_descriptor =
27031            getDescriptor().getMessageTypes().get(15);
27032          internal_static_hadoop_hdfs_qjournal_IsFormattedResponseProto_fieldAccessorTable = new
27033            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27034              internal_static_hadoop_hdfs_qjournal_IsFormattedResponseProto_descriptor,
27035              new java.lang.String[] { "IsFormatted", });
27036          internal_static_hadoop_hdfs_qjournal_DiscardSegmentsRequestProto_descriptor =
27037            getDescriptor().getMessageTypes().get(16);
27038          internal_static_hadoop_hdfs_qjournal_DiscardSegmentsRequestProto_fieldAccessorTable = new
27039            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27040              internal_static_hadoop_hdfs_qjournal_DiscardSegmentsRequestProto_descriptor,
27041              new java.lang.String[] { "Jid", "StartTxId", });
27042          internal_static_hadoop_hdfs_qjournal_DiscardSegmentsResponseProto_descriptor =
27043            getDescriptor().getMessageTypes().get(17);
27044          internal_static_hadoop_hdfs_qjournal_DiscardSegmentsResponseProto_fieldAccessorTable = new
27045            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27046              internal_static_hadoop_hdfs_qjournal_DiscardSegmentsResponseProto_descriptor,
27047              new java.lang.String[] { });
27048          internal_static_hadoop_hdfs_qjournal_GetJournalCTimeRequestProto_descriptor =
27049            getDescriptor().getMessageTypes().get(18);
27050          internal_static_hadoop_hdfs_qjournal_GetJournalCTimeRequestProto_fieldAccessorTable = new
27051            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27052              internal_static_hadoop_hdfs_qjournal_GetJournalCTimeRequestProto_descriptor,
27053              new java.lang.String[] { "Jid", });
27054          internal_static_hadoop_hdfs_qjournal_GetJournalCTimeResponseProto_descriptor =
27055            getDescriptor().getMessageTypes().get(19);
27056          internal_static_hadoop_hdfs_qjournal_GetJournalCTimeResponseProto_fieldAccessorTable = new
27057            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27058              internal_static_hadoop_hdfs_qjournal_GetJournalCTimeResponseProto_descriptor,
27059              new java.lang.String[] { "ResultCTime", });
27060          internal_static_hadoop_hdfs_qjournal_DoPreUpgradeRequestProto_descriptor =
27061            getDescriptor().getMessageTypes().get(20);
27062          internal_static_hadoop_hdfs_qjournal_DoPreUpgradeRequestProto_fieldAccessorTable = new
27063            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27064              internal_static_hadoop_hdfs_qjournal_DoPreUpgradeRequestProto_descriptor,
27065              new java.lang.String[] { "Jid", });
27066          internal_static_hadoop_hdfs_qjournal_DoPreUpgradeResponseProto_descriptor =
27067            getDescriptor().getMessageTypes().get(21);
27068          internal_static_hadoop_hdfs_qjournal_DoPreUpgradeResponseProto_fieldAccessorTable = new
27069            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27070              internal_static_hadoop_hdfs_qjournal_DoPreUpgradeResponseProto_descriptor,
27071              new java.lang.String[] { });
27072          internal_static_hadoop_hdfs_qjournal_DoUpgradeRequestProto_descriptor =
27073            getDescriptor().getMessageTypes().get(22);
27074          internal_static_hadoop_hdfs_qjournal_DoUpgradeRequestProto_fieldAccessorTable = new
27075            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27076              internal_static_hadoop_hdfs_qjournal_DoUpgradeRequestProto_descriptor,
27077              new java.lang.String[] { "Jid", "SInfo", });
27078          internal_static_hadoop_hdfs_qjournal_DoUpgradeResponseProto_descriptor =
27079            getDescriptor().getMessageTypes().get(23);
27080          internal_static_hadoop_hdfs_qjournal_DoUpgradeResponseProto_fieldAccessorTable = new
27081            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27082              internal_static_hadoop_hdfs_qjournal_DoUpgradeResponseProto_descriptor,
27083              new java.lang.String[] { });
27084          internal_static_hadoop_hdfs_qjournal_DoFinalizeRequestProto_descriptor =
27085            getDescriptor().getMessageTypes().get(24);
27086          internal_static_hadoop_hdfs_qjournal_DoFinalizeRequestProto_fieldAccessorTable = new
27087            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27088              internal_static_hadoop_hdfs_qjournal_DoFinalizeRequestProto_descriptor,
27089              new java.lang.String[] { "Jid", });
27090          internal_static_hadoop_hdfs_qjournal_DoFinalizeResponseProto_descriptor =
27091            getDescriptor().getMessageTypes().get(25);
27092          internal_static_hadoop_hdfs_qjournal_DoFinalizeResponseProto_fieldAccessorTable = new
27093            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27094              internal_static_hadoop_hdfs_qjournal_DoFinalizeResponseProto_descriptor,
27095              new java.lang.String[] { });
27096          internal_static_hadoop_hdfs_qjournal_CanRollBackRequestProto_descriptor =
27097            getDescriptor().getMessageTypes().get(26);
27098          internal_static_hadoop_hdfs_qjournal_CanRollBackRequestProto_fieldAccessorTable = new
27099            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27100              internal_static_hadoop_hdfs_qjournal_CanRollBackRequestProto_descriptor,
27101              new java.lang.String[] { "Jid", "Storage", "PrevStorage", "TargetLayoutVersion", });
27102          internal_static_hadoop_hdfs_qjournal_CanRollBackResponseProto_descriptor =
27103            getDescriptor().getMessageTypes().get(27);
27104          internal_static_hadoop_hdfs_qjournal_CanRollBackResponseProto_fieldAccessorTable = new
27105            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27106              internal_static_hadoop_hdfs_qjournal_CanRollBackResponseProto_descriptor,
27107              new java.lang.String[] { "CanRollBack", });
27108          internal_static_hadoop_hdfs_qjournal_DoRollbackRequestProto_descriptor =
27109            getDescriptor().getMessageTypes().get(28);
27110          internal_static_hadoop_hdfs_qjournal_DoRollbackRequestProto_fieldAccessorTable = new
27111            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27112              internal_static_hadoop_hdfs_qjournal_DoRollbackRequestProto_descriptor,
27113              new java.lang.String[] { "Jid", });
27114          internal_static_hadoop_hdfs_qjournal_DoRollbackResponseProto_descriptor =
27115            getDescriptor().getMessageTypes().get(29);
27116          internal_static_hadoop_hdfs_qjournal_DoRollbackResponseProto_fieldAccessorTable = new
27117            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27118              internal_static_hadoop_hdfs_qjournal_DoRollbackResponseProto_descriptor,
27119              new java.lang.String[] { });
27120          internal_static_hadoop_hdfs_qjournal_GetJournalStateRequestProto_descriptor =
27121            getDescriptor().getMessageTypes().get(30);
27122          internal_static_hadoop_hdfs_qjournal_GetJournalStateRequestProto_fieldAccessorTable = new
27123            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27124              internal_static_hadoop_hdfs_qjournal_GetJournalStateRequestProto_descriptor,
27125              new java.lang.String[] { "Jid", });
27126          internal_static_hadoop_hdfs_qjournal_GetJournalStateResponseProto_descriptor =
27127            getDescriptor().getMessageTypes().get(31);
27128          internal_static_hadoop_hdfs_qjournal_GetJournalStateResponseProto_fieldAccessorTable = new
27129            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27130              internal_static_hadoop_hdfs_qjournal_GetJournalStateResponseProto_descriptor,
27131              new java.lang.String[] { "LastPromisedEpoch", "HttpPort", "FromURL", });
27132          internal_static_hadoop_hdfs_qjournal_FormatRequestProto_descriptor =
27133            getDescriptor().getMessageTypes().get(32);
27134          internal_static_hadoop_hdfs_qjournal_FormatRequestProto_fieldAccessorTable = new
27135            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27136              internal_static_hadoop_hdfs_qjournal_FormatRequestProto_descriptor,
27137              new java.lang.String[] { "Jid", "NsInfo", });
27138          internal_static_hadoop_hdfs_qjournal_FormatResponseProto_descriptor =
27139            getDescriptor().getMessageTypes().get(33);
27140          internal_static_hadoop_hdfs_qjournal_FormatResponseProto_fieldAccessorTable = new
27141            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27142              internal_static_hadoop_hdfs_qjournal_FormatResponseProto_descriptor,
27143              new java.lang.String[] { });
27144          internal_static_hadoop_hdfs_qjournal_NewEpochRequestProto_descriptor =
27145            getDescriptor().getMessageTypes().get(34);
27146          internal_static_hadoop_hdfs_qjournal_NewEpochRequestProto_fieldAccessorTable = new
27147            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27148              internal_static_hadoop_hdfs_qjournal_NewEpochRequestProto_descriptor,
27149              new java.lang.String[] { "Jid", "NsInfo", "Epoch", });
27150          internal_static_hadoop_hdfs_qjournal_NewEpochResponseProto_descriptor =
27151            getDescriptor().getMessageTypes().get(35);
27152          internal_static_hadoop_hdfs_qjournal_NewEpochResponseProto_fieldAccessorTable = new
27153            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27154              internal_static_hadoop_hdfs_qjournal_NewEpochResponseProto_descriptor,
27155              new java.lang.String[] { "LastSegmentTxId", });
27156          internal_static_hadoop_hdfs_qjournal_GetEditLogManifestRequestProto_descriptor =
27157            getDescriptor().getMessageTypes().get(36);
27158          internal_static_hadoop_hdfs_qjournal_GetEditLogManifestRequestProto_fieldAccessorTable = new
27159            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27160              internal_static_hadoop_hdfs_qjournal_GetEditLogManifestRequestProto_descriptor,
27161              new java.lang.String[] { "Jid", "SinceTxId", "InProgressOk", });
27162          internal_static_hadoop_hdfs_qjournal_GetEditLogManifestResponseProto_descriptor =
27163            getDescriptor().getMessageTypes().get(37);
27164          internal_static_hadoop_hdfs_qjournal_GetEditLogManifestResponseProto_fieldAccessorTable = new
27165            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27166              internal_static_hadoop_hdfs_qjournal_GetEditLogManifestResponseProto_descriptor,
27167              new java.lang.String[] { "Manifest", "HttpPort", "FromURL", });
27168          internal_static_hadoop_hdfs_qjournal_PrepareRecoveryRequestProto_descriptor =
27169            getDescriptor().getMessageTypes().get(38);
27170          internal_static_hadoop_hdfs_qjournal_PrepareRecoveryRequestProto_fieldAccessorTable = new
27171            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27172              internal_static_hadoop_hdfs_qjournal_PrepareRecoveryRequestProto_descriptor,
27173              new java.lang.String[] { "ReqInfo", "SegmentTxId", });
27174          internal_static_hadoop_hdfs_qjournal_PrepareRecoveryResponseProto_descriptor =
27175            getDescriptor().getMessageTypes().get(39);
27176          internal_static_hadoop_hdfs_qjournal_PrepareRecoveryResponseProto_fieldAccessorTable = new
27177            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27178              internal_static_hadoop_hdfs_qjournal_PrepareRecoveryResponseProto_descriptor,
27179              new java.lang.String[] { "SegmentState", "AcceptedInEpoch", "LastWriterEpoch", "LastCommittedTxId", });
27180          internal_static_hadoop_hdfs_qjournal_AcceptRecoveryRequestProto_descriptor =
27181            getDescriptor().getMessageTypes().get(40);
27182          internal_static_hadoop_hdfs_qjournal_AcceptRecoveryRequestProto_fieldAccessorTable = new
27183            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27184              internal_static_hadoop_hdfs_qjournal_AcceptRecoveryRequestProto_descriptor,
27185              new java.lang.String[] { "ReqInfo", "StateToAccept", "FromURL", });
27186          internal_static_hadoop_hdfs_qjournal_AcceptRecoveryResponseProto_descriptor =
27187            getDescriptor().getMessageTypes().get(41);
27188          internal_static_hadoop_hdfs_qjournal_AcceptRecoveryResponseProto_fieldAccessorTable = new
27189            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
27190              internal_static_hadoop_hdfs_qjournal_AcceptRecoveryResponseProto_descriptor,
27191              new java.lang.String[] { });
27192          return null;
27193        }
27194      };
27195    com.google.protobuf.Descriptors.FileDescriptor
27196      .internalBuildGeneratedFileFrom(descriptorData,
27197        new com.google.protobuf.Descriptors.FileDescriptor[] {
27198          org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.getDescriptor(),
27199          org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.getDescriptor(),
27200        }, assigner);
27201  }
27202
27203  // @@protoc_insertion_point(outer_class_scope)
27204}