001// Generated by the protocol buffer compiler. DO NOT EDIT! 002// source: HAServiceProtocol.proto 003 004package org.apache.hadoop.ha.proto; 005 006public final class HAServiceProtocolProtos { 007 private HAServiceProtocolProtos() {} 008 public static void registerAllExtensions( 009 com.google.protobuf.ExtensionRegistry registry) { 010 } 011 /** 012 * Protobuf enum {@code hadoop.common.HAServiceStateProto} 013 */ 014 public enum HAServiceStateProto 015 implements com.google.protobuf.ProtocolMessageEnum { 016 /** 017 * <code>INITIALIZING = 0;</code> 018 */ 019 INITIALIZING(0, 0), 020 /** 021 * <code>ACTIVE = 1;</code> 022 */ 023 ACTIVE(1, 1), 024 /** 025 * <code>STANDBY = 2;</code> 026 */ 027 STANDBY(2, 2), 028 ; 029 030 /** 031 * <code>INITIALIZING = 0;</code> 032 */ 033 public static final int INITIALIZING_VALUE = 0; 034 /** 035 * <code>ACTIVE = 1;</code> 036 */ 037 public static final int ACTIVE_VALUE = 1; 038 /** 039 * <code>STANDBY = 2;</code> 040 */ 041 public static final int STANDBY_VALUE = 2; 042 043 044 public final int getNumber() { return value; } 045 046 public static HAServiceStateProto valueOf(int value) { 047 switch (value) { 048 case 0: return INITIALIZING; 049 case 1: return ACTIVE; 050 case 2: return STANDBY; 051 default: return null; 052 } 053 } 054 055 public static com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto> 056 internalGetValueMap() { 057 return internalValueMap; 058 } 059 private static com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto> 060 internalValueMap = 061 new com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto>() { 062 public HAServiceStateProto findValueByNumber(int number) { 063 return HAServiceStateProto.valueOf(number); 064 } 065 }; 066 067 public final com.google.protobuf.Descriptors.EnumValueDescriptor 068 getValueDescriptor() { 069 return getDescriptor().getValues().get(index); 070 } 071 public final com.google.protobuf.Descriptors.EnumDescriptor 072 getDescriptorForType() { 073 return getDescriptor(); 074 } 075 public static final com.google.protobuf.Descriptors.EnumDescriptor 076 getDescriptor() { 077 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getEnumTypes().get(0); 078 } 079 080 private static final HAServiceStateProto[] VALUES = values(); 081 082 public static HAServiceStateProto valueOf( 083 com.google.protobuf.Descriptors.EnumValueDescriptor desc) { 084 if (desc.getType() != getDescriptor()) { 085 throw new java.lang.IllegalArgumentException( 086 "EnumValueDescriptor is not for this type."); 087 } 088 return VALUES[desc.getIndex()]; 089 } 090 091 private final int index; 092 private final int value; 093 094 private HAServiceStateProto(int index, int value) { 095 this.index = index; 096 this.value = value; 097 } 098 099 // @@protoc_insertion_point(enum_scope:hadoop.common.HAServiceStateProto) 100 } 101 102 /** 103 * Protobuf enum {@code hadoop.common.HARequestSource} 104 */ 105 public enum HARequestSource 106 implements com.google.protobuf.ProtocolMessageEnum { 107 /** 108 * <code>REQUEST_BY_USER = 0;</code> 109 */ 110 REQUEST_BY_USER(0, 0), 111 /** 112 * <code>REQUEST_BY_USER_FORCED = 1;</code> 113 */ 114 REQUEST_BY_USER_FORCED(1, 1), 115 /** 116 * <code>REQUEST_BY_ZKFC = 2;</code> 117 */ 118 REQUEST_BY_ZKFC(2, 2), 119 ; 120 121 /** 122 * <code>REQUEST_BY_USER = 0;</code> 123 */ 124 public static final int REQUEST_BY_USER_VALUE = 0; 125 /** 126 * <code>REQUEST_BY_USER_FORCED = 1;</code> 127 */ 128 public static final int REQUEST_BY_USER_FORCED_VALUE = 1; 129 /** 130 * <code>REQUEST_BY_ZKFC = 2;</code> 131 */ 132 public static final int REQUEST_BY_ZKFC_VALUE = 2; 133 134 135 public final int getNumber() { return value; } 136 137 public static HARequestSource valueOf(int value) { 138 switch (value) { 139 case 0: return REQUEST_BY_USER; 140 case 1: return REQUEST_BY_USER_FORCED; 141 case 2: return REQUEST_BY_ZKFC; 142 default: return null; 143 } 144 } 145 146 public static com.google.protobuf.Internal.EnumLiteMap<HARequestSource> 147 internalGetValueMap() { 148 return internalValueMap; 149 } 150 private static com.google.protobuf.Internal.EnumLiteMap<HARequestSource> 151 internalValueMap = 152 new com.google.protobuf.Internal.EnumLiteMap<HARequestSource>() { 153 public HARequestSource findValueByNumber(int number) { 154 return HARequestSource.valueOf(number); 155 } 156 }; 157 158 public final com.google.protobuf.Descriptors.EnumValueDescriptor 159 getValueDescriptor() { 160 return getDescriptor().getValues().get(index); 161 } 162 public final com.google.protobuf.Descriptors.EnumDescriptor 163 getDescriptorForType() { 164 return getDescriptor(); 165 } 166 public static final com.google.protobuf.Descriptors.EnumDescriptor 167 getDescriptor() { 168 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getEnumTypes().get(1); 169 } 170 171 private static final HARequestSource[] VALUES = values(); 172 173 public static HARequestSource valueOf( 174 com.google.protobuf.Descriptors.EnumValueDescriptor desc) { 175 if (desc.getType() != getDescriptor()) { 176 throw new java.lang.IllegalArgumentException( 177 "EnumValueDescriptor is not for this type."); 178 } 179 return VALUES[desc.getIndex()]; 180 } 181 182 private final int index; 183 private final int value; 184 185 private HARequestSource(int index, int value) { 186 this.index = index; 187 this.value = value; 188 } 189 190 // @@protoc_insertion_point(enum_scope:hadoop.common.HARequestSource) 191 } 192 193 public interface HAStateChangeRequestInfoProtoOrBuilder 194 extends com.google.protobuf.MessageOrBuilder { 195 196 // required .hadoop.common.HARequestSource reqSource = 1; 197 /** 198 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code> 199 */ 200 boolean hasReqSource(); 201 /** 202 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code> 203 */ 204 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource(); 205 } 206 /** 207 * Protobuf type {@code hadoop.common.HAStateChangeRequestInfoProto} 208 */ 209 public static final class HAStateChangeRequestInfoProto extends 210 com.google.protobuf.GeneratedMessage 211 implements HAStateChangeRequestInfoProtoOrBuilder { 212 // Use HAStateChangeRequestInfoProto.newBuilder() to construct. 213 private HAStateChangeRequestInfoProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 214 super(builder); 215 this.unknownFields = builder.getUnknownFields(); 216 } 217 private HAStateChangeRequestInfoProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 218 219 private static final HAStateChangeRequestInfoProto defaultInstance; 220 public static HAStateChangeRequestInfoProto getDefaultInstance() { 221 return defaultInstance; 222 } 223 224 public HAStateChangeRequestInfoProto getDefaultInstanceForType() { 225 return defaultInstance; 226 } 227 228 private final com.google.protobuf.UnknownFieldSet unknownFields; 229 @java.lang.Override 230 public final com.google.protobuf.UnknownFieldSet 231 getUnknownFields() { 232 return this.unknownFields; 233 } 234 private HAStateChangeRequestInfoProto( 235 com.google.protobuf.CodedInputStream input, 236 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 237 throws com.google.protobuf.InvalidProtocolBufferException { 238 initFields(); 239 int mutable_bitField0_ = 0; 240 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 241 com.google.protobuf.UnknownFieldSet.newBuilder(); 242 try { 243 boolean done = false; 244 while (!done) { 245 int tag = input.readTag(); 246 switch (tag) { 247 case 0: 248 done = true; 249 break; 250 default: { 251 if (!parseUnknownField(input, unknownFields, 252 extensionRegistry, tag)) { 253 done = true; 254 } 255 break; 256 } 257 case 8: { 258 int rawValue = input.readEnum(); 259 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource value = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.valueOf(rawValue); 260 if (value == null) { 261 unknownFields.mergeVarintField(1, rawValue); 262 } else { 263 bitField0_ |= 0x00000001; 264 reqSource_ = value; 265 } 266 break; 267 } 268 } 269 } 270 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 271 throw e.setUnfinishedMessage(this); 272 } catch (java.io.IOException e) { 273 throw new com.google.protobuf.InvalidProtocolBufferException( 274 e.getMessage()).setUnfinishedMessage(this); 275 } finally { 276 this.unknownFields = unknownFields.build(); 277 makeExtensionsImmutable(); 278 } 279 } 280 public static final com.google.protobuf.Descriptors.Descriptor 281 getDescriptor() { 282 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor; 283 } 284 285 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 286 internalGetFieldAccessorTable() { 287 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable 288 .ensureFieldAccessorsInitialized( 289 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder.class); 290 } 291 292 public static com.google.protobuf.Parser<HAStateChangeRequestInfoProto> PARSER = 293 new com.google.protobuf.AbstractParser<HAStateChangeRequestInfoProto>() { 294 public HAStateChangeRequestInfoProto parsePartialFrom( 295 com.google.protobuf.CodedInputStream input, 296 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 297 throws com.google.protobuf.InvalidProtocolBufferException { 298 return new HAStateChangeRequestInfoProto(input, extensionRegistry); 299 } 300 }; 301 302 @java.lang.Override 303 public com.google.protobuf.Parser<HAStateChangeRequestInfoProto> getParserForType() { 304 return PARSER; 305 } 306 307 private int bitField0_; 308 // required .hadoop.common.HARequestSource reqSource = 1; 309 public static final int REQSOURCE_FIELD_NUMBER = 1; 310 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource reqSource_; 311 /** 312 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code> 313 */ 314 public boolean hasReqSource() { 315 return ((bitField0_ & 0x00000001) == 0x00000001); 316 } 317 /** 318 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code> 319 */ 320 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource() { 321 return reqSource_; 322 } 323 324 private void initFields() { 325 reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER; 326 } 327 private byte memoizedIsInitialized = -1; 328 public final boolean isInitialized() { 329 byte isInitialized = memoizedIsInitialized; 330 if (isInitialized != -1) return isInitialized == 1; 331 332 if (!hasReqSource()) { 333 memoizedIsInitialized = 0; 334 return false; 335 } 336 memoizedIsInitialized = 1; 337 return true; 338 } 339 340 public void writeTo(com.google.protobuf.CodedOutputStream output) 341 throws java.io.IOException { 342 getSerializedSize(); 343 if (((bitField0_ & 0x00000001) == 0x00000001)) { 344 output.writeEnum(1, reqSource_.getNumber()); 345 } 346 getUnknownFields().writeTo(output); 347 } 348 349 private int memoizedSerializedSize = -1; 350 public int getSerializedSize() { 351 int size = memoizedSerializedSize; 352 if (size != -1) return size; 353 354 size = 0; 355 if (((bitField0_ & 0x00000001) == 0x00000001)) { 356 size += com.google.protobuf.CodedOutputStream 357 .computeEnumSize(1, reqSource_.getNumber()); 358 } 359 size += getUnknownFields().getSerializedSize(); 360 memoizedSerializedSize = size; 361 return size; 362 } 363 364 private static final long serialVersionUID = 0L; 365 @java.lang.Override 366 protected java.lang.Object writeReplace() 367 throws java.io.ObjectStreamException { 368 return super.writeReplace(); 369 } 370 371 @java.lang.Override 372 public boolean equals(final java.lang.Object obj) { 373 if (obj == this) { 374 return true; 375 } 376 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto)) { 377 return super.equals(obj); 378 } 379 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto) obj; 380 381 boolean result = true; 382 result = result && (hasReqSource() == other.hasReqSource()); 383 if (hasReqSource()) { 384 result = result && 385 (getReqSource() == other.getReqSource()); 386 } 387 result = result && 388 getUnknownFields().equals(other.getUnknownFields()); 389 return result; 390 } 391 392 private int memoizedHashCode = 0; 393 @java.lang.Override 394 public int hashCode() { 395 if (memoizedHashCode != 0) { 396 return memoizedHashCode; 397 } 398 int hash = 41; 399 hash = (19 * hash) + getDescriptorForType().hashCode(); 400 if (hasReqSource()) { 401 hash = (37 * hash) + REQSOURCE_FIELD_NUMBER; 402 hash = (53 * hash) + hashEnum(getReqSource()); 403 } 404 hash = (29 * hash) + getUnknownFields().hashCode(); 405 memoizedHashCode = hash; 406 return hash; 407 } 408 409 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom( 410 com.google.protobuf.ByteString data) 411 throws com.google.protobuf.InvalidProtocolBufferException { 412 return PARSER.parseFrom(data); 413 } 414 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom( 415 com.google.protobuf.ByteString data, 416 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 417 throws com.google.protobuf.InvalidProtocolBufferException { 418 return PARSER.parseFrom(data, extensionRegistry); 419 } 420 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(byte[] data) 421 throws com.google.protobuf.InvalidProtocolBufferException { 422 return PARSER.parseFrom(data); 423 } 424 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom( 425 byte[] data, 426 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 427 throws com.google.protobuf.InvalidProtocolBufferException { 428 return PARSER.parseFrom(data, extensionRegistry); 429 } 430 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(java.io.InputStream input) 431 throws java.io.IOException { 432 return PARSER.parseFrom(input); 433 } 434 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom( 435 java.io.InputStream input, 436 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 437 throws java.io.IOException { 438 return PARSER.parseFrom(input, extensionRegistry); 439 } 440 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseDelimitedFrom(java.io.InputStream input) 441 throws java.io.IOException { 442 return PARSER.parseDelimitedFrom(input); 443 } 444 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseDelimitedFrom( 445 java.io.InputStream input, 446 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 447 throws java.io.IOException { 448 return PARSER.parseDelimitedFrom(input, extensionRegistry); 449 } 450 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom( 451 com.google.protobuf.CodedInputStream input) 452 throws java.io.IOException { 453 return PARSER.parseFrom(input); 454 } 455 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom( 456 com.google.protobuf.CodedInputStream input, 457 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 458 throws java.io.IOException { 459 return PARSER.parseFrom(input, extensionRegistry); 460 } 461 462 public static Builder newBuilder() { return Builder.create(); } 463 public Builder newBuilderForType() { return newBuilder(); } 464 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto prototype) { 465 return newBuilder().mergeFrom(prototype); 466 } 467 public Builder toBuilder() { return newBuilder(this); } 468 469 @java.lang.Override 470 protected Builder newBuilderForType( 471 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 472 Builder builder = new Builder(parent); 473 return builder; 474 } 475 /** 476 * Protobuf type {@code hadoop.common.HAStateChangeRequestInfoProto} 477 */ 478 public static final class Builder extends 479 com.google.protobuf.GeneratedMessage.Builder<Builder> 480 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder { 481 public static final com.google.protobuf.Descriptors.Descriptor 482 getDescriptor() { 483 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor; 484 } 485 486 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 487 internalGetFieldAccessorTable() { 488 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable 489 .ensureFieldAccessorsInitialized( 490 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder.class); 491 } 492 493 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder() 494 private Builder() { 495 maybeForceBuilderInitialization(); 496 } 497 498 private Builder( 499 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 500 super(parent); 501 maybeForceBuilderInitialization(); 502 } 503 private void maybeForceBuilderInitialization() { 504 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 505 } 506 } 507 private static Builder create() { 508 return new Builder(); 509 } 510 511 public Builder clear() { 512 super.clear(); 513 reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER; 514 bitField0_ = (bitField0_ & ~0x00000001); 515 return this; 516 } 517 518 public Builder clone() { 519 return create().mergeFrom(buildPartial()); 520 } 521 522 public com.google.protobuf.Descriptors.Descriptor 523 getDescriptorForType() { 524 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor; 525 } 526 527 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getDefaultInstanceForType() { 528 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance(); 529 } 530 531 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto build() { 532 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto result = buildPartial(); 533 if (!result.isInitialized()) { 534 throw newUninitializedMessageException(result); 535 } 536 return result; 537 } 538 539 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto buildPartial() { 540 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto(this); 541 int from_bitField0_ = bitField0_; 542 int to_bitField0_ = 0; 543 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 544 to_bitField0_ |= 0x00000001; 545 } 546 result.reqSource_ = reqSource_; 547 result.bitField0_ = to_bitField0_; 548 onBuilt(); 549 return result; 550 } 551 552 public Builder mergeFrom(com.google.protobuf.Message other) { 553 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto) { 554 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto)other); 555 } else { 556 super.mergeFrom(other); 557 return this; 558 } 559 } 560 561 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto other) { 562 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) return this; 563 if (other.hasReqSource()) { 564 setReqSource(other.getReqSource()); 565 } 566 this.mergeUnknownFields(other.getUnknownFields()); 567 return this; 568 } 569 570 public final boolean isInitialized() { 571 if (!hasReqSource()) { 572 573 return false; 574 } 575 return true; 576 } 577 578 public Builder mergeFrom( 579 com.google.protobuf.CodedInputStream input, 580 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 581 throws java.io.IOException { 582 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parsedMessage = null; 583 try { 584 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 585 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 586 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto) e.getUnfinishedMessage(); 587 throw e; 588 } finally { 589 if (parsedMessage != null) { 590 mergeFrom(parsedMessage); 591 } 592 } 593 return this; 594 } 595 private int bitField0_; 596 597 // required .hadoop.common.HARequestSource reqSource = 1; 598 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER; 599 /** 600 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code> 601 */ 602 public boolean hasReqSource() { 603 return ((bitField0_ & 0x00000001) == 0x00000001); 604 } 605 /** 606 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code> 607 */ 608 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource() { 609 return reqSource_; 610 } 611 /** 612 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code> 613 */ 614 public Builder setReqSource(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource value) { 615 if (value == null) { 616 throw new NullPointerException(); 617 } 618 bitField0_ |= 0x00000001; 619 reqSource_ = value; 620 onChanged(); 621 return this; 622 } 623 /** 624 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code> 625 */ 626 public Builder clearReqSource() { 627 bitField0_ = (bitField0_ & ~0x00000001); 628 reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER; 629 onChanged(); 630 return this; 631 } 632 633 // @@protoc_insertion_point(builder_scope:hadoop.common.HAStateChangeRequestInfoProto) 634 } 635 636 static { 637 defaultInstance = new HAStateChangeRequestInfoProto(true); 638 defaultInstance.initFields(); 639 } 640 641 // @@protoc_insertion_point(class_scope:hadoop.common.HAStateChangeRequestInfoProto) 642 } 643 644 public interface MonitorHealthRequestProtoOrBuilder 645 extends com.google.protobuf.MessageOrBuilder { 646 } 647 /** 648 * Protobuf type {@code hadoop.common.MonitorHealthRequestProto} 649 * 650 * <pre> 651 ** 652 * void request 653 * </pre> 654 */ 655 public static final class MonitorHealthRequestProto extends 656 com.google.protobuf.GeneratedMessage 657 implements MonitorHealthRequestProtoOrBuilder { 658 // Use MonitorHealthRequestProto.newBuilder() to construct. 659 private MonitorHealthRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 660 super(builder); 661 this.unknownFields = builder.getUnknownFields(); 662 } 663 private MonitorHealthRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 664 665 private static final MonitorHealthRequestProto defaultInstance; 666 public static MonitorHealthRequestProto getDefaultInstance() { 667 return defaultInstance; 668 } 669 670 public MonitorHealthRequestProto getDefaultInstanceForType() { 671 return defaultInstance; 672 } 673 674 private final com.google.protobuf.UnknownFieldSet unknownFields; 675 @java.lang.Override 676 public final com.google.protobuf.UnknownFieldSet 677 getUnknownFields() { 678 return this.unknownFields; 679 } 680 private MonitorHealthRequestProto( 681 com.google.protobuf.CodedInputStream input, 682 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 683 throws com.google.protobuf.InvalidProtocolBufferException { 684 initFields(); 685 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 686 com.google.protobuf.UnknownFieldSet.newBuilder(); 687 try { 688 boolean done = false; 689 while (!done) { 690 int tag = input.readTag(); 691 switch (tag) { 692 case 0: 693 done = true; 694 break; 695 default: { 696 if (!parseUnknownField(input, unknownFields, 697 extensionRegistry, tag)) { 698 done = true; 699 } 700 break; 701 } 702 } 703 } 704 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 705 throw e.setUnfinishedMessage(this); 706 } catch (java.io.IOException e) { 707 throw new com.google.protobuf.InvalidProtocolBufferException( 708 e.getMessage()).setUnfinishedMessage(this); 709 } finally { 710 this.unknownFields = unknownFields.build(); 711 makeExtensionsImmutable(); 712 } 713 } 714 public static final com.google.protobuf.Descriptors.Descriptor 715 getDescriptor() { 716 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_descriptor; 717 } 718 719 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 720 internalGetFieldAccessorTable() { 721 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable 722 .ensureFieldAccessorsInitialized( 723 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.Builder.class); 724 } 725 726 public static com.google.protobuf.Parser<MonitorHealthRequestProto> PARSER = 727 new com.google.protobuf.AbstractParser<MonitorHealthRequestProto>() { 728 public MonitorHealthRequestProto parsePartialFrom( 729 com.google.protobuf.CodedInputStream input, 730 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 731 throws com.google.protobuf.InvalidProtocolBufferException { 732 return new MonitorHealthRequestProto(input, extensionRegistry); 733 } 734 }; 735 736 @java.lang.Override 737 public com.google.protobuf.Parser<MonitorHealthRequestProto> getParserForType() { 738 return PARSER; 739 } 740 741 private void initFields() { 742 } 743 private byte memoizedIsInitialized = -1; 744 public final boolean isInitialized() { 745 byte isInitialized = memoizedIsInitialized; 746 if (isInitialized != -1) return isInitialized == 1; 747 748 memoizedIsInitialized = 1; 749 return true; 750 } 751 752 public void writeTo(com.google.protobuf.CodedOutputStream output) 753 throws java.io.IOException { 754 getSerializedSize(); 755 getUnknownFields().writeTo(output); 756 } 757 758 private int memoizedSerializedSize = -1; 759 public int getSerializedSize() { 760 int size = memoizedSerializedSize; 761 if (size != -1) return size; 762 763 size = 0; 764 size += getUnknownFields().getSerializedSize(); 765 memoizedSerializedSize = size; 766 return size; 767 } 768 769 private static final long serialVersionUID = 0L; 770 @java.lang.Override 771 protected java.lang.Object writeReplace() 772 throws java.io.ObjectStreamException { 773 return super.writeReplace(); 774 } 775 776 @java.lang.Override 777 public boolean equals(final java.lang.Object obj) { 778 if (obj == this) { 779 return true; 780 } 781 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)) { 782 return super.equals(obj); 783 } 784 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) obj; 785 786 boolean result = true; 787 result = result && 788 getUnknownFields().equals(other.getUnknownFields()); 789 return result; 790 } 791 792 private int memoizedHashCode = 0; 793 @java.lang.Override 794 public int hashCode() { 795 if (memoizedHashCode != 0) { 796 return memoizedHashCode; 797 } 798 int hash = 41; 799 hash = (19 * hash) + getDescriptorForType().hashCode(); 800 hash = (29 * hash) + getUnknownFields().hashCode(); 801 memoizedHashCode = hash; 802 return hash; 803 } 804 805 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 806 com.google.protobuf.ByteString data) 807 throws com.google.protobuf.InvalidProtocolBufferException { 808 return PARSER.parseFrom(data); 809 } 810 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 811 com.google.protobuf.ByteString data, 812 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 813 throws com.google.protobuf.InvalidProtocolBufferException { 814 return PARSER.parseFrom(data, extensionRegistry); 815 } 816 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(byte[] data) 817 throws com.google.protobuf.InvalidProtocolBufferException { 818 return PARSER.parseFrom(data); 819 } 820 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 821 byte[] data, 822 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 823 throws com.google.protobuf.InvalidProtocolBufferException { 824 return PARSER.parseFrom(data, extensionRegistry); 825 } 826 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(java.io.InputStream input) 827 throws java.io.IOException { 828 return PARSER.parseFrom(input); 829 } 830 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 831 java.io.InputStream input, 832 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 833 throws java.io.IOException { 834 return PARSER.parseFrom(input, extensionRegistry); 835 } 836 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseDelimitedFrom(java.io.InputStream input) 837 throws java.io.IOException { 838 return PARSER.parseDelimitedFrom(input); 839 } 840 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseDelimitedFrom( 841 java.io.InputStream input, 842 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 843 throws java.io.IOException { 844 return PARSER.parseDelimitedFrom(input, extensionRegistry); 845 } 846 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 847 com.google.protobuf.CodedInputStream input) 848 throws java.io.IOException { 849 return PARSER.parseFrom(input); 850 } 851 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 852 com.google.protobuf.CodedInputStream input, 853 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 854 throws java.io.IOException { 855 return PARSER.parseFrom(input, extensionRegistry); 856 } 857 858 public static Builder newBuilder() { return Builder.create(); } 859 public Builder newBuilderForType() { return newBuilder(); } 860 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto prototype) { 861 return newBuilder().mergeFrom(prototype); 862 } 863 public Builder toBuilder() { return newBuilder(this); } 864 865 @java.lang.Override 866 protected Builder newBuilderForType( 867 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 868 Builder builder = new Builder(parent); 869 return builder; 870 } 871 /** 872 * Protobuf type {@code hadoop.common.MonitorHealthRequestProto} 873 * 874 * <pre> 875 ** 876 * void request 877 * </pre> 878 */ 879 public static final class Builder extends 880 com.google.protobuf.GeneratedMessage.Builder<Builder> 881 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProtoOrBuilder { 882 public static final com.google.protobuf.Descriptors.Descriptor 883 getDescriptor() { 884 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_descriptor; 885 } 886 887 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 888 internalGetFieldAccessorTable() { 889 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable 890 .ensureFieldAccessorsInitialized( 891 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.Builder.class); 892 } 893 894 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.newBuilder() 895 private Builder() { 896 maybeForceBuilderInitialization(); 897 } 898 899 private Builder( 900 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 901 super(parent); 902 maybeForceBuilderInitialization(); 903 } 904 private void maybeForceBuilderInitialization() { 905 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 906 } 907 } 908 private static Builder create() { 909 return new Builder(); 910 } 911 912 public Builder clear() { 913 super.clear(); 914 return this; 915 } 916 917 public Builder clone() { 918 return create().mergeFrom(buildPartial()); 919 } 920 921 public com.google.protobuf.Descriptors.Descriptor 922 getDescriptorForType() { 923 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_descriptor; 924 } 925 926 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto getDefaultInstanceForType() { 927 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance(); 928 } 929 930 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto build() { 931 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = buildPartial(); 932 if (!result.isInitialized()) { 933 throw newUninitializedMessageException(result); 934 } 935 return result; 936 } 937 938 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto buildPartial() { 939 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto(this); 940 onBuilt(); 941 return result; 942 } 943 944 public Builder mergeFrom(com.google.protobuf.Message other) { 945 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) { 946 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)other); 947 } else { 948 super.mergeFrom(other); 949 return this; 950 } 951 } 952 953 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto other) { 954 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance()) return this; 955 this.mergeUnknownFields(other.getUnknownFields()); 956 return this; 957 } 958 959 public final boolean isInitialized() { 960 return true; 961 } 962 963 public Builder mergeFrom( 964 com.google.protobuf.CodedInputStream input, 965 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 966 throws java.io.IOException { 967 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parsedMessage = null; 968 try { 969 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 970 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 971 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) e.getUnfinishedMessage(); 972 throw e; 973 } finally { 974 if (parsedMessage != null) { 975 mergeFrom(parsedMessage); 976 } 977 } 978 return this; 979 } 980 981 // @@protoc_insertion_point(builder_scope:hadoop.common.MonitorHealthRequestProto) 982 } 983 984 static { 985 defaultInstance = new MonitorHealthRequestProto(true); 986 defaultInstance.initFields(); 987 } 988 989 // @@protoc_insertion_point(class_scope:hadoop.common.MonitorHealthRequestProto) 990 } 991 992 public interface MonitorHealthResponseProtoOrBuilder 993 extends com.google.protobuf.MessageOrBuilder { 994 } 995 /** 996 * Protobuf type {@code hadoop.common.MonitorHealthResponseProto} 997 * 998 * <pre> 999 ** 1000 * void response 1001 * </pre> 1002 */ 1003 public static final class MonitorHealthResponseProto extends 1004 com.google.protobuf.GeneratedMessage 1005 implements MonitorHealthResponseProtoOrBuilder { 1006 // Use MonitorHealthResponseProto.newBuilder() to construct. 1007 private MonitorHealthResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 1008 super(builder); 1009 this.unknownFields = builder.getUnknownFields(); 1010 } 1011 private MonitorHealthResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 1012 1013 private static final MonitorHealthResponseProto defaultInstance; 1014 public static MonitorHealthResponseProto getDefaultInstance() { 1015 return defaultInstance; 1016 } 1017 1018 public MonitorHealthResponseProto getDefaultInstanceForType() { 1019 return defaultInstance; 1020 } 1021 1022 private final com.google.protobuf.UnknownFieldSet unknownFields; 1023 @java.lang.Override 1024 public final com.google.protobuf.UnknownFieldSet 1025 getUnknownFields() { 1026 return this.unknownFields; 1027 } 1028 private MonitorHealthResponseProto( 1029 com.google.protobuf.CodedInputStream input, 1030 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1031 throws com.google.protobuf.InvalidProtocolBufferException { 1032 initFields(); 1033 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1034 com.google.protobuf.UnknownFieldSet.newBuilder(); 1035 try { 1036 boolean done = false; 1037 while (!done) { 1038 int tag = input.readTag(); 1039 switch (tag) { 1040 case 0: 1041 done = true; 1042 break; 1043 default: { 1044 if (!parseUnknownField(input, unknownFields, 1045 extensionRegistry, tag)) { 1046 done = true; 1047 } 1048 break; 1049 } 1050 } 1051 } 1052 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1053 throw e.setUnfinishedMessage(this); 1054 } catch (java.io.IOException e) { 1055 throw new com.google.protobuf.InvalidProtocolBufferException( 1056 e.getMessage()).setUnfinishedMessage(this); 1057 } finally { 1058 this.unknownFields = unknownFields.build(); 1059 makeExtensionsImmutable(); 1060 } 1061 } 1062 public static final com.google.protobuf.Descriptors.Descriptor 1063 getDescriptor() { 1064 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_descriptor; 1065 } 1066 1067 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1068 internalGetFieldAccessorTable() { 1069 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable 1070 .ensureFieldAccessorsInitialized( 1071 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.Builder.class); 1072 } 1073 1074 public static com.google.protobuf.Parser<MonitorHealthResponseProto> PARSER = 1075 new com.google.protobuf.AbstractParser<MonitorHealthResponseProto>() { 1076 public MonitorHealthResponseProto parsePartialFrom( 1077 com.google.protobuf.CodedInputStream input, 1078 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1079 throws com.google.protobuf.InvalidProtocolBufferException { 1080 return new MonitorHealthResponseProto(input, extensionRegistry); 1081 } 1082 }; 1083 1084 @java.lang.Override 1085 public com.google.protobuf.Parser<MonitorHealthResponseProto> getParserForType() { 1086 return PARSER; 1087 } 1088 1089 private void initFields() { 1090 } 1091 private byte memoizedIsInitialized = -1; 1092 public final boolean isInitialized() { 1093 byte isInitialized = memoizedIsInitialized; 1094 if (isInitialized != -1) return isInitialized == 1; 1095 1096 memoizedIsInitialized = 1; 1097 return true; 1098 } 1099 1100 public void writeTo(com.google.protobuf.CodedOutputStream output) 1101 throws java.io.IOException { 1102 getSerializedSize(); 1103 getUnknownFields().writeTo(output); 1104 } 1105 1106 private int memoizedSerializedSize = -1; 1107 public int getSerializedSize() { 1108 int size = memoizedSerializedSize; 1109 if (size != -1) return size; 1110 1111 size = 0; 1112 size += getUnknownFields().getSerializedSize(); 1113 memoizedSerializedSize = size; 1114 return size; 1115 } 1116 1117 private static final long serialVersionUID = 0L; 1118 @java.lang.Override 1119 protected java.lang.Object writeReplace() 1120 throws java.io.ObjectStreamException { 1121 return super.writeReplace(); 1122 } 1123 1124 @java.lang.Override 1125 public boolean equals(final java.lang.Object obj) { 1126 if (obj == this) { 1127 return true; 1128 } 1129 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto)) { 1130 return super.equals(obj); 1131 } 1132 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) obj; 1133 1134 boolean result = true; 1135 result = result && 1136 getUnknownFields().equals(other.getUnknownFields()); 1137 return result; 1138 } 1139 1140 private int memoizedHashCode = 0; 1141 @java.lang.Override 1142 public int hashCode() { 1143 if (memoizedHashCode != 0) { 1144 return memoizedHashCode; 1145 } 1146 int hash = 41; 1147 hash = (19 * hash) + getDescriptorForType().hashCode(); 1148 hash = (29 * hash) + getUnknownFields().hashCode(); 1149 memoizedHashCode = hash; 1150 return hash; 1151 } 1152 1153 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 1154 com.google.protobuf.ByteString data) 1155 throws com.google.protobuf.InvalidProtocolBufferException { 1156 return PARSER.parseFrom(data); 1157 } 1158 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 1159 com.google.protobuf.ByteString data, 1160 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1161 throws com.google.protobuf.InvalidProtocolBufferException { 1162 return PARSER.parseFrom(data, extensionRegistry); 1163 } 1164 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(byte[] data) 1165 throws com.google.protobuf.InvalidProtocolBufferException { 1166 return PARSER.parseFrom(data); 1167 } 1168 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 1169 byte[] data, 1170 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1171 throws com.google.protobuf.InvalidProtocolBufferException { 1172 return PARSER.parseFrom(data, extensionRegistry); 1173 } 1174 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(java.io.InputStream input) 1175 throws java.io.IOException { 1176 return PARSER.parseFrom(input); 1177 } 1178 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 1179 java.io.InputStream input, 1180 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1181 throws java.io.IOException { 1182 return PARSER.parseFrom(input, extensionRegistry); 1183 } 1184 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseDelimitedFrom(java.io.InputStream input) 1185 throws java.io.IOException { 1186 return PARSER.parseDelimitedFrom(input); 1187 } 1188 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseDelimitedFrom( 1189 java.io.InputStream input, 1190 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1191 throws java.io.IOException { 1192 return PARSER.parseDelimitedFrom(input, extensionRegistry); 1193 } 1194 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 1195 com.google.protobuf.CodedInputStream input) 1196 throws java.io.IOException { 1197 return PARSER.parseFrom(input); 1198 } 1199 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 1200 com.google.protobuf.CodedInputStream input, 1201 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1202 throws java.io.IOException { 1203 return PARSER.parseFrom(input, extensionRegistry); 1204 } 1205 1206 public static Builder newBuilder() { return Builder.create(); } 1207 public Builder newBuilderForType() { return newBuilder(); } 1208 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto prototype) { 1209 return newBuilder().mergeFrom(prototype); 1210 } 1211 public Builder toBuilder() { return newBuilder(this); } 1212 1213 @java.lang.Override 1214 protected Builder newBuilderForType( 1215 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1216 Builder builder = new Builder(parent); 1217 return builder; 1218 } 1219 /** 1220 * Protobuf type {@code hadoop.common.MonitorHealthResponseProto} 1221 * 1222 * <pre> 1223 ** 1224 * void response 1225 * </pre> 1226 */ 1227 public static final class Builder extends 1228 com.google.protobuf.GeneratedMessage.Builder<Builder> 1229 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProtoOrBuilder { 1230 public static final com.google.protobuf.Descriptors.Descriptor 1231 getDescriptor() { 1232 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_descriptor; 1233 } 1234 1235 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1236 internalGetFieldAccessorTable() { 1237 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable 1238 .ensureFieldAccessorsInitialized( 1239 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.Builder.class); 1240 } 1241 1242 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.newBuilder() 1243 private Builder() { 1244 maybeForceBuilderInitialization(); 1245 } 1246 1247 private Builder( 1248 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1249 super(parent); 1250 maybeForceBuilderInitialization(); 1251 } 1252 private void maybeForceBuilderInitialization() { 1253 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1254 } 1255 } 1256 private static Builder create() { 1257 return new Builder(); 1258 } 1259 1260 public Builder clear() { 1261 super.clear(); 1262 return this; 1263 } 1264 1265 public Builder clone() { 1266 return create().mergeFrom(buildPartial()); 1267 } 1268 1269 public com.google.protobuf.Descriptors.Descriptor 1270 getDescriptorForType() { 1271 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_descriptor; 1272 } 1273 1274 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto getDefaultInstanceForType() { 1275 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance(); 1276 } 1277 1278 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto build() { 1279 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = buildPartial(); 1280 if (!result.isInitialized()) { 1281 throw newUninitializedMessageException(result); 1282 } 1283 return result; 1284 } 1285 1286 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto buildPartial() { 1287 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto(this); 1288 onBuilt(); 1289 return result; 1290 } 1291 1292 public Builder mergeFrom(com.google.protobuf.Message other) { 1293 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) { 1294 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto)other); 1295 } else { 1296 super.mergeFrom(other); 1297 return this; 1298 } 1299 } 1300 1301 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto other) { 1302 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance()) return this; 1303 this.mergeUnknownFields(other.getUnknownFields()); 1304 return this; 1305 } 1306 1307 public final boolean isInitialized() { 1308 return true; 1309 } 1310 1311 public Builder mergeFrom( 1312 com.google.protobuf.CodedInputStream input, 1313 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1314 throws java.io.IOException { 1315 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parsedMessage = null; 1316 try { 1317 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 1318 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1319 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) e.getUnfinishedMessage(); 1320 throw e; 1321 } finally { 1322 if (parsedMessage != null) { 1323 mergeFrom(parsedMessage); 1324 } 1325 } 1326 return this; 1327 } 1328 1329 // @@protoc_insertion_point(builder_scope:hadoop.common.MonitorHealthResponseProto) 1330 } 1331 1332 static { 1333 defaultInstance = new MonitorHealthResponseProto(true); 1334 defaultInstance.initFields(); 1335 } 1336 1337 // @@protoc_insertion_point(class_scope:hadoop.common.MonitorHealthResponseProto) 1338 } 1339 1340 public interface TransitionToActiveRequestProtoOrBuilder 1341 extends com.google.protobuf.MessageOrBuilder { 1342 1343 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1; 1344 /** 1345 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1346 */ 1347 boolean hasReqInfo(); 1348 /** 1349 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1350 */ 1351 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo(); 1352 /** 1353 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1354 */ 1355 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder(); 1356 } 1357 /** 1358 * Protobuf type {@code hadoop.common.TransitionToActiveRequestProto} 1359 * 1360 * <pre> 1361 ** 1362 * void request 1363 * </pre> 1364 */ 1365 public static final class TransitionToActiveRequestProto extends 1366 com.google.protobuf.GeneratedMessage 1367 implements TransitionToActiveRequestProtoOrBuilder { 1368 // Use TransitionToActiveRequestProto.newBuilder() to construct. 1369 private TransitionToActiveRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 1370 super(builder); 1371 this.unknownFields = builder.getUnknownFields(); 1372 } 1373 private TransitionToActiveRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 1374 1375 private static final TransitionToActiveRequestProto defaultInstance; 1376 public static TransitionToActiveRequestProto getDefaultInstance() { 1377 return defaultInstance; 1378 } 1379 1380 public TransitionToActiveRequestProto getDefaultInstanceForType() { 1381 return defaultInstance; 1382 } 1383 1384 private final com.google.protobuf.UnknownFieldSet unknownFields; 1385 @java.lang.Override 1386 public final com.google.protobuf.UnknownFieldSet 1387 getUnknownFields() { 1388 return this.unknownFields; 1389 } 1390 private TransitionToActiveRequestProto( 1391 com.google.protobuf.CodedInputStream input, 1392 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1393 throws com.google.protobuf.InvalidProtocolBufferException { 1394 initFields(); 1395 int mutable_bitField0_ = 0; 1396 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1397 com.google.protobuf.UnknownFieldSet.newBuilder(); 1398 try { 1399 boolean done = false; 1400 while (!done) { 1401 int tag = input.readTag(); 1402 switch (tag) { 1403 case 0: 1404 done = true; 1405 break; 1406 default: { 1407 if (!parseUnknownField(input, unknownFields, 1408 extensionRegistry, tag)) { 1409 done = true; 1410 } 1411 break; 1412 } 1413 case 10: { 1414 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder subBuilder = null; 1415 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1416 subBuilder = reqInfo_.toBuilder(); 1417 } 1418 reqInfo_ = input.readMessage(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.PARSER, extensionRegistry); 1419 if (subBuilder != null) { 1420 subBuilder.mergeFrom(reqInfo_); 1421 reqInfo_ = subBuilder.buildPartial(); 1422 } 1423 bitField0_ |= 0x00000001; 1424 break; 1425 } 1426 } 1427 } 1428 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1429 throw e.setUnfinishedMessage(this); 1430 } catch (java.io.IOException e) { 1431 throw new com.google.protobuf.InvalidProtocolBufferException( 1432 e.getMessage()).setUnfinishedMessage(this); 1433 } finally { 1434 this.unknownFields = unknownFields.build(); 1435 makeExtensionsImmutable(); 1436 } 1437 } 1438 public static final com.google.protobuf.Descriptors.Descriptor 1439 getDescriptor() { 1440 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor; 1441 } 1442 1443 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1444 internalGetFieldAccessorTable() { 1445 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable 1446 .ensureFieldAccessorsInitialized( 1447 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.Builder.class); 1448 } 1449 1450 public static com.google.protobuf.Parser<TransitionToActiveRequestProto> PARSER = 1451 new com.google.protobuf.AbstractParser<TransitionToActiveRequestProto>() { 1452 public TransitionToActiveRequestProto parsePartialFrom( 1453 com.google.protobuf.CodedInputStream input, 1454 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1455 throws com.google.protobuf.InvalidProtocolBufferException { 1456 return new TransitionToActiveRequestProto(input, extensionRegistry); 1457 } 1458 }; 1459 1460 @java.lang.Override 1461 public com.google.protobuf.Parser<TransitionToActiveRequestProto> getParserForType() { 1462 return PARSER; 1463 } 1464 1465 private int bitField0_; 1466 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1; 1467 public static final int REQINFO_FIELD_NUMBER = 1; 1468 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_; 1469 /** 1470 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1471 */ 1472 public boolean hasReqInfo() { 1473 return ((bitField0_ & 0x00000001) == 0x00000001); 1474 } 1475 /** 1476 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1477 */ 1478 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() { 1479 return reqInfo_; 1480 } 1481 /** 1482 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1483 */ 1484 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() { 1485 return reqInfo_; 1486 } 1487 1488 private void initFields() { 1489 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance(); 1490 } 1491 private byte memoizedIsInitialized = -1; 1492 public final boolean isInitialized() { 1493 byte isInitialized = memoizedIsInitialized; 1494 if (isInitialized != -1) return isInitialized == 1; 1495 1496 if (!hasReqInfo()) { 1497 memoizedIsInitialized = 0; 1498 return false; 1499 } 1500 if (!getReqInfo().isInitialized()) { 1501 memoizedIsInitialized = 0; 1502 return false; 1503 } 1504 memoizedIsInitialized = 1; 1505 return true; 1506 } 1507 1508 public void writeTo(com.google.protobuf.CodedOutputStream output) 1509 throws java.io.IOException { 1510 getSerializedSize(); 1511 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1512 output.writeMessage(1, reqInfo_); 1513 } 1514 getUnknownFields().writeTo(output); 1515 } 1516 1517 private int memoizedSerializedSize = -1; 1518 public int getSerializedSize() { 1519 int size = memoizedSerializedSize; 1520 if (size != -1) return size; 1521 1522 size = 0; 1523 if (((bitField0_ & 0x00000001) == 0x00000001)) { 1524 size += com.google.protobuf.CodedOutputStream 1525 .computeMessageSize(1, reqInfo_); 1526 } 1527 size += getUnknownFields().getSerializedSize(); 1528 memoizedSerializedSize = size; 1529 return size; 1530 } 1531 1532 private static final long serialVersionUID = 0L; 1533 @java.lang.Override 1534 protected java.lang.Object writeReplace() 1535 throws java.io.ObjectStreamException { 1536 return super.writeReplace(); 1537 } 1538 1539 @java.lang.Override 1540 public boolean equals(final java.lang.Object obj) { 1541 if (obj == this) { 1542 return true; 1543 } 1544 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)) { 1545 return super.equals(obj); 1546 } 1547 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) obj; 1548 1549 boolean result = true; 1550 result = result && (hasReqInfo() == other.hasReqInfo()); 1551 if (hasReqInfo()) { 1552 result = result && getReqInfo() 1553 .equals(other.getReqInfo()); 1554 } 1555 result = result && 1556 getUnknownFields().equals(other.getUnknownFields()); 1557 return result; 1558 } 1559 1560 private int memoizedHashCode = 0; 1561 @java.lang.Override 1562 public int hashCode() { 1563 if (memoizedHashCode != 0) { 1564 return memoizedHashCode; 1565 } 1566 int hash = 41; 1567 hash = (19 * hash) + getDescriptorForType().hashCode(); 1568 if (hasReqInfo()) { 1569 hash = (37 * hash) + REQINFO_FIELD_NUMBER; 1570 hash = (53 * hash) + getReqInfo().hashCode(); 1571 } 1572 hash = (29 * hash) + getUnknownFields().hashCode(); 1573 memoizedHashCode = hash; 1574 return hash; 1575 } 1576 1577 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 1578 com.google.protobuf.ByteString data) 1579 throws com.google.protobuf.InvalidProtocolBufferException { 1580 return PARSER.parseFrom(data); 1581 } 1582 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 1583 com.google.protobuf.ByteString data, 1584 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1585 throws com.google.protobuf.InvalidProtocolBufferException { 1586 return PARSER.parseFrom(data, extensionRegistry); 1587 } 1588 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(byte[] data) 1589 throws com.google.protobuf.InvalidProtocolBufferException { 1590 return PARSER.parseFrom(data); 1591 } 1592 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 1593 byte[] data, 1594 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1595 throws com.google.protobuf.InvalidProtocolBufferException { 1596 return PARSER.parseFrom(data, extensionRegistry); 1597 } 1598 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(java.io.InputStream input) 1599 throws java.io.IOException { 1600 return PARSER.parseFrom(input); 1601 } 1602 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 1603 java.io.InputStream input, 1604 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1605 throws java.io.IOException { 1606 return PARSER.parseFrom(input, extensionRegistry); 1607 } 1608 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseDelimitedFrom(java.io.InputStream input) 1609 throws java.io.IOException { 1610 return PARSER.parseDelimitedFrom(input); 1611 } 1612 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseDelimitedFrom( 1613 java.io.InputStream input, 1614 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1615 throws java.io.IOException { 1616 return PARSER.parseDelimitedFrom(input, extensionRegistry); 1617 } 1618 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 1619 com.google.protobuf.CodedInputStream input) 1620 throws java.io.IOException { 1621 return PARSER.parseFrom(input); 1622 } 1623 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 1624 com.google.protobuf.CodedInputStream input, 1625 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1626 throws java.io.IOException { 1627 return PARSER.parseFrom(input, extensionRegistry); 1628 } 1629 1630 public static Builder newBuilder() { return Builder.create(); } 1631 public Builder newBuilderForType() { return newBuilder(); } 1632 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto prototype) { 1633 return newBuilder().mergeFrom(prototype); 1634 } 1635 public Builder toBuilder() { return newBuilder(this); } 1636 1637 @java.lang.Override 1638 protected Builder newBuilderForType( 1639 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1640 Builder builder = new Builder(parent); 1641 return builder; 1642 } 1643 /** 1644 * Protobuf type {@code hadoop.common.TransitionToActiveRequestProto} 1645 * 1646 * <pre> 1647 ** 1648 * void request 1649 * </pre> 1650 */ 1651 public static final class Builder extends 1652 com.google.protobuf.GeneratedMessage.Builder<Builder> 1653 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProtoOrBuilder { 1654 public static final com.google.protobuf.Descriptors.Descriptor 1655 getDescriptor() { 1656 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor; 1657 } 1658 1659 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1660 internalGetFieldAccessorTable() { 1661 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable 1662 .ensureFieldAccessorsInitialized( 1663 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.Builder.class); 1664 } 1665 1666 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.newBuilder() 1667 private Builder() { 1668 maybeForceBuilderInitialization(); 1669 } 1670 1671 private Builder( 1672 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1673 super(parent); 1674 maybeForceBuilderInitialization(); 1675 } 1676 private void maybeForceBuilderInitialization() { 1677 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1678 getReqInfoFieldBuilder(); 1679 } 1680 } 1681 private static Builder create() { 1682 return new Builder(); 1683 } 1684 1685 public Builder clear() { 1686 super.clear(); 1687 if (reqInfoBuilder_ == null) { 1688 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance(); 1689 } else { 1690 reqInfoBuilder_.clear(); 1691 } 1692 bitField0_ = (bitField0_ & ~0x00000001); 1693 return this; 1694 } 1695 1696 public Builder clone() { 1697 return create().mergeFrom(buildPartial()); 1698 } 1699 1700 public com.google.protobuf.Descriptors.Descriptor 1701 getDescriptorForType() { 1702 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor; 1703 } 1704 1705 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto getDefaultInstanceForType() { 1706 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance(); 1707 } 1708 1709 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto build() { 1710 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = buildPartial(); 1711 if (!result.isInitialized()) { 1712 throw newUninitializedMessageException(result); 1713 } 1714 return result; 1715 } 1716 1717 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto buildPartial() { 1718 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto(this); 1719 int from_bitField0_ = bitField0_; 1720 int to_bitField0_ = 0; 1721 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 1722 to_bitField0_ |= 0x00000001; 1723 } 1724 if (reqInfoBuilder_ == null) { 1725 result.reqInfo_ = reqInfo_; 1726 } else { 1727 result.reqInfo_ = reqInfoBuilder_.build(); 1728 } 1729 result.bitField0_ = to_bitField0_; 1730 onBuilt(); 1731 return result; 1732 } 1733 1734 public Builder mergeFrom(com.google.protobuf.Message other) { 1735 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) { 1736 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)other); 1737 } else { 1738 super.mergeFrom(other); 1739 return this; 1740 } 1741 } 1742 1743 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto other) { 1744 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance()) return this; 1745 if (other.hasReqInfo()) { 1746 mergeReqInfo(other.getReqInfo()); 1747 } 1748 this.mergeUnknownFields(other.getUnknownFields()); 1749 return this; 1750 } 1751 1752 public final boolean isInitialized() { 1753 if (!hasReqInfo()) { 1754 1755 return false; 1756 } 1757 if (!getReqInfo().isInitialized()) { 1758 1759 return false; 1760 } 1761 return true; 1762 } 1763 1764 public Builder mergeFrom( 1765 com.google.protobuf.CodedInputStream input, 1766 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1767 throws java.io.IOException { 1768 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parsedMessage = null; 1769 try { 1770 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 1771 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1772 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) e.getUnfinishedMessage(); 1773 throw e; 1774 } finally { 1775 if (parsedMessage != null) { 1776 mergeFrom(parsedMessage); 1777 } 1778 } 1779 return this; 1780 } 1781 private int bitField0_; 1782 1783 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1; 1784 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance(); 1785 private com.google.protobuf.SingleFieldBuilder< 1786 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> reqInfoBuilder_; 1787 /** 1788 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1789 */ 1790 public boolean hasReqInfo() { 1791 return ((bitField0_ & 0x00000001) == 0x00000001); 1792 } 1793 /** 1794 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1795 */ 1796 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() { 1797 if (reqInfoBuilder_ == null) { 1798 return reqInfo_; 1799 } else { 1800 return reqInfoBuilder_.getMessage(); 1801 } 1802 } 1803 /** 1804 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1805 */ 1806 public Builder setReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) { 1807 if (reqInfoBuilder_ == null) { 1808 if (value == null) { 1809 throw new NullPointerException(); 1810 } 1811 reqInfo_ = value; 1812 onChanged(); 1813 } else { 1814 reqInfoBuilder_.setMessage(value); 1815 } 1816 bitField0_ |= 0x00000001; 1817 return this; 1818 } 1819 /** 1820 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1821 */ 1822 public Builder setReqInfo( 1823 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder builderForValue) { 1824 if (reqInfoBuilder_ == null) { 1825 reqInfo_ = builderForValue.build(); 1826 onChanged(); 1827 } else { 1828 reqInfoBuilder_.setMessage(builderForValue.build()); 1829 } 1830 bitField0_ |= 0x00000001; 1831 return this; 1832 } 1833 /** 1834 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1835 */ 1836 public Builder mergeReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) { 1837 if (reqInfoBuilder_ == null) { 1838 if (((bitField0_ & 0x00000001) == 0x00000001) && 1839 reqInfo_ != org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) { 1840 reqInfo_ = 1841 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial(); 1842 } else { 1843 reqInfo_ = value; 1844 } 1845 onChanged(); 1846 } else { 1847 reqInfoBuilder_.mergeFrom(value); 1848 } 1849 bitField0_ |= 0x00000001; 1850 return this; 1851 } 1852 /** 1853 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1854 */ 1855 public Builder clearReqInfo() { 1856 if (reqInfoBuilder_ == null) { 1857 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance(); 1858 onChanged(); 1859 } else { 1860 reqInfoBuilder_.clear(); 1861 } 1862 bitField0_ = (bitField0_ & ~0x00000001); 1863 return this; 1864 } 1865 /** 1866 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1867 */ 1868 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder getReqInfoBuilder() { 1869 bitField0_ |= 0x00000001; 1870 onChanged(); 1871 return getReqInfoFieldBuilder().getBuilder(); 1872 } 1873 /** 1874 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1875 */ 1876 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() { 1877 if (reqInfoBuilder_ != null) { 1878 return reqInfoBuilder_.getMessageOrBuilder(); 1879 } else { 1880 return reqInfo_; 1881 } 1882 } 1883 /** 1884 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 1885 */ 1886 private com.google.protobuf.SingleFieldBuilder< 1887 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> 1888 getReqInfoFieldBuilder() { 1889 if (reqInfoBuilder_ == null) { 1890 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< 1891 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>( 1892 reqInfo_, 1893 getParentForChildren(), 1894 isClean()); 1895 reqInfo_ = null; 1896 } 1897 return reqInfoBuilder_; 1898 } 1899 1900 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToActiveRequestProto) 1901 } 1902 1903 static { 1904 defaultInstance = new TransitionToActiveRequestProto(true); 1905 defaultInstance.initFields(); 1906 } 1907 1908 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToActiveRequestProto) 1909 } 1910 1911 public interface TransitionToActiveResponseProtoOrBuilder 1912 extends com.google.protobuf.MessageOrBuilder { 1913 } 1914 /** 1915 * Protobuf type {@code hadoop.common.TransitionToActiveResponseProto} 1916 * 1917 * <pre> 1918 ** 1919 * void response 1920 * </pre> 1921 */ 1922 public static final class TransitionToActiveResponseProto extends 1923 com.google.protobuf.GeneratedMessage 1924 implements TransitionToActiveResponseProtoOrBuilder { 1925 // Use TransitionToActiveResponseProto.newBuilder() to construct. 1926 private TransitionToActiveResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 1927 super(builder); 1928 this.unknownFields = builder.getUnknownFields(); 1929 } 1930 private TransitionToActiveResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 1931 1932 private static final TransitionToActiveResponseProto defaultInstance; 1933 public static TransitionToActiveResponseProto getDefaultInstance() { 1934 return defaultInstance; 1935 } 1936 1937 public TransitionToActiveResponseProto getDefaultInstanceForType() { 1938 return defaultInstance; 1939 } 1940 1941 private final com.google.protobuf.UnknownFieldSet unknownFields; 1942 @java.lang.Override 1943 public final com.google.protobuf.UnknownFieldSet 1944 getUnknownFields() { 1945 return this.unknownFields; 1946 } 1947 private TransitionToActiveResponseProto( 1948 com.google.protobuf.CodedInputStream input, 1949 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1950 throws com.google.protobuf.InvalidProtocolBufferException { 1951 initFields(); 1952 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1953 com.google.protobuf.UnknownFieldSet.newBuilder(); 1954 try { 1955 boolean done = false; 1956 while (!done) { 1957 int tag = input.readTag(); 1958 switch (tag) { 1959 case 0: 1960 done = true; 1961 break; 1962 default: { 1963 if (!parseUnknownField(input, unknownFields, 1964 extensionRegistry, tag)) { 1965 done = true; 1966 } 1967 break; 1968 } 1969 } 1970 } 1971 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1972 throw e.setUnfinishedMessage(this); 1973 } catch (java.io.IOException e) { 1974 throw new com.google.protobuf.InvalidProtocolBufferException( 1975 e.getMessage()).setUnfinishedMessage(this); 1976 } finally { 1977 this.unknownFields = unknownFields.build(); 1978 makeExtensionsImmutable(); 1979 } 1980 } 1981 public static final com.google.protobuf.Descriptors.Descriptor 1982 getDescriptor() { 1983 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor; 1984 } 1985 1986 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1987 internalGetFieldAccessorTable() { 1988 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable 1989 .ensureFieldAccessorsInitialized( 1990 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.Builder.class); 1991 } 1992 1993 public static com.google.protobuf.Parser<TransitionToActiveResponseProto> PARSER = 1994 new com.google.protobuf.AbstractParser<TransitionToActiveResponseProto>() { 1995 public TransitionToActiveResponseProto parsePartialFrom( 1996 com.google.protobuf.CodedInputStream input, 1997 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1998 throws com.google.protobuf.InvalidProtocolBufferException { 1999 return new TransitionToActiveResponseProto(input, extensionRegistry); 2000 } 2001 }; 2002 2003 @java.lang.Override 2004 public com.google.protobuf.Parser<TransitionToActiveResponseProto> getParserForType() { 2005 return PARSER; 2006 } 2007 2008 private void initFields() { 2009 } 2010 private byte memoizedIsInitialized = -1; 2011 public final boolean isInitialized() { 2012 byte isInitialized = memoizedIsInitialized; 2013 if (isInitialized != -1) return isInitialized == 1; 2014 2015 memoizedIsInitialized = 1; 2016 return true; 2017 } 2018 2019 public void writeTo(com.google.protobuf.CodedOutputStream output) 2020 throws java.io.IOException { 2021 getSerializedSize(); 2022 getUnknownFields().writeTo(output); 2023 } 2024 2025 private int memoizedSerializedSize = -1; 2026 public int getSerializedSize() { 2027 int size = memoizedSerializedSize; 2028 if (size != -1) return size; 2029 2030 size = 0; 2031 size += getUnknownFields().getSerializedSize(); 2032 memoizedSerializedSize = size; 2033 return size; 2034 } 2035 2036 private static final long serialVersionUID = 0L; 2037 @java.lang.Override 2038 protected java.lang.Object writeReplace() 2039 throws java.io.ObjectStreamException { 2040 return super.writeReplace(); 2041 } 2042 2043 @java.lang.Override 2044 public boolean equals(final java.lang.Object obj) { 2045 if (obj == this) { 2046 return true; 2047 } 2048 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto)) { 2049 return super.equals(obj); 2050 } 2051 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) obj; 2052 2053 boolean result = true; 2054 result = result && 2055 getUnknownFields().equals(other.getUnknownFields()); 2056 return result; 2057 } 2058 2059 private int memoizedHashCode = 0; 2060 @java.lang.Override 2061 public int hashCode() { 2062 if (memoizedHashCode != 0) { 2063 return memoizedHashCode; 2064 } 2065 int hash = 41; 2066 hash = (19 * hash) + getDescriptorForType().hashCode(); 2067 hash = (29 * hash) + getUnknownFields().hashCode(); 2068 memoizedHashCode = hash; 2069 return hash; 2070 } 2071 2072 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 2073 com.google.protobuf.ByteString data) 2074 throws com.google.protobuf.InvalidProtocolBufferException { 2075 return PARSER.parseFrom(data); 2076 } 2077 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 2078 com.google.protobuf.ByteString data, 2079 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2080 throws com.google.protobuf.InvalidProtocolBufferException { 2081 return PARSER.parseFrom(data, extensionRegistry); 2082 } 2083 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(byte[] data) 2084 throws com.google.protobuf.InvalidProtocolBufferException { 2085 return PARSER.parseFrom(data); 2086 } 2087 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 2088 byte[] data, 2089 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2090 throws com.google.protobuf.InvalidProtocolBufferException { 2091 return PARSER.parseFrom(data, extensionRegistry); 2092 } 2093 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(java.io.InputStream input) 2094 throws java.io.IOException { 2095 return PARSER.parseFrom(input); 2096 } 2097 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 2098 java.io.InputStream input, 2099 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2100 throws java.io.IOException { 2101 return PARSER.parseFrom(input, extensionRegistry); 2102 } 2103 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseDelimitedFrom(java.io.InputStream input) 2104 throws java.io.IOException { 2105 return PARSER.parseDelimitedFrom(input); 2106 } 2107 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseDelimitedFrom( 2108 java.io.InputStream input, 2109 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2110 throws java.io.IOException { 2111 return PARSER.parseDelimitedFrom(input, extensionRegistry); 2112 } 2113 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 2114 com.google.protobuf.CodedInputStream input) 2115 throws java.io.IOException { 2116 return PARSER.parseFrom(input); 2117 } 2118 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 2119 com.google.protobuf.CodedInputStream input, 2120 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2121 throws java.io.IOException { 2122 return PARSER.parseFrom(input, extensionRegistry); 2123 } 2124 2125 public static Builder newBuilder() { return Builder.create(); } 2126 public Builder newBuilderForType() { return newBuilder(); } 2127 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto prototype) { 2128 return newBuilder().mergeFrom(prototype); 2129 } 2130 public Builder toBuilder() { return newBuilder(this); } 2131 2132 @java.lang.Override 2133 protected Builder newBuilderForType( 2134 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2135 Builder builder = new Builder(parent); 2136 return builder; 2137 } 2138 /** 2139 * Protobuf type {@code hadoop.common.TransitionToActiveResponseProto} 2140 * 2141 * <pre> 2142 ** 2143 * void response 2144 * </pre> 2145 */ 2146 public static final class Builder extends 2147 com.google.protobuf.GeneratedMessage.Builder<Builder> 2148 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProtoOrBuilder { 2149 public static final com.google.protobuf.Descriptors.Descriptor 2150 getDescriptor() { 2151 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor; 2152 } 2153 2154 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2155 internalGetFieldAccessorTable() { 2156 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable 2157 .ensureFieldAccessorsInitialized( 2158 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.Builder.class); 2159 } 2160 2161 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.newBuilder() 2162 private Builder() { 2163 maybeForceBuilderInitialization(); 2164 } 2165 2166 private Builder( 2167 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2168 super(parent); 2169 maybeForceBuilderInitialization(); 2170 } 2171 private void maybeForceBuilderInitialization() { 2172 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 2173 } 2174 } 2175 private static Builder create() { 2176 return new Builder(); 2177 } 2178 2179 public Builder clear() { 2180 super.clear(); 2181 return this; 2182 } 2183 2184 public Builder clone() { 2185 return create().mergeFrom(buildPartial()); 2186 } 2187 2188 public com.google.protobuf.Descriptors.Descriptor 2189 getDescriptorForType() { 2190 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor; 2191 } 2192 2193 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto getDefaultInstanceForType() { 2194 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance(); 2195 } 2196 2197 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto build() { 2198 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = buildPartial(); 2199 if (!result.isInitialized()) { 2200 throw newUninitializedMessageException(result); 2201 } 2202 return result; 2203 } 2204 2205 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto buildPartial() { 2206 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto(this); 2207 onBuilt(); 2208 return result; 2209 } 2210 2211 public Builder mergeFrom(com.google.protobuf.Message other) { 2212 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) { 2213 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto)other); 2214 } else { 2215 super.mergeFrom(other); 2216 return this; 2217 } 2218 } 2219 2220 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto other) { 2221 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance()) return this; 2222 this.mergeUnknownFields(other.getUnknownFields()); 2223 return this; 2224 } 2225 2226 public final boolean isInitialized() { 2227 return true; 2228 } 2229 2230 public Builder mergeFrom( 2231 com.google.protobuf.CodedInputStream input, 2232 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2233 throws java.io.IOException { 2234 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parsedMessage = null; 2235 try { 2236 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 2237 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2238 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) e.getUnfinishedMessage(); 2239 throw e; 2240 } finally { 2241 if (parsedMessage != null) { 2242 mergeFrom(parsedMessage); 2243 } 2244 } 2245 return this; 2246 } 2247 2248 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToActiveResponseProto) 2249 } 2250 2251 static { 2252 defaultInstance = new TransitionToActiveResponseProto(true); 2253 defaultInstance.initFields(); 2254 } 2255 2256 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToActiveResponseProto) 2257 } 2258 2259 public interface TransitionToStandbyRequestProtoOrBuilder 2260 extends com.google.protobuf.MessageOrBuilder { 2261 2262 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1; 2263 /** 2264 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2265 */ 2266 boolean hasReqInfo(); 2267 /** 2268 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2269 */ 2270 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo(); 2271 /** 2272 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2273 */ 2274 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder(); 2275 } 2276 /** 2277 * Protobuf type {@code hadoop.common.TransitionToStandbyRequestProto} 2278 * 2279 * <pre> 2280 ** 2281 * void request 2282 * </pre> 2283 */ 2284 public static final class TransitionToStandbyRequestProto extends 2285 com.google.protobuf.GeneratedMessage 2286 implements TransitionToStandbyRequestProtoOrBuilder { 2287 // Use TransitionToStandbyRequestProto.newBuilder() to construct. 2288 private TransitionToStandbyRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 2289 super(builder); 2290 this.unknownFields = builder.getUnknownFields(); 2291 } 2292 private TransitionToStandbyRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 2293 2294 private static final TransitionToStandbyRequestProto defaultInstance; 2295 public static TransitionToStandbyRequestProto getDefaultInstance() { 2296 return defaultInstance; 2297 } 2298 2299 public TransitionToStandbyRequestProto getDefaultInstanceForType() { 2300 return defaultInstance; 2301 } 2302 2303 private final com.google.protobuf.UnknownFieldSet unknownFields; 2304 @java.lang.Override 2305 public final com.google.protobuf.UnknownFieldSet 2306 getUnknownFields() { 2307 return this.unknownFields; 2308 } 2309 private TransitionToStandbyRequestProto( 2310 com.google.protobuf.CodedInputStream input, 2311 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2312 throws com.google.protobuf.InvalidProtocolBufferException { 2313 initFields(); 2314 int mutable_bitField0_ = 0; 2315 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2316 com.google.protobuf.UnknownFieldSet.newBuilder(); 2317 try { 2318 boolean done = false; 2319 while (!done) { 2320 int tag = input.readTag(); 2321 switch (tag) { 2322 case 0: 2323 done = true; 2324 break; 2325 default: { 2326 if (!parseUnknownField(input, unknownFields, 2327 extensionRegistry, tag)) { 2328 done = true; 2329 } 2330 break; 2331 } 2332 case 10: { 2333 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder subBuilder = null; 2334 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2335 subBuilder = reqInfo_.toBuilder(); 2336 } 2337 reqInfo_ = input.readMessage(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.PARSER, extensionRegistry); 2338 if (subBuilder != null) { 2339 subBuilder.mergeFrom(reqInfo_); 2340 reqInfo_ = subBuilder.buildPartial(); 2341 } 2342 bitField0_ |= 0x00000001; 2343 break; 2344 } 2345 } 2346 } 2347 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2348 throw e.setUnfinishedMessage(this); 2349 } catch (java.io.IOException e) { 2350 throw new com.google.protobuf.InvalidProtocolBufferException( 2351 e.getMessage()).setUnfinishedMessage(this); 2352 } finally { 2353 this.unknownFields = unknownFields.build(); 2354 makeExtensionsImmutable(); 2355 } 2356 } 2357 public static final com.google.protobuf.Descriptors.Descriptor 2358 getDescriptor() { 2359 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor; 2360 } 2361 2362 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2363 internalGetFieldAccessorTable() { 2364 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable 2365 .ensureFieldAccessorsInitialized( 2366 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.Builder.class); 2367 } 2368 2369 public static com.google.protobuf.Parser<TransitionToStandbyRequestProto> PARSER = 2370 new com.google.protobuf.AbstractParser<TransitionToStandbyRequestProto>() { 2371 public TransitionToStandbyRequestProto parsePartialFrom( 2372 com.google.protobuf.CodedInputStream input, 2373 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2374 throws com.google.protobuf.InvalidProtocolBufferException { 2375 return new TransitionToStandbyRequestProto(input, extensionRegistry); 2376 } 2377 }; 2378 2379 @java.lang.Override 2380 public com.google.protobuf.Parser<TransitionToStandbyRequestProto> getParserForType() { 2381 return PARSER; 2382 } 2383 2384 private int bitField0_; 2385 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1; 2386 public static final int REQINFO_FIELD_NUMBER = 1; 2387 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_; 2388 /** 2389 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2390 */ 2391 public boolean hasReqInfo() { 2392 return ((bitField0_ & 0x00000001) == 0x00000001); 2393 } 2394 /** 2395 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2396 */ 2397 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() { 2398 return reqInfo_; 2399 } 2400 /** 2401 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2402 */ 2403 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() { 2404 return reqInfo_; 2405 } 2406 2407 private void initFields() { 2408 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance(); 2409 } 2410 private byte memoizedIsInitialized = -1; 2411 public final boolean isInitialized() { 2412 byte isInitialized = memoizedIsInitialized; 2413 if (isInitialized != -1) return isInitialized == 1; 2414 2415 if (!hasReqInfo()) { 2416 memoizedIsInitialized = 0; 2417 return false; 2418 } 2419 if (!getReqInfo().isInitialized()) { 2420 memoizedIsInitialized = 0; 2421 return false; 2422 } 2423 memoizedIsInitialized = 1; 2424 return true; 2425 } 2426 2427 public void writeTo(com.google.protobuf.CodedOutputStream output) 2428 throws java.io.IOException { 2429 getSerializedSize(); 2430 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2431 output.writeMessage(1, reqInfo_); 2432 } 2433 getUnknownFields().writeTo(output); 2434 } 2435 2436 private int memoizedSerializedSize = -1; 2437 public int getSerializedSize() { 2438 int size = memoizedSerializedSize; 2439 if (size != -1) return size; 2440 2441 size = 0; 2442 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2443 size += com.google.protobuf.CodedOutputStream 2444 .computeMessageSize(1, reqInfo_); 2445 } 2446 size += getUnknownFields().getSerializedSize(); 2447 memoizedSerializedSize = size; 2448 return size; 2449 } 2450 2451 private static final long serialVersionUID = 0L; 2452 @java.lang.Override 2453 protected java.lang.Object writeReplace() 2454 throws java.io.ObjectStreamException { 2455 return super.writeReplace(); 2456 } 2457 2458 @java.lang.Override 2459 public boolean equals(final java.lang.Object obj) { 2460 if (obj == this) { 2461 return true; 2462 } 2463 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)) { 2464 return super.equals(obj); 2465 } 2466 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) obj; 2467 2468 boolean result = true; 2469 result = result && (hasReqInfo() == other.hasReqInfo()); 2470 if (hasReqInfo()) { 2471 result = result && getReqInfo() 2472 .equals(other.getReqInfo()); 2473 } 2474 result = result && 2475 getUnknownFields().equals(other.getUnknownFields()); 2476 return result; 2477 } 2478 2479 private int memoizedHashCode = 0; 2480 @java.lang.Override 2481 public int hashCode() { 2482 if (memoizedHashCode != 0) { 2483 return memoizedHashCode; 2484 } 2485 int hash = 41; 2486 hash = (19 * hash) + getDescriptorForType().hashCode(); 2487 if (hasReqInfo()) { 2488 hash = (37 * hash) + REQINFO_FIELD_NUMBER; 2489 hash = (53 * hash) + getReqInfo().hashCode(); 2490 } 2491 hash = (29 * hash) + getUnknownFields().hashCode(); 2492 memoizedHashCode = hash; 2493 return hash; 2494 } 2495 2496 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 2497 com.google.protobuf.ByteString data) 2498 throws com.google.protobuf.InvalidProtocolBufferException { 2499 return PARSER.parseFrom(data); 2500 } 2501 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 2502 com.google.protobuf.ByteString data, 2503 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2504 throws com.google.protobuf.InvalidProtocolBufferException { 2505 return PARSER.parseFrom(data, extensionRegistry); 2506 } 2507 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(byte[] data) 2508 throws com.google.protobuf.InvalidProtocolBufferException { 2509 return PARSER.parseFrom(data); 2510 } 2511 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 2512 byte[] data, 2513 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2514 throws com.google.protobuf.InvalidProtocolBufferException { 2515 return PARSER.parseFrom(data, extensionRegistry); 2516 } 2517 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(java.io.InputStream input) 2518 throws java.io.IOException { 2519 return PARSER.parseFrom(input); 2520 } 2521 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 2522 java.io.InputStream input, 2523 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2524 throws java.io.IOException { 2525 return PARSER.parseFrom(input, extensionRegistry); 2526 } 2527 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseDelimitedFrom(java.io.InputStream input) 2528 throws java.io.IOException { 2529 return PARSER.parseDelimitedFrom(input); 2530 } 2531 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseDelimitedFrom( 2532 java.io.InputStream input, 2533 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2534 throws java.io.IOException { 2535 return PARSER.parseDelimitedFrom(input, extensionRegistry); 2536 } 2537 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 2538 com.google.protobuf.CodedInputStream input) 2539 throws java.io.IOException { 2540 return PARSER.parseFrom(input); 2541 } 2542 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 2543 com.google.protobuf.CodedInputStream input, 2544 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2545 throws java.io.IOException { 2546 return PARSER.parseFrom(input, extensionRegistry); 2547 } 2548 2549 public static Builder newBuilder() { return Builder.create(); } 2550 public Builder newBuilderForType() { return newBuilder(); } 2551 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto prototype) { 2552 return newBuilder().mergeFrom(prototype); 2553 } 2554 public Builder toBuilder() { return newBuilder(this); } 2555 2556 @java.lang.Override 2557 protected Builder newBuilderForType( 2558 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2559 Builder builder = new Builder(parent); 2560 return builder; 2561 } 2562 /** 2563 * Protobuf type {@code hadoop.common.TransitionToStandbyRequestProto} 2564 * 2565 * <pre> 2566 ** 2567 * void request 2568 * </pre> 2569 */ 2570 public static final class Builder extends 2571 com.google.protobuf.GeneratedMessage.Builder<Builder> 2572 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProtoOrBuilder { 2573 public static final com.google.protobuf.Descriptors.Descriptor 2574 getDescriptor() { 2575 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor; 2576 } 2577 2578 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2579 internalGetFieldAccessorTable() { 2580 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable 2581 .ensureFieldAccessorsInitialized( 2582 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.Builder.class); 2583 } 2584 2585 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.newBuilder() 2586 private Builder() { 2587 maybeForceBuilderInitialization(); 2588 } 2589 2590 private Builder( 2591 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2592 super(parent); 2593 maybeForceBuilderInitialization(); 2594 } 2595 private void maybeForceBuilderInitialization() { 2596 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 2597 getReqInfoFieldBuilder(); 2598 } 2599 } 2600 private static Builder create() { 2601 return new Builder(); 2602 } 2603 2604 public Builder clear() { 2605 super.clear(); 2606 if (reqInfoBuilder_ == null) { 2607 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance(); 2608 } else { 2609 reqInfoBuilder_.clear(); 2610 } 2611 bitField0_ = (bitField0_ & ~0x00000001); 2612 return this; 2613 } 2614 2615 public Builder clone() { 2616 return create().mergeFrom(buildPartial()); 2617 } 2618 2619 public com.google.protobuf.Descriptors.Descriptor 2620 getDescriptorForType() { 2621 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor; 2622 } 2623 2624 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto getDefaultInstanceForType() { 2625 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance(); 2626 } 2627 2628 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto build() { 2629 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = buildPartial(); 2630 if (!result.isInitialized()) { 2631 throw newUninitializedMessageException(result); 2632 } 2633 return result; 2634 } 2635 2636 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto buildPartial() { 2637 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto(this); 2638 int from_bitField0_ = bitField0_; 2639 int to_bitField0_ = 0; 2640 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 2641 to_bitField0_ |= 0x00000001; 2642 } 2643 if (reqInfoBuilder_ == null) { 2644 result.reqInfo_ = reqInfo_; 2645 } else { 2646 result.reqInfo_ = reqInfoBuilder_.build(); 2647 } 2648 result.bitField0_ = to_bitField0_; 2649 onBuilt(); 2650 return result; 2651 } 2652 2653 public Builder mergeFrom(com.google.protobuf.Message other) { 2654 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) { 2655 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)other); 2656 } else { 2657 super.mergeFrom(other); 2658 return this; 2659 } 2660 } 2661 2662 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto other) { 2663 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance()) return this; 2664 if (other.hasReqInfo()) { 2665 mergeReqInfo(other.getReqInfo()); 2666 } 2667 this.mergeUnknownFields(other.getUnknownFields()); 2668 return this; 2669 } 2670 2671 public final boolean isInitialized() { 2672 if (!hasReqInfo()) { 2673 2674 return false; 2675 } 2676 if (!getReqInfo().isInitialized()) { 2677 2678 return false; 2679 } 2680 return true; 2681 } 2682 2683 public Builder mergeFrom( 2684 com.google.protobuf.CodedInputStream input, 2685 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2686 throws java.io.IOException { 2687 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parsedMessage = null; 2688 try { 2689 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 2690 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2691 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) e.getUnfinishedMessage(); 2692 throw e; 2693 } finally { 2694 if (parsedMessage != null) { 2695 mergeFrom(parsedMessage); 2696 } 2697 } 2698 return this; 2699 } 2700 private int bitField0_; 2701 2702 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1; 2703 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance(); 2704 private com.google.protobuf.SingleFieldBuilder< 2705 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> reqInfoBuilder_; 2706 /** 2707 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2708 */ 2709 public boolean hasReqInfo() { 2710 return ((bitField0_ & 0x00000001) == 0x00000001); 2711 } 2712 /** 2713 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2714 */ 2715 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() { 2716 if (reqInfoBuilder_ == null) { 2717 return reqInfo_; 2718 } else { 2719 return reqInfoBuilder_.getMessage(); 2720 } 2721 } 2722 /** 2723 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2724 */ 2725 public Builder setReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) { 2726 if (reqInfoBuilder_ == null) { 2727 if (value == null) { 2728 throw new NullPointerException(); 2729 } 2730 reqInfo_ = value; 2731 onChanged(); 2732 } else { 2733 reqInfoBuilder_.setMessage(value); 2734 } 2735 bitField0_ |= 0x00000001; 2736 return this; 2737 } 2738 /** 2739 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2740 */ 2741 public Builder setReqInfo( 2742 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder builderForValue) { 2743 if (reqInfoBuilder_ == null) { 2744 reqInfo_ = builderForValue.build(); 2745 onChanged(); 2746 } else { 2747 reqInfoBuilder_.setMessage(builderForValue.build()); 2748 } 2749 bitField0_ |= 0x00000001; 2750 return this; 2751 } 2752 /** 2753 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2754 */ 2755 public Builder mergeReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) { 2756 if (reqInfoBuilder_ == null) { 2757 if (((bitField0_ & 0x00000001) == 0x00000001) && 2758 reqInfo_ != org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) { 2759 reqInfo_ = 2760 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial(); 2761 } else { 2762 reqInfo_ = value; 2763 } 2764 onChanged(); 2765 } else { 2766 reqInfoBuilder_.mergeFrom(value); 2767 } 2768 bitField0_ |= 0x00000001; 2769 return this; 2770 } 2771 /** 2772 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2773 */ 2774 public Builder clearReqInfo() { 2775 if (reqInfoBuilder_ == null) { 2776 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance(); 2777 onChanged(); 2778 } else { 2779 reqInfoBuilder_.clear(); 2780 } 2781 bitField0_ = (bitField0_ & ~0x00000001); 2782 return this; 2783 } 2784 /** 2785 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2786 */ 2787 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder getReqInfoBuilder() { 2788 bitField0_ |= 0x00000001; 2789 onChanged(); 2790 return getReqInfoFieldBuilder().getBuilder(); 2791 } 2792 /** 2793 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2794 */ 2795 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() { 2796 if (reqInfoBuilder_ != null) { 2797 return reqInfoBuilder_.getMessageOrBuilder(); 2798 } else { 2799 return reqInfo_; 2800 } 2801 } 2802 /** 2803 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code> 2804 */ 2805 private com.google.protobuf.SingleFieldBuilder< 2806 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> 2807 getReqInfoFieldBuilder() { 2808 if (reqInfoBuilder_ == null) { 2809 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< 2810 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>( 2811 reqInfo_, 2812 getParentForChildren(), 2813 isClean()); 2814 reqInfo_ = null; 2815 } 2816 return reqInfoBuilder_; 2817 } 2818 2819 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToStandbyRequestProto) 2820 } 2821 2822 static { 2823 defaultInstance = new TransitionToStandbyRequestProto(true); 2824 defaultInstance.initFields(); 2825 } 2826 2827 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToStandbyRequestProto) 2828 } 2829 2830 public interface TransitionToStandbyResponseProtoOrBuilder 2831 extends com.google.protobuf.MessageOrBuilder { 2832 } 2833 /** 2834 * Protobuf type {@code hadoop.common.TransitionToStandbyResponseProto} 2835 * 2836 * <pre> 2837 ** 2838 * void response 2839 * </pre> 2840 */ 2841 public static final class TransitionToStandbyResponseProto extends 2842 com.google.protobuf.GeneratedMessage 2843 implements TransitionToStandbyResponseProtoOrBuilder { 2844 // Use TransitionToStandbyResponseProto.newBuilder() to construct. 2845 private TransitionToStandbyResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 2846 super(builder); 2847 this.unknownFields = builder.getUnknownFields(); 2848 } 2849 private TransitionToStandbyResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 2850 2851 private static final TransitionToStandbyResponseProto defaultInstance; 2852 public static TransitionToStandbyResponseProto getDefaultInstance() { 2853 return defaultInstance; 2854 } 2855 2856 public TransitionToStandbyResponseProto getDefaultInstanceForType() { 2857 return defaultInstance; 2858 } 2859 2860 private final com.google.protobuf.UnknownFieldSet unknownFields; 2861 @java.lang.Override 2862 public final com.google.protobuf.UnknownFieldSet 2863 getUnknownFields() { 2864 return this.unknownFields; 2865 } 2866 private TransitionToStandbyResponseProto( 2867 com.google.protobuf.CodedInputStream input, 2868 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2869 throws com.google.protobuf.InvalidProtocolBufferException { 2870 initFields(); 2871 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2872 com.google.protobuf.UnknownFieldSet.newBuilder(); 2873 try { 2874 boolean done = false; 2875 while (!done) { 2876 int tag = input.readTag(); 2877 switch (tag) { 2878 case 0: 2879 done = true; 2880 break; 2881 default: { 2882 if (!parseUnknownField(input, unknownFields, 2883 extensionRegistry, tag)) { 2884 done = true; 2885 } 2886 break; 2887 } 2888 } 2889 } 2890 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 2891 throw e.setUnfinishedMessage(this); 2892 } catch (java.io.IOException e) { 2893 throw new com.google.protobuf.InvalidProtocolBufferException( 2894 e.getMessage()).setUnfinishedMessage(this); 2895 } finally { 2896 this.unknownFields = unknownFields.build(); 2897 makeExtensionsImmutable(); 2898 } 2899 } 2900 public static final com.google.protobuf.Descriptors.Descriptor 2901 getDescriptor() { 2902 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor; 2903 } 2904 2905 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2906 internalGetFieldAccessorTable() { 2907 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable 2908 .ensureFieldAccessorsInitialized( 2909 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.Builder.class); 2910 } 2911 2912 public static com.google.protobuf.Parser<TransitionToStandbyResponseProto> PARSER = 2913 new com.google.protobuf.AbstractParser<TransitionToStandbyResponseProto>() { 2914 public TransitionToStandbyResponseProto parsePartialFrom( 2915 com.google.protobuf.CodedInputStream input, 2916 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2917 throws com.google.protobuf.InvalidProtocolBufferException { 2918 return new TransitionToStandbyResponseProto(input, extensionRegistry); 2919 } 2920 }; 2921 2922 @java.lang.Override 2923 public com.google.protobuf.Parser<TransitionToStandbyResponseProto> getParserForType() { 2924 return PARSER; 2925 } 2926 2927 private void initFields() { 2928 } 2929 private byte memoizedIsInitialized = -1; 2930 public final boolean isInitialized() { 2931 byte isInitialized = memoizedIsInitialized; 2932 if (isInitialized != -1) return isInitialized == 1; 2933 2934 memoizedIsInitialized = 1; 2935 return true; 2936 } 2937 2938 public void writeTo(com.google.protobuf.CodedOutputStream output) 2939 throws java.io.IOException { 2940 getSerializedSize(); 2941 getUnknownFields().writeTo(output); 2942 } 2943 2944 private int memoizedSerializedSize = -1; 2945 public int getSerializedSize() { 2946 int size = memoizedSerializedSize; 2947 if (size != -1) return size; 2948 2949 size = 0; 2950 size += getUnknownFields().getSerializedSize(); 2951 memoizedSerializedSize = size; 2952 return size; 2953 } 2954 2955 private static final long serialVersionUID = 0L; 2956 @java.lang.Override 2957 protected java.lang.Object writeReplace() 2958 throws java.io.ObjectStreamException { 2959 return super.writeReplace(); 2960 } 2961 2962 @java.lang.Override 2963 public boolean equals(final java.lang.Object obj) { 2964 if (obj == this) { 2965 return true; 2966 } 2967 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto)) { 2968 return super.equals(obj); 2969 } 2970 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) obj; 2971 2972 boolean result = true; 2973 result = result && 2974 getUnknownFields().equals(other.getUnknownFields()); 2975 return result; 2976 } 2977 2978 private int memoizedHashCode = 0; 2979 @java.lang.Override 2980 public int hashCode() { 2981 if (memoizedHashCode != 0) { 2982 return memoizedHashCode; 2983 } 2984 int hash = 41; 2985 hash = (19 * hash) + getDescriptorForType().hashCode(); 2986 hash = (29 * hash) + getUnknownFields().hashCode(); 2987 memoizedHashCode = hash; 2988 return hash; 2989 } 2990 2991 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 2992 com.google.protobuf.ByteString data) 2993 throws com.google.protobuf.InvalidProtocolBufferException { 2994 return PARSER.parseFrom(data); 2995 } 2996 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 2997 com.google.protobuf.ByteString data, 2998 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2999 throws com.google.protobuf.InvalidProtocolBufferException { 3000 return PARSER.parseFrom(data, extensionRegistry); 3001 } 3002 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(byte[] data) 3003 throws com.google.protobuf.InvalidProtocolBufferException { 3004 return PARSER.parseFrom(data); 3005 } 3006 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 3007 byte[] data, 3008 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3009 throws com.google.protobuf.InvalidProtocolBufferException { 3010 return PARSER.parseFrom(data, extensionRegistry); 3011 } 3012 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(java.io.InputStream input) 3013 throws java.io.IOException { 3014 return PARSER.parseFrom(input); 3015 } 3016 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 3017 java.io.InputStream input, 3018 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3019 throws java.io.IOException { 3020 return PARSER.parseFrom(input, extensionRegistry); 3021 } 3022 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseDelimitedFrom(java.io.InputStream input) 3023 throws java.io.IOException { 3024 return PARSER.parseDelimitedFrom(input); 3025 } 3026 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseDelimitedFrom( 3027 java.io.InputStream input, 3028 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3029 throws java.io.IOException { 3030 return PARSER.parseDelimitedFrom(input, extensionRegistry); 3031 } 3032 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 3033 com.google.protobuf.CodedInputStream input) 3034 throws java.io.IOException { 3035 return PARSER.parseFrom(input); 3036 } 3037 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 3038 com.google.protobuf.CodedInputStream input, 3039 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3040 throws java.io.IOException { 3041 return PARSER.parseFrom(input, extensionRegistry); 3042 } 3043 3044 public static Builder newBuilder() { return Builder.create(); } 3045 public Builder newBuilderForType() { return newBuilder(); } 3046 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto prototype) { 3047 return newBuilder().mergeFrom(prototype); 3048 } 3049 public Builder toBuilder() { return newBuilder(this); } 3050 3051 @java.lang.Override 3052 protected Builder newBuilderForType( 3053 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3054 Builder builder = new Builder(parent); 3055 return builder; 3056 } 3057 /** 3058 * Protobuf type {@code hadoop.common.TransitionToStandbyResponseProto} 3059 * 3060 * <pre> 3061 ** 3062 * void response 3063 * </pre> 3064 */ 3065 public static final class Builder extends 3066 com.google.protobuf.GeneratedMessage.Builder<Builder> 3067 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProtoOrBuilder { 3068 public static final com.google.protobuf.Descriptors.Descriptor 3069 getDescriptor() { 3070 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor; 3071 } 3072 3073 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3074 internalGetFieldAccessorTable() { 3075 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable 3076 .ensureFieldAccessorsInitialized( 3077 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.Builder.class); 3078 } 3079 3080 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.newBuilder() 3081 private Builder() { 3082 maybeForceBuilderInitialization(); 3083 } 3084 3085 private Builder( 3086 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3087 super(parent); 3088 maybeForceBuilderInitialization(); 3089 } 3090 private void maybeForceBuilderInitialization() { 3091 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 3092 } 3093 } 3094 private static Builder create() { 3095 return new Builder(); 3096 } 3097 3098 public Builder clear() { 3099 super.clear(); 3100 return this; 3101 } 3102 3103 public Builder clone() { 3104 return create().mergeFrom(buildPartial()); 3105 } 3106 3107 public com.google.protobuf.Descriptors.Descriptor 3108 getDescriptorForType() { 3109 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor; 3110 } 3111 3112 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto getDefaultInstanceForType() { 3113 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance(); 3114 } 3115 3116 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto build() { 3117 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = buildPartial(); 3118 if (!result.isInitialized()) { 3119 throw newUninitializedMessageException(result); 3120 } 3121 return result; 3122 } 3123 3124 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto buildPartial() { 3125 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto(this); 3126 onBuilt(); 3127 return result; 3128 } 3129 3130 public Builder mergeFrom(com.google.protobuf.Message other) { 3131 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) { 3132 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto)other); 3133 } else { 3134 super.mergeFrom(other); 3135 return this; 3136 } 3137 } 3138 3139 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto other) { 3140 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance()) return this; 3141 this.mergeUnknownFields(other.getUnknownFields()); 3142 return this; 3143 } 3144 3145 public final boolean isInitialized() { 3146 return true; 3147 } 3148 3149 public Builder mergeFrom( 3150 com.google.protobuf.CodedInputStream input, 3151 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3152 throws java.io.IOException { 3153 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parsedMessage = null; 3154 try { 3155 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 3156 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3157 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) e.getUnfinishedMessage(); 3158 throw e; 3159 } finally { 3160 if (parsedMessage != null) { 3161 mergeFrom(parsedMessage); 3162 } 3163 } 3164 return this; 3165 } 3166 3167 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToStandbyResponseProto) 3168 } 3169 3170 static { 3171 defaultInstance = new TransitionToStandbyResponseProto(true); 3172 defaultInstance.initFields(); 3173 } 3174 3175 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToStandbyResponseProto) 3176 } 3177 3178 public interface GetServiceStatusRequestProtoOrBuilder 3179 extends com.google.protobuf.MessageOrBuilder { 3180 } 3181 /** 3182 * Protobuf type {@code hadoop.common.GetServiceStatusRequestProto} 3183 * 3184 * <pre> 3185 ** 3186 * void request 3187 * </pre> 3188 */ 3189 public static final class GetServiceStatusRequestProto extends 3190 com.google.protobuf.GeneratedMessage 3191 implements GetServiceStatusRequestProtoOrBuilder { 3192 // Use GetServiceStatusRequestProto.newBuilder() to construct. 3193 private GetServiceStatusRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 3194 super(builder); 3195 this.unknownFields = builder.getUnknownFields(); 3196 } 3197 private GetServiceStatusRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 3198 3199 private static final GetServiceStatusRequestProto defaultInstance; 3200 public static GetServiceStatusRequestProto getDefaultInstance() { 3201 return defaultInstance; 3202 } 3203 3204 public GetServiceStatusRequestProto getDefaultInstanceForType() { 3205 return defaultInstance; 3206 } 3207 3208 private final com.google.protobuf.UnknownFieldSet unknownFields; 3209 @java.lang.Override 3210 public final com.google.protobuf.UnknownFieldSet 3211 getUnknownFields() { 3212 return this.unknownFields; 3213 } 3214 private GetServiceStatusRequestProto( 3215 com.google.protobuf.CodedInputStream input, 3216 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3217 throws com.google.protobuf.InvalidProtocolBufferException { 3218 initFields(); 3219 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 3220 com.google.protobuf.UnknownFieldSet.newBuilder(); 3221 try { 3222 boolean done = false; 3223 while (!done) { 3224 int tag = input.readTag(); 3225 switch (tag) { 3226 case 0: 3227 done = true; 3228 break; 3229 default: { 3230 if (!parseUnknownField(input, unknownFields, 3231 extensionRegistry, tag)) { 3232 done = true; 3233 } 3234 break; 3235 } 3236 } 3237 } 3238 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3239 throw e.setUnfinishedMessage(this); 3240 } catch (java.io.IOException e) { 3241 throw new com.google.protobuf.InvalidProtocolBufferException( 3242 e.getMessage()).setUnfinishedMessage(this); 3243 } finally { 3244 this.unknownFields = unknownFields.build(); 3245 makeExtensionsImmutable(); 3246 } 3247 } 3248 public static final com.google.protobuf.Descriptors.Descriptor 3249 getDescriptor() { 3250 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor; 3251 } 3252 3253 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3254 internalGetFieldAccessorTable() { 3255 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable 3256 .ensureFieldAccessorsInitialized( 3257 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.Builder.class); 3258 } 3259 3260 public static com.google.protobuf.Parser<GetServiceStatusRequestProto> PARSER = 3261 new com.google.protobuf.AbstractParser<GetServiceStatusRequestProto>() { 3262 public GetServiceStatusRequestProto parsePartialFrom( 3263 com.google.protobuf.CodedInputStream input, 3264 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3265 throws com.google.protobuf.InvalidProtocolBufferException { 3266 return new GetServiceStatusRequestProto(input, extensionRegistry); 3267 } 3268 }; 3269 3270 @java.lang.Override 3271 public com.google.protobuf.Parser<GetServiceStatusRequestProto> getParserForType() { 3272 return PARSER; 3273 } 3274 3275 private void initFields() { 3276 } 3277 private byte memoizedIsInitialized = -1; 3278 public final boolean isInitialized() { 3279 byte isInitialized = memoizedIsInitialized; 3280 if (isInitialized != -1) return isInitialized == 1; 3281 3282 memoizedIsInitialized = 1; 3283 return true; 3284 } 3285 3286 public void writeTo(com.google.protobuf.CodedOutputStream output) 3287 throws java.io.IOException { 3288 getSerializedSize(); 3289 getUnknownFields().writeTo(output); 3290 } 3291 3292 private int memoizedSerializedSize = -1; 3293 public int getSerializedSize() { 3294 int size = memoizedSerializedSize; 3295 if (size != -1) return size; 3296 3297 size = 0; 3298 size += getUnknownFields().getSerializedSize(); 3299 memoizedSerializedSize = size; 3300 return size; 3301 } 3302 3303 private static final long serialVersionUID = 0L; 3304 @java.lang.Override 3305 protected java.lang.Object writeReplace() 3306 throws java.io.ObjectStreamException { 3307 return super.writeReplace(); 3308 } 3309 3310 @java.lang.Override 3311 public boolean equals(final java.lang.Object obj) { 3312 if (obj == this) { 3313 return true; 3314 } 3315 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)) { 3316 return super.equals(obj); 3317 } 3318 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) obj; 3319 3320 boolean result = true; 3321 result = result && 3322 getUnknownFields().equals(other.getUnknownFields()); 3323 return result; 3324 } 3325 3326 private int memoizedHashCode = 0; 3327 @java.lang.Override 3328 public int hashCode() { 3329 if (memoizedHashCode != 0) { 3330 return memoizedHashCode; 3331 } 3332 int hash = 41; 3333 hash = (19 * hash) + getDescriptorForType().hashCode(); 3334 hash = (29 * hash) + getUnknownFields().hashCode(); 3335 memoizedHashCode = hash; 3336 return hash; 3337 } 3338 3339 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 3340 com.google.protobuf.ByteString data) 3341 throws com.google.protobuf.InvalidProtocolBufferException { 3342 return PARSER.parseFrom(data); 3343 } 3344 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 3345 com.google.protobuf.ByteString data, 3346 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3347 throws com.google.protobuf.InvalidProtocolBufferException { 3348 return PARSER.parseFrom(data, extensionRegistry); 3349 } 3350 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(byte[] data) 3351 throws com.google.protobuf.InvalidProtocolBufferException { 3352 return PARSER.parseFrom(data); 3353 } 3354 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 3355 byte[] data, 3356 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3357 throws com.google.protobuf.InvalidProtocolBufferException { 3358 return PARSER.parseFrom(data, extensionRegistry); 3359 } 3360 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(java.io.InputStream input) 3361 throws java.io.IOException { 3362 return PARSER.parseFrom(input); 3363 } 3364 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 3365 java.io.InputStream input, 3366 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3367 throws java.io.IOException { 3368 return PARSER.parseFrom(input, extensionRegistry); 3369 } 3370 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseDelimitedFrom(java.io.InputStream input) 3371 throws java.io.IOException { 3372 return PARSER.parseDelimitedFrom(input); 3373 } 3374 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseDelimitedFrom( 3375 java.io.InputStream input, 3376 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3377 throws java.io.IOException { 3378 return PARSER.parseDelimitedFrom(input, extensionRegistry); 3379 } 3380 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 3381 com.google.protobuf.CodedInputStream input) 3382 throws java.io.IOException { 3383 return PARSER.parseFrom(input); 3384 } 3385 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 3386 com.google.protobuf.CodedInputStream input, 3387 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3388 throws java.io.IOException { 3389 return PARSER.parseFrom(input, extensionRegistry); 3390 } 3391 3392 public static Builder newBuilder() { return Builder.create(); } 3393 public Builder newBuilderForType() { return newBuilder(); } 3394 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto prototype) { 3395 return newBuilder().mergeFrom(prototype); 3396 } 3397 public Builder toBuilder() { return newBuilder(this); } 3398 3399 @java.lang.Override 3400 protected Builder newBuilderForType( 3401 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3402 Builder builder = new Builder(parent); 3403 return builder; 3404 } 3405 /** 3406 * Protobuf type {@code hadoop.common.GetServiceStatusRequestProto} 3407 * 3408 * <pre> 3409 ** 3410 * void request 3411 * </pre> 3412 */ 3413 public static final class Builder extends 3414 com.google.protobuf.GeneratedMessage.Builder<Builder> 3415 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProtoOrBuilder { 3416 public static final com.google.protobuf.Descriptors.Descriptor 3417 getDescriptor() { 3418 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor; 3419 } 3420 3421 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3422 internalGetFieldAccessorTable() { 3423 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable 3424 .ensureFieldAccessorsInitialized( 3425 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.Builder.class); 3426 } 3427 3428 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.newBuilder() 3429 private Builder() { 3430 maybeForceBuilderInitialization(); 3431 } 3432 3433 private Builder( 3434 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3435 super(parent); 3436 maybeForceBuilderInitialization(); 3437 } 3438 private void maybeForceBuilderInitialization() { 3439 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 3440 } 3441 } 3442 private static Builder create() { 3443 return new Builder(); 3444 } 3445 3446 public Builder clear() { 3447 super.clear(); 3448 return this; 3449 } 3450 3451 public Builder clone() { 3452 return create().mergeFrom(buildPartial()); 3453 } 3454 3455 public com.google.protobuf.Descriptors.Descriptor 3456 getDescriptorForType() { 3457 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor; 3458 } 3459 3460 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto getDefaultInstanceForType() { 3461 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance(); 3462 } 3463 3464 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto build() { 3465 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = buildPartial(); 3466 if (!result.isInitialized()) { 3467 throw newUninitializedMessageException(result); 3468 } 3469 return result; 3470 } 3471 3472 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto buildPartial() { 3473 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto(this); 3474 onBuilt(); 3475 return result; 3476 } 3477 3478 public Builder mergeFrom(com.google.protobuf.Message other) { 3479 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) { 3480 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)other); 3481 } else { 3482 super.mergeFrom(other); 3483 return this; 3484 } 3485 } 3486 3487 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto other) { 3488 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance()) return this; 3489 this.mergeUnknownFields(other.getUnknownFields()); 3490 return this; 3491 } 3492 3493 public final boolean isInitialized() { 3494 return true; 3495 } 3496 3497 public Builder mergeFrom( 3498 com.google.protobuf.CodedInputStream input, 3499 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3500 throws java.io.IOException { 3501 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parsedMessage = null; 3502 try { 3503 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 3504 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3505 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) e.getUnfinishedMessage(); 3506 throw e; 3507 } finally { 3508 if (parsedMessage != null) { 3509 mergeFrom(parsedMessage); 3510 } 3511 } 3512 return this; 3513 } 3514 3515 // @@protoc_insertion_point(builder_scope:hadoop.common.GetServiceStatusRequestProto) 3516 } 3517 3518 static { 3519 defaultInstance = new GetServiceStatusRequestProto(true); 3520 defaultInstance.initFields(); 3521 } 3522 3523 // @@protoc_insertion_point(class_scope:hadoop.common.GetServiceStatusRequestProto) 3524 } 3525 3526 public interface GetServiceStatusResponseProtoOrBuilder 3527 extends com.google.protobuf.MessageOrBuilder { 3528 3529 // required .hadoop.common.HAServiceStateProto state = 1; 3530 /** 3531 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code> 3532 */ 3533 boolean hasState(); 3534 /** 3535 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code> 3536 */ 3537 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState(); 3538 3539 // optional bool readyToBecomeActive = 2; 3540 /** 3541 * <code>optional bool readyToBecomeActive = 2;</code> 3542 * 3543 * <pre> 3544 * If state is STANDBY, indicate whether it is 3545 * ready to become active. 3546 * </pre> 3547 */ 3548 boolean hasReadyToBecomeActive(); 3549 /** 3550 * <code>optional bool readyToBecomeActive = 2;</code> 3551 * 3552 * <pre> 3553 * If state is STANDBY, indicate whether it is 3554 * ready to become active. 3555 * </pre> 3556 */ 3557 boolean getReadyToBecomeActive(); 3558 3559 // optional string notReadyReason = 3; 3560 /** 3561 * <code>optional string notReadyReason = 3;</code> 3562 * 3563 * <pre> 3564 * If not ready to become active, a textual explanation of why not 3565 * </pre> 3566 */ 3567 boolean hasNotReadyReason(); 3568 /** 3569 * <code>optional string notReadyReason = 3;</code> 3570 * 3571 * <pre> 3572 * If not ready to become active, a textual explanation of why not 3573 * </pre> 3574 */ 3575 java.lang.String getNotReadyReason(); 3576 /** 3577 * <code>optional string notReadyReason = 3;</code> 3578 * 3579 * <pre> 3580 * If not ready to become active, a textual explanation of why not 3581 * </pre> 3582 */ 3583 com.google.protobuf.ByteString 3584 getNotReadyReasonBytes(); 3585 } 3586 /** 3587 * Protobuf type {@code hadoop.common.GetServiceStatusResponseProto} 3588 * 3589 * <pre> 3590 ** 3591 * Returns the state of the service 3592 * </pre> 3593 */ 3594 public static final class GetServiceStatusResponseProto extends 3595 com.google.protobuf.GeneratedMessage 3596 implements GetServiceStatusResponseProtoOrBuilder { 3597 // Use GetServiceStatusResponseProto.newBuilder() to construct. 3598 private GetServiceStatusResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 3599 super(builder); 3600 this.unknownFields = builder.getUnknownFields(); 3601 } 3602 private GetServiceStatusResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 3603 3604 private static final GetServiceStatusResponseProto defaultInstance; 3605 public static GetServiceStatusResponseProto getDefaultInstance() { 3606 return defaultInstance; 3607 } 3608 3609 public GetServiceStatusResponseProto getDefaultInstanceForType() { 3610 return defaultInstance; 3611 } 3612 3613 private final com.google.protobuf.UnknownFieldSet unknownFields; 3614 @java.lang.Override 3615 public final com.google.protobuf.UnknownFieldSet 3616 getUnknownFields() { 3617 return this.unknownFields; 3618 } 3619 private GetServiceStatusResponseProto( 3620 com.google.protobuf.CodedInputStream input, 3621 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3622 throws com.google.protobuf.InvalidProtocolBufferException { 3623 initFields(); 3624 int mutable_bitField0_ = 0; 3625 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 3626 com.google.protobuf.UnknownFieldSet.newBuilder(); 3627 try { 3628 boolean done = false; 3629 while (!done) { 3630 int tag = input.readTag(); 3631 switch (tag) { 3632 case 0: 3633 done = true; 3634 break; 3635 default: { 3636 if (!parseUnknownField(input, unknownFields, 3637 extensionRegistry, tag)) { 3638 done = true; 3639 } 3640 break; 3641 } 3642 case 8: { 3643 int rawValue = input.readEnum(); 3644 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto value = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.valueOf(rawValue); 3645 if (value == null) { 3646 unknownFields.mergeVarintField(1, rawValue); 3647 } else { 3648 bitField0_ |= 0x00000001; 3649 state_ = value; 3650 } 3651 break; 3652 } 3653 case 16: { 3654 bitField0_ |= 0x00000002; 3655 readyToBecomeActive_ = input.readBool(); 3656 break; 3657 } 3658 case 26: { 3659 bitField0_ |= 0x00000004; 3660 notReadyReason_ = input.readBytes(); 3661 break; 3662 } 3663 } 3664 } 3665 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 3666 throw e.setUnfinishedMessage(this); 3667 } catch (java.io.IOException e) { 3668 throw new com.google.protobuf.InvalidProtocolBufferException( 3669 e.getMessage()).setUnfinishedMessage(this); 3670 } finally { 3671 this.unknownFields = unknownFields.build(); 3672 makeExtensionsImmutable(); 3673 } 3674 } 3675 public static final com.google.protobuf.Descriptors.Descriptor 3676 getDescriptor() { 3677 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor; 3678 } 3679 3680 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 3681 internalGetFieldAccessorTable() { 3682 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable 3683 .ensureFieldAccessorsInitialized( 3684 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.Builder.class); 3685 } 3686 3687 public static com.google.protobuf.Parser<GetServiceStatusResponseProto> PARSER = 3688 new com.google.protobuf.AbstractParser<GetServiceStatusResponseProto>() { 3689 public GetServiceStatusResponseProto parsePartialFrom( 3690 com.google.protobuf.CodedInputStream input, 3691 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3692 throws com.google.protobuf.InvalidProtocolBufferException { 3693 return new GetServiceStatusResponseProto(input, extensionRegistry); 3694 } 3695 }; 3696 3697 @java.lang.Override 3698 public com.google.protobuf.Parser<GetServiceStatusResponseProto> getParserForType() { 3699 return PARSER; 3700 } 3701 3702 private int bitField0_; 3703 // required .hadoop.common.HAServiceStateProto state = 1; 3704 public static final int STATE_FIELD_NUMBER = 1; 3705 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto state_; 3706 /** 3707 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code> 3708 */ 3709 public boolean hasState() { 3710 return ((bitField0_ & 0x00000001) == 0x00000001); 3711 } 3712 /** 3713 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code> 3714 */ 3715 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState() { 3716 return state_; 3717 } 3718 3719 // optional bool readyToBecomeActive = 2; 3720 public static final int READYTOBECOMEACTIVE_FIELD_NUMBER = 2; 3721 private boolean readyToBecomeActive_; 3722 /** 3723 * <code>optional bool readyToBecomeActive = 2;</code> 3724 * 3725 * <pre> 3726 * If state is STANDBY, indicate whether it is 3727 * ready to become active. 3728 * </pre> 3729 */ 3730 public boolean hasReadyToBecomeActive() { 3731 return ((bitField0_ & 0x00000002) == 0x00000002); 3732 } 3733 /** 3734 * <code>optional bool readyToBecomeActive = 2;</code> 3735 * 3736 * <pre> 3737 * If state is STANDBY, indicate whether it is 3738 * ready to become active. 3739 * </pre> 3740 */ 3741 public boolean getReadyToBecomeActive() { 3742 return readyToBecomeActive_; 3743 } 3744 3745 // optional string notReadyReason = 3; 3746 public static final int NOTREADYREASON_FIELD_NUMBER = 3; 3747 private java.lang.Object notReadyReason_; 3748 /** 3749 * <code>optional string notReadyReason = 3;</code> 3750 * 3751 * <pre> 3752 * If not ready to become active, a textual explanation of why not 3753 * </pre> 3754 */ 3755 public boolean hasNotReadyReason() { 3756 return ((bitField0_ & 0x00000004) == 0x00000004); 3757 } 3758 /** 3759 * <code>optional string notReadyReason = 3;</code> 3760 * 3761 * <pre> 3762 * If not ready to become active, a textual explanation of why not 3763 * </pre> 3764 */ 3765 public java.lang.String getNotReadyReason() { 3766 java.lang.Object ref = notReadyReason_; 3767 if (ref instanceof java.lang.String) { 3768 return (java.lang.String) ref; 3769 } else { 3770 com.google.protobuf.ByteString bs = 3771 (com.google.protobuf.ByteString) ref; 3772 java.lang.String s = bs.toStringUtf8(); 3773 if (bs.isValidUtf8()) { 3774 notReadyReason_ = s; 3775 } 3776 return s; 3777 } 3778 } 3779 /** 3780 * <code>optional string notReadyReason = 3;</code> 3781 * 3782 * <pre> 3783 * If not ready to become active, a textual explanation of why not 3784 * </pre> 3785 */ 3786 public com.google.protobuf.ByteString 3787 getNotReadyReasonBytes() { 3788 java.lang.Object ref = notReadyReason_; 3789 if (ref instanceof java.lang.String) { 3790 com.google.protobuf.ByteString b = 3791 com.google.protobuf.ByteString.copyFromUtf8( 3792 (java.lang.String) ref); 3793 notReadyReason_ = b; 3794 return b; 3795 } else { 3796 return (com.google.protobuf.ByteString) ref; 3797 } 3798 } 3799 3800 private void initFields() { 3801 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING; 3802 readyToBecomeActive_ = false; 3803 notReadyReason_ = ""; 3804 } 3805 private byte memoizedIsInitialized = -1; 3806 public final boolean isInitialized() { 3807 byte isInitialized = memoizedIsInitialized; 3808 if (isInitialized != -1) return isInitialized == 1; 3809 3810 if (!hasState()) { 3811 memoizedIsInitialized = 0; 3812 return false; 3813 } 3814 memoizedIsInitialized = 1; 3815 return true; 3816 } 3817 3818 public void writeTo(com.google.protobuf.CodedOutputStream output) 3819 throws java.io.IOException { 3820 getSerializedSize(); 3821 if (((bitField0_ & 0x00000001) == 0x00000001)) { 3822 output.writeEnum(1, state_.getNumber()); 3823 } 3824 if (((bitField0_ & 0x00000002) == 0x00000002)) { 3825 output.writeBool(2, readyToBecomeActive_); 3826 } 3827 if (((bitField0_ & 0x00000004) == 0x00000004)) { 3828 output.writeBytes(3, getNotReadyReasonBytes()); 3829 } 3830 getUnknownFields().writeTo(output); 3831 } 3832 3833 private int memoizedSerializedSize = -1; 3834 public int getSerializedSize() { 3835 int size = memoizedSerializedSize; 3836 if (size != -1) return size; 3837 3838 size = 0; 3839 if (((bitField0_ & 0x00000001) == 0x00000001)) { 3840 size += com.google.protobuf.CodedOutputStream 3841 .computeEnumSize(1, state_.getNumber()); 3842 } 3843 if (((bitField0_ & 0x00000002) == 0x00000002)) { 3844 size += com.google.protobuf.CodedOutputStream 3845 .computeBoolSize(2, readyToBecomeActive_); 3846 } 3847 if (((bitField0_ & 0x00000004) == 0x00000004)) { 3848 size += com.google.protobuf.CodedOutputStream 3849 .computeBytesSize(3, getNotReadyReasonBytes()); 3850 } 3851 size += getUnknownFields().getSerializedSize(); 3852 memoizedSerializedSize = size; 3853 return size; 3854 } 3855 3856 private static final long serialVersionUID = 0L; 3857 @java.lang.Override 3858 protected java.lang.Object writeReplace() 3859 throws java.io.ObjectStreamException { 3860 return super.writeReplace(); 3861 } 3862 3863 @java.lang.Override 3864 public boolean equals(final java.lang.Object obj) { 3865 if (obj == this) { 3866 return true; 3867 } 3868 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto)) { 3869 return super.equals(obj); 3870 } 3871 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) obj; 3872 3873 boolean result = true; 3874 result = result && (hasState() == other.hasState()); 3875 if (hasState()) { 3876 result = result && 3877 (getState() == other.getState()); 3878 } 3879 result = result && (hasReadyToBecomeActive() == other.hasReadyToBecomeActive()); 3880 if (hasReadyToBecomeActive()) { 3881 result = result && (getReadyToBecomeActive() 3882 == other.getReadyToBecomeActive()); 3883 } 3884 result = result && (hasNotReadyReason() == other.hasNotReadyReason()); 3885 if (hasNotReadyReason()) { 3886 result = result && getNotReadyReason() 3887 .equals(other.getNotReadyReason()); 3888 } 3889 result = result && 3890 getUnknownFields().equals(other.getUnknownFields()); 3891 return result; 3892 } 3893 3894 private int memoizedHashCode = 0; 3895 @java.lang.Override 3896 public int hashCode() { 3897 if (memoizedHashCode != 0) { 3898 return memoizedHashCode; 3899 } 3900 int hash = 41; 3901 hash = (19 * hash) + getDescriptorForType().hashCode(); 3902 if (hasState()) { 3903 hash = (37 * hash) + STATE_FIELD_NUMBER; 3904 hash = (53 * hash) + hashEnum(getState()); 3905 } 3906 if (hasReadyToBecomeActive()) { 3907 hash = (37 * hash) + READYTOBECOMEACTIVE_FIELD_NUMBER; 3908 hash = (53 * hash) + hashBoolean(getReadyToBecomeActive()); 3909 } 3910 if (hasNotReadyReason()) { 3911 hash = (37 * hash) + NOTREADYREASON_FIELD_NUMBER; 3912 hash = (53 * hash) + getNotReadyReason().hashCode(); 3913 } 3914 hash = (29 * hash) + getUnknownFields().hashCode(); 3915 memoizedHashCode = hash; 3916 return hash; 3917 } 3918 3919 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 3920 com.google.protobuf.ByteString data) 3921 throws com.google.protobuf.InvalidProtocolBufferException { 3922 return PARSER.parseFrom(data); 3923 } 3924 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 3925 com.google.protobuf.ByteString data, 3926 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3927 throws com.google.protobuf.InvalidProtocolBufferException { 3928 return PARSER.parseFrom(data, extensionRegistry); 3929 } 3930 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(byte[] data) 3931 throws com.google.protobuf.InvalidProtocolBufferException { 3932 return PARSER.parseFrom(data); 3933 } 3934 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 3935 byte[] data, 3936 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3937 throws com.google.protobuf.InvalidProtocolBufferException { 3938 return PARSER.parseFrom(data, extensionRegistry); 3939 } 3940 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(java.io.InputStream input) 3941 throws java.io.IOException { 3942 return PARSER.parseFrom(input); 3943 } 3944 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 3945 java.io.InputStream input, 3946 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3947 throws java.io.IOException { 3948 return PARSER.parseFrom(input, extensionRegistry); 3949 } 3950 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseDelimitedFrom(java.io.InputStream input) 3951 throws java.io.IOException { 3952 return PARSER.parseDelimitedFrom(input); 3953 } 3954 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseDelimitedFrom( 3955 java.io.InputStream input, 3956 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3957 throws java.io.IOException { 3958 return PARSER.parseDelimitedFrom(input, extensionRegistry); 3959 } 3960 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 3961 com.google.protobuf.CodedInputStream input) 3962 throws java.io.IOException { 3963 return PARSER.parseFrom(input); 3964 } 3965 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 3966 com.google.protobuf.CodedInputStream input, 3967 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 3968 throws java.io.IOException { 3969 return PARSER.parseFrom(input, extensionRegistry); 3970 } 3971 3972 public static Builder newBuilder() { return Builder.create(); } 3973 public Builder newBuilderForType() { return newBuilder(); } 3974 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto prototype) { 3975 return newBuilder().mergeFrom(prototype); 3976 } 3977 public Builder toBuilder() { return newBuilder(this); } 3978 3979 @java.lang.Override 3980 protected Builder newBuilderForType( 3981 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 3982 Builder builder = new Builder(parent); 3983 return builder; 3984 } 3985 /** 3986 * Protobuf type {@code hadoop.common.GetServiceStatusResponseProto} 3987 * 3988 * <pre> 3989 ** 3990 * Returns the state of the service 3991 * </pre> 3992 */ 3993 public static final class Builder extends 3994 com.google.protobuf.GeneratedMessage.Builder<Builder> 3995 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProtoOrBuilder { 3996 public static final com.google.protobuf.Descriptors.Descriptor 3997 getDescriptor() { 3998 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor; 3999 } 4000 4001 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 4002 internalGetFieldAccessorTable() { 4003 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable 4004 .ensureFieldAccessorsInitialized( 4005 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.Builder.class); 4006 } 4007 4008 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.newBuilder() 4009 private Builder() { 4010 maybeForceBuilderInitialization(); 4011 } 4012 4013 private Builder( 4014 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 4015 super(parent); 4016 maybeForceBuilderInitialization(); 4017 } 4018 private void maybeForceBuilderInitialization() { 4019 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 4020 } 4021 } 4022 private static Builder create() { 4023 return new Builder(); 4024 } 4025 4026 public Builder clear() { 4027 super.clear(); 4028 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING; 4029 bitField0_ = (bitField0_ & ~0x00000001); 4030 readyToBecomeActive_ = false; 4031 bitField0_ = (bitField0_ & ~0x00000002); 4032 notReadyReason_ = ""; 4033 bitField0_ = (bitField0_ & ~0x00000004); 4034 return this; 4035 } 4036 4037 public Builder clone() { 4038 return create().mergeFrom(buildPartial()); 4039 } 4040 4041 public com.google.protobuf.Descriptors.Descriptor 4042 getDescriptorForType() { 4043 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor; 4044 } 4045 4046 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getDefaultInstanceForType() { 4047 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance(); 4048 } 4049 4050 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto build() { 4051 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = buildPartial(); 4052 if (!result.isInitialized()) { 4053 throw newUninitializedMessageException(result); 4054 } 4055 return result; 4056 } 4057 4058 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto buildPartial() { 4059 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto(this); 4060 int from_bitField0_ = bitField0_; 4061 int to_bitField0_ = 0; 4062 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 4063 to_bitField0_ |= 0x00000001; 4064 } 4065 result.state_ = state_; 4066 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 4067 to_bitField0_ |= 0x00000002; 4068 } 4069 result.readyToBecomeActive_ = readyToBecomeActive_; 4070 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 4071 to_bitField0_ |= 0x00000004; 4072 } 4073 result.notReadyReason_ = notReadyReason_; 4074 result.bitField0_ = to_bitField0_; 4075 onBuilt(); 4076 return result; 4077 } 4078 4079 public Builder mergeFrom(com.google.protobuf.Message other) { 4080 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) { 4081 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto)other); 4082 } else { 4083 super.mergeFrom(other); 4084 return this; 4085 } 4086 } 4087 4088 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto other) { 4089 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance()) return this; 4090 if (other.hasState()) { 4091 setState(other.getState()); 4092 } 4093 if (other.hasReadyToBecomeActive()) { 4094 setReadyToBecomeActive(other.getReadyToBecomeActive()); 4095 } 4096 if (other.hasNotReadyReason()) { 4097 bitField0_ |= 0x00000004; 4098 notReadyReason_ = other.notReadyReason_; 4099 onChanged(); 4100 } 4101 this.mergeUnknownFields(other.getUnknownFields()); 4102 return this; 4103 } 4104 4105 public final boolean isInitialized() { 4106 if (!hasState()) { 4107 4108 return false; 4109 } 4110 return true; 4111 } 4112 4113 public Builder mergeFrom( 4114 com.google.protobuf.CodedInputStream input, 4115 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 4116 throws java.io.IOException { 4117 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parsedMessage = null; 4118 try { 4119 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 4120 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 4121 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) e.getUnfinishedMessage(); 4122 throw e; 4123 } finally { 4124 if (parsedMessage != null) { 4125 mergeFrom(parsedMessage); 4126 } 4127 } 4128 return this; 4129 } 4130 private int bitField0_; 4131 4132 // required .hadoop.common.HAServiceStateProto state = 1; 4133 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING; 4134 /** 4135 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code> 4136 */ 4137 public boolean hasState() { 4138 return ((bitField0_ & 0x00000001) == 0x00000001); 4139 } 4140 /** 4141 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code> 4142 */ 4143 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState() { 4144 return state_; 4145 } 4146 /** 4147 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code> 4148 */ 4149 public Builder setState(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto value) { 4150 if (value == null) { 4151 throw new NullPointerException(); 4152 } 4153 bitField0_ |= 0x00000001; 4154 state_ = value; 4155 onChanged(); 4156 return this; 4157 } 4158 /** 4159 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code> 4160 */ 4161 public Builder clearState() { 4162 bitField0_ = (bitField0_ & ~0x00000001); 4163 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING; 4164 onChanged(); 4165 return this; 4166 } 4167 4168 // optional bool readyToBecomeActive = 2; 4169 private boolean readyToBecomeActive_ ; 4170 /** 4171 * <code>optional bool readyToBecomeActive = 2;</code> 4172 * 4173 * <pre> 4174 * If state is STANDBY, indicate whether it is 4175 * ready to become active. 4176 * </pre> 4177 */ 4178 public boolean hasReadyToBecomeActive() { 4179 return ((bitField0_ & 0x00000002) == 0x00000002); 4180 } 4181 /** 4182 * <code>optional bool readyToBecomeActive = 2;</code> 4183 * 4184 * <pre> 4185 * If state is STANDBY, indicate whether it is 4186 * ready to become active. 4187 * </pre> 4188 */ 4189 public boolean getReadyToBecomeActive() { 4190 return readyToBecomeActive_; 4191 } 4192 /** 4193 * <code>optional bool readyToBecomeActive = 2;</code> 4194 * 4195 * <pre> 4196 * If state is STANDBY, indicate whether it is 4197 * ready to become active. 4198 * </pre> 4199 */ 4200 public Builder setReadyToBecomeActive(boolean value) { 4201 bitField0_ |= 0x00000002; 4202 readyToBecomeActive_ = value; 4203 onChanged(); 4204 return this; 4205 } 4206 /** 4207 * <code>optional bool readyToBecomeActive = 2;</code> 4208 * 4209 * <pre> 4210 * If state is STANDBY, indicate whether it is 4211 * ready to become active. 4212 * </pre> 4213 */ 4214 public Builder clearReadyToBecomeActive() { 4215 bitField0_ = (bitField0_ & ~0x00000002); 4216 readyToBecomeActive_ = false; 4217 onChanged(); 4218 return this; 4219 } 4220 4221 // optional string notReadyReason = 3; 4222 private java.lang.Object notReadyReason_ = ""; 4223 /** 4224 * <code>optional string notReadyReason = 3;</code> 4225 * 4226 * <pre> 4227 * If not ready to become active, a textual explanation of why not 4228 * </pre> 4229 */ 4230 public boolean hasNotReadyReason() { 4231 return ((bitField0_ & 0x00000004) == 0x00000004); 4232 } 4233 /** 4234 * <code>optional string notReadyReason = 3;</code> 4235 * 4236 * <pre> 4237 * If not ready to become active, a textual explanation of why not 4238 * </pre> 4239 */ 4240 public java.lang.String getNotReadyReason() { 4241 java.lang.Object ref = notReadyReason_; 4242 if (!(ref instanceof java.lang.String)) { 4243 java.lang.String s = ((com.google.protobuf.ByteString) ref) 4244 .toStringUtf8(); 4245 notReadyReason_ = s; 4246 return s; 4247 } else { 4248 return (java.lang.String) ref; 4249 } 4250 } 4251 /** 4252 * <code>optional string notReadyReason = 3;</code> 4253 * 4254 * <pre> 4255 * If not ready to become active, a textual explanation of why not 4256 * </pre> 4257 */ 4258 public com.google.protobuf.ByteString 4259 getNotReadyReasonBytes() { 4260 java.lang.Object ref = notReadyReason_; 4261 if (ref instanceof String) { 4262 com.google.protobuf.ByteString b = 4263 com.google.protobuf.ByteString.copyFromUtf8( 4264 (java.lang.String) ref); 4265 notReadyReason_ = b; 4266 return b; 4267 } else { 4268 return (com.google.protobuf.ByteString) ref; 4269 } 4270 } 4271 /** 4272 * <code>optional string notReadyReason = 3;</code> 4273 * 4274 * <pre> 4275 * If not ready to become active, a textual explanation of why not 4276 * </pre> 4277 */ 4278 public Builder setNotReadyReason( 4279 java.lang.String value) { 4280 if (value == null) { 4281 throw new NullPointerException(); 4282 } 4283 bitField0_ |= 0x00000004; 4284 notReadyReason_ = value; 4285 onChanged(); 4286 return this; 4287 } 4288 /** 4289 * <code>optional string notReadyReason = 3;</code> 4290 * 4291 * <pre> 4292 * If not ready to become active, a textual explanation of why not 4293 * </pre> 4294 */ 4295 public Builder clearNotReadyReason() { 4296 bitField0_ = (bitField0_ & ~0x00000004); 4297 notReadyReason_ = getDefaultInstance().getNotReadyReason(); 4298 onChanged(); 4299 return this; 4300 } 4301 /** 4302 * <code>optional string notReadyReason = 3;</code> 4303 * 4304 * <pre> 4305 * If not ready to become active, a textual explanation of why not 4306 * </pre> 4307 */ 4308 public Builder setNotReadyReasonBytes( 4309 com.google.protobuf.ByteString value) { 4310 if (value == null) { 4311 throw new NullPointerException(); 4312 } 4313 bitField0_ |= 0x00000004; 4314 notReadyReason_ = value; 4315 onChanged(); 4316 return this; 4317 } 4318 4319 // @@protoc_insertion_point(builder_scope:hadoop.common.GetServiceStatusResponseProto) 4320 } 4321 4322 static { 4323 defaultInstance = new GetServiceStatusResponseProto(true); 4324 defaultInstance.initFields(); 4325 } 4326 4327 // @@protoc_insertion_point(class_scope:hadoop.common.GetServiceStatusResponseProto) 4328 } 4329 4330 /** 4331 * Protobuf service {@code hadoop.common.HAServiceProtocolService} 4332 * 4333 * <pre> 4334 ** 4335 * Protocol interface provides High availability related 4336 * primitives to monitor and failover a service. 4337 * 4338 * For details see o.a.h.ha.HAServiceProtocol. 4339 * </pre> 4340 */ 4341 public static abstract class HAServiceProtocolService 4342 implements com.google.protobuf.Service { 4343 protected HAServiceProtocolService() {} 4344 4345 public interface Interface { 4346 /** 4347 * <code>rpc monitorHealth(.hadoop.common.MonitorHealthRequestProto) returns (.hadoop.common.MonitorHealthResponseProto);</code> 4348 * 4349 * <pre> 4350 ** 4351 * Monitor the health of a service. 4352 * </pre> 4353 */ 4354 public abstract void monitorHealth( 4355 com.google.protobuf.RpcController controller, 4356 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request, 4357 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done); 4358 4359 /** 4360 * <code>rpc transitionToActive(.hadoop.common.TransitionToActiveRequestProto) returns (.hadoop.common.TransitionToActiveResponseProto);</code> 4361 * 4362 * <pre> 4363 ** 4364 * Request service to tranisition to active state. 4365 * </pre> 4366 */ 4367 public abstract void transitionToActive( 4368 com.google.protobuf.RpcController controller, 4369 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request, 4370 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done); 4371 4372 /** 4373 * <code>rpc transitionToStandby(.hadoop.common.TransitionToStandbyRequestProto) returns (.hadoop.common.TransitionToStandbyResponseProto);</code> 4374 * 4375 * <pre> 4376 ** 4377 * Request service to transition to standby state. 4378 * </pre> 4379 */ 4380 public abstract void transitionToStandby( 4381 com.google.protobuf.RpcController controller, 4382 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request, 4383 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done); 4384 4385 /** 4386 * <code>rpc getServiceStatus(.hadoop.common.GetServiceStatusRequestProto) returns (.hadoop.common.GetServiceStatusResponseProto);</code> 4387 * 4388 * <pre> 4389 ** 4390 * Get the current status of the service. 4391 * </pre> 4392 */ 4393 public abstract void getServiceStatus( 4394 com.google.protobuf.RpcController controller, 4395 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request, 4396 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done); 4397 4398 } 4399 4400 public static com.google.protobuf.Service newReflectiveService( 4401 final Interface impl) { 4402 return new HAServiceProtocolService() { 4403 @java.lang.Override 4404 public void monitorHealth( 4405 com.google.protobuf.RpcController controller, 4406 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request, 4407 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done) { 4408 impl.monitorHealth(controller, request, done); 4409 } 4410 4411 @java.lang.Override 4412 public void transitionToActive( 4413 com.google.protobuf.RpcController controller, 4414 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request, 4415 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done) { 4416 impl.transitionToActive(controller, request, done); 4417 } 4418 4419 @java.lang.Override 4420 public void transitionToStandby( 4421 com.google.protobuf.RpcController controller, 4422 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request, 4423 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done) { 4424 impl.transitionToStandby(controller, request, done); 4425 } 4426 4427 @java.lang.Override 4428 public void getServiceStatus( 4429 com.google.protobuf.RpcController controller, 4430 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request, 4431 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done) { 4432 impl.getServiceStatus(controller, request, done); 4433 } 4434 4435 }; 4436 } 4437 4438 public static com.google.protobuf.BlockingService 4439 newReflectiveBlockingService(final BlockingInterface impl) { 4440 return new com.google.protobuf.BlockingService() { 4441 public final com.google.protobuf.Descriptors.ServiceDescriptor 4442 getDescriptorForType() { 4443 return getDescriptor(); 4444 } 4445 4446 public final com.google.protobuf.Message callBlockingMethod( 4447 com.google.protobuf.Descriptors.MethodDescriptor method, 4448 com.google.protobuf.RpcController controller, 4449 com.google.protobuf.Message request) 4450 throws com.google.protobuf.ServiceException { 4451 if (method.getService() != getDescriptor()) { 4452 throw new java.lang.IllegalArgumentException( 4453 "Service.callBlockingMethod() given method descriptor for " + 4454 "wrong service type."); 4455 } 4456 switch(method.getIndex()) { 4457 case 0: 4458 return impl.monitorHealth(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)request); 4459 case 1: 4460 return impl.transitionToActive(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)request); 4461 case 2: 4462 return impl.transitionToStandby(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)request); 4463 case 3: 4464 return impl.getServiceStatus(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)request); 4465 default: 4466 throw new java.lang.AssertionError("Can't get here."); 4467 } 4468 } 4469 4470 public final com.google.protobuf.Message 4471 getRequestPrototype( 4472 com.google.protobuf.Descriptors.MethodDescriptor method) { 4473 if (method.getService() != getDescriptor()) { 4474 throw new java.lang.IllegalArgumentException( 4475 "Service.getRequestPrototype() given method " + 4476 "descriptor for wrong service type."); 4477 } 4478 switch(method.getIndex()) { 4479 case 0: 4480 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance(); 4481 case 1: 4482 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance(); 4483 case 2: 4484 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance(); 4485 case 3: 4486 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance(); 4487 default: 4488 throw new java.lang.AssertionError("Can't get here."); 4489 } 4490 } 4491 4492 public final com.google.protobuf.Message 4493 getResponsePrototype( 4494 com.google.protobuf.Descriptors.MethodDescriptor method) { 4495 if (method.getService() != getDescriptor()) { 4496 throw new java.lang.IllegalArgumentException( 4497 "Service.getResponsePrototype() given method " + 4498 "descriptor for wrong service type."); 4499 } 4500 switch(method.getIndex()) { 4501 case 0: 4502 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance(); 4503 case 1: 4504 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance(); 4505 case 2: 4506 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance(); 4507 case 3: 4508 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance(); 4509 default: 4510 throw new java.lang.AssertionError("Can't get here."); 4511 } 4512 } 4513 4514 }; 4515 } 4516 4517 /** 4518 * <code>rpc monitorHealth(.hadoop.common.MonitorHealthRequestProto) returns (.hadoop.common.MonitorHealthResponseProto);</code> 4519 * 4520 * <pre> 4521 ** 4522 * Monitor the health of a service. 4523 * </pre> 4524 */ 4525 public abstract void monitorHealth( 4526 com.google.protobuf.RpcController controller, 4527 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request, 4528 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done); 4529 4530 /** 4531 * <code>rpc transitionToActive(.hadoop.common.TransitionToActiveRequestProto) returns (.hadoop.common.TransitionToActiveResponseProto);</code> 4532 * 4533 * <pre> 4534 ** 4535 * Request service to tranisition to active state. 4536 * </pre> 4537 */ 4538 public abstract void transitionToActive( 4539 com.google.protobuf.RpcController controller, 4540 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request, 4541 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done); 4542 4543 /** 4544 * <code>rpc transitionToStandby(.hadoop.common.TransitionToStandbyRequestProto) returns (.hadoop.common.TransitionToStandbyResponseProto);</code> 4545 * 4546 * <pre> 4547 ** 4548 * Request service to transition to standby state. 4549 * </pre> 4550 */ 4551 public abstract void transitionToStandby( 4552 com.google.protobuf.RpcController controller, 4553 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request, 4554 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done); 4555 4556 /** 4557 * <code>rpc getServiceStatus(.hadoop.common.GetServiceStatusRequestProto) returns (.hadoop.common.GetServiceStatusResponseProto);</code> 4558 * 4559 * <pre> 4560 ** 4561 * Get the current status of the service. 4562 * </pre> 4563 */ 4564 public abstract void getServiceStatus( 4565 com.google.protobuf.RpcController controller, 4566 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request, 4567 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done); 4568 4569 public static final 4570 com.google.protobuf.Descriptors.ServiceDescriptor 4571 getDescriptor() { 4572 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getServices().get(0); 4573 } 4574 public final com.google.protobuf.Descriptors.ServiceDescriptor 4575 getDescriptorForType() { 4576 return getDescriptor(); 4577 } 4578 4579 public final void callMethod( 4580 com.google.protobuf.Descriptors.MethodDescriptor method, 4581 com.google.protobuf.RpcController controller, 4582 com.google.protobuf.Message request, 4583 com.google.protobuf.RpcCallback< 4584 com.google.protobuf.Message> done) { 4585 if (method.getService() != getDescriptor()) { 4586 throw new java.lang.IllegalArgumentException( 4587 "Service.callMethod() given method descriptor for wrong " + 4588 "service type."); 4589 } 4590 switch(method.getIndex()) { 4591 case 0: 4592 this.monitorHealth(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)request, 4593 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto>specializeCallback( 4594 done)); 4595 return; 4596 case 1: 4597 this.transitionToActive(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)request, 4598 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto>specializeCallback( 4599 done)); 4600 return; 4601 case 2: 4602 this.transitionToStandby(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)request, 4603 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto>specializeCallback( 4604 done)); 4605 return; 4606 case 3: 4607 this.getServiceStatus(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)request, 4608 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto>specializeCallback( 4609 done)); 4610 return; 4611 default: 4612 throw new java.lang.AssertionError("Can't get here."); 4613 } 4614 } 4615 4616 public final com.google.protobuf.Message 4617 getRequestPrototype( 4618 com.google.protobuf.Descriptors.MethodDescriptor method) { 4619 if (method.getService() != getDescriptor()) { 4620 throw new java.lang.IllegalArgumentException( 4621 "Service.getRequestPrototype() given method " + 4622 "descriptor for wrong service type."); 4623 } 4624 switch(method.getIndex()) { 4625 case 0: 4626 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance(); 4627 case 1: 4628 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance(); 4629 case 2: 4630 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance(); 4631 case 3: 4632 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance(); 4633 default: 4634 throw new java.lang.AssertionError("Can't get here."); 4635 } 4636 } 4637 4638 public final com.google.protobuf.Message 4639 getResponsePrototype( 4640 com.google.protobuf.Descriptors.MethodDescriptor method) { 4641 if (method.getService() != getDescriptor()) { 4642 throw new java.lang.IllegalArgumentException( 4643 "Service.getResponsePrototype() given method " + 4644 "descriptor for wrong service type."); 4645 } 4646 switch(method.getIndex()) { 4647 case 0: 4648 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance(); 4649 case 1: 4650 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance(); 4651 case 2: 4652 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance(); 4653 case 3: 4654 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance(); 4655 default: 4656 throw new java.lang.AssertionError("Can't get here."); 4657 } 4658 } 4659 4660 public static Stub newStub( 4661 com.google.protobuf.RpcChannel channel) { 4662 return new Stub(channel); 4663 } 4664 4665 public static final class Stub extends org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceProtocolService implements Interface { 4666 private Stub(com.google.protobuf.RpcChannel channel) { 4667 this.channel = channel; 4668 } 4669 4670 private final com.google.protobuf.RpcChannel channel; 4671 4672 public com.google.protobuf.RpcChannel getChannel() { 4673 return channel; 4674 } 4675 4676 public void monitorHealth( 4677 com.google.protobuf.RpcController controller, 4678 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request, 4679 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done) { 4680 channel.callMethod( 4681 getDescriptor().getMethods().get(0), 4682 controller, 4683 request, 4684 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance(), 4685 com.google.protobuf.RpcUtil.generalizeCallback( 4686 done, 4687 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class, 4688 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance())); 4689 } 4690 4691 public void transitionToActive( 4692 com.google.protobuf.RpcController controller, 4693 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request, 4694 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done) { 4695 channel.callMethod( 4696 getDescriptor().getMethods().get(1), 4697 controller, 4698 request, 4699 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance(), 4700 com.google.protobuf.RpcUtil.generalizeCallback( 4701 done, 4702 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class, 4703 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance())); 4704 } 4705 4706 public void transitionToStandby( 4707 com.google.protobuf.RpcController controller, 4708 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request, 4709 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done) { 4710 channel.callMethod( 4711 getDescriptor().getMethods().get(2), 4712 controller, 4713 request, 4714 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance(), 4715 com.google.protobuf.RpcUtil.generalizeCallback( 4716 done, 4717 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class, 4718 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance())); 4719 } 4720 4721 public void getServiceStatus( 4722 com.google.protobuf.RpcController controller, 4723 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request, 4724 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done) { 4725 channel.callMethod( 4726 getDescriptor().getMethods().get(3), 4727 controller, 4728 request, 4729 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance(), 4730 com.google.protobuf.RpcUtil.generalizeCallback( 4731 done, 4732 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class, 4733 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance())); 4734 } 4735 } 4736 4737 public static BlockingInterface newBlockingStub( 4738 com.google.protobuf.BlockingRpcChannel channel) { 4739 return new BlockingStub(channel); 4740 } 4741 4742 public interface BlockingInterface { 4743 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto monitorHealth( 4744 com.google.protobuf.RpcController controller, 4745 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request) 4746 throws com.google.protobuf.ServiceException; 4747 4748 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto transitionToActive( 4749 com.google.protobuf.RpcController controller, 4750 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request) 4751 throws com.google.protobuf.ServiceException; 4752 4753 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto transitionToStandby( 4754 com.google.protobuf.RpcController controller, 4755 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request) 4756 throws com.google.protobuf.ServiceException; 4757 4758 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getServiceStatus( 4759 com.google.protobuf.RpcController controller, 4760 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request) 4761 throws com.google.protobuf.ServiceException; 4762 } 4763 4764 private static final class BlockingStub implements BlockingInterface { 4765 private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { 4766 this.channel = channel; 4767 } 4768 4769 private final com.google.protobuf.BlockingRpcChannel channel; 4770 4771 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto monitorHealth( 4772 com.google.protobuf.RpcController controller, 4773 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request) 4774 throws com.google.protobuf.ServiceException { 4775 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) channel.callBlockingMethod( 4776 getDescriptor().getMethods().get(0), 4777 controller, 4778 request, 4779 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance()); 4780 } 4781 4782 4783 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto transitionToActive( 4784 com.google.protobuf.RpcController controller, 4785 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request) 4786 throws com.google.protobuf.ServiceException { 4787 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) channel.callBlockingMethod( 4788 getDescriptor().getMethods().get(1), 4789 controller, 4790 request, 4791 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance()); 4792 } 4793 4794 4795 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto transitionToStandby( 4796 com.google.protobuf.RpcController controller, 4797 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request) 4798 throws com.google.protobuf.ServiceException { 4799 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) channel.callBlockingMethod( 4800 getDescriptor().getMethods().get(2), 4801 controller, 4802 request, 4803 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance()); 4804 } 4805 4806 4807 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getServiceStatus( 4808 com.google.protobuf.RpcController controller, 4809 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request) 4810 throws com.google.protobuf.ServiceException { 4811 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) channel.callBlockingMethod( 4812 getDescriptor().getMethods().get(3), 4813 controller, 4814 request, 4815 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance()); 4816 } 4817 4818 } 4819 4820 // @@protoc_insertion_point(class_scope:hadoop.common.HAServiceProtocolService) 4821 } 4822 4823 private static com.google.protobuf.Descriptors.Descriptor 4824 internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor; 4825 private static 4826 com.google.protobuf.GeneratedMessage.FieldAccessorTable 4827 internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable; 4828 private static com.google.protobuf.Descriptors.Descriptor 4829 internal_static_hadoop_common_MonitorHealthRequestProto_descriptor; 4830 private static 4831 com.google.protobuf.GeneratedMessage.FieldAccessorTable 4832 internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable; 4833 private static com.google.protobuf.Descriptors.Descriptor 4834 internal_static_hadoop_common_MonitorHealthResponseProto_descriptor; 4835 private static 4836 com.google.protobuf.GeneratedMessage.FieldAccessorTable 4837 internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable; 4838 private static com.google.protobuf.Descriptors.Descriptor 4839 internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor; 4840 private static 4841 com.google.protobuf.GeneratedMessage.FieldAccessorTable 4842 internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable; 4843 private static com.google.protobuf.Descriptors.Descriptor 4844 internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor; 4845 private static 4846 com.google.protobuf.GeneratedMessage.FieldAccessorTable 4847 internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable; 4848 private static com.google.protobuf.Descriptors.Descriptor 4849 internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor; 4850 private static 4851 com.google.protobuf.GeneratedMessage.FieldAccessorTable 4852 internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable; 4853 private static com.google.protobuf.Descriptors.Descriptor 4854 internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor; 4855 private static 4856 com.google.protobuf.GeneratedMessage.FieldAccessorTable 4857 internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable; 4858 private static com.google.protobuf.Descriptors.Descriptor 4859 internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor; 4860 private static 4861 com.google.protobuf.GeneratedMessage.FieldAccessorTable 4862 internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable; 4863 private static com.google.protobuf.Descriptors.Descriptor 4864 internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor; 4865 private static 4866 com.google.protobuf.GeneratedMessage.FieldAccessorTable 4867 internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable; 4868 4869 public static com.google.protobuf.Descriptors.FileDescriptor 4870 getDescriptor() { 4871 return descriptor; 4872 } 4873 private static com.google.protobuf.Descriptors.FileDescriptor 4874 descriptor; 4875 static { 4876 java.lang.String[] descriptorData = { 4877 "\n\027HAServiceProtocol.proto\022\rhadoop.common" + 4878 "\"R\n\035HAStateChangeRequestInfoProto\0221\n\treq" + 4879 "Source\030\001 \002(\0162\036.hadoop.common.HARequestSo" + 4880 "urce\"\033\n\031MonitorHealthRequestProto\"\034\n\032Mon" + 4881 "itorHealthResponseProto\"_\n\036TransitionToA" + 4882 "ctiveRequestProto\022=\n\007reqInfo\030\001 \002(\0132,.had" + 4883 "oop.common.HAStateChangeRequestInfoProto" + 4884 "\"!\n\037TransitionToActiveResponseProto\"`\n\037T" + 4885 "ransitionToStandbyRequestProto\022=\n\007reqInf" + 4886 "o\030\001 \002(\0132,.hadoop.common.HAStateChangeReq", 4887 "uestInfoProto\"\"\n TransitionToStandbyResp" + 4888 "onseProto\"\036\n\034GetServiceStatusRequestProt" + 4889 "o\"\207\001\n\035GetServiceStatusResponseProto\0221\n\005s" + 4890 "tate\030\001 \002(\0162\".hadoop.common.HAServiceStat" + 4891 "eProto\022\033\n\023readyToBecomeActive\030\002 \001(\010\022\026\n\016n" + 4892 "otReadyReason\030\003 \001(\t*@\n\023HAServiceStatePro" + 4893 "to\022\020\n\014INITIALIZING\020\000\022\n\n\006ACTIVE\020\001\022\013\n\007STAN" + 4894 "DBY\020\002*W\n\017HARequestSource\022\023\n\017REQUEST_BY_U" + 4895 "SER\020\000\022\032\n\026REQUEST_BY_USER_FORCED\020\001\022\023\n\017REQ" + 4896 "UEST_BY_ZKFC\020\0022\334\003\n\030HAServiceProtocolServ", 4897 "ice\022d\n\rmonitorHealth\022(.hadoop.common.Mon" + 4898 "itorHealthRequestProto\032).hadoop.common.M" + 4899 "onitorHealthResponseProto\022s\n\022transitionT" + 4900 "oActive\022-.hadoop.common.TransitionToActi" + 4901 "veRequestProto\032..hadoop.common.Transitio" + 4902 "nToActiveResponseProto\022v\n\023transitionToSt" + 4903 "andby\022..hadoop.common.TransitionToStandb" + 4904 "yRequestProto\032/.hadoop.common.Transition" + 4905 "ToStandbyResponseProto\022m\n\020getServiceStat" + 4906 "us\022+.hadoop.common.GetServiceStatusReque", 4907 "stProto\032,.hadoop.common.GetServiceStatus" + 4908 "ResponseProtoB;\n\032org.apache.hadoop.ha.pr" + 4909 "otoB\027HAServiceProtocolProtos\210\001\001\240\001\001" 4910 }; 4911 com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = 4912 new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { 4913 public com.google.protobuf.ExtensionRegistry assignDescriptors( 4914 com.google.protobuf.Descriptors.FileDescriptor root) { 4915 descriptor = root; 4916 internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor = 4917 getDescriptor().getMessageTypes().get(0); 4918 internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable = new 4919 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 4920 internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor, 4921 new java.lang.String[] { "ReqSource", }); 4922 internal_static_hadoop_common_MonitorHealthRequestProto_descriptor = 4923 getDescriptor().getMessageTypes().get(1); 4924 internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable = new 4925 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 4926 internal_static_hadoop_common_MonitorHealthRequestProto_descriptor, 4927 new java.lang.String[] { }); 4928 internal_static_hadoop_common_MonitorHealthResponseProto_descriptor = 4929 getDescriptor().getMessageTypes().get(2); 4930 internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable = new 4931 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 4932 internal_static_hadoop_common_MonitorHealthResponseProto_descriptor, 4933 new java.lang.String[] { }); 4934 internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor = 4935 getDescriptor().getMessageTypes().get(3); 4936 internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable = new 4937 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 4938 internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor, 4939 new java.lang.String[] { "ReqInfo", }); 4940 internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor = 4941 getDescriptor().getMessageTypes().get(4); 4942 internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable = new 4943 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 4944 internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor, 4945 new java.lang.String[] { }); 4946 internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor = 4947 getDescriptor().getMessageTypes().get(5); 4948 internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable = new 4949 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 4950 internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor, 4951 new java.lang.String[] { "ReqInfo", }); 4952 internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor = 4953 getDescriptor().getMessageTypes().get(6); 4954 internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable = new 4955 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 4956 internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor, 4957 new java.lang.String[] { }); 4958 internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor = 4959 getDescriptor().getMessageTypes().get(7); 4960 internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable = new 4961 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 4962 internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor, 4963 new java.lang.String[] { }); 4964 internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor = 4965 getDescriptor().getMessageTypes().get(8); 4966 internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable = new 4967 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 4968 internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor, 4969 new java.lang.String[] { "State", "ReadyToBecomeActive", "NotReadyReason", }); 4970 return null; 4971 } 4972 }; 4973 com.google.protobuf.Descriptors.FileDescriptor 4974 .internalBuildGeneratedFileFrom(descriptorData, 4975 new com.google.protobuf.Descriptors.FileDescriptor[] { 4976 }, assigner); 4977 } 4978 4979 // @@protoc_insertion_point(outer_class_scope) 4980}