001// Generated by the protocol buffer compiler. DO NOT EDIT! 002// source: ZKFCProtocol.proto 003 004package org.apache.hadoop.ha.proto; 005 006public final class ZKFCProtocolProtos { 007 private ZKFCProtocolProtos() {} 008 public static void registerAllExtensions( 009 com.google.protobuf.ExtensionRegistry registry) { 010 } 011 public interface CedeActiveRequestProtoOrBuilder 012 extends com.google.protobuf.MessageOrBuilder { 013 014 // required uint32 millisToCede = 1; 015 /** 016 * <code>required uint32 millisToCede = 1;</code> 017 */ 018 boolean hasMillisToCede(); 019 /** 020 * <code>required uint32 millisToCede = 1;</code> 021 */ 022 int getMillisToCede(); 023 } 024 /** 025 * Protobuf type {@code hadoop.common.CedeActiveRequestProto} 026 */ 027 public static final class CedeActiveRequestProto extends 028 com.google.protobuf.GeneratedMessage 029 implements CedeActiveRequestProtoOrBuilder { 030 // Use CedeActiveRequestProto.newBuilder() to construct. 031 private CedeActiveRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 032 super(builder); 033 this.unknownFields = builder.getUnknownFields(); 034 } 035 private CedeActiveRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 036 037 private static final CedeActiveRequestProto defaultInstance; 038 public static CedeActiveRequestProto getDefaultInstance() { 039 return defaultInstance; 040 } 041 042 public CedeActiveRequestProto getDefaultInstanceForType() { 043 return defaultInstance; 044 } 045 046 private final com.google.protobuf.UnknownFieldSet unknownFields; 047 @java.lang.Override 048 public final com.google.protobuf.UnknownFieldSet 049 getUnknownFields() { 050 return this.unknownFields; 051 } 052 private CedeActiveRequestProto( 053 com.google.protobuf.CodedInputStream input, 054 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 055 throws com.google.protobuf.InvalidProtocolBufferException { 056 initFields(); 057 int mutable_bitField0_ = 0; 058 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 059 com.google.protobuf.UnknownFieldSet.newBuilder(); 060 try { 061 boolean done = false; 062 while (!done) { 063 int tag = input.readTag(); 064 switch (tag) { 065 case 0: 066 done = true; 067 break; 068 default: { 069 if (!parseUnknownField(input, unknownFields, 070 extensionRegistry, tag)) { 071 done = true; 072 } 073 break; 074 } 075 case 8: { 076 bitField0_ |= 0x00000001; 077 millisToCede_ = input.readUInt32(); 078 break; 079 } 080 } 081 } 082 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 083 throw e.setUnfinishedMessage(this); 084 } catch (java.io.IOException e) { 085 throw new com.google.protobuf.InvalidProtocolBufferException( 086 e.getMessage()).setUnfinishedMessage(this); 087 } finally { 088 this.unknownFields = unknownFields.build(); 089 makeExtensionsImmutable(); 090 } 091 } 092 public static final com.google.protobuf.Descriptors.Descriptor 093 getDescriptor() { 094 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_descriptor; 095 } 096 097 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 098 internalGetFieldAccessorTable() { 099 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable 100 .ensureFieldAccessorsInitialized( 101 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.Builder.class); 102 } 103 104 public static com.google.protobuf.Parser<CedeActiveRequestProto> PARSER = 105 new com.google.protobuf.AbstractParser<CedeActiveRequestProto>() { 106 public CedeActiveRequestProto parsePartialFrom( 107 com.google.protobuf.CodedInputStream input, 108 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 109 throws com.google.protobuf.InvalidProtocolBufferException { 110 return new CedeActiveRequestProto(input, extensionRegistry); 111 } 112 }; 113 114 @java.lang.Override 115 public com.google.protobuf.Parser<CedeActiveRequestProto> getParserForType() { 116 return PARSER; 117 } 118 119 private int bitField0_; 120 // required uint32 millisToCede = 1; 121 public static final int MILLISTOCEDE_FIELD_NUMBER = 1; 122 private int millisToCede_; 123 /** 124 * <code>required uint32 millisToCede = 1;</code> 125 */ 126 public boolean hasMillisToCede() { 127 return ((bitField0_ & 0x00000001) == 0x00000001); 128 } 129 /** 130 * <code>required uint32 millisToCede = 1;</code> 131 */ 132 public int getMillisToCede() { 133 return millisToCede_; 134 } 135 136 private void initFields() { 137 millisToCede_ = 0; 138 } 139 private byte memoizedIsInitialized = -1; 140 public final boolean isInitialized() { 141 byte isInitialized = memoizedIsInitialized; 142 if (isInitialized != -1) return isInitialized == 1; 143 144 if (!hasMillisToCede()) { 145 memoizedIsInitialized = 0; 146 return false; 147 } 148 memoizedIsInitialized = 1; 149 return true; 150 } 151 152 public void writeTo(com.google.protobuf.CodedOutputStream output) 153 throws java.io.IOException { 154 getSerializedSize(); 155 if (((bitField0_ & 0x00000001) == 0x00000001)) { 156 output.writeUInt32(1, millisToCede_); 157 } 158 getUnknownFields().writeTo(output); 159 } 160 161 private int memoizedSerializedSize = -1; 162 public int getSerializedSize() { 163 int size = memoizedSerializedSize; 164 if (size != -1) return size; 165 166 size = 0; 167 if (((bitField0_ & 0x00000001) == 0x00000001)) { 168 size += com.google.protobuf.CodedOutputStream 169 .computeUInt32Size(1, millisToCede_); 170 } 171 size += getUnknownFields().getSerializedSize(); 172 memoizedSerializedSize = size; 173 return size; 174 } 175 176 private static final long serialVersionUID = 0L; 177 @java.lang.Override 178 protected java.lang.Object writeReplace() 179 throws java.io.ObjectStreamException { 180 return super.writeReplace(); 181 } 182 183 @java.lang.Override 184 public boolean equals(final java.lang.Object obj) { 185 if (obj == this) { 186 return true; 187 } 188 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)) { 189 return super.equals(obj); 190 } 191 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto) obj; 192 193 boolean result = true; 194 result = result && (hasMillisToCede() == other.hasMillisToCede()); 195 if (hasMillisToCede()) { 196 result = result && (getMillisToCede() 197 == other.getMillisToCede()); 198 } 199 result = result && 200 getUnknownFields().equals(other.getUnknownFields()); 201 return result; 202 } 203 204 private int memoizedHashCode = 0; 205 @java.lang.Override 206 public int hashCode() { 207 if (memoizedHashCode != 0) { 208 return memoizedHashCode; 209 } 210 int hash = 41; 211 hash = (19 * hash) + getDescriptorForType().hashCode(); 212 if (hasMillisToCede()) { 213 hash = (37 * hash) + MILLISTOCEDE_FIELD_NUMBER; 214 hash = (53 * hash) + getMillisToCede(); 215 } 216 hash = (29 * hash) + getUnknownFields().hashCode(); 217 memoizedHashCode = hash; 218 return hash; 219 } 220 221 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 222 com.google.protobuf.ByteString data) 223 throws com.google.protobuf.InvalidProtocolBufferException { 224 return PARSER.parseFrom(data); 225 } 226 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 227 com.google.protobuf.ByteString data, 228 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 229 throws com.google.protobuf.InvalidProtocolBufferException { 230 return PARSER.parseFrom(data, extensionRegistry); 231 } 232 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(byte[] data) 233 throws com.google.protobuf.InvalidProtocolBufferException { 234 return PARSER.parseFrom(data); 235 } 236 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 237 byte[] data, 238 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 239 throws com.google.protobuf.InvalidProtocolBufferException { 240 return PARSER.parseFrom(data, extensionRegistry); 241 } 242 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(java.io.InputStream input) 243 throws java.io.IOException { 244 return PARSER.parseFrom(input); 245 } 246 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 247 java.io.InputStream input, 248 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 249 throws java.io.IOException { 250 return PARSER.parseFrom(input, extensionRegistry); 251 } 252 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseDelimitedFrom(java.io.InputStream input) 253 throws java.io.IOException { 254 return PARSER.parseDelimitedFrom(input); 255 } 256 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseDelimitedFrom( 257 java.io.InputStream input, 258 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 259 throws java.io.IOException { 260 return PARSER.parseDelimitedFrom(input, extensionRegistry); 261 } 262 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 263 com.google.protobuf.CodedInputStream input) 264 throws java.io.IOException { 265 return PARSER.parseFrom(input); 266 } 267 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 268 com.google.protobuf.CodedInputStream input, 269 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 270 throws java.io.IOException { 271 return PARSER.parseFrom(input, extensionRegistry); 272 } 273 274 public static Builder newBuilder() { return Builder.create(); } 275 public Builder newBuilderForType() { return newBuilder(); } 276 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto prototype) { 277 return newBuilder().mergeFrom(prototype); 278 } 279 public Builder toBuilder() { return newBuilder(this); } 280 281 @java.lang.Override 282 protected Builder newBuilderForType( 283 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 284 Builder builder = new Builder(parent); 285 return builder; 286 } 287 /** 288 * Protobuf type {@code hadoop.common.CedeActiveRequestProto} 289 */ 290 public static final class Builder extends 291 com.google.protobuf.GeneratedMessage.Builder<Builder> 292 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProtoOrBuilder { 293 public static final com.google.protobuf.Descriptors.Descriptor 294 getDescriptor() { 295 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_descriptor; 296 } 297 298 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 299 internalGetFieldAccessorTable() { 300 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable 301 .ensureFieldAccessorsInitialized( 302 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.Builder.class); 303 } 304 305 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.newBuilder() 306 private Builder() { 307 maybeForceBuilderInitialization(); 308 } 309 310 private Builder( 311 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 312 super(parent); 313 maybeForceBuilderInitialization(); 314 } 315 private void maybeForceBuilderInitialization() { 316 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 317 } 318 } 319 private static Builder create() { 320 return new Builder(); 321 } 322 323 public Builder clear() { 324 super.clear(); 325 millisToCede_ = 0; 326 bitField0_ = (bitField0_ & ~0x00000001); 327 return this; 328 } 329 330 public Builder clone() { 331 return create().mergeFrom(buildPartial()); 332 } 333 334 public com.google.protobuf.Descriptors.Descriptor 335 getDescriptorForType() { 336 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_descriptor; 337 } 338 339 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto getDefaultInstanceForType() { 340 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance(); 341 } 342 343 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto build() { 344 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto result = buildPartial(); 345 if (!result.isInitialized()) { 346 throw newUninitializedMessageException(result); 347 } 348 return result; 349 } 350 351 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto buildPartial() { 352 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto(this); 353 int from_bitField0_ = bitField0_; 354 int to_bitField0_ = 0; 355 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 356 to_bitField0_ |= 0x00000001; 357 } 358 result.millisToCede_ = millisToCede_; 359 result.bitField0_ = to_bitField0_; 360 onBuilt(); 361 return result; 362 } 363 364 public Builder mergeFrom(com.google.protobuf.Message other) { 365 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto) { 366 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)other); 367 } else { 368 super.mergeFrom(other); 369 return this; 370 } 371 } 372 373 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto other) { 374 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance()) return this; 375 if (other.hasMillisToCede()) { 376 setMillisToCede(other.getMillisToCede()); 377 } 378 this.mergeUnknownFields(other.getUnknownFields()); 379 return this; 380 } 381 382 public final boolean isInitialized() { 383 if (!hasMillisToCede()) { 384 385 return false; 386 } 387 return true; 388 } 389 390 public Builder mergeFrom( 391 com.google.protobuf.CodedInputStream input, 392 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 393 throws java.io.IOException { 394 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parsedMessage = null; 395 try { 396 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 397 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 398 parsedMessage = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto) e.getUnfinishedMessage(); 399 throw e; 400 } finally { 401 if (parsedMessage != null) { 402 mergeFrom(parsedMessage); 403 } 404 } 405 return this; 406 } 407 private int bitField0_; 408 409 // required uint32 millisToCede = 1; 410 private int millisToCede_ ; 411 /** 412 * <code>required uint32 millisToCede = 1;</code> 413 */ 414 public boolean hasMillisToCede() { 415 return ((bitField0_ & 0x00000001) == 0x00000001); 416 } 417 /** 418 * <code>required uint32 millisToCede = 1;</code> 419 */ 420 public int getMillisToCede() { 421 return millisToCede_; 422 } 423 /** 424 * <code>required uint32 millisToCede = 1;</code> 425 */ 426 public Builder setMillisToCede(int value) { 427 bitField0_ |= 0x00000001; 428 millisToCede_ = value; 429 onChanged(); 430 return this; 431 } 432 /** 433 * <code>required uint32 millisToCede = 1;</code> 434 */ 435 public Builder clearMillisToCede() { 436 bitField0_ = (bitField0_ & ~0x00000001); 437 millisToCede_ = 0; 438 onChanged(); 439 return this; 440 } 441 442 // @@protoc_insertion_point(builder_scope:hadoop.common.CedeActiveRequestProto) 443 } 444 445 static { 446 defaultInstance = new CedeActiveRequestProto(true); 447 defaultInstance.initFields(); 448 } 449 450 // @@protoc_insertion_point(class_scope:hadoop.common.CedeActiveRequestProto) 451 } 452 453 public interface CedeActiveResponseProtoOrBuilder 454 extends com.google.protobuf.MessageOrBuilder { 455 } 456 /** 457 * Protobuf type {@code hadoop.common.CedeActiveResponseProto} 458 */ 459 public static final class CedeActiveResponseProto extends 460 com.google.protobuf.GeneratedMessage 461 implements CedeActiveResponseProtoOrBuilder { 462 // Use CedeActiveResponseProto.newBuilder() to construct. 463 private CedeActiveResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 464 super(builder); 465 this.unknownFields = builder.getUnknownFields(); 466 } 467 private CedeActiveResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 468 469 private static final CedeActiveResponseProto defaultInstance; 470 public static CedeActiveResponseProto getDefaultInstance() { 471 return defaultInstance; 472 } 473 474 public CedeActiveResponseProto getDefaultInstanceForType() { 475 return defaultInstance; 476 } 477 478 private final com.google.protobuf.UnknownFieldSet unknownFields; 479 @java.lang.Override 480 public final com.google.protobuf.UnknownFieldSet 481 getUnknownFields() { 482 return this.unknownFields; 483 } 484 private CedeActiveResponseProto( 485 com.google.protobuf.CodedInputStream input, 486 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 487 throws com.google.protobuf.InvalidProtocolBufferException { 488 initFields(); 489 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 490 com.google.protobuf.UnknownFieldSet.newBuilder(); 491 try { 492 boolean done = false; 493 while (!done) { 494 int tag = input.readTag(); 495 switch (tag) { 496 case 0: 497 done = true; 498 break; 499 default: { 500 if (!parseUnknownField(input, unknownFields, 501 extensionRegistry, tag)) { 502 done = true; 503 } 504 break; 505 } 506 } 507 } 508 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 509 throw e.setUnfinishedMessage(this); 510 } catch (java.io.IOException e) { 511 throw new com.google.protobuf.InvalidProtocolBufferException( 512 e.getMessage()).setUnfinishedMessage(this); 513 } finally { 514 this.unknownFields = unknownFields.build(); 515 makeExtensionsImmutable(); 516 } 517 } 518 public static final com.google.protobuf.Descriptors.Descriptor 519 getDescriptor() { 520 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_descriptor; 521 } 522 523 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 524 internalGetFieldAccessorTable() { 525 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable 526 .ensureFieldAccessorsInitialized( 527 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.Builder.class); 528 } 529 530 public static com.google.protobuf.Parser<CedeActiveResponseProto> PARSER = 531 new com.google.protobuf.AbstractParser<CedeActiveResponseProto>() { 532 public CedeActiveResponseProto parsePartialFrom( 533 com.google.protobuf.CodedInputStream input, 534 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 535 throws com.google.protobuf.InvalidProtocolBufferException { 536 return new CedeActiveResponseProto(input, extensionRegistry); 537 } 538 }; 539 540 @java.lang.Override 541 public com.google.protobuf.Parser<CedeActiveResponseProto> getParserForType() { 542 return PARSER; 543 } 544 545 private void initFields() { 546 } 547 private byte memoizedIsInitialized = -1; 548 public final boolean isInitialized() { 549 byte isInitialized = memoizedIsInitialized; 550 if (isInitialized != -1) return isInitialized == 1; 551 552 memoizedIsInitialized = 1; 553 return true; 554 } 555 556 public void writeTo(com.google.protobuf.CodedOutputStream output) 557 throws java.io.IOException { 558 getSerializedSize(); 559 getUnknownFields().writeTo(output); 560 } 561 562 private int memoizedSerializedSize = -1; 563 public int getSerializedSize() { 564 int size = memoizedSerializedSize; 565 if (size != -1) return size; 566 567 size = 0; 568 size += getUnknownFields().getSerializedSize(); 569 memoizedSerializedSize = size; 570 return size; 571 } 572 573 private static final long serialVersionUID = 0L; 574 @java.lang.Override 575 protected java.lang.Object writeReplace() 576 throws java.io.ObjectStreamException { 577 return super.writeReplace(); 578 } 579 580 @java.lang.Override 581 public boolean equals(final java.lang.Object obj) { 582 if (obj == this) { 583 return true; 584 } 585 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto)) { 586 return super.equals(obj); 587 } 588 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) obj; 589 590 boolean result = true; 591 result = result && 592 getUnknownFields().equals(other.getUnknownFields()); 593 return result; 594 } 595 596 private int memoizedHashCode = 0; 597 @java.lang.Override 598 public int hashCode() { 599 if (memoizedHashCode != 0) { 600 return memoizedHashCode; 601 } 602 int hash = 41; 603 hash = (19 * hash) + getDescriptorForType().hashCode(); 604 hash = (29 * hash) + getUnknownFields().hashCode(); 605 memoizedHashCode = hash; 606 return hash; 607 } 608 609 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 610 com.google.protobuf.ByteString data) 611 throws com.google.protobuf.InvalidProtocolBufferException { 612 return PARSER.parseFrom(data); 613 } 614 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 615 com.google.protobuf.ByteString data, 616 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 617 throws com.google.protobuf.InvalidProtocolBufferException { 618 return PARSER.parseFrom(data, extensionRegistry); 619 } 620 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(byte[] data) 621 throws com.google.protobuf.InvalidProtocolBufferException { 622 return PARSER.parseFrom(data); 623 } 624 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 625 byte[] data, 626 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 627 throws com.google.protobuf.InvalidProtocolBufferException { 628 return PARSER.parseFrom(data, extensionRegistry); 629 } 630 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(java.io.InputStream input) 631 throws java.io.IOException { 632 return PARSER.parseFrom(input); 633 } 634 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 635 java.io.InputStream input, 636 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 637 throws java.io.IOException { 638 return PARSER.parseFrom(input, extensionRegistry); 639 } 640 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseDelimitedFrom(java.io.InputStream input) 641 throws java.io.IOException { 642 return PARSER.parseDelimitedFrom(input); 643 } 644 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseDelimitedFrom( 645 java.io.InputStream input, 646 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 647 throws java.io.IOException { 648 return PARSER.parseDelimitedFrom(input, extensionRegistry); 649 } 650 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 651 com.google.protobuf.CodedInputStream input) 652 throws java.io.IOException { 653 return PARSER.parseFrom(input); 654 } 655 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 656 com.google.protobuf.CodedInputStream input, 657 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 658 throws java.io.IOException { 659 return PARSER.parseFrom(input, extensionRegistry); 660 } 661 662 public static Builder newBuilder() { return Builder.create(); } 663 public Builder newBuilderForType() { return newBuilder(); } 664 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto prototype) { 665 return newBuilder().mergeFrom(prototype); 666 } 667 public Builder toBuilder() { return newBuilder(this); } 668 669 @java.lang.Override 670 protected Builder newBuilderForType( 671 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 672 Builder builder = new Builder(parent); 673 return builder; 674 } 675 /** 676 * Protobuf type {@code hadoop.common.CedeActiveResponseProto} 677 */ 678 public static final class Builder extends 679 com.google.protobuf.GeneratedMessage.Builder<Builder> 680 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProtoOrBuilder { 681 public static final com.google.protobuf.Descriptors.Descriptor 682 getDescriptor() { 683 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_descriptor; 684 } 685 686 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 687 internalGetFieldAccessorTable() { 688 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable 689 .ensureFieldAccessorsInitialized( 690 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.Builder.class); 691 } 692 693 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.newBuilder() 694 private Builder() { 695 maybeForceBuilderInitialization(); 696 } 697 698 private Builder( 699 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 700 super(parent); 701 maybeForceBuilderInitialization(); 702 } 703 private void maybeForceBuilderInitialization() { 704 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 705 } 706 } 707 private static Builder create() { 708 return new Builder(); 709 } 710 711 public Builder clear() { 712 super.clear(); 713 return this; 714 } 715 716 public Builder clone() { 717 return create().mergeFrom(buildPartial()); 718 } 719 720 public com.google.protobuf.Descriptors.Descriptor 721 getDescriptorForType() { 722 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_descriptor; 723 } 724 725 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto getDefaultInstanceForType() { 726 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance(); 727 } 728 729 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto build() { 730 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto result = buildPartial(); 731 if (!result.isInitialized()) { 732 throw newUninitializedMessageException(result); 733 } 734 return result; 735 } 736 737 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto buildPartial() { 738 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto(this); 739 onBuilt(); 740 return result; 741 } 742 743 public Builder mergeFrom(com.google.protobuf.Message other) { 744 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) { 745 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto)other); 746 } else { 747 super.mergeFrom(other); 748 return this; 749 } 750 } 751 752 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto other) { 753 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance()) return this; 754 this.mergeUnknownFields(other.getUnknownFields()); 755 return this; 756 } 757 758 public final boolean isInitialized() { 759 return true; 760 } 761 762 public Builder mergeFrom( 763 com.google.protobuf.CodedInputStream input, 764 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 765 throws java.io.IOException { 766 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parsedMessage = null; 767 try { 768 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 769 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 770 parsedMessage = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) e.getUnfinishedMessage(); 771 throw e; 772 } finally { 773 if (parsedMessage != null) { 774 mergeFrom(parsedMessage); 775 } 776 } 777 return this; 778 } 779 780 // @@protoc_insertion_point(builder_scope:hadoop.common.CedeActiveResponseProto) 781 } 782 783 static { 784 defaultInstance = new CedeActiveResponseProto(true); 785 defaultInstance.initFields(); 786 } 787 788 // @@protoc_insertion_point(class_scope:hadoop.common.CedeActiveResponseProto) 789 } 790 791 public interface GracefulFailoverRequestProtoOrBuilder 792 extends com.google.protobuf.MessageOrBuilder { 793 } 794 /** 795 * Protobuf type {@code hadoop.common.GracefulFailoverRequestProto} 796 */ 797 public static final class GracefulFailoverRequestProto extends 798 com.google.protobuf.GeneratedMessage 799 implements GracefulFailoverRequestProtoOrBuilder { 800 // Use GracefulFailoverRequestProto.newBuilder() to construct. 801 private GracefulFailoverRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 802 super(builder); 803 this.unknownFields = builder.getUnknownFields(); 804 } 805 private GracefulFailoverRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 806 807 private static final GracefulFailoverRequestProto defaultInstance; 808 public static GracefulFailoverRequestProto getDefaultInstance() { 809 return defaultInstance; 810 } 811 812 public GracefulFailoverRequestProto getDefaultInstanceForType() { 813 return defaultInstance; 814 } 815 816 private final com.google.protobuf.UnknownFieldSet unknownFields; 817 @java.lang.Override 818 public final com.google.protobuf.UnknownFieldSet 819 getUnknownFields() { 820 return this.unknownFields; 821 } 822 private GracefulFailoverRequestProto( 823 com.google.protobuf.CodedInputStream input, 824 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 825 throws com.google.protobuf.InvalidProtocolBufferException { 826 initFields(); 827 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 828 com.google.protobuf.UnknownFieldSet.newBuilder(); 829 try { 830 boolean done = false; 831 while (!done) { 832 int tag = input.readTag(); 833 switch (tag) { 834 case 0: 835 done = true; 836 break; 837 default: { 838 if (!parseUnknownField(input, unknownFields, 839 extensionRegistry, tag)) { 840 done = true; 841 } 842 break; 843 } 844 } 845 } 846 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 847 throw e.setUnfinishedMessage(this); 848 } catch (java.io.IOException e) { 849 throw new com.google.protobuf.InvalidProtocolBufferException( 850 e.getMessage()).setUnfinishedMessage(this); 851 } finally { 852 this.unknownFields = unknownFields.build(); 853 makeExtensionsImmutable(); 854 } 855 } 856 public static final com.google.protobuf.Descriptors.Descriptor 857 getDescriptor() { 858 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor; 859 } 860 861 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 862 internalGetFieldAccessorTable() { 863 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable 864 .ensureFieldAccessorsInitialized( 865 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.Builder.class); 866 } 867 868 public static com.google.protobuf.Parser<GracefulFailoverRequestProto> PARSER = 869 new com.google.protobuf.AbstractParser<GracefulFailoverRequestProto>() { 870 public GracefulFailoverRequestProto parsePartialFrom( 871 com.google.protobuf.CodedInputStream input, 872 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 873 throws com.google.protobuf.InvalidProtocolBufferException { 874 return new GracefulFailoverRequestProto(input, extensionRegistry); 875 } 876 }; 877 878 @java.lang.Override 879 public com.google.protobuf.Parser<GracefulFailoverRequestProto> getParserForType() { 880 return PARSER; 881 } 882 883 private void initFields() { 884 } 885 private byte memoizedIsInitialized = -1; 886 public final boolean isInitialized() { 887 byte isInitialized = memoizedIsInitialized; 888 if (isInitialized != -1) return isInitialized == 1; 889 890 memoizedIsInitialized = 1; 891 return true; 892 } 893 894 public void writeTo(com.google.protobuf.CodedOutputStream output) 895 throws java.io.IOException { 896 getSerializedSize(); 897 getUnknownFields().writeTo(output); 898 } 899 900 private int memoizedSerializedSize = -1; 901 public int getSerializedSize() { 902 int size = memoizedSerializedSize; 903 if (size != -1) return size; 904 905 size = 0; 906 size += getUnknownFields().getSerializedSize(); 907 memoizedSerializedSize = size; 908 return size; 909 } 910 911 private static final long serialVersionUID = 0L; 912 @java.lang.Override 913 protected java.lang.Object writeReplace() 914 throws java.io.ObjectStreamException { 915 return super.writeReplace(); 916 } 917 918 @java.lang.Override 919 public boolean equals(final java.lang.Object obj) { 920 if (obj == this) { 921 return true; 922 } 923 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)) { 924 return super.equals(obj); 925 } 926 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto) obj; 927 928 boolean result = true; 929 result = result && 930 getUnknownFields().equals(other.getUnknownFields()); 931 return result; 932 } 933 934 private int memoizedHashCode = 0; 935 @java.lang.Override 936 public int hashCode() { 937 if (memoizedHashCode != 0) { 938 return memoizedHashCode; 939 } 940 int hash = 41; 941 hash = (19 * hash) + getDescriptorForType().hashCode(); 942 hash = (29 * hash) + getUnknownFields().hashCode(); 943 memoizedHashCode = hash; 944 return hash; 945 } 946 947 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 948 com.google.protobuf.ByteString data) 949 throws com.google.protobuf.InvalidProtocolBufferException { 950 return PARSER.parseFrom(data); 951 } 952 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 953 com.google.protobuf.ByteString data, 954 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 955 throws com.google.protobuf.InvalidProtocolBufferException { 956 return PARSER.parseFrom(data, extensionRegistry); 957 } 958 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(byte[] data) 959 throws com.google.protobuf.InvalidProtocolBufferException { 960 return PARSER.parseFrom(data); 961 } 962 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 963 byte[] data, 964 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 965 throws com.google.protobuf.InvalidProtocolBufferException { 966 return PARSER.parseFrom(data, extensionRegistry); 967 } 968 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(java.io.InputStream input) 969 throws java.io.IOException { 970 return PARSER.parseFrom(input); 971 } 972 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 973 java.io.InputStream input, 974 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 975 throws java.io.IOException { 976 return PARSER.parseFrom(input, extensionRegistry); 977 } 978 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseDelimitedFrom(java.io.InputStream input) 979 throws java.io.IOException { 980 return PARSER.parseDelimitedFrom(input); 981 } 982 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseDelimitedFrom( 983 java.io.InputStream input, 984 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 985 throws java.io.IOException { 986 return PARSER.parseDelimitedFrom(input, extensionRegistry); 987 } 988 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 989 com.google.protobuf.CodedInputStream input) 990 throws java.io.IOException { 991 return PARSER.parseFrom(input); 992 } 993 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 994 com.google.protobuf.CodedInputStream input, 995 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 996 throws java.io.IOException { 997 return PARSER.parseFrom(input, extensionRegistry); 998 } 999 1000 public static Builder newBuilder() { return Builder.create(); } 1001 public Builder newBuilderForType() { return newBuilder(); } 1002 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto prototype) { 1003 return newBuilder().mergeFrom(prototype); 1004 } 1005 public Builder toBuilder() { return newBuilder(this); } 1006 1007 @java.lang.Override 1008 protected Builder newBuilderForType( 1009 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1010 Builder builder = new Builder(parent); 1011 return builder; 1012 } 1013 /** 1014 * Protobuf type {@code hadoop.common.GracefulFailoverRequestProto} 1015 */ 1016 public static final class Builder extends 1017 com.google.protobuf.GeneratedMessage.Builder<Builder> 1018 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProtoOrBuilder { 1019 public static final com.google.protobuf.Descriptors.Descriptor 1020 getDescriptor() { 1021 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor; 1022 } 1023 1024 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1025 internalGetFieldAccessorTable() { 1026 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable 1027 .ensureFieldAccessorsInitialized( 1028 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.Builder.class); 1029 } 1030 1031 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.newBuilder() 1032 private Builder() { 1033 maybeForceBuilderInitialization(); 1034 } 1035 1036 private Builder( 1037 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1038 super(parent); 1039 maybeForceBuilderInitialization(); 1040 } 1041 private void maybeForceBuilderInitialization() { 1042 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1043 } 1044 } 1045 private static Builder create() { 1046 return new Builder(); 1047 } 1048 1049 public Builder clear() { 1050 super.clear(); 1051 return this; 1052 } 1053 1054 public Builder clone() { 1055 return create().mergeFrom(buildPartial()); 1056 } 1057 1058 public com.google.protobuf.Descriptors.Descriptor 1059 getDescriptorForType() { 1060 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor; 1061 } 1062 1063 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto getDefaultInstanceForType() { 1064 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance(); 1065 } 1066 1067 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto build() { 1068 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto result = buildPartial(); 1069 if (!result.isInitialized()) { 1070 throw newUninitializedMessageException(result); 1071 } 1072 return result; 1073 } 1074 1075 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto buildPartial() { 1076 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto(this); 1077 onBuilt(); 1078 return result; 1079 } 1080 1081 public Builder mergeFrom(com.google.protobuf.Message other) { 1082 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto) { 1083 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)other); 1084 } else { 1085 super.mergeFrom(other); 1086 return this; 1087 } 1088 } 1089 1090 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto other) { 1091 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance()) return this; 1092 this.mergeUnknownFields(other.getUnknownFields()); 1093 return this; 1094 } 1095 1096 public final boolean isInitialized() { 1097 return true; 1098 } 1099 1100 public Builder mergeFrom( 1101 com.google.protobuf.CodedInputStream input, 1102 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1103 throws java.io.IOException { 1104 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parsedMessage = null; 1105 try { 1106 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 1107 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1108 parsedMessage = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto) e.getUnfinishedMessage(); 1109 throw e; 1110 } finally { 1111 if (parsedMessage != null) { 1112 mergeFrom(parsedMessage); 1113 } 1114 } 1115 return this; 1116 } 1117 1118 // @@protoc_insertion_point(builder_scope:hadoop.common.GracefulFailoverRequestProto) 1119 } 1120 1121 static { 1122 defaultInstance = new GracefulFailoverRequestProto(true); 1123 defaultInstance.initFields(); 1124 } 1125 1126 // @@protoc_insertion_point(class_scope:hadoop.common.GracefulFailoverRequestProto) 1127 } 1128 1129 public interface GracefulFailoverResponseProtoOrBuilder 1130 extends com.google.protobuf.MessageOrBuilder { 1131 } 1132 /** 1133 * Protobuf type {@code hadoop.common.GracefulFailoverResponseProto} 1134 */ 1135 public static final class GracefulFailoverResponseProto extends 1136 com.google.protobuf.GeneratedMessage 1137 implements GracefulFailoverResponseProtoOrBuilder { 1138 // Use GracefulFailoverResponseProto.newBuilder() to construct. 1139 private GracefulFailoverResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { 1140 super(builder); 1141 this.unknownFields = builder.getUnknownFields(); 1142 } 1143 private GracefulFailoverResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } 1144 1145 private static final GracefulFailoverResponseProto defaultInstance; 1146 public static GracefulFailoverResponseProto getDefaultInstance() { 1147 return defaultInstance; 1148 } 1149 1150 public GracefulFailoverResponseProto getDefaultInstanceForType() { 1151 return defaultInstance; 1152 } 1153 1154 private final com.google.protobuf.UnknownFieldSet unknownFields; 1155 @java.lang.Override 1156 public final com.google.protobuf.UnknownFieldSet 1157 getUnknownFields() { 1158 return this.unknownFields; 1159 } 1160 private GracefulFailoverResponseProto( 1161 com.google.protobuf.CodedInputStream input, 1162 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1163 throws com.google.protobuf.InvalidProtocolBufferException { 1164 initFields(); 1165 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1166 com.google.protobuf.UnknownFieldSet.newBuilder(); 1167 try { 1168 boolean done = false; 1169 while (!done) { 1170 int tag = input.readTag(); 1171 switch (tag) { 1172 case 0: 1173 done = true; 1174 break; 1175 default: { 1176 if (!parseUnknownField(input, unknownFields, 1177 extensionRegistry, tag)) { 1178 done = true; 1179 } 1180 break; 1181 } 1182 } 1183 } 1184 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1185 throw e.setUnfinishedMessage(this); 1186 } catch (java.io.IOException e) { 1187 throw new com.google.protobuf.InvalidProtocolBufferException( 1188 e.getMessage()).setUnfinishedMessage(this); 1189 } finally { 1190 this.unknownFields = unknownFields.build(); 1191 makeExtensionsImmutable(); 1192 } 1193 } 1194 public static final com.google.protobuf.Descriptors.Descriptor 1195 getDescriptor() { 1196 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor; 1197 } 1198 1199 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1200 internalGetFieldAccessorTable() { 1201 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable 1202 .ensureFieldAccessorsInitialized( 1203 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.Builder.class); 1204 } 1205 1206 public static com.google.protobuf.Parser<GracefulFailoverResponseProto> PARSER = 1207 new com.google.protobuf.AbstractParser<GracefulFailoverResponseProto>() { 1208 public GracefulFailoverResponseProto parsePartialFrom( 1209 com.google.protobuf.CodedInputStream input, 1210 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1211 throws com.google.protobuf.InvalidProtocolBufferException { 1212 return new GracefulFailoverResponseProto(input, extensionRegistry); 1213 } 1214 }; 1215 1216 @java.lang.Override 1217 public com.google.protobuf.Parser<GracefulFailoverResponseProto> getParserForType() { 1218 return PARSER; 1219 } 1220 1221 private void initFields() { 1222 } 1223 private byte memoizedIsInitialized = -1; 1224 public final boolean isInitialized() { 1225 byte isInitialized = memoizedIsInitialized; 1226 if (isInitialized != -1) return isInitialized == 1; 1227 1228 memoizedIsInitialized = 1; 1229 return true; 1230 } 1231 1232 public void writeTo(com.google.protobuf.CodedOutputStream output) 1233 throws java.io.IOException { 1234 getSerializedSize(); 1235 getUnknownFields().writeTo(output); 1236 } 1237 1238 private int memoizedSerializedSize = -1; 1239 public int getSerializedSize() { 1240 int size = memoizedSerializedSize; 1241 if (size != -1) return size; 1242 1243 size = 0; 1244 size += getUnknownFields().getSerializedSize(); 1245 memoizedSerializedSize = size; 1246 return size; 1247 } 1248 1249 private static final long serialVersionUID = 0L; 1250 @java.lang.Override 1251 protected java.lang.Object writeReplace() 1252 throws java.io.ObjectStreamException { 1253 return super.writeReplace(); 1254 } 1255 1256 @java.lang.Override 1257 public boolean equals(final java.lang.Object obj) { 1258 if (obj == this) { 1259 return true; 1260 } 1261 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto)) { 1262 return super.equals(obj); 1263 } 1264 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) obj; 1265 1266 boolean result = true; 1267 result = result && 1268 getUnknownFields().equals(other.getUnknownFields()); 1269 return result; 1270 } 1271 1272 private int memoizedHashCode = 0; 1273 @java.lang.Override 1274 public int hashCode() { 1275 if (memoizedHashCode != 0) { 1276 return memoizedHashCode; 1277 } 1278 int hash = 41; 1279 hash = (19 * hash) + getDescriptorForType().hashCode(); 1280 hash = (29 * hash) + getUnknownFields().hashCode(); 1281 memoizedHashCode = hash; 1282 return hash; 1283 } 1284 1285 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1286 com.google.protobuf.ByteString data) 1287 throws com.google.protobuf.InvalidProtocolBufferException { 1288 return PARSER.parseFrom(data); 1289 } 1290 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1291 com.google.protobuf.ByteString data, 1292 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1293 throws com.google.protobuf.InvalidProtocolBufferException { 1294 return PARSER.parseFrom(data, extensionRegistry); 1295 } 1296 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(byte[] data) 1297 throws com.google.protobuf.InvalidProtocolBufferException { 1298 return PARSER.parseFrom(data); 1299 } 1300 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1301 byte[] data, 1302 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1303 throws com.google.protobuf.InvalidProtocolBufferException { 1304 return PARSER.parseFrom(data, extensionRegistry); 1305 } 1306 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(java.io.InputStream input) 1307 throws java.io.IOException { 1308 return PARSER.parseFrom(input); 1309 } 1310 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1311 java.io.InputStream input, 1312 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1313 throws java.io.IOException { 1314 return PARSER.parseFrom(input, extensionRegistry); 1315 } 1316 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseDelimitedFrom(java.io.InputStream input) 1317 throws java.io.IOException { 1318 return PARSER.parseDelimitedFrom(input); 1319 } 1320 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseDelimitedFrom( 1321 java.io.InputStream input, 1322 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1323 throws java.io.IOException { 1324 return PARSER.parseDelimitedFrom(input, extensionRegistry); 1325 } 1326 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1327 com.google.protobuf.CodedInputStream input) 1328 throws java.io.IOException { 1329 return PARSER.parseFrom(input); 1330 } 1331 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1332 com.google.protobuf.CodedInputStream input, 1333 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1334 throws java.io.IOException { 1335 return PARSER.parseFrom(input, extensionRegistry); 1336 } 1337 1338 public static Builder newBuilder() { return Builder.create(); } 1339 public Builder newBuilderForType() { return newBuilder(); } 1340 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto prototype) { 1341 return newBuilder().mergeFrom(prototype); 1342 } 1343 public Builder toBuilder() { return newBuilder(this); } 1344 1345 @java.lang.Override 1346 protected Builder newBuilderForType( 1347 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1348 Builder builder = new Builder(parent); 1349 return builder; 1350 } 1351 /** 1352 * Protobuf type {@code hadoop.common.GracefulFailoverResponseProto} 1353 */ 1354 public static final class Builder extends 1355 com.google.protobuf.GeneratedMessage.Builder<Builder> 1356 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProtoOrBuilder { 1357 public static final com.google.protobuf.Descriptors.Descriptor 1358 getDescriptor() { 1359 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor; 1360 } 1361 1362 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1363 internalGetFieldAccessorTable() { 1364 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable 1365 .ensureFieldAccessorsInitialized( 1366 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.Builder.class); 1367 } 1368 1369 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.newBuilder() 1370 private Builder() { 1371 maybeForceBuilderInitialization(); 1372 } 1373 1374 private Builder( 1375 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1376 super(parent); 1377 maybeForceBuilderInitialization(); 1378 } 1379 private void maybeForceBuilderInitialization() { 1380 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1381 } 1382 } 1383 private static Builder create() { 1384 return new Builder(); 1385 } 1386 1387 public Builder clear() { 1388 super.clear(); 1389 return this; 1390 } 1391 1392 public Builder clone() { 1393 return create().mergeFrom(buildPartial()); 1394 } 1395 1396 public com.google.protobuf.Descriptors.Descriptor 1397 getDescriptorForType() { 1398 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor; 1399 } 1400 1401 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto getDefaultInstanceForType() { 1402 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance(); 1403 } 1404 1405 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto build() { 1406 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto result = buildPartial(); 1407 if (!result.isInitialized()) { 1408 throw newUninitializedMessageException(result); 1409 } 1410 return result; 1411 } 1412 1413 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto buildPartial() { 1414 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto(this); 1415 onBuilt(); 1416 return result; 1417 } 1418 1419 public Builder mergeFrom(com.google.protobuf.Message other) { 1420 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) { 1421 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto)other); 1422 } else { 1423 super.mergeFrom(other); 1424 return this; 1425 } 1426 } 1427 1428 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto other) { 1429 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance()) return this; 1430 this.mergeUnknownFields(other.getUnknownFields()); 1431 return this; 1432 } 1433 1434 public final boolean isInitialized() { 1435 return true; 1436 } 1437 1438 public Builder mergeFrom( 1439 com.google.protobuf.CodedInputStream input, 1440 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1441 throws java.io.IOException { 1442 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parsedMessage = null; 1443 try { 1444 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); 1445 } catch (com.google.protobuf.InvalidProtocolBufferException e) { 1446 parsedMessage = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) e.getUnfinishedMessage(); 1447 throw e; 1448 } finally { 1449 if (parsedMessage != null) { 1450 mergeFrom(parsedMessage); 1451 } 1452 } 1453 return this; 1454 } 1455 1456 // @@protoc_insertion_point(builder_scope:hadoop.common.GracefulFailoverResponseProto) 1457 } 1458 1459 static { 1460 defaultInstance = new GracefulFailoverResponseProto(true); 1461 defaultInstance.initFields(); 1462 } 1463 1464 // @@protoc_insertion_point(class_scope:hadoop.common.GracefulFailoverResponseProto) 1465 } 1466 1467 /** 1468 * Protobuf service {@code hadoop.common.ZKFCProtocolService} 1469 * 1470 * <pre> 1471 ** 1472 * Protocol provides manual control of the ZK Failover Controllers 1473 * </pre> 1474 */ 1475 public static abstract class ZKFCProtocolService 1476 implements com.google.protobuf.Service { 1477 protected ZKFCProtocolService() {} 1478 1479 public interface Interface { 1480 /** 1481 * <code>rpc cedeActive(.hadoop.common.CedeActiveRequestProto) returns (.hadoop.common.CedeActiveResponseProto);</code> 1482 * 1483 * <pre> 1484 ** 1485 * Request that the service cede its active state, and quit the election 1486 * for some amount of time 1487 * </pre> 1488 */ 1489 public abstract void cedeActive( 1490 com.google.protobuf.RpcController controller, 1491 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request, 1492 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done); 1493 1494 /** 1495 * <code>rpc gracefulFailover(.hadoop.common.GracefulFailoverRequestProto) returns (.hadoop.common.GracefulFailoverResponseProto);</code> 1496 */ 1497 public abstract void gracefulFailover( 1498 com.google.protobuf.RpcController controller, 1499 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request, 1500 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done); 1501 1502 } 1503 1504 public static com.google.protobuf.Service newReflectiveService( 1505 final Interface impl) { 1506 return new ZKFCProtocolService() { 1507 @java.lang.Override 1508 public void cedeActive( 1509 com.google.protobuf.RpcController controller, 1510 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request, 1511 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done) { 1512 impl.cedeActive(controller, request, done); 1513 } 1514 1515 @java.lang.Override 1516 public void gracefulFailover( 1517 com.google.protobuf.RpcController controller, 1518 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request, 1519 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done) { 1520 impl.gracefulFailover(controller, request, done); 1521 } 1522 1523 }; 1524 } 1525 1526 public static com.google.protobuf.BlockingService 1527 newReflectiveBlockingService(final BlockingInterface impl) { 1528 return new com.google.protobuf.BlockingService() { 1529 public final com.google.protobuf.Descriptors.ServiceDescriptor 1530 getDescriptorForType() { 1531 return getDescriptor(); 1532 } 1533 1534 public final com.google.protobuf.Message callBlockingMethod( 1535 com.google.protobuf.Descriptors.MethodDescriptor method, 1536 com.google.protobuf.RpcController controller, 1537 com.google.protobuf.Message request) 1538 throws com.google.protobuf.ServiceException { 1539 if (method.getService() != getDescriptor()) { 1540 throw new java.lang.IllegalArgumentException( 1541 "Service.callBlockingMethod() given method descriptor for " + 1542 "wrong service type."); 1543 } 1544 switch(method.getIndex()) { 1545 case 0: 1546 return impl.cedeActive(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)request); 1547 case 1: 1548 return impl.gracefulFailover(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)request); 1549 default: 1550 throw new java.lang.AssertionError("Can't get here."); 1551 } 1552 } 1553 1554 public final com.google.protobuf.Message 1555 getRequestPrototype( 1556 com.google.protobuf.Descriptors.MethodDescriptor method) { 1557 if (method.getService() != getDescriptor()) { 1558 throw new java.lang.IllegalArgumentException( 1559 "Service.getRequestPrototype() given method " + 1560 "descriptor for wrong service type."); 1561 } 1562 switch(method.getIndex()) { 1563 case 0: 1564 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance(); 1565 case 1: 1566 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance(); 1567 default: 1568 throw new java.lang.AssertionError("Can't get here."); 1569 } 1570 } 1571 1572 public final com.google.protobuf.Message 1573 getResponsePrototype( 1574 com.google.protobuf.Descriptors.MethodDescriptor method) { 1575 if (method.getService() != getDescriptor()) { 1576 throw new java.lang.IllegalArgumentException( 1577 "Service.getResponsePrototype() given method " + 1578 "descriptor for wrong service type."); 1579 } 1580 switch(method.getIndex()) { 1581 case 0: 1582 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance(); 1583 case 1: 1584 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance(); 1585 default: 1586 throw new java.lang.AssertionError("Can't get here."); 1587 } 1588 } 1589 1590 }; 1591 } 1592 1593 /** 1594 * <code>rpc cedeActive(.hadoop.common.CedeActiveRequestProto) returns (.hadoop.common.CedeActiveResponseProto);</code> 1595 * 1596 * <pre> 1597 ** 1598 * Request that the service cede its active state, and quit the election 1599 * for some amount of time 1600 * </pre> 1601 */ 1602 public abstract void cedeActive( 1603 com.google.protobuf.RpcController controller, 1604 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request, 1605 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done); 1606 1607 /** 1608 * <code>rpc gracefulFailover(.hadoop.common.GracefulFailoverRequestProto) returns (.hadoop.common.GracefulFailoverResponseProto);</code> 1609 */ 1610 public abstract void gracefulFailover( 1611 com.google.protobuf.RpcController controller, 1612 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request, 1613 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done); 1614 1615 public static final 1616 com.google.protobuf.Descriptors.ServiceDescriptor 1617 getDescriptor() { 1618 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.getDescriptor().getServices().get(0); 1619 } 1620 public final com.google.protobuf.Descriptors.ServiceDescriptor 1621 getDescriptorForType() { 1622 return getDescriptor(); 1623 } 1624 1625 public final void callMethod( 1626 com.google.protobuf.Descriptors.MethodDescriptor method, 1627 com.google.protobuf.RpcController controller, 1628 com.google.protobuf.Message request, 1629 com.google.protobuf.RpcCallback< 1630 com.google.protobuf.Message> done) { 1631 if (method.getService() != getDescriptor()) { 1632 throw new java.lang.IllegalArgumentException( 1633 "Service.callMethod() given method descriptor for wrong " + 1634 "service type."); 1635 } 1636 switch(method.getIndex()) { 1637 case 0: 1638 this.cedeActive(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)request, 1639 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto>specializeCallback( 1640 done)); 1641 return; 1642 case 1: 1643 this.gracefulFailover(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)request, 1644 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto>specializeCallback( 1645 done)); 1646 return; 1647 default: 1648 throw new java.lang.AssertionError("Can't get here."); 1649 } 1650 } 1651 1652 public final com.google.protobuf.Message 1653 getRequestPrototype( 1654 com.google.protobuf.Descriptors.MethodDescriptor method) { 1655 if (method.getService() != getDescriptor()) { 1656 throw new java.lang.IllegalArgumentException( 1657 "Service.getRequestPrototype() given method " + 1658 "descriptor for wrong service type."); 1659 } 1660 switch(method.getIndex()) { 1661 case 0: 1662 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance(); 1663 case 1: 1664 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance(); 1665 default: 1666 throw new java.lang.AssertionError("Can't get here."); 1667 } 1668 } 1669 1670 public final com.google.protobuf.Message 1671 getResponsePrototype( 1672 com.google.protobuf.Descriptors.MethodDescriptor method) { 1673 if (method.getService() != getDescriptor()) { 1674 throw new java.lang.IllegalArgumentException( 1675 "Service.getResponsePrototype() given method " + 1676 "descriptor for wrong service type."); 1677 } 1678 switch(method.getIndex()) { 1679 case 0: 1680 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance(); 1681 case 1: 1682 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance(); 1683 default: 1684 throw new java.lang.AssertionError("Can't get here."); 1685 } 1686 } 1687 1688 public static Stub newStub( 1689 com.google.protobuf.RpcChannel channel) { 1690 return new Stub(channel); 1691 } 1692 1693 public static final class Stub extends org.apache.hadoop.ha.proto.ZKFCProtocolProtos.ZKFCProtocolService implements Interface { 1694 private Stub(com.google.protobuf.RpcChannel channel) { 1695 this.channel = channel; 1696 } 1697 1698 private final com.google.protobuf.RpcChannel channel; 1699 1700 public com.google.protobuf.RpcChannel getChannel() { 1701 return channel; 1702 } 1703 1704 public void cedeActive( 1705 com.google.protobuf.RpcController controller, 1706 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request, 1707 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done) { 1708 channel.callMethod( 1709 getDescriptor().getMethods().get(0), 1710 controller, 1711 request, 1712 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance(), 1713 com.google.protobuf.RpcUtil.generalizeCallback( 1714 done, 1715 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.class, 1716 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance())); 1717 } 1718 1719 public void gracefulFailover( 1720 com.google.protobuf.RpcController controller, 1721 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request, 1722 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done) { 1723 channel.callMethod( 1724 getDescriptor().getMethods().get(1), 1725 controller, 1726 request, 1727 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance(), 1728 com.google.protobuf.RpcUtil.generalizeCallback( 1729 done, 1730 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.class, 1731 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance())); 1732 } 1733 } 1734 1735 public static BlockingInterface newBlockingStub( 1736 com.google.protobuf.BlockingRpcChannel channel) { 1737 return new BlockingStub(channel); 1738 } 1739 1740 public interface BlockingInterface { 1741 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto cedeActive( 1742 com.google.protobuf.RpcController controller, 1743 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request) 1744 throws com.google.protobuf.ServiceException; 1745 1746 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto gracefulFailover( 1747 com.google.protobuf.RpcController controller, 1748 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request) 1749 throws com.google.protobuf.ServiceException; 1750 } 1751 1752 private static final class BlockingStub implements BlockingInterface { 1753 private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { 1754 this.channel = channel; 1755 } 1756 1757 private final com.google.protobuf.BlockingRpcChannel channel; 1758 1759 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto cedeActive( 1760 com.google.protobuf.RpcController controller, 1761 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request) 1762 throws com.google.protobuf.ServiceException { 1763 return (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) channel.callBlockingMethod( 1764 getDescriptor().getMethods().get(0), 1765 controller, 1766 request, 1767 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance()); 1768 } 1769 1770 1771 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto gracefulFailover( 1772 com.google.protobuf.RpcController controller, 1773 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request) 1774 throws com.google.protobuf.ServiceException { 1775 return (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) channel.callBlockingMethod( 1776 getDescriptor().getMethods().get(1), 1777 controller, 1778 request, 1779 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance()); 1780 } 1781 1782 } 1783 1784 // @@protoc_insertion_point(class_scope:hadoop.common.ZKFCProtocolService) 1785 } 1786 1787 private static com.google.protobuf.Descriptors.Descriptor 1788 internal_static_hadoop_common_CedeActiveRequestProto_descriptor; 1789 private static 1790 com.google.protobuf.GeneratedMessage.FieldAccessorTable 1791 internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable; 1792 private static com.google.protobuf.Descriptors.Descriptor 1793 internal_static_hadoop_common_CedeActiveResponseProto_descriptor; 1794 private static 1795 com.google.protobuf.GeneratedMessage.FieldAccessorTable 1796 internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable; 1797 private static com.google.protobuf.Descriptors.Descriptor 1798 internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor; 1799 private static 1800 com.google.protobuf.GeneratedMessage.FieldAccessorTable 1801 internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable; 1802 private static com.google.protobuf.Descriptors.Descriptor 1803 internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor; 1804 private static 1805 com.google.protobuf.GeneratedMessage.FieldAccessorTable 1806 internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable; 1807 1808 public static com.google.protobuf.Descriptors.FileDescriptor 1809 getDescriptor() { 1810 return descriptor; 1811 } 1812 private static com.google.protobuf.Descriptors.FileDescriptor 1813 descriptor; 1814 static { 1815 java.lang.String[] descriptorData = { 1816 "\n\022ZKFCProtocol.proto\022\rhadoop.common\".\n\026C" + 1817 "edeActiveRequestProto\022\024\n\014millisToCede\030\001 " + 1818 "\002(\r\"\031\n\027CedeActiveResponseProto\"\036\n\034Gracef" + 1819 "ulFailoverRequestProto\"\037\n\035GracefulFailov" + 1820 "erResponseProto2\341\001\n\023ZKFCProtocolService\022" + 1821 "[\n\ncedeActive\022%.hadoop.common.CedeActive" + 1822 "RequestProto\032&.hadoop.common.CedeActiveR" + 1823 "esponseProto\022m\n\020gracefulFailover\022+.hadoo" + 1824 "p.common.GracefulFailoverRequestProto\032,." + 1825 "hadoop.common.GracefulFailoverResponsePr", 1826 "otoB6\n\032org.apache.hadoop.ha.protoB\022ZKFCP" + 1827 "rotocolProtos\210\001\001\240\001\001" 1828 }; 1829 com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = 1830 new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { 1831 public com.google.protobuf.ExtensionRegistry assignDescriptors( 1832 com.google.protobuf.Descriptors.FileDescriptor root) { 1833 descriptor = root; 1834 internal_static_hadoop_common_CedeActiveRequestProto_descriptor = 1835 getDescriptor().getMessageTypes().get(0); 1836 internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable = new 1837 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 1838 internal_static_hadoop_common_CedeActiveRequestProto_descriptor, 1839 new java.lang.String[] { "MillisToCede", }); 1840 internal_static_hadoop_common_CedeActiveResponseProto_descriptor = 1841 getDescriptor().getMessageTypes().get(1); 1842 internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable = new 1843 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 1844 internal_static_hadoop_common_CedeActiveResponseProto_descriptor, 1845 new java.lang.String[] { }); 1846 internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor = 1847 getDescriptor().getMessageTypes().get(2); 1848 internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable = new 1849 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 1850 internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor, 1851 new java.lang.String[] { }); 1852 internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor = 1853 getDescriptor().getMessageTypes().get(3); 1854 internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable = new 1855 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 1856 internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor, 1857 new java.lang.String[] { }); 1858 return null; 1859 } 1860 }; 1861 com.google.protobuf.Descriptors.FileDescriptor 1862 .internalBuildGeneratedFileFrom(descriptorData, 1863 new com.google.protobuf.Descriptors.FileDescriptor[] { 1864 }, assigner); 1865 } 1866 1867 // @@protoc_insertion_point(outer_class_scope) 1868}