Class ClusterLoadAssignment.Policy.Builder
java.lang.Object
com.google.protobuf.AbstractMessageLite.Builder
com.google.protobuf.AbstractMessage.Builder<ClusterLoadAssignment.Policy.Builder>
com.google.protobuf.GeneratedMessage.Builder<ClusterLoadAssignment.Policy.Builder>
io.envoyproxy.envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.Builder
- All Implemented Interfaces:
com.google.protobuf.Message.Builder,com.google.protobuf.MessageLite.Builder,com.google.protobuf.MessageLiteOrBuilder,com.google.protobuf.MessageOrBuilder,ClusterLoadAssignment.PolicyOrBuilder,Cloneable
- Enclosing class:
ClusterLoadAssignment.Policy
public static final class ClusterLoadAssignment.Policy.Builder
extends com.google.protobuf.GeneratedMessage.Builder<ClusterLoadAssignment.Policy.Builder>
implements ClusterLoadAssignment.PolicyOrBuilder
Load balancing policy settings. [#next-free-field: 7]Protobuf type
envoy.config.endpoint.v3.ClusterLoadAssignment.Policy-
Field Summary
FieldsModifier and TypeFieldDescriptionprivate intprivate com.google.protobuf.RepeatedFieldBuilder<ClusterLoadAssignment.Policy.DropOverload, ClusterLoadAssignment.Policy.DropOverload.Builder, ClusterLoadAssignment.Policy.DropOverloadOrBuilder> private com.google.protobuf.Durationprivate com.google.protobuf.SingleFieldBuilder<com.google.protobuf.Duration, com.google.protobuf.Duration.Builder, com.google.protobuf.DurationOrBuilder> private com.google.protobuf.UInt32Valueprivate com.google.protobuf.SingleFieldBuilder<com.google.protobuf.UInt32Value, com.google.protobuf.UInt32Value.Builder, com.google.protobuf.UInt32ValueOrBuilder> private boolean -
Constructor Summary
Constructors -
Method Summary
Modifier and TypeMethodDescriptionaddAllDropOverloads(Iterable<? extends ClusterLoadAssignment.Policy.DropOverload> values) Action to trim the overall incoming traffic to protect the upstream hosts.addDropOverloads(int index, ClusterLoadAssignment.Policy.DropOverload value) Action to trim the overall incoming traffic to protect the upstream hosts.addDropOverloads(int index, ClusterLoadAssignment.Policy.DropOverload.Builder builderForValue) Action to trim the overall incoming traffic to protect the upstream hosts.Action to trim the overall incoming traffic to protect the upstream hosts.addDropOverloads(ClusterLoadAssignment.Policy.DropOverload.Builder builderForValue) Action to trim the overall incoming traffic to protect the upstream hosts.Action to trim the overall incoming traffic to protect the upstream hosts.addDropOverloadsBuilder(int index) Action to trim the overall incoming traffic to protect the upstream hosts.build()private voidprivate voidclear()Action to trim the overall incoming traffic to protect the upstream hosts.The max time until which the endpoints from this assignment can be used.Priority levels and localities are considered overprovisioned with this factor (in percentage).If true, use the :ref:`load balancing weight <envoy_v3_api_field_config.endpoint.v3.LbEndpoint.load_balancing_weight>` of healthy and unhealthy hosts to determine the health of the priority level.private voidstatic final com.google.protobuf.Descriptors.Descriptorcom.google.protobuf.Descriptors.DescriptorgetDropOverloads(int index) Action to trim the overall incoming traffic to protect the upstream hosts.getDropOverloadsBuilder(int index) Action to trim the overall incoming traffic to protect the upstream hosts.Action to trim the overall incoming traffic to protect the upstream hosts.intAction to trim the overall incoming traffic to protect the upstream hosts.Action to trim the overall incoming traffic to protect the upstream hosts.getDropOverloadsOrBuilder(int index) Action to trim the overall incoming traffic to protect the upstream hosts.Action to trim the overall incoming traffic to protect the upstream hosts.com.google.protobuf.DurationThe max time until which the endpoints from this assignment can be used.com.google.protobuf.Duration.BuilderThe max time until which the endpoints from this assignment can be used.com.google.protobuf.DurationOrBuilderThe max time until which the endpoints from this assignment can be used.com.google.protobuf.UInt32ValuePriority levels and localities are considered overprovisioned with this factor (in percentage).com.google.protobuf.UInt32Value.BuilderPriority levels and localities are considered overprovisioned with this factor (in percentage).com.google.protobuf.UInt32ValueOrBuilderPriority levels and localities are considered overprovisioned with this factor (in percentage).booleanIf true, use the :ref:`load balancing weight <envoy_v3_api_field_config.endpoint.v3.LbEndpoint.load_balancing_weight>` of healthy and unhealthy hosts to determine the health of the priority level.booleanThe max time until which the endpoints from this assignment can be used.booleanPriority levels and localities are considered overprovisioned with this factor (in percentage).private com.google.protobuf.RepeatedFieldBuilder<ClusterLoadAssignment.Policy.DropOverload, ClusterLoadAssignment.Policy.DropOverload.Builder, ClusterLoadAssignment.Policy.DropOverloadOrBuilder> private com.google.protobuf.SingleFieldBuilder<com.google.protobuf.Duration, com.google.protobuf.Duration.Builder, com.google.protobuf.DurationOrBuilder> The max time until which the endpoints from this assignment can be used.protected com.google.protobuf.GeneratedMessage.FieldAccessorTableprivate com.google.protobuf.SingleFieldBuilder<com.google.protobuf.UInt32Value, com.google.protobuf.UInt32Value.Builder, com.google.protobuf.UInt32ValueOrBuilder> Priority levels and localities are considered overprovisioned with this factor (in percentage).final booleanprivate voidmergeEndpointStaleAfter(com.google.protobuf.Duration value) The max time until which the endpoints from this assignment can be used.mergeFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) mergeFrom(com.google.protobuf.Message other) mergeOverprovisioningFactor(com.google.protobuf.UInt32Value value) Priority levels and localities are considered overprovisioned with this factor (in percentage).removeDropOverloads(int index) Action to trim the overall incoming traffic to protect the upstream hosts.setDropOverloads(int index, ClusterLoadAssignment.Policy.DropOverload value) Action to trim the overall incoming traffic to protect the upstream hosts.setDropOverloads(int index, ClusterLoadAssignment.Policy.DropOverload.Builder builderForValue) Action to trim the overall incoming traffic to protect the upstream hosts.setEndpointStaleAfter(com.google.protobuf.Duration value) The max time until which the endpoints from this assignment can be used.setEndpointStaleAfter(com.google.protobuf.Duration.Builder builderForValue) The max time until which the endpoints from this assignment can be used.setOverprovisioningFactor(com.google.protobuf.UInt32Value value) Priority levels and localities are considered overprovisioned with this factor (in percentage).setOverprovisioningFactor(com.google.protobuf.UInt32Value.Builder builderForValue) Priority levels and localities are considered overprovisioned with this factor (in percentage).setWeightedPriorityHealth(boolean value) If true, use the :ref:`load balancing weight <envoy_v3_api_field_config.endpoint.v3.LbEndpoint.load_balancing_weight>` of healthy and unhealthy hosts to determine the health of the priority level.Methods inherited from class com.google.protobuf.GeneratedMessage.Builder
addRepeatedField, clearField, clearOneof, clone, getAllFields, getField, getFieldBuilder, getOneofFieldDescriptor, getParentForChildren, getRepeatedField, getRepeatedFieldBuilder, getRepeatedFieldCount, getUnknownFields, getUnknownFieldSetBuilder, hasField, hasOneof, internalGetMapField, internalGetMapFieldReflection, internalGetMutableMapField, internalGetMutableMapFieldReflection, isClean, markClean, mergeUnknownFields, mergeUnknownLengthDelimitedField, mergeUnknownVarintField, newBuilderForField, onBuilt, onChanged, parseUnknownField, setField, setRepeatedField, setUnknownFields, setUnknownFieldSetBuilder, setUnknownFieldsProto3Methods inherited from class com.google.protobuf.AbstractMessage.Builder
findInitializationErrors, getInitializationErrorString, internalMergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, mergeFrom, newUninitializedMessageException, toStringMethods inherited from class com.google.protobuf.AbstractMessageLite.Builder
addAll, addAll, mergeDelimitedFrom, mergeDelimitedFrom, mergeFrom, newUninitializedMessageExceptionMethods inherited from class java.lang.Object
equals, finalize, getClass, hashCode, notify, notifyAll, wait, wait, waitMethods inherited from interface com.google.protobuf.Message.Builder
mergeDelimitedFrom, mergeDelimitedFromMethods inherited from interface com.google.protobuf.MessageLite.Builder
mergeFromMethods inherited from interface com.google.protobuf.MessageOrBuilder
findInitializationErrors, getAllFields, getField, getInitializationErrorString, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, getUnknownFields, hasField, hasOneof
-
Field Details
-
bitField0_
private int bitField0_ -
dropOverloads_
-
dropOverloadsBuilder_
private com.google.protobuf.RepeatedFieldBuilder<ClusterLoadAssignment.Policy.DropOverload,ClusterLoadAssignment.Policy.DropOverload.Builder, dropOverloadsBuilder_ClusterLoadAssignment.Policy.DropOverloadOrBuilder> -
overprovisioningFactor_
private com.google.protobuf.UInt32Value overprovisioningFactor_ -
overprovisioningFactorBuilder_
private com.google.protobuf.SingleFieldBuilder<com.google.protobuf.UInt32Value,com.google.protobuf.UInt32Value.Builder, overprovisioningFactorBuilder_com.google.protobuf.UInt32ValueOrBuilder> -
endpointStaleAfter_
private com.google.protobuf.Duration endpointStaleAfter_ -
endpointStaleAfterBuilder_
private com.google.protobuf.SingleFieldBuilder<com.google.protobuf.Duration,com.google.protobuf.Duration.Builder, endpointStaleAfterBuilder_com.google.protobuf.DurationOrBuilder> -
weightedPriorityHealth_
private boolean weightedPriorityHealth_
-
-
Constructor Details
-
Builder
private Builder() -
Builder
private Builder(com.google.protobuf.AbstractMessage.BuilderParent parent)
-
-
Method Details
-
getDescriptor
public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() -
internalGetFieldAccessorTable
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable()- Specified by:
internalGetFieldAccessorTablein classcom.google.protobuf.GeneratedMessage.Builder<ClusterLoadAssignment.Policy.Builder>
-
maybeForceBuilderInitialization
private void maybeForceBuilderInitialization() -
clear
- Specified by:
clearin interfacecom.google.protobuf.Message.Builder- Specified by:
clearin interfacecom.google.protobuf.MessageLite.Builder- Overrides:
clearin classcom.google.protobuf.GeneratedMessage.Builder<ClusterLoadAssignment.Policy.Builder>
-
getDescriptorForType
public com.google.protobuf.Descriptors.Descriptor getDescriptorForType()- Specified by:
getDescriptorForTypein interfacecom.google.protobuf.Message.Builder- Specified by:
getDescriptorForTypein interfacecom.google.protobuf.MessageOrBuilder- Overrides:
getDescriptorForTypein classcom.google.protobuf.GeneratedMessage.Builder<ClusterLoadAssignment.Policy.Builder>
-
getDefaultInstanceForType
- Specified by:
getDefaultInstanceForTypein interfacecom.google.protobuf.MessageLiteOrBuilder- Specified by:
getDefaultInstanceForTypein interfacecom.google.protobuf.MessageOrBuilder
-
build
- Specified by:
buildin interfacecom.google.protobuf.Message.Builder- Specified by:
buildin interfacecom.google.protobuf.MessageLite.Builder
-
buildPartial
- Specified by:
buildPartialin interfacecom.google.protobuf.Message.Builder- Specified by:
buildPartialin interfacecom.google.protobuf.MessageLite.Builder
-
buildPartialRepeatedFields
-
buildPartial0
-
mergeFrom
- Specified by:
mergeFromin interfacecom.google.protobuf.Message.Builder- Overrides:
mergeFromin classcom.google.protobuf.AbstractMessage.Builder<ClusterLoadAssignment.Policy.Builder>
-
mergeFrom
-
isInitialized
public final boolean isInitialized()- Specified by:
isInitializedin interfacecom.google.protobuf.MessageLiteOrBuilder- Overrides:
isInitializedin classcom.google.protobuf.GeneratedMessage.Builder<ClusterLoadAssignment.Policy.Builder>
-
mergeFrom
public ClusterLoadAssignment.Policy.Builder mergeFrom(com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws IOException - Specified by:
mergeFromin interfacecom.google.protobuf.Message.Builder- Specified by:
mergeFromin interfacecom.google.protobuf.MessageLite.Builder- Overrides:
mergeFromin classcom.google.protobuf.AbstractMessage.Builder<ClusterLoadAssignment.Policy.Builder>- Throws:
IOException
-
ensureDropOverloadsIsMutable
private void ensureDropOverloadsIsMutable() -
getDropOverloadsList
Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2;- Specified by:
getDropOverloadsListin interfaceClusterLoadAssignment.PolicyOrBuilder
-
getDropOverloadsCount
public int getDropOverloadsCount()Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2;- Specified by:
getDropOverloadsCountin interfaceClusterLoadAssignment.PolicyOrBuilder
-
getDropOverloads
Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2;- Specified by:
getDropOverloadsin interfaceClusterLoadAssignment.PolicyOrBuilder
-
setDropOverloads
public ClusterLoadAssignment.Policy.Builder setDropOverloads(int index, ClusterLoadAssignment.Policy.DropOverload value) Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
setDropOverloads
public ClusterLoadAssignment.Policy.Builder setDropOverloads(int index, ClusterLoadAssignment.Policy.DropOverload.Builder builderForValue) Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
addDropOverloads
public ClusterLoadAssignment.Policy.Builder addDropOverloads(ClusterLoadAssignment.Policy.DropOverload value) Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
addDropOverloads
public ClusterLoadAssignment.Policy.Builder addDropOverloads(int index, ClusterLoadAssignment.Policy.DropOverload value) Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
addDropOverloads
public ClusterLoadAssignment.Policy.Builder addDropOverloads(ClusterLoadAssignment.Policy.DropOverload.Builder builderForValue) Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
addDropOverloads
public ClusterLoadAssignment.Policy.Builder addDropOverloads(int index, ClusterLoadAssignment.Policy.DropOverload.Builder builderForValue) Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
addAllDropOverloads
public ClusterLoadAssignment.Policy.Builder addAllDropOverloads(Iterable<? extends ClusterLoadAssignment.Policy.DropOverload> values) Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
clearDropOverloads
Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
removeDropOverloads
Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
getDropOverloadsBuilder
Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
getDropOverloadsOrBuilder
Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2;- Specified by:
getDropOverloadsOrBuilderin interfaceClusterLoadAssignment.PolicyOrBuilder
-
getDropOverloadsOrBuilderList
public List<? extends ClusterLoadAssignment.Policy.DropOverloadOrBuilder> getDropOverloadsOrBuilderList()Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2;- Specified by:
getDropOverloadsOrBuilderListin interfaceClusterLoadAssignment.PolicyOrBuilder
-
addDropOverloadsBuilder
Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
addDropOverloadsBuilder
Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
getDropOverloadsBuilderList
Action to trim the overall incoming traffic to protect the upstream hosts. This action allows protection in case the hosts are unable to recover from an outage, or unable to autoscale or unable to handle incoming traffic volume for any reason. At the client each category is applied one after the other to generate the 'actual' drop percentage on all outgoing traffic. For example: .. code-block:: json { "drop_overloads": [ { "category": "throttle", "drop_percentage": 60 } { "category": "lb", "drop_percentage": 50 } ]} The actual drop percentages applied to the traffic at the clients will be "throttle"_drop = 60% "lb"_drop = 20% // 50% of the remaining 'actual' load, which is 40%. actual_outgoing_load = 20% // remaining after applying all categories. Envoy supports only one element and will NACK if more than one element is present. Other xDS-capable data planes will not necessarily have this limitation. In Envoy, this ``drop_overloads`` config can be overridden by a runtime key "load_balancing_policy.drop_overload_limit" setting. This runtime key can be set to any integer number between 0 and 100. 0 means drop 0%. 100 means drop 100%. When both ``drop_overloads`` config and "load_balancing_policy.drop_overload_limit" setting are in place, the min of these two wins.repeated .envoy.config.endpoint.v3.ClusterLoadAssignment.Policy.DropOverload drop_overloads = 2; -
internalGetDropOverloadsFieldBuilder
private com.google.protobuf.RepeatedFieldBuilder<ClusterLoadAssignment.Policy.DropOverload,ClusterLoadAssignment.Policy.DropOverload.Builder, internalGetDropOverloadsFieldBuilder()ClusterLoadAssignment.Policy.DropOverloadOrBuilder> -
hasOverprovisioningFactor
public boolean hasOverprovisioningFactor()Priority levels and localities are considered overprovisioned with this factor (in percentage). This means that we don't consider a priority level or locality unhealthy until the fraction of healthy hosts multiplied by the overprovisioning factor drops below 100. With the default value 140(1.4), Envoy doesn't consider a priority level or a locality unhealthy until their percentage of healthy hosts drops below 72%. For example: .. code-block:: json { "overprovisioning_factor": 100 } Read more at :ref:`priority levels <arch_overview_load_balancing_priority_levels>` and :ref:`localities <arch_overview_load_balancing_locality_weighted_lb>`..google.protobuf.UInt32Value overprovisioning_factor = 3 [(.validate.rules) = { ... }- Specified by:
hasOverprovisioningFactorin interfaceClusterLoadAssignment.PolicyOrBuilder- Returns:
- Whether the overprovisioningFactor field is set.
-
getOverprovisioningFactor
public com.google.protobuf.UInt32Value getOverprovisioningFactor()Priority levels and localities are considered overprovisioned with this factor (in percentage). This means that we don't consider a priority level or locality unhealthy until the fraction of healthy hosts multiplied by the overprovisioning factor drops below 100. With the default value 140(1.4), Envoy doesn't consider a priority level or a locality unhealthy until their percentage of healthy hosts drops below 72%. For example: .. code-block:: json { "overprovisioning_factor": 100 } Read more at :ref:`priority levels <arch_overview_load_balancing_priority_levels>` and :ref:`localities <arch_overview_load_balancing_locality_weighted_lb>`..google.protobuf.UInt32Value overprovisioning_factor = 3 [(.validate.rules) = { ... }- Specified by:
getOverprovisioningFactorin interfaceClusterLoadAssignment.PolicyOrBuilder- Returns:
- The overprovisioningFactor.
-
setOverprovisioningFactor
public ClusterLoadAssignment.Policy.Builder setOverprovisioningFactor(com.google.protobuf.UInt32Value value) Priority levels and localities are considered overprovisioned with this factor (in percentage). This means that we don't consider a priority level or locality unhealthy until the fraction of healthy hosts multiplied by the overprovisioning factor drops below 100. With the default value 140(1.4), Envoy doesn't consider a priority level or a locality unhealthy until their percentage of healthy hosts drops below 72%. For example: .. code-block:: json { "overprovisioning_factor": 100 } Read more at :ref:`priority levels <arch_overview_load_balancing_priority_levels>` and :ref:`localities <arch_overview_load_balancing_locality_weighted_lb>`..google.protobuf.UInt32Value overprovisioning_factor = 3 [(.validate.rules) = { ... } -
setOverprovisioningFactor
public ClusterLoadAssignment.Policy.Builder setOverprovisioningFactor(com.google.protobuf.UInt32Value.Builder builderForValue) Priority levels and localities are considered overprovisioned with this factor (in percentage). This means that we don't consider a priority level or locality unhealthy until the fraction of healthy hosts multiplied by the overprovisioning factor drops below 100. With the default value 140(1.4), Envoy doesn't consider a priority level or a locality unhealthy until their percentage of healthy hosts drops below 72%. For example: .. code-block:: json { "overprovisioning_factor": 100 } Read more at :ref:`priority levels <arch_overview_load_balancing_priority_levels>` and :ref:`localities <arch_overview_load_balancing_locality_weighted_lb>`..google.protobuf.UInt32Value overprovisioning_factor = 3 [(.validate.rules) = { ... } -
mergeOverprovisioningFactor
public ClusterLoadAssignment.Policy.Builder mergeOverprovisioningFactor(com.google.protobuf.UInt32Value value) Priority levels and localities are considered overprovisioned with this factor (in percentage). This means that we don't consider a priority level or locality unhealthy until the fraction of healthy hosts multiplied by the overprovisioning factor drops below 100. With the default value 140(1.4), Envoy doesn't consider a priority level or a locality unhealthy until their percentage of healthy hosts drops below 72%. For example: .. code-block:: json { "overprovisioning_factor": 100 } Read more at :ref:`priority levels <arch_overview_load_balancing_priority_levels>` and :ref:`localities <arch_overview_load_balancing_locality_weighted_lb>`..google.protobuf.UInt32Value overprovisioning_factor = 3 [(.validate.rules) = { ... } -
clearOverprovisioningFactor
Priority levels and localities are considered overprovisioned with this factor (in percentage). This means that we don't consider a priority level or locality unhealthy until the fraction of healthy hosts multiplied by the overprovisioning factor drops below 100. With the default value 140(1.4), Envoy doesn't consider a priority level or a locality unhealthy until their percentage of healthy hosts drops below 72%. For example: .. code-block:: json { "overprovisioning_factor": 100 } Read more at :ref:`priority levels <arch_overview_load_balancing_priority_levels>` and :ref:`localities <arch_overview_load_balancing_locality_weighted_lb>`..google.protobuf.UInt32Value overprovisioning_factor = 3 [(.validate.rules) = { ... } -
getOverprovisioningFactorBuilder
public com.google.protobuf.UInt32Value.Builder getOverprovisioningFactorBuilder()Priority levels and localities are considered overprovisioned with this factor (in percentage). This means that we don't consider a priority level or locality unhealthy until the fraction of healthy hosts multiplied by the overprovisioning factor drops below 100. With the default value 140(1.4), Envoy doesn't consider a priority level or a locality unhealthy until their percentage of healthy hosts drops below 72%. For example: .. code-block:: json { "overprovisioning_factor": 100 } Read more at :ref:`priority levels <arch_overview_load_balancing_priority_levels>` and :ref:`localities <arch_overview_load_balancing_locality_weighted_lb>`..google.protobuf.UInt32Value overprovisioning_factor = 3 [(.validate.rules) = { ... } -
getOverprovisioningFactorOrBuilder
public com.google.protobuf.UInt32ValueOrBuilder getOverprovisioningFactorOrBuilder()Priority levels and localities are considered overprovisioned with this factor (in percentage). This means that we don't consider a priority level or locality unhealthy until the fraction of healthy hosts multiplied by the overprovisioning factor drops below 100. With the default value 140(1.4), Envoy doesn't consider a priority level or a locality unhealthy until their percentage of healthy hosts drops below 72%. For example: .. code-block:: json { "overprovisioning_factor": 100 } Read more at :ref:`priority levels <arch_overview_load_balancing_priority_levels>` and :ref:`localities <arch_overview_load_balancing_locality_weighted_lb>`..google.protobuf.UInt32Value overprovisioning_factor = 3 [(.validate.rules) = { ... }- Specified by:
getOverprovisioningFactorOrBuilderin interfaceClusterLoadAssignment.PolicyOrBuilder
-
internalGetOverprovisioningFactorFieldBuilder
private com.google.protobuf.SingleFieldBuilder<com.google.protobuf.UInt32Value,com.google.protobuf.UInt32Value.Builder, internalGetOverprovisioningFactorFieldBuilder()com.google.protobuf.UInt32ValueOrBuilder> Priority levels and localities are considered overprovisioned with this factor (in percentage). This means that we don't consider a priority level or locality unhealthy until the fraction of healthy hosts multiplied by the overprovisioning factor drops below 100. With the default value 140(1.4), Envoy doesn't consider a priority level or a locality unhealthy until their percentage of healthy hosts drops below 72%. For example: .. code-block:: json { "overprovisioning_factor": 100 } Read more at :ref:`priority levels <arch_overview_load_balancing_priority_levels>` and :ref:`localities <arch_overview_load_balancing_locality_weighted_lb>`..google.protobuf.UInt32Value overprovisioning_factor = 3 [(.validate.rules) = { ... } -
hasEndpointStaleAfter
public boolean hasEndpointStaleAfter()The max time until which the endpoints from this assignment can be used. If no new assignments are received before this time expires the endpoints are considered stale and should be marked unhealthy. Defaults to 0 which means endpoints never go stale.
.google.protobuf.Duration endpoint_stale_after = 4 [(.validate.rules) = { ... }- Specified by:
hasEndpointStaleAfterin interfaceClusterLoadAssignment.PolicyOrBuilder- Returns:
- Whether the endpointStaleAfter field is set.
-
getEndpointStaleAfter
public com.google.protobuf.Duration getEndpointStaleAfter()The max time until which the endpoints from this assignment can be used. If no new assignments are received before this time expires the endpoints are considered stale and should be marked unhealthy. Defaults to 0 which means endpoints never go stale.
.google.protobuf.Duration endpoint_stale_after = 4 [(.validate.rules) = { ... }- Specified by:
getEndpointStaleAfterin interfaceClusterLoadAssignment.PolicyOrBuilder- Returns:
- The endpointStaleAfter.
-
setEndpointStaleAfter
public ClusterLoadAssignment.Policy.Builder setEndpointStaleAfter(com.google.protobuf.Duration value) The max time until which the endpoints from this assignment can be used. If no new assignments are received before this time expires the endpoints are considered stale and should be marked unhealthy. Defaults to 0 which means endpoints never go stale.
.google.protobuf.Duration endpoint_stale_after = 4 [(.validate.rules) = { ... } -
setEndpointStaleAfter
public ClusterLoadAssignment.Policy.Builder setEndpointStaleAfter(com.google.protobuf.Duration.Builder builderForValue) The max time until which the endpoints from this assignment can be used. If no new assignments are received before this time expires the endpoints are considered stale and should be marked unhealthy. Defaults to 0 which means endpoints never go stale.
.google.protobuf.Duration endpoint_stale_after = 4 [(.validate.rules) = { ... } -
mergeEndpointStaleAfter
public ClusterLoadAssignment.Policy.Builder mergeEndpointStaleAfter(com.google.protobuf.Duration value) The max time until which the endpoints from this assignment can be used. If no new assignments are received before this time expires the endpoints are considered stale and should be marked unhealthy. Defaults to 0 which means endpoints never go stale.
.google.protobuf.Duration endpoint_stale_after = 4 [(.validate.rules) = { ... } -
clearEndpointStaleAfter
The max time until which the endpoints from this assignment can be used. If no new assignments are received before this time expires the endpoints are considered stale and should be marked unhealthy. Defaults to 0 which means endpoints never go stale.
.google.protobuf.Duration endpoint_stale_after = 4 [(.validate.rules) = { ... } -
getEndpointStaleAfterBuilder
public com.google.protobuf.Duration.Builder getEndpointStaleAfterBuilder()The max time until which the endpoints from this assignment can be used. If no new assignments are received before this time expires the endpoints are considered stale and should be marked unhealthy. Defaults to 0 which means endpoints never go stale.
.google.protobuf.Duration endpoint_stale_after = 4 [(.validate.rules) = { ... } -
getEndpointStaleAfterOrBuilder
public com.google.protobuf.DurationOrBuilder getEndpointStaleAfterOrBuilder()The max time until which the endpoints from this assignment can be used. If no new assignments are received before this time expires the endpoints are considered stale and should be marked unhealthy. Defaults to 0 which means endpoints never go stale.
.google.protobuf.Duration endpoint_stale_after = 4 [(.validate.rules) = { ... }- Specified by:
getEndpointStaleAfterOrBuilderin interfaceClusterLoadAssignment.PolicyOrBuilder
-
internalGetEndpointStaleAfterFieldBuilder
private com.google.protobuf.SingleFieldBuilder<com.google.protobuf.Duration,com.google.protobuf.Duration.Builder, internalGetEndpointStaleAfterFieldBuilder()com.google.protobuf.DurationOrBuilder> The max time until which the endpoints from this assignment can be used. If no new assignments are received before this time expires the endpoints are considered stale and should be marked unhealthy. Defaults to 0 which means endpoints never go stale.
.google.protobuf.Duration endpoint_stale_after = 4 [(.validate.rules) = { ... } -
getWeightedPriorityHealth
public boolean getWeightedPriorityHealth()If true, use the :ref:`load balancing weight <envoy_v3_api_field_config.endpoint.v3.LbEndpoint.load_balancing_weight>` of healthy and unhealthy hosts to determine the health of the priority level. If false, use the number of healthy and unhealthy hosts to determine the health of the priority level, or in other words assume each host has a weight of 1 for this calculation. Note: this is not currently implemented for :ref:`locality weighted load balancing <arch_overview_load_balancing_locality_weighted_lb>`.
bool weighted_priority_health = 6;- Specified by:
getWeightedPriorityHealthin interfaceClusterLoadAssignment.PolicyOrBuilder- Returns:
- The weightedPriorityHealth.
-
setWeightedPriorityHealth
If true, use the :ref:`load balancing weight <envoy_v3_api_field_config.endpoint.v3.LbEndpoint.load_balancing_weight>` of healthy and unhealthy hosts to determine the health of the priority level. If false, use the number of healthy and unhealthy hosts to determine the health of the priority level, or in other words assume each host has a weight of 1 for this calculation. Note: this is not currently implemented for :ref:`locality weighted load balancing <arch_overview_load_balancing_locality_weighted_lb>`.
bool weighted_priority_health = 6;- Parameters:
value- The weightedPriorityHealth to set.- Returns:
- This builder for chaining.
-
clearWeightedPriorityHealth
If true, use the :ref:`load balancing weight <envoy_v3_api_field_config.endpoint.v3.LbEndpoint.load_balancing_weight>` of healthy and unhealthy hosts to determine the health of the priority level. If false, use the number of healthy and unhealthy hosts to determine the health of the priority level, or in other words assume each host has a weight of 1 for this calculation. Note: this is not currently implemented for :ref:`locality weighted load balancing <arch_overview_load_balancing_locality_weighted_lb>`.
bool weighted_priority_health = 6;- Returns:
- This builder for chaining.
-