grpc 1.49.1-x86_64-linux → 1.50.0.pre1-x86_64-linux
Sign up to get free protection for your applications and to get access to all the features.
Potentially problematic release.
This version of grpc might be problematic. Click here for more details.
- checksums.yaml +4 -4
- data/Makefile +54 -153
- data/include/grpc/event_engine/endpoint_config.h +11 -5
- data/include/grpc/event_engine/event_engine.h +1 -1
- data/include/grpc/impl/codegen/atm_gcc_atomic.h +19 -28
- data/include/grpc/impl/codegen/atm_gcc_sync.h +0 -2
- data/include/grpc/impl/codegen/atm_windows.h +0 -2
- data/include/grpc/impl/codegen/grpc_types.h +6 -0
- data/src/core/ext/filters/channel_idle/channel_idle_filter.cc +3 -3
- data/src/core/ext/filters/client_channel/backup_poller.cc +4 -6
- data/src/core/ext/filters/client_channel/client_channel.cc +33 -22
- data/src/core/ext/filters/client_channel/client_channel.h +1 -1
- data/src/core/ext/filters/client_channel/client_channel_plugin.cc +0 -16
- data/src/core/ext/filters/client_channel/http_proxy.cc +12 -19
- data/src/core/ext/filters/client_channel/http_proxy.h +3 -2
- data/src/core/ext/filters/client_channel/lb_policy/child_policy_handler.cc +6 -4
- data/src/core/ext/filters/client_channel/lb_policy/child_policy_handler.h +5 -4
- data/src/core/ext/filters/client_channel/lb_policy/grpclb/client_load_reporting_filter.cc +0 -2
- data/src/core/ext/filters/client_channel/lb_policy/grpclb/grpclb.cc +112 -96
- data/src/core/ext/filters/client_channel/lb_policy/oob_backend_metric.cc +20 -11
- data/src/core/ext/filters/client_channel/lb_policy/outlier_detection/outlier_detection.cc +106 -108
- data/src/core/ext/filters/client_channel/lb_policy/outlier_detection/outlier_detection.h +16 -0
- data/src/core/ext/filters/client_channel/lb_policy/pick_first/pick_first.cc +20 -13
- data/src/core/ext/filters/client_channel/lb_policy/priority/priority.cc +165 -257
- data/src/core/ext/filters/client_channel/lb_policy/ring_hash/ring_hash.cc +218 -231
- data/src/core/ext/filters/client_channel/lb_policy/ring_hash/ring_hash.h +10 -6
- data/src/core/ext/filters/client_channel/lb_policy/rls/rls.cc +389 -444
- data/src/core/ext/filters/client_channel/lb_policy/round_robin/round_robin.cc +16 -16
- data/src/core/ext/filters/client_channel/lb_policy/subchannel_list.h +8 -13
- data/src/core/ext/filters/client_channel/lb_policy/weighted_target/weighted_target.cc +84 -96
- data/src/core/ext/filters/client_channel/lb_policy/xds/cds.cc +38 -37
- data/src/core/ext/filters/client_channel/lb_policy/xds/xds_cluster_impl.cc +106 -186
- data/src/core/ext/filters/client_channel/lb_policy/xds/xds_cluster_manager.cc +106 -93
- data/src/core/ext/filters/client_channel/lb_policy/xds/xds_cluster_resolver.cc +170 -218
- data/src/core/ext/filters/client_channel/resolver/dns/c_ares/grpc_ares_wrapper.cc +2 -2
- data/src/core/ext/filters/client_channel/resolver/dns/native/dns_resolver.cc +1 -1
- data/src/core/ext/filters/client_channel/resolver/google_c2p/google_c2p_resolver.cc +13 -15
- data/src/core/ext/filters/client_channel/resolver/polling_resolver.cc +84 -37
- data/src/core/ext/filters/client_channel/resolver/polling_resolver.h +11 -0
- data/src/core/ext/filters/client_channel/resolver/sockaddr/sockaddr_resolver.cc +1 -0
- data/src/core/ext/filters/client_channel/resolver/xds/xds_resolver.cc +5 -3
- data/src/core/ext/filters/client_channel/resolver_result_parsing.cc +5 -4
- data/src/core/ext/filters/client_channel/retry_filter.cc +25 -29
- data/src/core/ext/filters/client_channel/subchannel.cc +38 -33
- data/src/core/ext/filters/client_channel/subchannel.h +12 -3
- data/src/core/ext/filters/client_channel/subchannel_stream_client.cc +1 -2
- data/src/core/ext/filters/fault_injection/fault_injection_filter.cc +23 -16
- data/src/core/ext/filters/fault_injection/fault_injection_filter.h +8 -0
- data/src/core/ext/filters/http/client/http_client_filter.cc +1 -2
- data/src/core/ext/filters/http/message_compress/message_compress_filter.cc +2 -4
- data/src/core/ext/filters/http/message_compress/message_decompress_filter.cc +0 -2
- data/src/core/ext/filters/http/server/http_server_filter.cc +1 -2
- data/src/core/ext/transport/chttp2/client/chttp2_connector.cc +12 -8
- data/src/core/ext/transport/chttp2/server/chttp2_server.cc +32 -26
- data/src/core/ext/transport/chttp2/transport/bin_encoder.cc +1 -1
- data/src/core/ext/transport/chttp2/transport/chttp2_transport.cc +25 -130
- data/src/core/ext/transport/chttp2/transport/decode_huff.cc +287 -0
- data/src/core/ext/transport/chttp2/transport/decode_huff.h +1018 -0
- data/src/core/ext/transport/chttp2/transport/flow_control.cc +83 -51
- data/src/core/ext/transport/chttp2/transport/flow_control.h +11 -6
- data/src/core/ext/transport/chttp2/transport/frame_ping.cc +1 -2
- data/src/core/ext/transport/chttp2/transport/hpack_encoder.cc +2 -20
- data/src/core/ext/transport/chttp2/transport/hpack_parser.cc +28 -28
- data/src/core/ext/transport/chttp2/transport/hpack_parser_table.cc +1 -10
- data/src/core/ext/transport/chttp2/transport/hpack_parser_table.h +11 -6
- data/src/core/ext/transport/chttp2/transport/internal.h +2 -0
- data/src/core/ext/transport/chttp2/transport/parsing.cc +44 -0
- data/src/core/ext/transport/chttp2/transport/writing.cc +3 -14
- data/src/core/ext/transport/inproc/inproc_transport.cc +1 -3
- data/src/core/ext/xds/certificate_provider_store.cc +63 -3
- data/src/core/ext/xds/certificate_provider_store.h +9 -1
- data/src/core/ext/xds/file_watcher_certificate_provider_factory.cc +5 -5
- data/src/core/ext/xds/file_watcher_certificate_provider_factory.h +1 -1
- data/src/core/ext/xds/xds_api.cc +21 -17
- data/src/core/ext/xds/xds_api.h +7 -0
- data/src/core/ext/xds/xds_bootstrap.cc +5 -537
- data/src/core/ext/xds/xds_bootstrap.h +39 -111
- data/src/core/ext/xds/xds_bootstrap_grpc.cc +370 -0
- data/src/core/ext/xds/xds_bootstrap_grpc.h +169 -0
- data/src/core/ext/xds/xds_client.cc +219 -145
- data/src/core/ext/xds/xds_client.h +19 -17
- data/src/core/ext/xds/xds_client_grpc.cc +18 -80
- data/src/core/ext/xds/xds_client_grpc.h +2 -25
- data/src/core/ext/xds/xds_client_stats.cc +4 -4
- data/src/core/ext/xds/xds_cluster.cc +87 -79
- data/src/core/ext/xds/xds_cluster.h +5 -5
- data/src/core/ext/xds/xds_cluster_specifier_plugin.cc +3 -1
- data/src/core/ext/xds/xds_common_types.cc +13 -5
- data/src/core/ext/xds/xds_endpoint.cc +8 -6
- data/src/core/ext/xds/xds_endpoint.h +3 -4
- data/src/core/ext/xds/xds_lb_policy_registry.cc +4 -2
- data/src/core/ext/xds/xds_listener.cc +25 -20
- data/src/core/ext/xds/xds_listener.h +3 -4
- data/src/core/ext/xds/xds_resource_type.h +11 -8
- data/src/core/ext/xds/xds_route_config.cc +15 -16
- data/src/core/ext/xds/xds_route_config.h +3 -3
- data/src/core/ext/xds/xds_server_config_fetcher.cc +7 -5
- data/src/core/ext/xds/xds_transport_grpc.cc +15 -7
- data/src/core/lib/backoff/backoff.cc +2 -4
- data/src/core/lib/channel/call_finalization.h +1 -3
- data/src/core/lib/channel/channel_args.h +114 -14
- data/src/core/lib/channel/channel_trace.cc +3 -4
- data/src/core/lib/channel/promise_based_filter.cc +18 -19
- data/src/core/lib/channel/status_util.cc +27 -0
- data/src/core/lib/channel/status_util.h +10 -0
- data/src/core/lib/config/core_configuration.cc +5 -1
- data/src/core/lib/config/core_configuration.h +33 -0
- data/src/core/lib/debug/stats.cc +26 -30
- data/src/core/lib/debug/stats.h +2 -12
- data/src/core/lib/debug/stats_data.cc +118 -614
- data/src/core/lib/debug/stats_data.h +67 -465
- data/src/core/lib/debug/trace.cc +0 -2
- data/src/core/lib/event_engine/channel_args_endpoint_config.cc +12 -20
- data/src/core/lib/event_engine/channel_args_endpoint_config.h +13 -7
- data/src/core/lib/event_engine/forkable.cc +1 -1
- data/src/core/lib/event_engine/poller.h +14 -12
- data/src/core/lib/event_engine/posix_engine/timer_manager.cc +53 -32
- data/src/core/lib/event_engine/posix_engine/timer_manager.h +23 -1
- data/src/core/lib/event_engine/thread_pool.cc +131 -94
- data/src/core/lib/event_engine/thread_pool.h +56 -23
- data/src/core/lib/event_engine/time_util.cc +30 -0
- data/src/core/lib/event_engine/time_util.h +32 -0
- data/src/core/lib/event_engine/utils.cc +0 -5
- data/src/core/lib/event_engine/utils.h +0 -4
- data/src/core/lib/event_engine/windows/iocp.cc +13 -7
- data/src/core/lib/event_engine/windows/iocp.h +2 -1
- data/src/core/lib/event_engine/windows/win_socket.cc +1 -1
- data/src/core/lib/experiments/config.cc +146 -0
- data/src/core/lib/experiments/config.h +43 -0
- data/src/core/lib/experiments/experiments.cc +75 -0
- data/src/core/lib/experiments/experiments.h +56 -0
- data/src/core/lib/gpr/alloc.cc +1 -9
- data/src/core/lib/gpr/log_windows.cc +0 -1
- data/src/core/lib/gpr/string_util_windows.cc +3 -30
- data/src/core/lib/gpr/sync_abseil.cc +0 -14
- data/src/core/lib/gpr/sync_posix.cc +0 -14
- data/src/core/lib/gpr/time_posix.cc +0 -6
- data/src/core/lib/gpr/time_precise.h +1 -1
- data/src/core/lib/gpr/tmpfile_windows.cc +5 -7
- data/src/core/lib/gpr/useful.h +11 -0
- data/src/core/lib/{gpr → gprpp}/env.h +25 -12
- data/src/core/lib/{gpr → gprpp}/env_linux.cc +20 -15
- data/src/core/lib/{gpr → gprpp}/env_posix.cc +11 -10
- data/src/core/lib/gprpp/env_windows.cc +56 -0
- data/src/core/lib/gprpp/fork.cc +14 -22
- data/src/core/lib/gprpp/fork.h +0 -8
- data/src/core/lib/gprpp/global_config_env.cc +7 -6
- data/src/core/lib/gprpp/notification.h +67 -0
- data/src/core/lib/gprpp/packed_table.h +40 -0
- data/src/core/lib/gprpp/ref_counted_ptr.h +20 -33
- data/src/core/lib/gprpp/sorted_pack.h +98 -0
- data/src/core/lib/gprpp/status_helper.h +6 -0
- data/src/core/lib/gprpp/table.h +9 -1
- data/src/core/lib/gprpp/tchar.cc +49 -0
- data/src/core/lib/gprpp/tchar.h +33 -0
- data/src/core/lib/gprpp/time.cc +21 -0
- data/src/core/lib/gprpp/time.h +55 -0
- data/src/core/lib/gprpp/validation_errors.cc +61 -0
- data/src/core/lib/gprpp/validation_errors.h +110 -0
- data/src/core/{ext/filters/client_channel → lib/handshaker}/proxy_mapper.h +3 -3
- data/src/core/{ext/filters/client_channel → lib/handshaker}/proxy_mapper_registry.cc +14 -36
- data/src/core/lib/handshaker/proxy_mapper_registry.h +75 -0
- data/src/core/lib/iomgr/call_combiner.cc +0 -8
- data/src/core/lib/iomgr/closure.h +0 -1
- data/src/core/lib/iomgr/endpoint_pair_posix.cc +14 -10
- data/src/core/lib/iomgr/endpoint_pair_windows.cc +2 -2
- data/src/core/lib/iomgr/ev_epoll1_linux.cc +1 -38
- data/src/core/lib/iomgr/ev_poll_posix.cc +2 -17
- data/src/core/lib/iomgr/exec_ctx.cc +0 -10
- data/src/core/lib/iomgr/exec_ctx.h +7 -31
- data/src/core/lib/iomgr/iocp_windows.cc +1 -2
- data/src/core/lib/iomgr/iomgr.cc +6 -8
- data/src/core/lib/iomgr/iomgr_fwd.h +1 -0
- data/src/core/lib/iomgr/pollset.h +1 -1
- data/src/core/lib/iomgr/pollset_set.h +0 -1
- data/src/core/lib/iomgr/resolve_address.h +1 -0
- data/src/core/lib/iomgr/resolve_address_impl.h +1 -0
- data/src/core/lib/iomgr/resolve_address_posix.cc +1 -0
- data/src/core/lib/iomgr/resolve_address_windows.cc +1 -0
- data/src/core/lib/iomgr/sockaddr_utils_posix.cc +2 -1
- data/src/core/lib/iomgr/socket_utils_common_posix.cc +12 -34
- data/src/core/lib/iomgr/socket_utils_posix.cc +83 -1
- data/src/core/lib/iomgr/socket_utils_posix.h +98 -6
- data/src/core/lib/iomgr/tcp_client.cc +6 -7
- data/src/core/lib/iomgr/tcp_client.h +11 -11
- data/src/core/lib/iomgr/tcp_client_cfstream.cc +6 -6
- data/src/core/lib/iomgr/tcp_client_posix.cc +33 -29
- data/src/core/lib/iomgr/tcp_client_posix.h +12 -9
- data/src/core/lib/iomgr/tcp_client_windows.cc +6 -6
- data/src/core/lib/iomgr/tcp_posix.cc +131 -114
- data/src/core/lib/iomgr/tcp_posix.h +3 -1
- data/src/core/lib/iomgr/tcp_server.cc +5 -4
- data/src/core/lib/iomgr/tcp_server.h +9 -6
- data/src/core/lib/iomgr/tcp_server_posix.cc +17 -28
- data/src/core/lib/iomgr/tcp_server_utils_posix.h +2 -2
- data/src/core/lib/iomgr/tcp_server_utils_posix_common.cc +3 -3
- data/src/core/lib/iomgr/tcp_server_windows.cc +6 -7
- data/src/core/lib/iomgr/tcp_windows.cc +0 -1
- data/src/core/lib/iomgr/tcp_windows.h +0 -1
- data/src/core/lib/iomgr/timer_generic.cc +4 -4
- data/src/core/lib/iomgr/timer_manager.cc +1 -2
- data/src/core/lib/iomgr/wakeup_fd_eventfd.cc +0 -2
- data/src/core/lib/json/json_object_loader.cc +21 -52
- data/src/core/lib/json/json_object_loader.h +56 -76
- data/src/core/lib/json/json_util.cc +2 -1
- data/src/core/lib/load_balancing/lb_policy.h +5 -5
- data/src/core/lib/load_balancing/lb_policy_registry.cc +29 -55
- data/src/core/lib/load_balancing/lb_policy_registry.h +23 -11
- data/src/core/lib/promise/activity.h +2 -3
- data/src/core/lib/promise/context.h +1 -1
- data/src/core/lib/promise/sleep.cc +16 -4
- data/src/core/lib/promise/sleep.h +8 -2
- data/src/core/lib/resolver/resolver.h +13 -3
- data/src/core/lib/resource_quota/api.cc +9 -0
- data/src/core/lib/resource_quota/api.h +6 -0
- data/src/core/lib/resource_quota/arena.cc +1 -3
- data/src/core/lib/resource_quota/memory_quota.cc +8 -24
- data/src/core/lib/resource_quota/memory_quota.h +6 -19
- data/src/core/lib/resource_quota/periodic_update.cc +2 -3
- data/src/core/{ext/xds → lib/security/certificate_provider}/certificate_provider_factory.h +3 -3
- data/src/core/lib/security/certificate_provider/certificate_provider_registry.cc +60 -0
- data/src/core/lib/security/certificate_provider/certificate_provider_registry.h +70 -0
- data/src/core/lib/security/credentials/channel_creds_registry_init.cc +1 -0
- data/src/core/lib/security/credentials/external/aws_external_account_credentials.cc +15 -16
- data/src/core/lib/security/credentials/external/external_account_credentials.cc +2 -1
- data/src/core/lib/security/credentials/google_default/credentials_generic.cc +5 -8
- data/src/core/lib/security/credentials/google_default/google_default_credentials.cc +6 -6
- data/src/core/lib/security/credentials/jwt/jwt_verifier.cc +3 -2
- data/src/core/lib/security/credentials/jwt/jwt_verifier.h +1 -1
- data/src/core/lib/security/credentials/oauth2/oauth2_credentials.cc +1 -2
- data/src/core/lib/security/credentials/tls/grpc_tls_certificate_provider.cc +4 -3
- data/src/core/lib/security/credentials/tls/grpc_tls_certificate_provider.h +4 -2
- data/src/core/lib/security/credentials/tls/tls_utils.cc +3 -1
- data/src/core/lib/security/transport/client_auth_filter.cc +12 -1
- data/src/core/lib/security/transport/secure_endpoint.cc +0 -4
- data/src/core/lib/surface/call.cc +1 -11
- data/src/core/lib/surface/channel.cc +3 -2
- data/src/core/lib/surface/completion_queue.cc +16 -28
- data/src/core/lib/surface/completion_queue.h +1 -1
- data/src/core/lib/surface/completion_queue_factory.cc +5 -0
- data/src/core/lib/surface/init.cc +16 -11
- data/src/core/lib/surface/init_internally.cc +24 -0
- data/src/core/lib/surface/init_internally.h +28 -0
- data/src/core/lib/surface/server.cc +1 -7
- data/src/core/lib/surface/server.h +4 -6
- data/src/core/lib/surface/version.cc +2 -2
- data/src/core/lib/transport/bdp_estimator.cc +1 -3
- data/src/core/lib/transport/metadata_batch.cc +2 -3
- data/src/core/lib/transport/metadata_batch.h +9 -7
- data/src/core/lib/transport/parsed_metadata.h +4 -2
- data/src/core/lib/transport/status_conversion.cc +1 -3
- data/src/core/lib/transport/tcp_connect_handshaker.cc +9 -5
- data/src/core/lib/transport/transport.h +0 -1
- data/src/core/lib/transport/transport_impl.h +0 -1
- data/src/core/plugin_registry/grpc_plugin_registry.cc +23 -46
- data/src/core/plugin_registry/grpc_plugin_registry_extra.cc +13 -25
- data/src/ruby/lib/grpc/2.6/grpc_c.so +0 -0
- data/src/ruby/lib/grpc/2.7/grpc_c.so +0 -0
- data/src/ruby/lib/grpc/3.0/grpc_c.so +0 -0
- data/src/ruby/lib/grpc/3.1/grpc_c.so +0 -0
- data/src/ruby/lib/grpc/grpc_c.so +0 -0
- data/src/ruby/lib/grpc/version.rb +1 -1
- data/src/ruby/spec/channel_spec.rb +5 -0
- data/src/ruby/spec/generic/server_interceptors_spec.rb +1 -1
- data/src/ruby/spec/user_agent_spec.rb +1 -1
- metadata +33 -19
- data/src/core/ext/filters/client_channel/proxy_mapper_registry.h +0 -56
- data/src/core/ext/xds/certificate_provider_registry.cc +0 -103
- data/src/core/ext/xds/certificate_provider_registry.h +0 -59
- data/src/core/lib/event_engine/promise.h +0 -78
- data/src/core/lib/gpr/env_windows.cc +0 -74
- data/src/core/lib/gpr/string_windows.h +0 -32
- data/src/core/lib/profiling/basic_timers.cc +0 -295
- data/src/core/lib/profiling/stap_timers.cc +0 -50
- data/src/core/lib/profiling/timers.h +0 -94
@@ -20,10 +20,13 @@
|
|
20
20
|
|
21
21
|
#include <inttypes.h>
|
22
22
|
|
23
|
+
#include <functional>
|
23
24
|
#include <utility>
|
25
|
+
#include <vector>
|
24
26
|
|
25
27
|
#include "absl/status/status.h"
|
26
28
|
#include "absl/status/statusor.h"
|
29
|
+
#include "absl/strings/str_cat.h"
|
27
30
|
#include "absl/strings/string_view.h"
|
28
31
|
#include "absl/strings/strip.h"
|
29
32
|
|
@@ -36,6 +39,7 @@
|
|
36
39
|
#include "src/core/lib/gprpp/work_serializer.h"
|
37
40
|
#include "src/core/lib/iomgr/exec_ctx.h"
|
38
41
|
#include "src/core/lib/iomgr/timer.h"
|
42
|
+
#include "src/core/lib/service_config/service_config.h"
|
39
43
|
#include "src/core/lib/uri/uri_parser.h"
|
40
44
|
|
41
45
|
namespace grpc_core {
|
@@ -69,7 +73,16 @@ void PollingResolver::StartLocked() { MaybeStartResolvingLocked(); }
|
|
69
73
|
|
70
74
|
void PollingResolver::RequestReresolutionLocked() {
|
71
75
|
if (request_ == nullptr) {
|
72
|
-
|
76
|
+
// If we're still waiting for a result-health callback from the last
|
77
|
+
// result we reported, don't trigger the re-resolution until we get
|
78
|
+
// that callback.
|
79
|
+
if (result_status_state_ ==
|
80
|
+
ResultStatusState::kResultHealthCallbackPending) {
|
81
|
+
result_status_state_ =
|
82
|
+
ResultStatusState::kReresolutionRequestedWhileCallbackWasPending;
|
83
|
+
} else {
|
84
|
+
MaybeStartResolvingLocked();
|
85
|
+
}
|
73
86
|
}
|
74
87
|
}
|
75
88
|
|
@@ -126,44 +139,78 @@ void PollingResolver::OnRequestCompleteLocked(Result result) {
|
|
126
139
|
}
|
127
140
|
request_.reset();
|
128
141
|
if (!shutdown_) {
|
129
|
-
if (
|
130
|
-
|
131
|
-
|
132
|
-
|
133
|
-
|
134
|
-
|
135
|
-
|
136
|
-
|
137
|
-
|
138
|
-
|
139
|
-
|
140
|
-
|
141
|
-
|
142
|
-
|
143
|
-
|
144
|
-
// Also see https://github.com/grpc/grpc/issues/26079.
|
145
|
-
ExecCtx::Get()->InvalidateNow();
|
146
|
-
Timestamp next_try = backoff_.NextAttemptTime();
|
147
|
-
Duration timeout = next_try - ExecCtx::Get()->Now();
|
148
|
-
GPR_ASSERT(!have_next_resolution_timer_);
|
149
|
-
have_next_resolution_timer_ = true;
|
150
|
-
if (GPR_UNLIKELY(tracer_ != nullptr && tracer_->enabled())) {
|
151
|
-
if (timeout > Duration::Zero()) {
|
152
|
-
gpr_log(GPR_INFO, "[polling resolver %p] retrying in %" PRId64 " ms",
|
153
|
-
this, timeout.millis());
|
154
|
-
} else {
|
155
|
-
gpr_log(GPR_INFO, "[polling resolver %p] retrying immediately", this);
|
156
|
-
}
|
157
|
-
}
|
158
|
-
Ref(DEBUG_LOCATION, "next_resolution_timer").release();
|
159
|
-
GRPC_CLOSURE_INIT(&on_next_resolution_, OnNextResolution, this, nullptr);
|
160
|
-
grpc_timer_init(&next_resolution_timer_, next_try, &on_next_resolution_);
|
142
|
+
if (GPR_UNLIKELY(tracer_ != nullptr && tracer_->enabled())) {
|
143
|
+
gpr_log(GPR_INFO,
|
144
|
+
"[polling resolver %p] returning result: "
|
145
|
+
"addresses=%s, service_config=%s",
|
146
|
+
this,
|
147
|
+
result.addresses.ok()
|
148
|
+
? absl::StrCat("<", result.addresses->size(), " addresses>")
|
149
|
+
.c_str()
|
150
|
+
: result.addresses.status().ToString().c_str(),
|
151
|
+
result.service_config.ok()
|
152
|
+
? (*result.service_config == nullptr
|
153
|
+
? "<null>"
|
154
|
+
: std::string((*result.service_config)->json_string())
|
155
|
+
.c_str())
|
156
|
+
: result.service_config.status().ToString().c_str());
|
161
157
|
}
|
158
|
+
GPR_ASSERT(result.result_health_callback == nullptr);
|
159
|
+
RefCountedPtr<PollingResolver> self =
|
160
|
+
Ref(DEBUG_LOCATION, "result_health_callback");
|
161
|
+
result.result_health_callback = [self =
|
162
|
+
std::move(self)](absl::Status status) {
|
163
|
+
self->GetResultStatus(std::move(status));
|
164
|
+
};
|
165
|
+
result_status_state_ = ResultStatusState::kResultHealthCallbackPending;
|
162
166
|
result_handler_->ReportResult(std::move(result));
|
163
167
|
}
|
164
168
|
Unref(DEBUG_LOCATION, "OnRequestComplete");
|
165
169
|
}
|
166
170
|
|
171
|
+
void PollingResolver::GetResultStatus(absl::Status status) {
|
172
|
+
if (GPR_UNLIKELY(tracer_ != nullptr && tracer_->enabled())) {
|
173
|
+
gpr_log(GPR_INFO, "[polling resolver %p] result status from channel: %s",
|
174
|
+
this, status.ToString().c_str());
|
175
|
+
}
|
176
|
+
if (status.ok()) {
|
177
|
+
// Reset backoff state so that we start from the beginning when the
|
178
|
+
// next request gets triggered.
|
179
|
+
backoff_.Reset();
|
180
|
+
// If a re-resolution attempt was requested while the result-status
|
181
|
+
// callback was pending, trigger a new request now.
|
182
|
+
if (std::exchange(result_status_state_, ResultStatusState::kNone) ==
|
183
|
+
ResultStatusState::kReresolutionRequestedWhileCallbackWasPending) {
|
184
|
+
MaybeStartResolvingLocked();
|
185
|
+
}
|
186
|
+
} else {
|
187
|
+
// Set up for retry.
|
188
|
+
// InvalidateNow to avoid getting stuck re-initializing this timer
|
189
|
+
// in a loop while draining the currently-held WorkSerializer.
|
190
|
+
// Also see https://github.com/grpc/grpc/issues/26079.
|
191
|
+
ExecCtx::Get()->InvalidateNow();
|
192
|
+
Timestamp next_try = backoff_.NextAttemptTime();
|
193
|
+
Duration timeout = next_try - Timestamp::Now();
|
194
|
+
GPR_ASSERT(!have_next_resolution_timer_);
|
195
|
+
have_next_resolution_timer_ = true;
|
196
|
+
if (GPR_UNLIKELY(tracer_ != nullptr && tracer_->enabled())) {
|
197
|
+
if (timeout > Duration::Zero()) {
|
198
|
+
gpr_log(GPR_INFO, "[polling resolver %p] retrying in %" PRId64 " ms",
|
199
|
+
this, timeout.millis());
|
200
|
+
} else {
|
201
|
+
gpr_log(GPR_INFO, "[polling resolver %p] retrying immediately", this);
|
202
|
+
}
|
203
|
+
}
|
204
|
+
Ref(DEBUG_LOCATION, "next_resolution_timer").release();
|
205
|
+
GRPC_CLOSURE_INIT(&on_next_resolution_, OnNextResolution, this, nullptr);
|
206
|
+
grpc_timer_init(&next_resolution_timer_, next_try, &on_next_resolution_);
|
207
|
+
// Reset result_status_state_. Note that even if re-resolution was
|
208
|
+
// requested while the result-health callback was pending, we can
|
209
|
+
// ignore it here, because we are in backoff to re-resolve anyway.
|
210
|
+
result_status_state_ = ResultStatusState::kNone;
|
211
|
+
}
|
212
|
+
}
|
213
|
+
|
167
214
|
void PollingResolver::MaybeStartResolvingLocked() {
|
168
215
|
// If there is an existing timer, the time it fires is the earliest time we
|
169
216
|
// can start the next resolution.
|
@@ -176,11 +223,11 @@ void PollingResolver::MaybeStartResolvingLocked() {
|
|
176
223
|
const Timestamp earliest_next_resolution =
|
177
224
|
*last_resolution_timestamp_ + min_time_between_resolutions_;
|
178
225
|
const Duration time_until_next_resolution =
|
179
|
-
earliest_next_resolution -
|
226
|
+
earliest_next_resolution - Timestamp::Now();
|
180
227
|
if (time_until_next_resolution > Duration::Zero()) {
|
181
228
|
if (GPR_UNLIKELY(tracer_ != nullptr && tracer_->enabled())) {
|
182
229
|
const Duration last_resolution_ago =
|
183
|
-
|
230
|
+
Timestamp::Now() - *last_resolution_timestamp_;
|
184
231
|
gpr_log(GPR_INFO,
|
185
232
|
"[polling resolver %p] in cooldown from last resolution "
|
186
233
|
"(from %" PRId64 " ms ago); will resolve again in %" PRId64
|
@@ -192,7 +239,7 @@ void PollingResolver::MaybeStartResolvingLocked() {
|
|
192
239
|
Ref(DEBUG_LOCATION, "next_resolution_timer_cooldown").release();
|
193
240
|
GRPC_CLOSURE_INIT(&on_next_resolution_, OnNextResolution, this, nullptr);
|
194
241
|
grpc_timer_init(&next_resolution_timer_,
|
195
|
-
|
242
|
+
Timestamp::Now() + time_until_next_resolution,
|
196
243
|
&on_next_resolution_);
|
197
244
|
return;
|
198
245
|
}
|
@@ -202,7 +249,7 @@ void PollingResolver::MaybeStartResolvingLocked() {
|
|
202
249
|
|
203
250
|
void PollingResolver::StartResolvingLocked() {
|
204
251
|
request_ = StartRequest();
|
205
|
-
last_resolution_timestamp_ =
|
252
|
+
last_resolution_timestamp_ = Timestamp::Now();
|
206
253
|
if (GPR_UNLIKELY(tracer_ != nullptr && tracer_->enabled())) {
|
207
254
|
gpr_log(GPR_INFO, "[polling resolver %p] starting resolution, request_=%p",
|
208
255
|
this, request_.get());
|
@@ -22,6 +22,7 @@
|
|
22
22
|
#include <memory>
|
23
23
|
#include <string>
|
24
24
|
|
25
|
+
#include "absl/status/status.h"
|
25
26
|
#include "absl/types/optional.h"
|
26
27
|
|
27
28
|
#include "src/core/lib/backoff/backoff.h"
|
@@ -77,6 +78,8 @@ class PollingResolver : public Resolver {
|
|
77
78
|
|
78
79
|
void OnRequestCompleteLocked(Result result);
|
79
80
|
|
81
|
+
void GetResultStatus(absl::Status status);
|
82
|
+
|
80
83
|
static void OnNextResolution(void* arg, grpc_error_handle error);
|
81
84
|
void OnNextResolutionLocked(grpc_error_handle error);
|
82
85
|
|
@@ -105,6 +108,14 @@ class PollingResolver : public Resolver {
|
|
105
108
|
absl::optional<Timestamp> last_resolution_timestamp_;
|
106
109
|
/// retry backoff state
|
107
110
|
BackOff backoff_;
|
111
|
+
/// state for handling interactions between re-resolution requests and
|
112
|
+
/// result health callbacks
|
113
|
+
enum class ResultStatusState {
|
114
|
+
kNone,
|
115
|
+
kResultHealthCallbackPending,
|
116
|
+
kReresolutionRequestedWhileCallbackWasPending,
|
117
|
+
};
|
118
|
+
ResultStatusState result_status_state_ = ResultStatusState::kNone;
|
108
119
|
};
|
109
120
|
|
110
121
|
} // namespace grpc_core
|
@@ -56,6 +56,7 @@
|
|
56
56
|
#include "src/core/ext/filters/client_channel/config_selector.h"
|
57
57
|
#include "src/core/ext/filters/client_channel/lb_policy/ring_hash/ring_hash.h"
|
58
58
|
#include "src/core/ext/xds/xds_bootstrap.h"
|
59
|
+
#include "src/core/ext/xds/xds_bootstrap_grpc.h"
|
59
60
|
#include "src/core/ext/xds/xds_client.h"
|
60
61
|
#include "src/core/ext/xds/xds_client_grpc.h"
|
61
62
|
#include "src/core/ext/xds/xds_http_filters.h"
|
@@ -813,7 +814,8 @@ void XdsResolver::StartLocked() {
|
|
813
814
|
if (!uri_.authority().empty()) {
|
814
815
|
// target_uri.authority is set case
|
815
816
|
const auto* authority_config =
|
816
|
-
|
817
|
+
static_cast<const GrpcXdsBootstrap::GrpcAuthority*>(
|
818
|
+
xds_client_->bootstrap().LookupAuthority(uri_.authority()));
|
817
819
|
if (authority_config == nullptr) {
|
818
820
|
absl::Status status = absl::UnavailableError(
|
819
821
|
absl::StrCat("Invalid target URI -- authority not found for ",
|
@@ -826,7 +828,7 @@ void XdsResolver::StartLocked() {
|
|
826
828
|
return;
|
827
829
|
}
|
828
830
|
std::string name_template =
|
829
|
-
authority_config->client_listener_resource_name_template;
|
831
|
+
authority_config->client_listener_resource_name_template();
|
830
832
|
if (name_template.empty()) {
|
831
833
|
name_template = absl::StrCat(
|
832
834
|
"xdstp://", URI::PercentEncodeAuthority(uri_.authority()),
|
@@ -838,7 +840,7 @@ void XdsResolver::StartLocked() {
|
|
838
840
|
} else {
|
839
841
|
// target_uri.authority not set
|
840
842
|
absl::string_view name_template =
|
841
|
-
xds_client_->bootstrap()
|
843
|
+
static_cast<const GrpcXdsBootstrap&>(xds_client_->bootstrap())
|
842
844
|
.client_default_listener_resource_name_template();
|
843
845
|
if (name_template.empty()) {
|
844
846
|
name_template = "%s";
|
@@ -85,12 +85,13 @@ absl::StatusOr<std::unique_ptr<ServiceConfigParser::ParsedConfig>>
|
|
85
85
|
ClientChannelServiceConfigParser::ParseGlobalParams(const ChannelArgs& /*args*/,
|
86
86
|
const Json& json) {
|
87
87
|
std::vector<grpc_error_handle> error_list;
|
88
|
+
const auto& lb_policy_registry =
|
89
|
+
CoreConfiguration::Get().lb_policy_registry();
|
88
90
|
// Parse LB config.
|
89
91
|
RefCountedPtr<LoadBalancingPolicy::Config> parsed_lb_config;
|
90
92
|
auto it = json.object_value().find("loadBalancingConfig");
|
91
93
|
if (it != json.object_value().end()) {
|
92
|
-
auto config =
|
93
|
-
LoadBalancingPolicyRegistry::ParseLoadBalancingConfig(it->second);
|
94
|
+
auto config = lb_policy_registry.ParseLoadBalancingConfig(it->second);
|
94
95
|
if (!config.ok()) {
|
95
96
|
error_list.push_back(GRPC_ERROR_CREATE_FROM_CPP_STRING(absl::StrCat(
|
96
97
|
"field:loadBalancingConfig error:", config.status().message())));
|
@@ -111,8 +112,8 @@ ClientChannelServiceConfigParser::ParseGlobalParams(const ChannelArgs& /*args*/,
|
|
111
112
|
lb_policy_name[i] = tolower(lb_policy_name[i]);
|
112
113
|
}
|
113
114
|
bool requires_config = false;
|
114
|
-
if (!
|
115
|
-
|
115
|
+
if (!lb_policy_registry.LoadBalancingPolicyExists(lb_policy_name.c_str(),
|
116
|
+
&requires_config)) {
|
116
117
|
error_list.push_back(GRPC_ERROR_CREATE_FROM_STATIC_STRING(
|
117
118
|
"field:loadBalancingPolicy error:Unknown lb policy"));
|
118
119
|
} else if (requires_config) {
|
@@ -33,7 +33,6 @@
|
|
33
33
|
#include "absl/strings/string_view.h"
|
34
34
|
#include "absl/strings/strip.h"
|
35
35
|
#include "absl/types/optional.h"
|
36
|
-
#include "absl/utility/utility.h"
|
37
36
|
|
38
37
|
#include <grpc/impl/codegen/grpc_types.h>
|
39
38
|
#include <grpc/slice.h>
|
@@ -351,7 +350,11 @@ class RetryFilter::CallData {
|
|
351
350
|
// cancel_stream op.
|
352
351
|
static void OnCompleteForCancelOp(void* arg, grpc_error_handle error);
|
353
352
|
|
354
|
-
RefCountedPtr
|
353
|
+
// This DOES hold a ref, but it cannot be a RefCountedPtr<>, because
|
354
|
+
// our dtor unrefs the owning call, which may delete the arena in
|
355
|
+
// which we are allocated, which means that running the dtor of any
|
356
|
+
// data members after that would cause a crash.
|
357
|
+
CallAttempt* call_attempt_;
|
355
358
|
// The batch to use in the LB call.
|
356
359
|
// Its payload field points to CallAttempt::batch_payload_.
|
357
360
|
grpc_transport_stream_op_batch batch_;
|
@@ -722,8 +725,7 @@ RetryFilter::CallData::CallAttempt::CallAttempt(CallData* calld,
|
|
722
725
|
if (calld->retry_policy_ != nullptr &&
|
723
726
|
calld->retry_policy_->per_attempt_recv_timeout().has_value()) {
|
724
727
|
Timestamp per_attempt_recv_deadline =
|
725
|
-
|
726
|
-
*calld->retry_policy_->per_attempt_recv_timeout();
|
728
|
+
Timestamp::Now() + *calld->retry_policy_->per_attempt_recv_timeout();
|
727
729
|
if (GRPC_TRACE_FLAG_ENABLED(grpc_retry_trace)) {
|
728
730
|
gpr_log(GPR_INFO,
|
729
731
|
"chand=%p calld=%p attempt=%p: per-attempt timeout in %" PRId64
|
@@ -1330,11 +1332,11 @@ RetryFilter::CallData::CallAttempt::BatchData::BatchData(
|
|
1330
1332
|
: RefCounted(
|
1331
1333
|
GRPC_TRACE_FLAG_ENABLED(grpc_retry_trace) ? "BatchData" : nullptr,
|
1332
1334
|
refcount),
|
1333
|
-
call_attempt_(
|
1335
|
+
call_attempt_(attempt.release()) {
|
1334
1336
|
if (GRPC_TRACE_FLAG_ENABLED(grpc_retry_trace)) {
|
1335
1337
|
gpr_log(GPR_INFO, "chand=%p calld=%p attempt=%p: creating batch %p",
|
1336
|
-
call_attempt_->calld_->chand_, call_attempt_->calld_,
|
1337
|
-
|
1338
|
+
call_attempt_->calld_->chand_, call_attempt_->calld_, call_attempt_,
|
1339
|
+
this);
|
1338
1340
|
}
|
1339
1341
|
// We hold a ref to the call stack for every batch sent on a call attempt.
|
1340
1342
|
// This is because some batches on the call attempt may not complete
|
@@ -1354,11 +1356,12 @@ RetryFilter::CallData::CallAttempt::BatchData::BatchData(
|
|
1354
1356
|
RetryFilter::CallData::CallAttempt::BatchData::~BatchData() {
|
1355
1357
|
if (GRPC_TRACE_FLAG_ENABLED(grpc_retry_trace)) {
|
1356
1358
|
gpr_log(GPR_INFO, "chand=%p calld=%p attempt=%p: destroying batch %p",
|
1357
|
-
call_attempt_->calld_->chand_, call_attempt_->calld_,
|
1358
|
-
|
1359
|
+
call_attempt_->calld_->chand_, call_attempt_->calld_, call_attempt_,
|
1360
|
+
this);
|
1359
1361
|
}
|
1360
|
-
|
1361
|
-
|
1362
|
+
CallAttempt* call_attempt = std::exchange(call_attempt_, nullptr);
|
1363
|
+
GRPC_CALL_STACK_UNREF(call_attempt->calld_->owning_call_, "Retry BatchData");
|
1364
|
+
call_attempt->Unref(DEBUG_LOCATION, "~BatchData");
|
1362
1365
|
}
|
1363
1366
|
|
1364
1367
|
void RetryFilter::CallData::CallAttempt::BatchData::
|
@@ -1422,7 +1425,7 @@ void RetryFilter::CallData::CallAttempt::BatchData::
|
|
1422
1425
|
void RetryFilter::CallData::CallAttempt::BatchData::RecvInitialMetadataReady(
|
1423
1426
|
void* arg, grpc_error_handle error) {
|
1424
1427
|
RefCountedPtr<BatchData> batch_data(static_cast<BatchData*>(arg));
|
1425
|
-
CallAttempt* call_attempt = batch_data->call_attempt_
|
1428
|
+
CallAttempt* call_attempt = batch_data->call_attempt_;
|
1426
1429
|
CallData* calld = call_attempt->calld_;
|
1427
1430
|
if (GRPC_TRACE_FLAG_ENABLED(grpc_retry_trace)) {
|
1428
1431
|
gpr_log(GPR_INFO,
|
@@ -1524,7 +1527,7 @@ void RetryFilter::CallData::CallAttempt::BatchData::
|
|
1524
1527
|
void RetryFilter::CallData::CallAttempt::BatchData::RecvMessageReady(
|
1525
1528
|
void* arg, grpc_error_handle error) {
|
1526
1529
|
RefCountedPtr<BatchData> batch_data(static_cast<BatchData*>(arg));
|
1527
|
-
CallAttempt* call_attempt = batch_data->call_attempt_
|
1530
|
+
CallAttempt* call_attempt = batch_data->call_attempt_;
|
1528
1531
|
CallData* calld = call_attempt->calld_;
|
1529
1532
|
if (GRPC_TRACE_FLAG_ENABLED(grpc_retry_trace)) {
|
1530
1533
|
gpr_log(GPR_INFO,
|
@@ -1722,15 +1725,8 @@ void RetryFilter::CallData::CallAttempt::BatchData::RunClosuresForCompletedCall(
|
|
1722
1725
|
|
1723
1726
|
void RetryFilter::CallData::CallAttempt::BatchData::RecvTrailingMetadataReady(
|
1724
1727
|
void* arg, grpc_error_handle error) {
|
1725
|
-
// Add a ref to the outer call stack. This works around a TSAN reported
|
1726
|
-
// failure that we do not understand, but since we expect the lifetime of
|
1727
|
-
// this code to be relatively short we think it's OK to work around for now.
|
1728
|
-
// TSAN failure:
|
1729
|
-
// https://source.cloud.google.com/results/invocations/5b122974-4977-4862-beb1-dc1af9fbbd1d/targets/%2F%2Ftest%2Fcore%2Fend2end:h2_census_test@max_concurrent_streams@poller%3Dpoll/log
|
1730
|
-
RefCountedPtr<grpc_call_stack> owning_call_stack =
|
1731
|
-
static_cast<BatchData*>(arg)->call_attempt_->calld_->owning_call_->Ref();
|
1732
1728
|
RefCountedPtr<BatchData> batch_data(static_cast<BatchData*>(arg));
|
1733
|
-
CallAttempt* call_attempt = batch_data->call_attempt_
|
1729
|
+
CallAttempt* call_attempt = batch_data->call_attempt_;
|
1734
1730
|
CallData* calld = call_attempt->calld_;
|
1735
1731
|
if (GRPC_TRACE_FLAG_ENABLED(grpc_retry_trace)) {
|
1736
1732
|
gpr_log(GPR_INFO,
|
@@ -1886,7 +1882,7 @@ void RetryFilter::CallData::CallAttempt::BatchData::
|
|
1886
1882
|
gpr_log(GPR_INFO,
|
1887
1883
|
"chand=%p calld=%p attempt=%p: starting next batch for pending "
|
1888
1884
|
"send op(s)",
|
1889
|
-
calld->chand_, calld, call_attempt_
|
1885
|
+
calld->chand_, calld, call_attempt_);
|
1890
1886
|
}
|
1891
1887
|
call_attempt_->AddRetriableBatches(closures);
|
1892
1888
|
}
|
@@ -1895,7 +1891,7 @@ void RetryFilter::CallData::CallAttempt::BatchData::
|
|
1895
1891
|
void RetryFilter::CallData::CallAttempt::BatchData::OnComplete(
|
1896
1892
|
void* arg, grpc_error_handle error) {
|
1897
1893
|
RefCountedPtr<BatchData> batch_data(static_cast<BatchData*>(arg));
|
1898
|
-
CallAttempt* call_attempt = batch_data->call_attempt_
|
1894
|
+
CallAttempt* call_attempt = batch_data->call_attempt_;
|
1899
1895
|
CallData* calld = call_attempt->calld_;
|
1900
1896
|
if (GRPC_TRACE_FLAG_ENABLED(grpc_retry_trace)) {
|
1901
1897
|
gpr_log(GPR_INFO,
|
@@ -1971,7 +1967,7 @@ void RetryFilter::CallData::CallAttempt::BatchData::OnComplete(
|
|
1971
1967
|
void RetryFilter::CallData::CallAttempt::BatchData::OnCompleteForCancelOp(
|
1972
1968
|
void* arg, grpc_error_handle error) {
|
1973
1969
|
RefCountedPtr<BatchData> batch_data(static_cast<BatchData*>(arg));
|
1974
|
-
CallAttempt* call_attempt = batch_data->call_attempt_
|
1970
|
+
CallAttempt* call_attempt = batch_data->call_attempt_;
|
1975
1971
|
CallData* calld = call_attempt->calld_;
|
1976
1972
|
if (GRPC_TRACE_FLAG_ENABLED(grpc_retry_trace)) {
|
1977
1973
|
gpr_log(GPR_INFO,
|
@@ -2022,7 +2018,7 @@ void RetryFilter::CallData::CallAttempt::BatchData::
|
|
2022
2018
|
GPR_INFO,
|
2023
2019
|
"chand=%p calld=%p attempt=%p: starting calld->send_messages[%" PRIuPTR
|
2024
2020
|
"]",
|
2025
|
-
calld->chand_, calld, call_attempt_
|
2021
|
+
calld->chand_, calld, call_attempt_,
|
2026
2022
|
call_attempt_->started_send_message_count_);
|
2027
2023
|
}
|
2028
2024
|
CachedSendMessage cache =
|
@@ -2367,7 +2363,7 @@ void RetryFilter::CallData::MaybeCacheSendOpsForBatch(PendingBatch* pending) {
|
|
2367
2363
|
// Set up cache for send_message ops.
|
2368
2364
|
if (batch->send_message) {
|
2369
2365
|
SliceBuffer* cache = arena_->New<SliceBuffer>(std::move(
|
2370
|
-
*
|
2366
|
+
*std::exchange(batch->payload->send_message.send_message, nullptr)));
|
2371
2367
|
send_messages_.push_back({cache, batch->payload->send_message.flags});
|
2372
2368
|
}
|
2373
2369
|
// Save metadata batch for send_trailing_metadata ops.
|
@@ -2394,7 +2390,7 @@ void RetryFilter::CallData::FreeCachedSendMessage(size_t idx) {
|
|
2394
2390
|
"chand=%p calld=%p: destroying send_messages[%" PRIuPTR "]",
|
2395
2391
|
chand_, this, idx);
|
2396
2392
|
}
|
2397
|
-
Destruct(
|
2393
|
+
Destruct(std::exchange(send_messages_[idx].slices, nullptr));
|
2398
2394
|
}
|
2399
2395
|
}
|
2400
2396
|
|
@@ -2607,7 +2603,7 @@ void RetryFilter::CallData::StartRetryTimer(
|
|
2607
2603
|
Timestamp next_attempt_time;
|
2608
2604
|
if (server_pushback.has_value()) {
|
2609
2605
|
GPR_ASSERT(*server_pushback >= Duration::Zero());
|
2610
|
-
next_attempt_time =
|
2606
|
+
next_attempt_time = Timestamp::Now() + *server_pushback;
|
2611
2607
|
retry_backoff_.Reset();
|
2612
2608
|
} else {
|
2613
2609
|
next_attempt_time = retry_backoff_.NextAttemptTime();
|
@@ -2615,7 +2611,7 @@ void RetryFilter::CallData::StartRetryTimer(
|
|
2615
2611
|
if (GRPC_TRACE_FLAG_ENABLED(grpc_retry_trace)) {
|
2616
2612
|
gpr_log(GPR_INFO,
|
2617
2613
|
"chand=%p calld=%p: retrying failed call in %" PRId64 " ms", chand_,
|
2618
|
-
this, (next_attempt_time -
|
2614
|
+
this, (next_attempt_time - Timestamp::Now()).millis());
|
2619
2615
|
}
|
2620
2616
|
// Schedule retry after computed delay.
|
2621
2617
|
GRPC_CLOSURE_INIT(&retry_closure_, OnRetryTimer, this, nullptr);
|
@@ -27,15 +27,16 @@
|
|
27
27
|
#include <utility>
|
28
28
|
|
29
29
|
#include "absl/status/statusor.h"
|
30
|
+
#include "absl/strings/cord.h"
|
31
|
+
#include "absl/strings/str_cat.h"
|
32
|
+
#include "absl/strings/string_view.h"
|
30
33
|
|
31
|
-
#include <grpc/grpc.h>
|
32
34
|
#include <grpc/impl/codegen/grpc_types.h>
|
33
35
|
#include <grpc/slice.h>
|
34
36
|
#include <grpc/status.h>
|
35
37
|
#include <grpc/support/log.h>
|
36
38
|
|
37
39
|
#include "src/core/ext/filters/client_channel/health/health_check_client.h"
|
38
|
-
#include "src/core/ext/filters/client_channel/proxy_mapper_registry.h"
|
39
40
|
#include "src/core/ext/filters/client_channel/subchannel_pool_interface.h"
|
40
41
|
#include "src/core/ext/filters/client_channel/subchannel_stream_client.h"
|
41
42
|
#include "src/core/lib/address_utils/sockaddr_utils.h"
|
@@ -55,18 +56,15 @@
|
|
55
56
|
#include "src/core/lib/gprpp/debug_location.h"
|
56
57
|
#include "src/core/lib/gprpp/ref_counted_ptr.h"
|
57
58
|
#include "src/core/lib/gprpp/sync.h"
|
59
|
+
#include "src/core/lib/handshaker/proxy_mapper_registry.h"
|
58
60
|
#include "src/core/lib/iomgr/exec_ctx.h"
|
59
61
|
#include "src/core/lib/iomgr/pollset_set.h"
|
60
|
-
#include "src/core/lib/profiling/timers.h"
|
61
62
|
#include "src/core/lib/surface/channel_init.h"
|
62
63
|
#include "src/core/lib/surface/channel_stack_type.h"
|
64
|
+
#include "src/core/lib/surface/init_internally.h"
|
63
65
|
#include "src/core/lib/transport/connectivity_state.h"
|
64
66
|
#include "src/core/lib/transport/error_utils.h"
|
65
67
|
|
66
|
-
// Strong and weak refs.
|
67
|
-
#define INTERNAL_REF_BITS 16
|
68
|
-
#define STRONG_REF_MASK (~(gpr_atm)((1 << INTERNAL_REF_BITS) - 1))
|
69
|
-
|
70
68
|
// Backoff parameters.
|
71
69
|
#define GRPC_SUBCHANNEL_INITIAL_CONNECT_BACKOFF_SECONDS 1
|
72
70
|
#define GRPC_SUBCHANNEL_RECONNECT_BACKOFF_MULTIPLIER 1.6
|
@@ -175,7 +173,6 @@ SubchannelCall::SubchannelCall(Args args, grpc_error_handle* error)
|
|
175
173
|
|
176
174
|
void SubchannelCall::StartTransportStreamOpBatch(
|
177
175
|
grpc_transport_stream_op_batch* batch) {
|
178
|
-
GPR_TIMER_SCOPE("subchannel_call_process_op", 0);
|
179
176
|
MaybeInterceptRecvTrailingMetadata(batch);
|
180
177
|
grpc_call_stack* call_stack = SUBCHANNEL_CALL_TO_CALL_STACK(this);
|
181
178
|
grpc_call_element* top_elem = grpc_call_stack_element(call_stack, 0);
|
@@ -214,7 +211,6 @@ void SubchannelCall::Unref(const DebugLocation& /*location*/,
|
|
214
211
|
}
|
215
212
|
|
216
213
|
void SubchannelCall::Destroy(void* arg, grpc_error_handle /*error*/) {
|
217
|
-
GPR_TIMER_SCOPE("subchannel_call_destroy", 0);
|
218
214
|
SubchannelCall* self = static_cast<SubchannelCall*>(arg);
|
219
215
|
// Keep some members before destroying the subchannel call.
|
220
216
|
grpc_closure* after_call_stack_destroy = self->after_call_stack_destroy_;
|
@@ -640,13 +636,15 @@ Subchannel::Subchannel(SubchannelKey key,
|
|
640
636
|
// before the last subchannel destruction, then there maybe race conditions
|
641
637
|
// triggering segmentation faults. To prevent this issue, we call a grpc_init
|
642
638
|
// here and a grpc_shutdown in the subchannel destructor.
|
643
|
-
|
639
|
+
InitInternally();
|
644
640
|
GRPC_STATS_INC_CLIENT_SUBCHANNELS_CREATED();
|
645
641
|
GRPC_CLOSURE_INIT(&on_connecting_finished_, OnConnectingFinished, this,
|
646
642
|
grpc_schedule_on_exec_ctx);
|
647
643
|
// Check proxy mapper to determine address to connect to and channel
|
648
644
|
// args to use.
|
649
|
-
address_for_connect_ =
|
645
|
+
address_for_connect_ = CoreConfiguration::Get()
|
646
|
+
.proxy_mapper_registry()
|
647
|
+
.MapAddress(key_.address(), &args_)
|
650
648
|
.value_or(key_.address());
|
651
649
|
// Initialize channelz.
|
652
650
|
const bool channelz_enabled = args_.GetBool(GRPC_ARG_ENABLE_CHANNELZ)
|
@@ -676,7 +674,7 @@ Subchannel::~Subchannel() {
|
|
676
674
|
connector_.reset();
|
677
675
|
grpc_pollset_set_destroy(pollset_set_);
|
678
676
|
// grpc_shutdown is called here because grpc_init is called in the ctor.
|
679
|
-
|
677
|
+
ShutdownInternally();
|
680
678
|
}
|
681
679
|
|
682
680
|
RefCountedPtr<Subchannel> Subchannel::Create(
|
@@ -769,7 +767,7 @@ void Subchannel::ResetBackoff() {
|
|
769
767
|
GetDefaultEventEngine()->Cancel(retry_timer_handle_)) {
|
770
768
|
OnRetryTimerLocked();
|
771
769
|
} else if (state_ == GRPC_CHANNEL_CONNECTING) {
|
772
|
-
next_attempt_time_ =
|
770
|
+
next_attempt_time_ = Timestamp::Now();
|
773
771
|
}
|
774
772
|
}
|
775
773
|
|
@@ -788,27 +786,20 @@ void Subchannel::Orphan() {
|
|
788
786
|
health_watcher_map_.ShutdownLocked();
|
789
787
|
}
|
790
788
|
|
791
|
-
void Subchannel::
|
789
|
+
void Subchannel::GetOrAddDataProducer(
|
790
|
+
UniqueTypeName type,
|
791
|
+
std::function<void(DataProducerInterface**)> get_or_add) {
|
792
792
|
MutexLock lock(&mu_);
|
793
|
-
auto
|
794
|
-
|
795
|
-
entry = data_producer;
|
793
|
+
auto it = data_producer_map_.emplace(type, nullptr).first;
|
794
|
+
get_or_add(&it->second);
|
796
795
|
}
|
797
796
|
|
798
797
|
void Subchannel::RemoveDataProducer(DataProducerInterface* data_producer) {
|
799
798
|
MutexLock lock(&mu_);
|
800
799
|
auto it = data_producer_map_.find(data_producer->type());
|
801
|
-
|
802
|
-
|
803
|
-
|
804
|
-
}
|
805
|
-
|
806
|
-
Subchannel::DataProducerInterface* Subchannel::GetDataProducer(
|
807
|
-
UniqueTypeName type) {
|
808
|
-
MutexLock lock(&mu_);
|
809
|
-
auto it = data_producer_map_.find(type);
|
810
|
-
if (it == data_producer_map_.end()) return nullptr;
|
811
|
-
return it->second;
|
800
|
+
if (it != data_producer_map_.end() && it->second == data_producer) {
|
801
|
+
data_producer_map_.erase(it);
|
802
|
+
}
|
812
803
|
}
|
813
804
|
|
814
805
|
namespace {
|
@@ -838,7 +829,21 @@ const char* SubchannelConnectivityStateChangeString(
|
|
838
829
|
void Subchannel::SetConnectivityStateLocked(grpc_connectivity_state state,
|
839
830
|
const absl::Status& status) {
|
840
831
|
state_ = state;
|
841
|
-
|
832
|
+
if (status.ok()) {
|
833
|
+
status_ = status;
|
834
|
+
} else {
|
835
|
+
// Augment status message to include IP address.
|
836
|
+
status_ = absl::Status(status.code(),
|
837
|
+
absl::StrCat(grpc_sockaddr_to_uri(&key_.address())
|
838
|
+
.value_or("<unknown address type>"),
|
839
|
+
": ", status.message()));
|
840
|
+
status.ForEachPayload(
|
841
|
+
[this](absl::string_view key, const absl::Cord& value)
|
842
|
+
// Want to use ABSL_EXCLUSIVE_LOCKS_REQUIRED(&mu_) here,
|
843
|
+
// but that won't work, because we can't pass the lock
|
844
|
+
// annotation through absl::Status::ForEachPayload().
|
845
|
+
ABSL_NO_THREAD_SAFETY_ANALYSIS { status_.SetPayload(key, value); });
|
846
|
+
}
|
842
847
|
if (channelz_node_ != nullptr) {
|
843
848
|
channelz_node_->UpdateConnectivityState(state);
|
844
849
|
channelz_node_->AddTraceEvent(
|
@@ -847,9 +852,9 @@ void Subchannel::SetConnectivityStateLocked(grpc_connectivity_state state,
|
|
847
852
|
SubchannelConnectivityStateChangeString(state)));
|
848
853
|
}
|
849
854
|
// Notify non-health watchers.
|
850
|
-
watcher_list_.NotifyLocked(state,
|
855
|
+
watcher_list_.NotifyLocked(state, status_);
|
851
856
|
// Notify health watchers.
|
852
|
-
health_watcher_map_.NotifyLocked(state,
|
857
|
+
health_watcher_map_.NotifyLocked(state, status_);
|
853
858
|
}
|
854
859
|
|
855
860
|
void Subchannel::OnRetryTimer() {
|
@@ -866,7 +871,7 @@ void Subchannel::OnRetryTimerLocked() {
|
|
866
871
|
|
867
872
|
void Subchannel::StartConnectingLocked() {
|
868
873
|
// Set next attempt time.
|
869
|
-
const Timestamp min_deadline = min_connect_timeout_ +
|
874
|
+
const Timestamp min_deadline = min_connect_timeout_ + Timestamp::Now();
|
870
875
|
next_attempt_time_ = backoff_.NextAttemptTime();
|
871
876
|
// Report CONNECTING.
|
872
877
|
SetConnectivityStateLocked(GRPC_CHANNEL_CONNECTING, absl::OkStatus());
|
@@ -901,7 +906,7 @@ void Subchannel::OnConnectingFinishedLocked(grpc_error_handle error) {
|
|
901
906
|
// transition back to IDLE.
|
902
907
|
if (connecting_result_.transport == nullptr || !PublishTransportLocked()) {
|
903
908
|
const Duration time_until_next_attempt =
|
904
|
-
next_attempt_time_ -
|
909
|
+
next_attempt_time_ - Timestamp::Now();
|
905
910
|
gpr_log(GPR_INFO,
|
906
911
|
"subchannel %p %s: connect failed (%s), backing off for %" PRId64
|
907
912
|
" ms",
|
@@ -22,6 +22,7 @@
|
|
22
22
|
#include <stddef.h>
|
23
23
|
|
24
24
|
#include <deque>
|
25
|
+
#include <functional>
|
25
26
|
#include <map>
|
26
27
|
#include <string>
|
27
28
|
|
@@ -278,12 +279,20 @@ class Subchannel : public DualRefCounted<Subchannel> {
|
|
278
279
|
// We do not hold refs to the data producer; the implementation is
|
279
280
|
// expected to register itself upon construction and remove itself
|
280
281
|
// upon destruction.
|
281
|
-
|
282
|
+
//
|
283
|
+
// Looks up the current data producer for type and invokes get_or_add()
|
284
|
+
// with a pointer to that producer in the map. The get_or_add() function
|
285
|
+
// can modify the pointed-to value to update the map. This provides a
|
286
|
+
// way to either re-use an existing producer or register a new one in
|
287
|
+
// a non-racy way.
|
288
|
+
void GetOrAddDataProducer(
|
289
|
+
UniqueTypeName type,
|
290
|
+
std::function<void(DataProducerInterface**)> get_or_add)
|
282
291
|
ABSL_LOCKS_EXCLUDED(mu_);
|
292
|
+
// Removes the data producer from the map, if the current producer for
|
293
|
+
// this type is the specified producer.
|
283
294
|
void RemoveDataProducer(DataProducerInterface* data_producer)
|
284
295
|
ABSL_LOCKS_EXCLUDED(mu_);
|
285
|
-
DataProducerInterface* GetDataProducer(UniqueTypeName type)
|
286
|
-
ABSL_LOCKS_EXCLUDED(mu_);
|
287
296
|
|
288
297
|
private:
|
289
298
|
// A linked list of ConnectivityStateWatcherInterfaces that are monitoring
|