grpc 1.66.0.pre3 → 1.66.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/Makefile +1 -1
- data/src/core/handshaker/security/secure_endpoint.cc +15 -3
- data/src/core/lib/surface/call.cc +7 -0
- data/src/core/lib/surface/call.h +19 -0
- data/src/core/load_balancing/rls/rls.cc +99 -48
- data/src/ruby/ext/grpc/rb_call.c +1 -1
- data/src/ruby/ext/grpc/rb_call_credentials.c +1 -1
- data/src/ruby/ext/grpc/rb_grpc.c +1 -1
- data/src/ruby/ext/grpc/rb_server.c +3 -3
- data/src/ruby/lib/grpc/version.rb +1 -1
- metadata +3 -3
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: ce7eaaad7b69540dd61b0af07af801394a945137a41d75342533bef620fe1ad9
|
4
|
+
data.tar.gz: f186be6e589c40ca83cb6fdbb3d90f72b7ebb87115148a6f07f1698e3c5174e5
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: a931175742993f7d1c5c630ab2902313d386da09f36267bb31dd2c4fd8300dbf5b3f08caf805b3c71336028872790a79637d95be8e4714a13904c2fc1cade859
|
7
|
+
data.tar.gz: bd0dcbacaeb42f15cb8a733ccf59d105702818a8496cf13d8521d294faeaeaaddad45a80ea26b651ff503d40b4de46eef89f0954977e03a4081ffb9447c33873
|
data/Makefile
CHANGED
@@ -108,7 +108,6 @@ struct secure_endpoint : public grpc_endpoint {
|
|
108
108
|
}
|
109
109
|
|
110
110
|
~secure_endpoint() {
|
111
|
-
memory_owner.Reset();
|
112
111
|
tsi_frame_protector_destroy(protector);
|
113
112
|
tsi_zero_copy_grpc_protector_destroy(zero_copy_protector);
|
114
113
|
grpc_slice_buffer_destroy(&source_buffer);
|
@@ -254,6 +253,13 @@ static void on_read(void* user_data, grpc_error_handle error) {
|
|
254
253
|
|
255
254
|
{
|
256
255
|
grpc_core::MutexLock l(&ep->read_mu);
|
256
|
+
|
257
|
+
// If we were shut down after this callback was scheduled with OK
|
258
|
+
// status but before it was invoked, we need to treat that as an error.
|
259
|
+
if (ep->wrapped_ep == nullptr && error.ok()) {
|
260
|
+
error = absl::CancelledError("secure endpoint shutdown");
|
261
|
+
}
|
262
|
+
|
257
263
|
uint8_t* cur = GRPC_SLICE_START_PTR(ep->read_staging_buffer);
|
258
264
|
uint8_t* end = GRPC_SLICE_END_PTR(ep->read_staging_buffer);
|
259
265
|
|
@@ -380,9 +386,12 @@ static void flush_write_staging_buffer(secure_endpoint* ep, uint8_t** cur,
|
|
380
386
|
|
381
387
|
static void on_write(void* user_data, grpc_error_handle error) {
|
382
388
|
secure_endpoint* ep = static_cast<secure_endpoint*>(user_data);
|
383
|
-
|
384
|
-
|
389
|
+
grpc_closure* cb = ep->write_cb;
|
390
|
+
ep->write_cb = nullptr;
|
385
391
|
SECURE_ENDPOINT_UNREF(ep, "write");
|
392
|
+
grpc_core::EnsureRunInExecCtx([cb, error = std::move(error)]() {
|
393
|
+
grpc_core::Closure::Run(DEBUG_LOCATION, cb, error);
|
394
|
+
});
|
386
395
|
}
|
387
396
|
|
388
397
|
static void endpoint_write(grpc_endpoint* secure_ep, grpc_slice_buffer* slices,
|
@@ -503,7 +512,10 @@ static void endpoint_write(grpc_endpoint* secure_ep, grpc_slice_buffer* slices,
|
|
503
512
|
|
504
513
|
static void endpoint_destroy(grpc_endpoint* secure_ep) {
|
505
514
|
secure_endpoint* ep = reinterpret_cast<secure_endpoint*>(secure_ep);
|
515
|
+
ep->read_mu.Lock();
|
506
516
|
ep->wrapped_ep.reset();
|
517
|
+
ep->memory_owner.Reset();
|
518
|
+
ep->read_mu.Unlock();
|
507
519
|
SECURE_ENDPOINT_UNREF(ep, "destroy");
|
508
520
|
}
|
509
521
|
|
@@ -497,6 +497,13 @@ void grpc_call_tracer_set(grpc_call* call,
|
|
497
497
|
return arena->SetContext<grpc_core::CallTracerAnnotationInterface>(tracer);
|
498
498
|
}
|
499
499
|
|
500
|
+
void grpc_call_tracer_set_and_manage(grpc_call* call,
|
501
|
+
grpc_core::ClientCallTracer* tracer) {
|
502
|
+
grpc_core::Arena* arena = grpc_call_get_arena(call);
|
503
|
+
arena->ManagedNew<ClientCallTracerWrapper>(tracer);
|
504
|
+
return arena->SetContext<grpc_core::CallTracerAnnotationInterface>(tracer);
|
505
|
+
}
|
506
|
+
|
500
507
|
void* grpc_call_tracer_get(grpc_call* call) {
|
501
508
|
grpc_core::Arena* arena = grpc_call_get_arena(call);
|
502
509
|
auto* call_tracer =
|
data/src/core/lib/surface/call.h
CHANGED
@@ -265,6 +265,16 @@ void grpc_call_log_batch(const char* file, int line, const grpc_op* ops,
|
|
265
265
|
|
266
266
|
void grpc_call_tracer_set(grpc_call* call, grpc_core::ClientCallTracer* tracer);
|
267
267
|
|
268
|
+
// Sets call tracer on the call and manages its life by using the call's arena.
|
269
|
+
// When using this API, the tracer will be destroyed by grpc_call arena when
|
270
|
+
// grpc_call is about to be destroyed. The caller of this API SHOULD NOT
|
271
|
+
// manually destroy the tracer. This API is used by Python as a way of using
|
272
|
+
// Arena to manage the lifetime of the call tracer. Python needs this API
|
273
|
+
// because the tracer was created within a separate shared object library which
|
274
|
+
// doesn't have access to core functions like arena->ManagedNew<>.
|
275
|
+
void grpc_call_tracer_set_and_manage(grpc_call* call,
|
276
|
+
grpc_core::ClientCallTracer* tracer);
|
277
|
+
|
268
278
|
void* grpc_call_tracer_get(grpc_call* call);
|
269
279
|
|
270
280
|
#define GRPC_CALL_LOG_BATCH(ops, nops) \
|
@@ -276,6 +286,15 @@ void* grpc_call_tracer_get(grpc_call* call);
|
|
276
286
|
|
277
287
|
uint8_t grpc_call_is_client(grpc_call* call);
|
278
288
|
|
289
|
+
class ClientCallTracerWrapper {
|
290
|
+
public:
|
291
|
+
explicit ClientCallTracerWrapper(grpc_core::ClientCallTracer* tracer)
|
292
|
+
: tracer_(tracer) {}
|
293
|
+
|
294
|
+
private:
|
295
|
+
std::unique_ptr<grpc_core::ClientCallTracer> tracer_;
|
296
|
+
};
|
297
|
+
|
279
298
|
// Return an appropriate compression algorithm for the requested compression \a
|
280
299
|
// level in the context of \a call.
|
281
300
|
grpc_compression_algorithm grpc_call_compression_for_level(
|
@@ -353,7 +353,8 @@ class RlsLb final : public LoadBalancingPolicy {
|
|
353
353
|
// is called after releasing it.
|
354
354
|
//
|
355
355
|
// Both methods grab the data they need from the parent object.
|
356
|
-
void StartUpdate(
|
356
|
+
void StartUpdate(OrphanablePtr<ChildPolicyHandler>* child_policy_to_delete)
|
357
|
+
ABSL_EXCLUSIVE_LOCKS_REQUIRED(&RlsLb::mu_);
|
357
358
|
absl::Status MaybeFinishUpdate() ABSL_LOCKS_EXCLUDED(&RlsLb::mu_);
|
358
359
|
|
359
360
|
void ExitIdleLocked() {
|
@@ -397,14 +398,14 @@ class RlsLb final : public LoadBalancingPolicy {
|
|
397
398
|
};
|
398
399
|
|
399
400
|
// Note: We are forced to disable lock analysis here because
|
400
|
-
//
|
401
|
+
// Orphaned() is called by Unref() which is called by RefCountedPtr<>, which
|
401
402
|
// cannot have lock annotations for this particular caller.
|
402
403
|
void Orphaned() override ABSL_NO_THREAD_SAFETY_ANALYSIS;
|
403
404
|
|
404
405
|
RefCountedPtr<RlsLb> lb_policy_;
|
405
406
|
std::string target_;
|
406
407
|
|
407
|
-
bool is_shutdown_ = false;
|
408
|
+
bool is_shutdown_ = false; // Protected by WorkSerializer
|
408
409
|
|
409
410
|
OrphanablePtr<ChildPolicyHandler> child_policy_;
|
410
411
|
RefCountedPtr<LoadBalancingPolicy::Config> pending_config_;
|
@@ -503,12 +504,25 @@ class RlsLb final : public LoadBalancingPolicy {
|
|
503
504
|
// Returns a list of child policy wrappers on which FinishUpdate()
|
504
505
|
// needs to be called after releasing the lock.
|
505
506
|
std::vector<ChildPolicyWrapper*> OnRlsResponseLocked(
|
506
|
-
ResponseInfo response, std::unique_ptr<BackOff> backoff_state
|
507
|
+
ResponseInfo response, std::unique_ptr<BackOff> backoff_state,
|
508
|
+
OrphanablePtr<ChildPolicyHandler>* child_policy_to_delete)
|
507
509
|
ABSL_EXCLUSIVE_LOCKS_REQUIRED(&RlsLb::mu_);
|
508
510
|
|
509
511
|
// Moves entry to the end of the LRU list.
|
510
512
|
void MarkUsed() ABSL_EXCLUSIVE_LOCKS_REQUIRED(&RlsLb::mu_);
|
511
513
|
|
514
|
+
// Takes entries from child_policy_wrappers_ and appends them to the end
|
515
|
+
// of \a child_policy_wrappers.
|
516
|
+
void TakeChildPolicyWrappers(
|
517
|
+
std::vector<RefCountedPtr<ChildPolicyWrapper>>* child_policy_wrappers)
|
518
|
+
ABSL_EXCLUSIVE_LOCKS_REQUIRED(&RlsLb::mu_) {
|
519
|
+
child_policy_wrappers->insert(
|
520
|
+
child_policy_wrappers->end(),
|
521
|
+
std::make_move_iterator(child_policy_wrappers_.begin()),
|
522
|
+
std::make_move_iterator(child_policy_wrappers_.end()));
|
523
|
+
child_policy_wrappers_.clear();
|
524
|
+
}
|
525
|
+
|
512
526
|
private:
|
513
527
|
class BackoffTimer final : public InternallyRefCounted<BackoffTimer> {
|
514
528
|
public:
|
@@ -566,19 +580,24 @@ class RlsLb final : public LoadBalancingPolicy {
|
|
566
580
|
// the caller. Otherwise, the entry found is returned to the caller. The
|
567
581
|
// entry returned to the user is considered recently used and its order in
|
568
582
|
// the LRU list of the cache is updated.
|
569
|
-
Entry* FindOrInsert(const RequestKey& key
|
583
|
+
Entry* FindOrInsert(const RequestKey& key,
|
584
|
+
std::vector<RefCountedPtr<ChildPolicyWrapper>>*
|
585
|
+
child_policy_wrappers_to_delete)
|
570
586
|
ABSL_EXCLUSIVE_LOCKS_REQUIRED(&RlsLb::mu_);
|
571
587
|
|
572
588
|
// Resizes the cache. If the new cache size is greater than the current size
|
573
589
|
// of the cache, do nothing. Otherwise, evict the oldest entries that
|
574
590
|
// exceed the new size limit of the cache.
|
575
|
-
void Resize(size_t bytes
|
591
|
+
void Resize(size_t bytes, std::vector<RefCountedPtr<ChildPolicyWrapper>>*
|
592
|
+
child_policy_wrappers_to_delete)
|
593
|
+
ABSL_EXCLUSIVE_LOCKS_REQUIRED(&RlsLb::mu_);
|
576
594
|
|
577
595
|
// Resets backoff of all the cache entries.
|
578
596
|
void ResetAllBackoff() ABSL_EXCLUSIVE_LOCKS_REQUIRED(&RlsLb::mu_);
|
579
597
|
|
580
598
|
// Shutdown the cache; clean-up and orphan all the stored cache entries.
|
581
|
-
|
599
|
+
GRPC_MUST_USE_RESULT std::vector<RefCountedPtr<ChildPolicyWrapper>>
|
600
|
+
Shutdown() ABSL_EXCLUSIVE_LOCKS_REQUIRED(&RlsLb::mu_);
|
582
601
|
|
583
602
|
void ReportMetricsLocked(CallbackMetricReporter& reporter)
|
584
603
|
ABSL_EXCLUSIVE_LOCKS_REQUIRED(&RlsLb::mu_);
|
@@ -594,7 +613,9 @@ class RlsLb final : public LoadBalancingPolicy {
|
|
594
613
|
|
595
614
|
// Evicts oversized cache elements when the current size is greater than
|
596
615
|
// the specified limit.
|
597
|
-
void MaybeShrinkSize(size_t bytes
|
616
|
+
void MaybeShrinkSize(size_t bytes,
|
617
|
+
std::vector<RefCountedPtr<ChildPolicyWrapper>>*
|
618
|
+
child_policy_wrappers_to_delete)
|
598
619
|
ABSL_EXCLUSIVE_LOCKS_REQUIRED(&RlsLb::mu_);
|
599
620
|
|
600
621
|
RlsLb* lb_policy_;
|
@@ -859,7 +880,8 @@ absl::optional<Json> InsertOrUpdateChildPolicyField(const std::string& field,
|
|
859
880
|
return Json::FromArray(std::move(array));
|
860
881
|
}
|
861
882
|
|
862
|
-
void RlsLb::ChildPolicyWrapper::StartUpdate(
|
883
|
+
void RlsLb::ChildPolicyWrapper::StartUpdate(
|
884
|
+
OrphanablePtr<ChildPolicyHandler>* child_policy_to_delete) {
|
863
885
|
ValidationErrors errors;
|
864
886
|
auto child_policy_config = InsertOrUpdateChildPolicyField(
|
865
887
|
lb_policy_->config_->child_policy_config_target_field_name(), target_,
|
@@ -884,7 +906,7 @@ void RlsLb::ChildPolicyWrapper::StartUpdate() {
|
|
884
906
|
pending_config_.reset();
|
885
907
|
picker_ = MakeRefCounted<TransientFailurePicker>(
|
886
908
|
absl::UnavailableError(config.status().message()));
|
887
|
-
child_policy_
|
909
|
+
*child_policy_to_delete = std::move(child_policy_);
|
888
910
|
} else {
|
889
911
|
pending_config_ = std::move(*config);
|
890
912
|
}
|
@@ -939,9 +961,9 @@ void RlsLb::ChildPolicyWrapper::ChildPolicyHelper::UpdateState(
|
|
939
961
|
<< ": UpdateState(state=" << ConnectivityStateName(state)
|
940
962
|
<< ", status=" << status << ", picker=" << picker.get() << ")";
|
941
963
|
}
|
964
|
+
if (wrapper_->is_shutdown_) return;
|
942
965
|
{
|
943
966
|
MutexLock lock(&wrapper_->lb_policy_->mu_);
|
944
|
-
if (wrapper_->is_shutdown_) return;
|
945
967
|
// TODO(roth): It looks like this ignores subsequent TF updates that
|
946
968
|
// might change the status used to fail picks, which seems wrong.
|
947
969
|
if (wrapper_->connectivity_state_ == GRPC_CHANNEL_TRANSIENT_FAILURE &&
|
@@ -951,7 +973,8 @@ void RlsLb::ChildPolicyWrapper::ChildPolicyHelper::UpdateState(
|
|
951
973
|
wrapper_->connectivity_state_ = state;
|
952
974
|
DCHECK(picker != nullptr);
|
953
975
|
if (picker != nullptr) {
|
954
|
-
|
976
|
+
// We want to unref the picker after we release the lock.
|
977
|
+
wrapper_->picker_.swap(picker);
|
955
978
|
}
|
956
979
|
}
|
957
980
|
wrapper_->lb_policy_->UpdatePickerLocked();
|
@@ -1204,19 +1227,19 @@ RlsLb::Cache::Entry::Entry(RefCountedPtr<RlsLb> lb_policy,
|
|
1204
1227
|
lb_policy_->cache_.lru_list_.end(), key)) {}
|
1205
1228
|
|
1206
1229
|
void RlsLb::Cache::Entry::Orphan() {
|
1207
|
-
|
1208
|
-
|
1209
|
-
|
1210
|
-
|
1230
|
+
// We should be holding RlsLB::mu_.
|
1231
|
+
GRPC_TRACE_LOG(rls_lb, INFO)
|
1232
|
+
<< "[rlslb " << lb_policy_.get() << "] cache entry=" << this << " "
|
1233
|
+
<< lru_iterator_->ToString() << ": cache entry evicted";
|
1211
1234
|
is_shutdown_ = true;
|
1212
1235
|
lb_policy_->cache_.lru_list_.erase(lru_iterator_);
|
1213
1236
|
lru_iterator_ = lb_policy_->cache_.lru_list_.end(); // Just in case.
|
1237
|
+
CHECK(child_policy_wrappers_.empty());
|
1214
1238
|
backoff_state_.reset();
|
1215
1239
|
if (backoff_timer_ != nullptr) {
|
1216
1240
|
backoff_timer_.reset();
|
1217
1241
|
lb_policy_->UpdatePickerAsync();
|
1218
1242
|
}
|
1219
|
-
child_policy_wrappers_.clear();
|
1220
1243
|
Unref(DEBUG_LOCATION, "Orphan");
|
1221
1244
|
}
|
1222
1245
|
|
@@ -1295,7 +1318,8 @@ void RlsLb::Cache::Entry::MarkUsed() {
|
|
1295
1318
|
|
1296
1319
|
std::vector<RlsLb::ChildPolicyWrapper*>
|
1297
1320
|
RlsLb::Cache::Entry::OnRlsResponseLocked(
|
1298
|
-
ResponseInfo response, std::unique_ptr<BackOff> backoff_state
|
1321
|
+
ResponseInfo response, std::unique_ptr<BackOff> backoff_state,
|
1322
|
+
OrphanablePtr<ChildPolicyHandler>* child_policy_to_delete) {
|
1299
1323
|
// Move the entry to the end of the LRU list.
|
1300
1324
|
MarkUsed();
|
1301
1325
|
// If the request failed, store the failed status and update the
|
@@ -1356,7 +1380,7 @@ RlsLb::Cache::Entry::OnRlsResponseLocked(
|
|
1356
1380
|
if (it == lb_policy_->child_policy_map_.end()) {
|
1357
1381
|
auto new_child = MakeRefCounted<ChildPolicyWrapper>(
|
1358
1382
|
lb_policy_.Ref(DEBUG_LOCATION, "ChildPolicyWrapper"), target);
|
1359
|
-
new_child->StartUpdate();
|
1383
|
+
new_child->StartUpdate(child_policy_to_delete);
|
1360
1384
|
child_policies_to_finish_update.push_back(new_child.get());
|
1361
1385
|
new_child_policy_wrappers.emplace_back(std::move(new_child));
|
1362
1386
|
} else {
|
@@ -1393,12 +1417,15 @@ RlsLb::Cache::Entry* RlsLb::Cache::Find(const RequestKey& key) {
|
|
1393
1417
|
return it->second.get();
|
1394
1418
|
}
|
1395
1419
|
|
1396
|
-
RlsLb::Cache::Entry* RlsLb::Cache::FindOrInsert(
|
1420
|
+
RlsLb::Cache::Entry* RlsLb::Cache::FindOrInsert(
|
1421
|
+
const RequestKey& key, std::vector<RefCountedPtr<ChildPolicyWrapper>>*
|
1422
|
+
child_policy_wrappers_to_delete) {
|
1397
1423
|
auto it = map_.find(key);
|
1398
1424
|
// If not found, create new entry.
|
1399
1425
|
if (it == map_.end()) {
|
1400
1426
|
size_t entry_size = EntrySizeForKey(key);
|
1401
|
-
MaybeShrinkSize(size_limit_ - std::min(size_limit_, entry_size)
|
1427
|
+
MaybeShrinkSize(size_limit_ - std::min(size_limit_, entry_size),
|
1428
|
+
child_policy_wrappers_to_delete);
|
1402
1429
|
Entry* entry = new Entry(
|
1403
1430
|
lb_policy_->RefAsSubclass<RlsLb>(DEBUG_LOCATION, "CacheEntry"), key);
|
1404
1431
|
map_.emplace(key, OrphanablePtr<Entry>(entry));
|
@@ -1418,13 +1445,13 @@ RlsLb::Cache::Entry* RlsLb::Cache::FindOrInsert(const RequestKey& key) {
|
|
1418
1445
|
return it->second.get();
|
1419
1446
|
}
|
1420
1447
|
|
1421
|
-
void RlsLb::Cache::Resize(size_t bytes
|
1422
|
-
|
1423
|
-
|
1424
|
-
|
1425
|
-
|
1448
|
+
void RlsLb::Cache::Resize(size_t bytes,
|
1449
|
+
std::vector<RefCountedPtr<ChildPolicyWrapper>>*
|
1450
|
+
child_policy_wrappers_to_delete) {
|
1451
|
+
GRPC_TRACE_LOG(rls_lb, INFO)
|
1452
|
+
<< "[rlslb " << lb_policy_ << "] resizing cache to " << bytes << " bytes";
|
1426
1453
|
size_limit_ = bytes;
|
1427
|
-
MaybeShrinkSize(size_limit_);
|
1454
|
+
MaybeShrinkSize(size_limit_, child_policy_wrappers_to_delete);
|
1428
1455
|
}
|
1429
1456
|
|
1430
1457
|
void RlsLb::Cache::ResetAllBackoff() {
|
@@ -1434,7 +1461,12 @@ void RlsLb::Cache::ResetAllBackoff() {
|
|
1434
1461
|
lb_policy_->UpdatePickerAsync();
|
1435
1462
|
}
|
1436
1463
|
|
1437
|
-
|
1464
|
+
std::vector<RefCountedPtr<RlsLb::ChildPolicyWrapper>> RlsLb::Cache::Shutdown() {
|
1465
|
+
std::vector<RefCountedPtr<ChildPolicyWrapper>>
|
1466
|
+
child_policy_wrappers_to_delete;
|
1467
|
+
for (auto& entry : map_) {
|
1468
|
+
entry.second->TakeChildPolicyWrappers(&child_policy_wrappers_to_delete);
|
1469
|
+
}
|
1438
1470
|
map_.clear();
|
1439
1471
|
lru_list_.clear();
|
1440
1472
|
if (cleanup_timer_handle_.has_value() &&
|
@@ -1445,6 +1477,7 @@ void RlsLb::Cache::Shutdown() {
|
|
1445
1477
|
}
|
1446
1478
|
}
|
1447
1479
|
cleanup_timer_handle_.reset();
|
1480
|
+
return child_policy_wrappers_to_delete;
|
1448
1481
|
}
|
1449
1482
|
|
1450
1483
|
void RlsLb::Cache::ReportMetricsLocked(CallbackMetricReporter& reporter) {
|
@@ -1478,15 +1511,17 @@ void RlsLb::Cache::StartCleanupTimer() {
|
|
1478
1511
|
}
|
1479
1512
|
|
1480
1513
|
void RlsLb::Cache::OnCleanupTimer() {
|
1481
|
-
|
1482
|
-
|
1483
|
-
|
1514
|
+
GRPC_TRACE_LOG(rls_lb, INFO)
|
1515
|
+
<< "[rlslb " << lb_policy_ << "] cache cleanup timer fired";
|
1516
|
+
std::vector<RefCountedPtr<ChildPolicyWrapper>>
|
1517
|
+
child_policy_wrappers_to_delete;
|
1484
1518
|
MutexLock lock(&lb_policy_->mu_);
|
1485
1519
|
if (!cleanup_timer_handle_.has_value()) return;
|
1486
1520
|
if (lb_policy_->is_shutdown_) return;
|
1487
1521
|
for (auto it = map_.begin(); it != map_.end();) {
|
1488
1522
|
if (GPR_UNLIKELY(it->second->ShouldRemove() && it->second->CanEvict())) {
|
1489
1523
|
size_ -= it->second->Size();
|
1524
|
+
it->second->TakeChildPolicyWrappers(&child_policy_wrappers_to_delete);
|
1490
1525
|
it = map_.erase(it);
|
1491
1526
|
} else {
|
1492
1527
|
++it;
|
@@ -1500,7 +1535,9 @@ size_t RlsLb::Cache::EntrySizeForKey(const RequestKey& key) {
|
|
1500
1535
|
return (key.Size() * 2) + sizeof(Entry);
|
1501
1536
|
}
|
1502
1537
|
|
1503
|
-
void RlsLb::Cache::MaybeShrinkSize(
|
1538
|
+
void RlsLb::Cache::MaybeShrinkSize(
|
1539
|
+
size_t bytes, std::vector<RefCountedPtr<ChildPolicyWrapper>>*
|
1540
|
+
child_policy_wrappers_to_delete) {
|
1504
1541
|
while (size_ > bytes) {
|
1505
1542
|
auto lru_it = lru_list_.begin();
|
1506
1543
|
if (GPR_UNLIKELY(lru_it == lru_list_.end())) break;
|
@@ -1512,6 +1549,7 @@ void RlsLb::Cache::MaybeShrinkSize(size_t bytes) {
|
|
1512
1549
|
<< map_it->second.get() << " " << lru_it->ToString();
|
1513
1550
|
}
|
1514
1551
|
size_ -= map_it->second->Size();
|
1552
|
+
map_it->second->TakeChildPolicyWrappers(child_policy_wrappers_to_delete);
|
1515
1553
|
map_.erase(map_it);
|
1516
1554
|
}
|
1517
1555
|
if (GRPC_TRACE_FLAG_ENABLED(rls_lb)) {
|
@@ -1841,13 +1879,18 @@ void RlsLb::RlsRequest::OnRlsCallCompleteLocked(grpc_error_handle error) {
|
|
1841
1879
|
<< ": response info: " << response.ToString();
|
1842
1880
|
}
|
1843
1881
|
std::vector<ChildPolicyWrapper*> child_policies_to_finish_update;
|
1882
|
+
std::vector<RefCountedPtr<ChildPolicyWrapper>>
|
1883
|
+
child_policy_wrappers_to_delete;
|
1884
|
+
OrphanablePtr<ChildPolicyHandler> child_policy_to_delete;
|
1844
1885
|
{
|
1845
1886
|
MutexLock lock(&lb_policy_->mu_);
|
1846
1887
|
if (lb_policy_->is_shutdown_) return;
|
1847
1888
|
rls_channel_->ReportResponseLocked(response.status.ok());
|
1848
|
-
Cache::Entry* cache_entry =
|
1889
|
+
Cache::Entry* cache_entry =
|
1890
|
+
lb_policy_->cache_.FindOrInsert(key_, &child_policy_wrappers_to_delete);
|
1849
1891
|
child_policies_to_finish_update = cache_entry->OnRlsResponseLocked(
|
1850
|
-
std::move(response), std::move(backoff_state_)
|
1892
|
+
std::move(response), std::move(backoff_state_),
|
1893
|
+
&child_policy_to_delete);
|
1851
1894
|
lb_policy_->request_map_.erase(key_);
|
1852
1895
|
}
|
1853
1896
|
// Now that we've released the lock, finish the update on any newly
|
@@ -2041,6 +2084,9 @@ absl::Status RlsLb::UpdateLocked(UpdateArgs args) {
|
|
2041
2084
|
}
|
2042
2085
|
}
|
2043
2086
|
// Now grab the lock to swap out the state it guards.
|
2087
|
+
std::vector<RefCountedPtr<ChildPolicyWrapper>>
|
2088
|
+
child_policy_wrappers_to_delete;
|
2089
|
+
OrphanablePtr<ChildPolicyHandler> child_policy_to_delete;
|
2044
2090
|
{
|
2045
2091
|
MutexLock lock(&mu_);
|
2046
2092
|
// Swap out RLS channel if needed.
|
@@ -2052,7 +2098,8 @@ absl::Status RlsLb::UpdateLocked(UpdateArgs args) {
|
|
2052
2098
|
// Resize cache if needed.
|
2053
2099
|
if (old_config == nullptr ||
|
2054
2100
|
config_->cache_size_bytes() != old_config->cache_size_bytes()) {
|
2055
|
-
cache_.Resize(static_cast<size_t>(config_->cache_size_bytes())
|
2101
|
+
cache_.Resize(static_cast<size_t>(config_->cache_size_bytes()),
|
2102
|
+
&child_policy_wrappers_to_delete);
|
2056
2103
|
}
|
2057
2104
|
// Start update of child policies if needed.
|
2058
2105
|
if (update_child_policies) {
|
@@ -2060,14 +2107,12 @@ absl::Status RlsLb::UpdateLocked(UpdateArgs args) {
|
|
2060
2107
|
LOG(INFO) << "[rlslb " << this << "] starting child policy updates";
|
2061
2108
|
}
|
2062
2109
|
for (auto& p : child_policy_map_) {
|
2063
|
-
p.second->StartUpdate();
|
2110
|
+
p.second->StartUpdate(&child_policy_to_delete);
|
2064
2111
|
}
|
2065
2112
|
} else if (created_default_child) {
|
2066
|
-
|
2067
|
-
|
2068
|
-
|
2069
|
-
}
|
2070
|
-
default_child_policy_->StartUpdate();
|
2113
|
+
GRPC_TRACE_LOG(rls_lb, INFO)
|
2114
|
+
<< "[rlslb " << this << "] starting default child policy update";
|
2115
|
+
default_child_policy_->StartUpdate(&child_policy_to_delete);
|
2071
2116
|
}
|
2072
2117
|
}
|
2073
2118
|
// Now that we've released the lock, finish update of child policies.
|
@@ -2133,14 +2178,20 @@ void RlsLb::ShutdownLocked() {
|
|
2133
2178
|
LOG(INFO) << "[rlslb " << this << "] policy shutdown";
|
2134
2179
|
}
|
2135
2180
|
registered_metric_callback_.reset();
|
2136
|
-
|
2137
|
-
|
2138
|
-
|
2181
|
+
RefCountedPtr<ChildPolicyWrapper> child_policy_to_delete;
|
2182
|
+
std::vector<RefCountedPtr<ChildPolicyWrapper>>
|
2183
|
+
child_policy_wrappers_to_delete;
|
2184
|
+
OrphanablePtr<RlsChannel> rls_channel_to_delete;
|
2185
|
+
{
|
2186
|
+
MutexLock lock(&mu_);
|
2187
|
+
is_shutdown_ = true;
|
2188
|
+
config_.reset(DEBUG_LOCATION, "ShutdownLocked");
|
2189
|
+
child_policy_wrappers_to_delete = cache_.Shutdown();
|
2190
|
+
request_map_.clear();
|
2191
|
+
rls_channel_to_delete = std::move(rls_channel_);
|
2192
|
+
child_policy_to_delete = std::move(default_child_policy_);
|
2193
|
+
}
|
2139
2194
|
channel_args_ = ChannelArgs();
|
2140
|
-
cache_.Shutdown();
|
2141
|
-
request_map_.clear();
|
2142
|
-
rls_channel_.reset();
|
2143
|
-
default_child_policy_.reset();
|
2144
2195
|
}
|
2145
2196
|
|
2146
2197
|
void RlsLb::UpdatePickerAsync() {
|
data/src/ruby/ext/grpc/rb_call.c
CHANGED
@@ -809,7 +809,7 @@ struct call_run_batch_args {
|
|
809
809
|
};
|
810
810
|
|
811
811
|
static void cancel_call_unblock_func(void* arg) {
|
812
|
-
gpr_log(
|
812
|
+
gpr_log(GPR_DEBUG, "GRPC_RUBY: cancel_call_unblock_func");
|
813
813
|
grpc_call* call = (grpc_call*)arg;
|
814
814
|
grpc_call_cancel(call, NULL);
|
815
815
|
}
|
@@ -112,7 +112,7 @@ static VALUE grpc_rb_call_credentials_callback_rescue(VALUE args,
|
|
112
112
|
VALUE rb_exception_info =
|
113
113
|
rb_funcall(exception_object, rb_intern("inspect"), 0);
|
114
114
|
(void)args;
|
115
|
-
gpr_log(
|
115
|
+
gpr_log(GPR_DEBUG,
|
116
116
|
"GRPC_RUBY call credentials callback failed, exception inspect:|%s| "
|
117
117
|
"backtrace:|%s|",
|
118
118
|
StringValueCStr(rb_exception_info), StringValueCStr(backtrace_str));
|
data/src/ruby/ext/grpc/rb_grpc.c
CHANGED
@@ -328,7 +328,7 @@ static void grpc_ruby_init_threads() {
|
|
328
328
|
// in gpr_once_init. In general, it appears to be unsafe to call
|
329
329
|
// into the ruby library while holding a non-ruby mutex, because a gil yield
|
330
330
|
// could end up trying to lock onto that same mutex and deadlocking.
|
331
|
-
gpr_log(
|
331
|
+
gpr_log(GPR_DEBUG,
|
332
332
|
"GRPC_RUBY: grpc_ruby_init_threads g_bg_thread_init_done=%d",
|
333
333
|
g_bg_thread_init_done);
|
334
334
|
rb_mutex_lock(g_bg_thread_init_rb_mu);
|
@@ -68,7 +68,7 @@ static void grpc_rb_server_shutdown_and_notify_internal(grpc_rb_server* server,
|
|
68
68
|
server->queue, tag, gpr_inf_future(GPR_CLOCK_REALTIME), NULL, NULL);
|
69
69
|
}
|
70
70
|
if (ev.type != GRPC_OP_COMPLETE) {
|
71
|
-
gpr_log(
|
71
|
+
gpr_log(GPR_DEBUG,
|
72
72
|
"GRPC_RUBY: bad grpc_server_shutdown_and_notify result:%d",
|
73
73
|
ev.type);
|
74
74
|
}
|
@@ -192,7 +192,7 @@ struct server_request_call_args {
|
|
192
192
|
|
193
193
|
static void shutdown_server_unblock_func(void* arg) {
|
194
194
|
grpc_rb_server* server = (grpc_rb_server*)arg;
|
195
|
-
gpr_log(
|
195
|
+
gpr_log(GPR_DEBUG, "GRPC_RUBY: shutdown_server_unblock_func");
|
196
196
|
GRPC_RUBY_ASSERT(server->wrapped != NULL);
|
197
197
|
grpc_event event;
|
198
198
|
void* tag = &event;
|
@@ -202,7 +202,7 @@ static void shutdown_server_unblock_func(void* arg) {
|
|
202
202
|
// cancelled all calls.
|
203
203
|
event = grpc_completion_queue_pluck(server->queue, tag,
|
204
204
|
gpr_inf_future(GPR_CLOCK_REALTIME), NULL);
|
205
|
-
gpr_log(
|
205
|
+
gpr_log(GPR_DEBUG,
|
206
206
|
"GRPC_RUBY: shutdown_server_unblock_func pluck event.type: %d "
|
207
207
|
"event.success: %d",
|
208
208
|
event.type, event.success);
|
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: grpc
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.66.0
|
4
|
+
version: 1.66.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- gRPC Authors
|
8
8
|
autorequire:
|
9
9
|
bindir: src/ruby/bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2024-08-
|
11
|
+
date: 2024-08-22 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: google-protobuf
|
@@ -3595,7 +3595,7 @@ required_ruby_version: !ruby/object:Gem::Requirement
|
|
3595
3595
|
requirements:
|
3596
3596
|
- - ">="
|
3597
3597
|
- !ruby/object:Gem::Version
|
3598
|
-
version:
|
3598
|
+
version: '3.0'
|
3599
3599
|
required_rubygems_version: !ruby/object:Gem::Requirement
|
3600
3600
|
requirements:
|
3601
3601
|
- - ">="
|