|  | @@ -0,0 +1,359 @@
 | 
	
		
			
				|  |  | +/*
 | 
	
		
			
				|  |  | + *
 | 
	
		
			
				|  |  | + * Copyright 2020 gRPC authors.
 | 
	
		
			
				|  |  | + *
 | 
	
		
			
				|  |  | + * Licensed under the Apache License, Version 2.0 (the "License");
 | 
	
		
			
				|  |  | + * you may not use this file except in compliance with the License.
 | 
	
		
			
				|  |  | + * You may obtain a copy of the License at
 | 
	
		
			
				|  |  | + *
 | 
	
		
			
				|  |  | + *     http://www.apache.org/licenses/LICENSE-2.0
 | 
	
		
			
				|  |  | + *
 | 
	
		
			
				|  |  | + * Unless required by applicable law or agreed to in writing, software
 | 
	
		
			
				|  |  | + * distributed under the License is distributed on an "AS IS" BASIS,
 | 
	
		
			
				|  |  | + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
	
		
			
				|  |  | + * See the License for the specific language governing permissions and
 | 
	
		
			
				|  |  | + * limitations under the License.
 | 
	
		
			
				|  |  | + *
 | 
	
		
			
				|  |  | + */
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +#include <grpc/support/port_platform.h>
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +#include <stdlib.h>
 | 
	
		
			
				|  |  | +#include <string.h>
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +#include <functional>
 | 
	
		
			
				|  |  | +#include <set>
 | 
	
		
			
				|  |  | +#include <thread>
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +#include <gmock/gmock.h>
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +#include <grpc/grpc.h>
 | 
	
		
			
				|  |  | +#include <grpc/impl/codegen/grpc_types.h>
 | 
	
		
			
				|  |  | +#include <grpc/slice.h>
 | 
	
		
			
				|  |  | +#include <grpc/support/alloc.h>
 | 
	
		
			
				|  |  | +#include <grpc/support/log.h>
 | 
	
		
			
				|  |  | +#include <grpc/support/string_util.h>
 | 
	
		
			
				|  |  | +#include <grpc/support/time.h>
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +#include "src/core/ext/filters/client_channel/backup_poller.h"
 | 
	
		
			
				|  |  | +#include "src/core/ext/transport/chttp2/transport/flow_control.h"
 | 
	
		
			
				|  |  | +#include "src/core/lib/channel/channel_args.h"
 | 
	
		
			
				|  |  | +#include "src/core/lib/gprpp/host_port.h"
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +#include "test/core/util/port.h"
 | 
	
		
			
				|  |  | +#include "test/core/util/test_config.h"
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +namespace {
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +class TransportTargetWindowEstimatesMocker
 | 
	
		
			
				|  |  | +    : public grpc_core::chttp2::TestOnlyTransportTargetWindowEstimatesMocker {
 | 
	
		
			
				|  |  | + public:
 | 
	
		
			
				|  |  | +  explicit TransportTargetWindowEstimatesMocker() {}
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +  double ComputeNextTargetInitialWindowSizeFromPeriodicUpdate(
 | 
	
		
			
				|  |  | +      double current_target) override {
 | 
	
		
			
				|  |  | +    const double kTinyWindow = 512;
 | 
	
		
			
				|  |  | +    const double kSmallWindow = 8192;
 | 
	
		
			
				|  |  | +    // The goal is to bounce back and forth between 512 and 8192 initial window
 | 
	
		
			
				|  |  | +    // sizes, in order to get the following to happen at the server (in order):
 | 
	
		
			
				|  |  | +    //
 | 
	
		
			
				|  |  | +    // 1) Stall the server-side RPC's outgoing message on stream window flow
 | 
	
		
			
				|  |  | +    // control.
 | 
	
		
			
				|  |  | +    //
 | 
	
		
			
				|  |  | +    // 2) Send another settings frame with a change in initial window
 | 
	
		
			
				|  |  | +    // size setting, which will make the server-side call go writable.
 | 
	
		
			
				|  |  | +    if (current_target > kTinyWindow) {
 | 
	
		
			
				|  |  | +      return kTinyWindow;
 | 
	
		
			
				|  |  | +    } else {
 | 
	
		
			
				|  |  | +      return kSmallWindow;
 | 
	
		
			
				|  |  | +    }
 | 
	
		
			
				|  |  | +  }
 | 
	
		
			
				|  |  | +};
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +void StartCall(grpc_call* call, grpc_completion_queue* cq) {
 | 
	
		
			
				|  |  | +  grpc_op ops[1];
 | 
	
		
			
				|  |  | +  grpc_op* op;
 | 
	
		
			
				|  |  | +  memset(ops, 0, sizeof(ops));
 | 
	
		
			
				|  |  | +  op = ops;
 | 
	
		
			
				|  |  | +  op->op = GRPC_OP_SEND_INITIAL_METADATA;
 | 
	
		
			
				|  |  | +  op->data.send_initial_metadata.count = 0;
 | 
	
		
			
				|  |  | +  op->flags = GRPC_INITIAL_METADATA_WAIT_FOR_READY;
 | 
	
		
			
				|  |  | +  op->reserved = nullptr;
 | 
	
		
			
				|  |  | +  op++;
 | 
	
		
			
				|  |  | +  void* tag = call;
 | 
	
		
			
				|  |  | +  grpc_call_error error = grpc_call_start_batch(
 | 
	
		
			
				|  |  | +      call, ops, static_cast<size_t>(op - ops), tag, nullptr);
 | 
	
		
			
				|  |  | +  GPR_ASSERT(GRPC_CALL_OK == error);
 | 
	
		
			
				|  |  | +  grpc_event event = grpc_completion_queue_next(
 | 
	
		
			
				|  |  | +      cq, gpr_inf_future(GPR_CLOCK_REALTIME), nullptr);
 | 
	
		
			
				|  |  | +  GPR_ASSERT(event.type == GRPC_OP_COMPLETE);
 | 
	
		
			
				|  |  | +  GPR_ASSERT(event.success);
 | 
	
		
			
				|  |  | +  GPR_ASSERT(event.tag == tag);
 | 
	
		
			
				|  |  | +}
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +void FinishCall(grpc_call* call, grpc_completion_queue* cq) {
 | 
	
		
			
				|  |  | +  grpc_op ops[4];
 | 
	
		
			
				|  |  | +  grpc_op* op;
 | 
	
		
			
				|  |  | +  grpc_metadata_array initial_metadata_recv;
 | 
	
		
			
				|  |  | +  grpc_metadata_array_init(&initial_metadata_recv);
 | 
	
		
			
				|  |  | +  grpc_metadata_array trailing_metadata_recv;
 | 
	
		
			
				|  |  | +  grpc_metadata_array_init(&trailing_metadata_recv);
 | 
	
		
			
				|  |  | +  grpc_status_code status;
 | 
	
		
			
				|  |  | +  grpc_slice details;
 | 
	
		
			
				|  |  | +  grpc_byte_buffer* recv_payload = nullptr;
 | 
	
		
			
				|  |  | +  memset(ops, 0, sizeof(ops));
 | 
	
		
			
				|  |  | +  op = ops;
 | 
	
		
			
				|  |  | +  op->op = GRPC_OP_SEND_CLOSE_FROM_CLIENT;
 | 
	
		
			
				|  |  | +  op->flags = 0;
 | 
	
		
			
				|  |  | +  op->reserved = nullptr;
 | 
	
		
			
				|  |  | +  op++;
 | 
	
		
			
				|  |  | +  op->op = GRPC_OP_RECV_INITIAL_METADATA;
 | 
	
		
			
				|  |  | +  op->data.recv_initial_metadata.recv_initial_metadata = &initial_metadata_recv;
 | 
	
		
			
				|  |  | +  op->flags = 0;
 | 
	
		
			
				|  |  | +  op->reserved = nullptr;
 | 
	
		
			
				|  |  | +  op++;
 | 
	
		
			
				|  |  | +  op->op = GRPC_OP_RECV_MESSAGE;
 | 
	
		
			
				|  |  | +  op->data.recv_message.recv_message = &recv_payload;
 | 
	
		
			
				|  |  | +  op->flags = 0;
 | 
	
		
			
				|  |  | +  op->reserved = nullptr;
 | 
	
		
			
				|  |  | +  op++;
 | 
	
		
			
				|  |  | +  op->op = GRPC_OP_RECV_STATUS_ON_CLIENT;
 | 
	
		
			
				|  |  | +  op->data.recv_status_on_client.trailing_metadata = &trailing_metadata_recv;
 | 
	
		
			
				|  |  | +  op->data.recv_status_on_client.status = &status;
 | 
	
		
			
				|  |  | +  op->data.recv_status_on_client.status_details = &details;
 | 
	
		
			
				|  |  | +  op->flags = 0;
 | 
	
		
			
				|  |  | +  op->reserved = nullptr;
 | 
	
		
			
				|  |  | +  op++;
 | 
	
		
			
				|  |  | +  void* tag = call;
 | 
	
		
			
				|  |  | +  grpc_call_error error = grpc_call_start_batch(
 | 
	
		
			
				|  |  | +      call, ops, static_cast<size_t>(op - ops), tag, nullptr);
 | 
	
		
			
				|  |  | +  GPR_ASSERT(GRPC_CALL_OK == error);
 | 
	
		
			
				|  |  | +  grpc_event event = grpc_completion_queue_next(
 | 
	
		
			
				|  |  | +      cq, gpr_inf_future(GPR_CLOCK_REALTIME), nullptr);
 | 
	
		
			
				|  |  | +  GPR_ASSERT(event.type == GRPC_OP_COMPLETE);
 | 
	
		
			
				|  |  | +  GPR_ASSERT(event.success);
 | 
	
		
			
				|  |  | +  GPR_ASSERT(event.tag == tag);
 | 
	
		
			
				|  |  | +  grpc_metadata_array_destroy(&initial_metadata_recv);
 | 
	
		
			
				|  |  | +  grpc_metadata_array_destroy(&trailing_metadata_recv);
 | 
	
		
			
				|  |  | +  grpc_byte_buffer_destroy(recv_payload);
 | 
	
		
			
				|  |  | +  grpc_slice_unref(details);
 | 
	
		
			
				|  |  | +}
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +class TestServer {
 | 
	
		
			
				|  |  | + public:
 | 
	
		
			
				|  |  | +  explicit TestServer() {
 | 
	
		
			
				|  |  | +    cq_ = grpc_completion_queue_create_for_next(nullptr);
 | 
	
		
			
				|  |  | +    server_ = grpc_server_create(nullptr, nullptr);
 | 
	
		
			
				|  |  | +    address_ = grpc_core::JoinHostPort("[::1]", grpc_pick_unused_port_or_die());
 | 
	
		
			
				|  |  | +    grpc_server_register_completion_queue(server_, cq_, nullptr);
 | 
	
		
			
				|  |  | +    GPR_ASSERT(grpc_server_add_insecure_http2_port(server_, address_.c_str()));
 | 
	
		
			
				|  |  | +    grpc_server_start(server_);
 | 
	
		
			
				|  |  | +    accept_thread_ = std::thread(std::bind(&TestServer::AcceptThread, this));
 | 
	
		
			
				|  |  | +  }
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +  int ShutdownAndGetNumCallsHandled() {
 | 
	
		
			
				|  |  | +    {
 | 
	
		
			
				|  |  | +      // prevent the server from requesting any more calls
 | 
	
		
			
				|  |  | +      grpc_core::MutexLock lock(&shutdown_mu_);
 | 
	
		
			
				|  |  | +      shutdown_ = true;
 | 
	
		
			
				|  |  | +    }
 | 
	
		
			
				|  |  | +    grpc_server_shutdown_and_notify(server_, cq_, this /* tag */);
 | 
	
		
			
				|  |  | +    accept_thread_.join();
 | 
	
		
			
				|  |  | +    grpc_server_destroy(server_);
 | 
	
		
			
				|  |  | +    grpc_completion_queue_shutdown(cq_);
 | 
	
		
			
				|  |  | +    while (grpc_completion_queue_next(cq_, gpr_inf_future(GPR_CLOCK_REALTIME),
 | 
	
		
			
				|  |  | +                                      nullptr)
 | 
	
		
			
				|  |  | +               .type != GRPC_QUEUE_SHUTDOWN)
 | 
	
		
			
				|  |  | +      ;
 | 
	
		
			
				|  |  | +    grpc_completion_queue_destroy(cq_);
 | 
	
		
			
				|  |  | +    return num_calls_handled_;
 | 
	
		
			
				|  |  | +  }
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +  std::string address() const { return address_; }
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | + private:
 | 
	
		
			
				|  |  | +  void AcceptThread() {
 | 
	
		
			
				|  |  | +    std::vector<std::thread> rpc_threads;
 | 
	
		
			
				|  |  | +    bool got_shutdown_and_notify_tag = false;
 | 
	
		
			
				|  |  | +    while (!got_shutdown_and_notify_tag) {
 | 
	
		
			
				|  |  | +      void* request_call_tag = &rpc_threads;
 | 
	
		
			
				|  |  | +      grpc_call_details call_details;
 | 
	
		
			
				|  |  | +      grpc_call_details_init(&call_details);
 | 
	
		
			
				|  |  | +      grpc_call* call = nullptr;
 | 
	
		
			
				|  |  | +      grpc_completion_queue* call_cq = nullptr;
 | 
	
		
			
				|  |  | +      grpc_metadata_array request_metadata_recv;
 | 
	
		
			
				|  |  | +      grpc_metadata_array_init(&request_metadata_recv);
 | 
	
		
			
				|  |  | +      {
 | 
	
		
			
				|  |  | +        grpc_core::MutexLock lock(&shutdown_mu_);
 | 
	
		
			
				|  |  | +        if (!shutdown_) {
 | 
	
		
			
				|  |  | +          call_cq = grpc_completion_queue_create_for_next(nullptr);
 | 
	
		
			
				|  |  | +          grpc_call_error error = grpc_server_request_call(
 | 
	
		
			
				|  |  | +              server_, &call, &call_details, &request_metadata_recv, call_cq,
 | 
	
		
			
				|  |  | +              cq_, request_call_tag);
 | 
	
		
			
				|  |  | +          GPR_ASSERT(error == GRPC_CALL_OK);
 | 
	
		
			
				|  |  | +        }
 | 
	
		
			
				|  |  | +      }
 | 
	
		
			
				|  |  | +      grpc_event event = grpc_completion_queue_next(
 | 
	
		
			
				|  |  | +          cq_, gpr_inf_future(GPR_CLOCK_REALTIME), nullptr);
 | 
	
		
			
				|  |  | +      GPR_ASSERT(event.type == GRPC_OP_COMPLETE);
 | 
	
		
			
				|  |  | +      grpc_call_details_destroy(&call_details);
 | 
	
		
			
				|  |  | +      grpc_metadata_array_destroy(&request_metadata_recv);
 | 
	
		
			
				|  |  | +      if (event.success) {
 | 
	
		
			
				|  |  | +        if (event.tag == request_call_tag) {
 | 
	
		
			
				|  |  | +          // HandleOneRpc takes ownership of its parameters
 | 
	
		
			
				|  |  | +          num_calls_handled_++;
 | 
	
		
			
				|  |  | +          rpc_threads.push_back(
 | 
	
		
			
				|  |  | +              std::thread(std::bind(&TestServer::HandleOneRpc, call, call_cq)));
 | 
	
		
			
				|  |  | +        } else if (event.tag == this /* shutdown_and_notify tag */) {
 | 
	
		
			
				|  |  | +          grpc_core::MutexLock lock(&shutdown_mu_);
 | 
	
		
			
				|  |  | +          GPR_ASSERT(shutdown_);
 | 
	
		
			
				|  |  | +          GPR_ASSERT(call_cq == nullptr);
 | 
	
		
			
				|  |  | +          got_shutdown_and_notify_tag = true;
 | 
	
		
			
				|  |  | +        } else {
 | 
	
		
			
				|  |  | +          GPR_ASSERT(0);
 | 
	
		
			
				|  |  | +        }
 | 
	
		
			
				|  |  | +      } else {
 | 
	
		
			
				|  |  | +        grpc_core::MutexLock lock(&shutdown_mu_);
 | 
	
		
			
				|  |  | +        GPR_ASSERT(shutdown_);
 | 
	
		
			
				|  |  | +        grpc_completion_queue_destroy(call_cq);
 | 
	
		
			
				|  |  | +      }
 | 
	
		
			
				|  |  | +    }
 | 
	
		
			
				|  |  | +    gpr_log(GPR_INFO, "test server shutdown, joining RPC threads...");
 | 
	
		
			
				|  |  | +    for (auto& t : rpc_threads) {
 | 
	
		
			
				|  |  | +      t.join();
 | 
	
		
			
				|  |  | +    }
 | 
	
		
			
				|  |  | +    gpr_log(GPR_INFO, "test server threads all finished!");
 | 
	
		
			
				|  |  | +  }
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +  static void HandleOneRpc(grpc_call* call, grpc_completion_queue* call_cq) {
 | 
	
		
			
				|  |  | +    // Send a large enough payload to get us stalled on outgoing flow control
 | 
	
		
			
				|  |  | +    std::string send_payload = "";
 | 
	
		
			
				|  |  | +    for (int i = 0; i < 4 * 1e6; i++) {
 | 
	
		
			
				|  |  | +      send_payload += "a";
 | 
	
		
			
				|  |  | +    }
 | 
	
		
			
				|  |  | +    grpc_slice request_payload_slice =
 | 
	
		
			
				|  |  | +        grpc_slice_from_copied_string(send_payload.c_str());
 | 
	
		
			
				|  |  | +    grpc_byte_buffer* request_payload =
 | 
	
		
			
				|  |  | +        grpc_raw_byte_buffer_create(&request_payload_slice, 1);
 | 
	
		
			
				|  |  | +    void* tag = call_cq;
 | 
	
		
			
				|  |  | +    grpc_op ops[2];
 | 
	
		
			
				|  |  | +    grpc_op* op;
 | 
	
		
			
				|  |  | +    memset(ops, 0, sizeof(ops));
 | 
	
		
			
				|  |  | +    op = ops;
 | 
	
		
			
				|  |  | +    op->op = GRPC_OP_SEND_INITIAL_METADATA;
 | 
	
		
			
				|  |  | +    op->data.send_initial_metadata.count = 0;
 | 
	
		
			
				|  |  | +    op->reserved = nullptr;
 | 
	
		
			
				|  |  | +    op++;
 | 
	
		
			
				|  |  | +    op->op = GRPC_OP_SEND_MESSAGE;
 | 
	
		
			
				|  |  | +    op->data.send_message.send_message = request_payload;
 | 
	
		
			
				|  |  | +    op->reserved = nullptr;
 | 
	
		
			
				|  |  | +    op++;
 | 
	
		
			
				|  |  | +    grpc_call_error error = grpc_call_start_batch(
 | 
	
		
			
				|  |  | +        call, ops, static_cast<size_t>(op - ops), tag, nullptr);
 | 
	
		
			
				|  |  | +    GPR_ASSERT(GRPC_CALL_OK == error);
 | 
	
		
			
				|  |  | +    std::thread poller([call_cq]() {
 | 
	
		
			
				|  |  | +      // poll the connection so that we actively pick up bytes off the wire,
 | 
	
		
			
				|  |  | +      // including settings frames with window size increases
 | 
	
		
			
				|  |  | +      while (grpc_completion_queue_next(
 | 
	
		
			
				|  |  | +                 call_cq, gpr_inf_future(GPR_CLOCK_REALTIME), nullptr)
 | 
	
		
			
				|  |  | +                 .type != GRPC_QUEUE_SHUTDOWN)
 | 
	
		
			
				|  |  | +        ;
 | 
	
		
			
				|  |  | +    });
 | 
	
		
			
				|  |  | +    grpc_call_cancel(call, nullptr);
 | 
	
		
			
				|  |  | +    grpc_call_unref(call);
 | 
	
		
			
				|  |  | +    grpc_completion_queue_shutdown(call_cq);
 | 
	
		
			
				|  |  | +    poller.join();
 | 
	
		
			
				|  |  | +    grpc_completion_queue_destroy(call_cq);
 | 
	
		
			
				|  |  | +    grpc_byte_buffer_destroy(request_payload);
 | 
	
		
			
				|  |  | +    grpc_slice_unref(request_payload_slice);
 | 
	
		
			
				|  |  | +  }
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +  grpc_server* server_;
 | 
	
		
			
				|  |  | +  grpc_completion_queue* cq_;
 | 
	
		
			
				|  |  | +  std::string address_;
 | 
	
		
			
				|  |  | +  std::thread accept_thread_;
 | 
	
		
			
				|  |  | +  int num_calls_handled_ = 0;
 | 
	
		
			
				|  |  | +  grpc_core::Mutex shutdown_mu_;
 | 
	
		
			
				|  |  | +  bool shutdown_ = false;
 | 
	
		
			
				|  |  | +};
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +// Perform a simple RPC where the server cancels the request with
 | 
	
		
			
				|  |  | +// grpc_call_cancel_with_status
 | 
	
		
			
				|  |  | +TEST(Pollers, TestDontCrashWhenTryingToReproIssueFixedBy23984) {
 | 
	
		
			
				|  |  | +  // 64 threads is arbitrary but chosen because, experimentally it's enough to
 | 
	
		
			
				|  |  | +  // repro the targetted crash crash (which is then fixed by
 | 
	
		
			
				|  |  | +  // https://github.com/grpc/grpc/pull/23984) at a very high rate.
 | 
	
		
			
				|  |  | +  const int kNumCalls = 64;
 | 
	
		
			
				|  |  | +  std::vector<std::thread> threads;
 | 
	
		
			
				|  |  | +  threads.reserve(kNumCalls);
 | 
	
		
			
				|  |  | +  std::unique_ptr<TestServer> test_server = absl::make_unique<TestServer>();
 | 
	
		
			
				|  |  | +  const std::string server_address = test_server->address();
 | 
	
		
			
				|  |  | +  for (int i = 0; i < kNumCalls; i++) {
 | 
	
		
			
				|  |  | +    threads.push_back(std::thread([server_address]() {
 | 
	
		
			
				|  |  | +      std::vector<grpc_arg> args;
 | 
	
		
			
				|  |  | +      // this test is meant to create one connection to the server for each
 | 
	
		
			
				|  |  | +      // of these threads
 | 
	
		
			
				|  |  | +      args.push_back(grpc_channel_arg_integer_create(
 | 
	
		
			
				|  |  | +          const_cast<char*>(GRPC_ARG_USE_LOCAL_SUBCHANNEL_POOL), true));
 | 
	
		
			
				|  |  | +      grpc_channel_args* channel_args =
 | 
	
		
			
				|  |  | +          grpc_channel_args_copy_and_add(nullptr, args.data(), args.size());
 | 
	
		
			
				|  |  | +      grpc_channel* channel = grpc_insecure_channel_create(
 | 
	
		
			
				|  |  | +          std::string("ipv6:" + server_address).c_str(), channel_args, nullptr);
 | 
	
		
			
				|  |  | +      grpc_channel_args_destroy(channel_args);
 | 
	
		
			
				|  |  | +      grpc_completion_queue* cq =
 | 
	
		
			
				|  |  | +          grpc_completion_queue_create_for_next(nullptr);
 | 
	
		
			
				|  |  | +      grpc_call* call = grpc_channel_create_call(
 | 
	
		
			
				|  |  | +          channel, nullptr, GRPC_PROPAGATE_DEFAULTS, cq,
 | 
	
		
			
				|  |  | +          grpc_slice_from_static_string("/foo"), nullptr,
 | 
	
		
			
				|  |  | +          gpr_inf_future(GPR_CLOCK_REALTIME), nullptr);
 | 
	
		
			
				|  |  | +      StartCall(call, cq);
 | 
	
		
			
				|  |  | +      // Explicitly avoid reading on this RPC for a period of time. The
 | 
	
		
			
				|  |  | +      // goal is to get the server side RPC to stall on it's outgoing stream
 | 
	
		
			
				|  |  | +      // flow control window, as the first step in trying to trigger a bug.
 | 
	
		
			
				|  |  | +      gpr_sleep_until(gpr_time_add(gpr_now(GPR_CLOCK_REALTIME),
 | 
	
		
			
				|  |  | +                                   gpr_time_from_seconds(1, GPR_TIMESPAN)));
 | 
	
		
			
				|  |  | +      // Note that this test doesn't really care what the status of the RPC was,
 | 
	
		
			
				|  |  | +      // because we're just trying to make sure that we don't crash.
 | 
	
		
			
				|  |  | +      FinishCall(call, cq);
 | 
	
		
			
				|  |  | +      grpc_call_unref(call);
 | 
	
		
			
				|  |  | +      grpc_channel_destroy(channel);
 | 
	
		
			
				|  |  | +      grpc_completion_queue_shutdown(cq);
 | 
	
		
			
				|  |  | +      while (grpc_completion_queue_next(cq, gpr_inf_future(GPR_CLOCK_REALTIME),
 | 
	
		
			
				|  |  | +                                        nullptr)
 | 
	
		
			
				|  |  | +                 .type != GRPC_QUEUE_SHUTDOWN)
 | 
	
		
			
				|  |  | +        ;
 | 
	
		
			
				|  |  | +      grpc_completion_queue_destroy(cq);
 | 
	
		
			
				|  |  | +    }));
 | 
	
		
			
				|  |  | +  }
 | 
	
		
			
				|  |  | +  for (auto& thread : threads) {
 | 
	
		
			
				|  |  | +    thread.join();
 | 
	
		
			
				|  |  | +  }
 | 
	
		
			
				|  |  | +  gpr_log(GPR_DEBUG, "All RPCs completed!");
 | 
	
		
			
				|  |  | +  int num_calls_seen_at_server = test_server->ShutdownAndGetNumCallsHandled();
 | 
	
		
			
				|  |  | +  if (num_calls_seen_at_server != kNumCalls) {
 | 
	
		
			
				|  |  | +    gpr_log(GPR_ERROR,
 | 
	
		
			
				|  |  | +            "Expected server to handle %d calls, but instead it only handled "
 | 
	
		
			
				|  |  | +            "%d. This suggests some or all RPCs didn't make it to the server, "
 | 
	
		
			
				|  |  | +            "which means "
 | 
	
		
			
				|  |  | +            "that this test likely isn't doing what it's meant to be doing.",
 | 
	
		
			
				|  |  | +            kNumCalls, num_calls_seen_at_server);
 | 
	
		
			
				|  |  | +    GPR_ASSERT(0);
 | 
	
		
			
				|  |  | +  }
 | 
	
		
			
				|  |  | +}
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +}  // namespace
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +int main(int argc, char** argv) {
 | 
	
		
			
				|  |  | +  ::testing::InitGoogleTest(&argc, argv);
 | 
	
		
			
				|  |  | +  // Make sure that we will have an active poller on all client-side fd's that
 | 
	
		
			
				|  |  | +  // are capable of sending settings frames with window updates etc., even in
 | 
	
		
			
				|  |  | +  // the case that we don't have an active RPC operation on the fd.
 | 
	
		
			
				|  |  | +  GPR_GLOBAL_CONFIG_SET(grpc_client_channel_backup_poll_interval_ms, 1);
 | 
	
		
			
				|  |  | +  grpc_core::chttp2::g_test_only_transport_target_window_estimates_mocker =
 | 
	
		
			
				|  |  | +      new TransportTargetWindowEstimatesMocker();
 | 
	
		
			
				|  |  | +  grpc::testing::TestEnvironment env(argc, argv);
 | 
	
		
			
				|  |  | +  grpc_init();
 | 
	
		
			
				|  |  | +  auto result = RUN_ALL_TESTS();
 | 
	
		
			
				|  |  | +  grpc_shutdown();
 | 
	
		
			
				|  |  | +  return result;
 | 
	
		
			
				|  |  | +}
 |