| 12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550 | 
							- /*
 
-  *
 
-  * Copyright 2018 gRPC authors.
 
-  *
 
-  * Licensed under the Apache License, Version 2.0 (the "License");
 
-  * you may not use this file except in compliance with the License.
 
-  * You may obtain a copy of the License at
 
-  *
 
-  *     http://www.apache.org/licenses/LICENSE-2.0
 
-  *
 
-  * Unless required by applicable law or agreed to in writing, software
 
-  * distributed under the License is distributed on an "AS IS" BASIS,
 
-  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 
-  * See the License for the specific language governing permissions and
 
-  * limitations under the License.
 
-  *
 
-  */
 
- #include <grpcpp/channel.h>
 
- #include <grpcpp/client_context.h>
 
- #include <grpcpp/create_channel.h>
 
- #include <grpcpp/generic/generic_stub.h>
 
- #include <grpcpp/impl/codegen/proto_utils.h>
 
- #include <grpcpp/server.h>
 
- #include <grpcpp/server_builder.h>
 
- #include <grpcpp/server_context.h>
 
- #include <grpcpp/support/client_callback.h>
 
- #include <gtest/gtest.h>
 
- #include <algorithm>
 
- #include <condition_variable>
 
- #include <functional>
 
- #include <mutex>
 
- #include <sstream>
 
- #include <thread>
 
- #include "src/core/lib/gpr/env.h"
 
- #include "src/core/lib/iomgr/iomgr.h"
 
- #include "src/proto/grpc/testing/echo.grpc.pb.h"
 
- #include "test/core/util/port.h"
 
- #include "test/core/util/test_config.h"
 
- #include "test/cpp/end2end/interceptors_util.h"
 
- #include "test/cpp/end2end/test_service_impl.h"
 
- #include "test/cpp/util/byte_buffer_proto_helper.h"
 
- #include "test/cpp/util/string_ref_helper.h"
 
- #include "test/cpp/util/test_credentials_provider.h"
 
- // MAYBE_SKIP_TEST is a macro to determine if this particular test configuration
 
- // should be skipped based on a decision made at SetUp time. In particular, any
 
- // callback tests can only be run if the iomgr can run in the background or if
 
- // the transport is in-process.
 
- #define MAYBE_SKIP_TEST \
 
-   do {                  \
 
-     if (do_not_test_) { \
 
-       return;           \
 
-     }                   \
 
-   } while (0)
 
- namespace grpc {
 
- namespace testing {
 
- namespace {
 
- enum class Protocol { INPROC, TCP };
 
- class TestScenario {
 
-  public:
 
-   TestScenario(bool serve_callback, Protocol protocol, bool intercept,
 
-                const grpc::string& creds_type)
 
-       : callback_server(serve_callback),
 
-         protocol(protocol),
 
-         use_interceptors(intercept),
 
-         credentials_type(creds_type) {}
 
-   void Log() const;
 
-   bool callback_server;
 
-   Protocol protocol;
 
-   bool use_interceptors;
 
-   const grpc::string credentials_type;
 
- };
 
- static std::ostream& operator<<(std::ostream& out,
 
-                                 const TestScenario& scenario) {
 
-   return out << "TestScenario{callback_server="
 
-              << (scenario.callback_server ? "true" : "false") << ",protocol="
 
-              << (scenario.protocol == Protocol::INPROC ? "INPROC" : "TCP")
 
-              << ",intercept=" << (scenario.use_interceptors ? "true" : "false")
 
-              << ",creds=" << scenario.credentials_type << "}";
 
- }
 
- void TestScenario::Log() const {
 
-   std::ostringstream out;
 
-   out << *this;
 
-   gpr_log(GPR_DEBUG, "%s", out.str().c_str());
 
- }
 
- class ClientCallbackEnd2endTest
 
-     : public ::testing::TestWithParam<TestScenario> {
 
-  protected:
 
-   ClientCallbackEnd2endTest() { GetParam().Log(); }
 
-   void SetUp() override {
 
-     ServerBuilder builder;
 
-     auto server_creds = GetCredentialsProvider()->GetServerCredentials(
 
-         GetParam().credentials_type);
 
-     // TODO(vjpai): Support testing of AuthMetadataProcessor
 
-     if (GetParam().protocol == Protocol::TCP) {
 
-       picked_port_ = grpc_pick_unused_port_or_die();
 
-       server_address_ << "localhost:" << picked_port_;
 
-       builder.AddListeningPort(server_address_.str(), server_creds);
 
-     }
 
-     if (!GetParam().callback_server) {
 
-       builder.RegisterService(&service_);
 
-     } else {
 
-       builder.RegisterService(&callback_service_);
 
-     }
 
-     if (GetParam().use_interceptors) {
 
-       std::vector<
 
-           std::unique_ptr<experimental::ServerInterceptorFactoryInterface>>
 
-           creators;
 
-       // Add 20 dummy server interceptors
 
-       creators.reserve(20);
 
-       for (auto i = 0; i < 20; i++) {
 
-         creators.push_back(std::unique_ptr<DummyInterceptorFactory>(
 
-             new DummyInterceptorFactory()));
 
-       }
 
-       builder.experimental().SetInterceptorCreators(std::move(creators));
 
-     }
 
-     server_ = builder.BuildAndStart();
 
-     is_server_started_ = true;
 
-     if (GetParam().protocol == Protocol::TCP &&
 
-         !grpc_iomgr_run_in_background()) {
 
-       do_not_test_ = true;
 
-     }
 
-   }
 
-   void ResetStub() {
 
-     ChannelArguments args;
 
-     auto channel_creds = GetCredentialsProvider()->GetChannelCredentials(
 
-         GetParam().credentials_type, &args);
 
-     switch (GetParam().protocol) {
 
-       case Protocol::TCP:
 
-         if (!GetParam().use_interceptors) {
 
-           channel_ = ::grpc::CreateCustomChannel(server_address_.str(),
 
-                                                  channel_creds, args);
 
-         } else {
 
-           channel_ = CreateCustomChannelWithInterceptors(
 
-               server_address_.str(), channel_creds, args,
 
-               CreateDummyClientInterceptors());
 
-         }
 
-         break;
 
-       case Protocol::INPROC:
 
-         if (!GetParam().use_interceptors) {
 
-           channel_ = server_->InProcessChannel(args);
 
-         } else {
 
-           channel_ = server_->experimental().InProcessChannelWithInterceptors(
 
-               args, CreateDummyClientInterceptors());
 
-         }
 
-         break;
 
-       default:
 
-         assert(false);
 
-     }
 
-     stub_ = grpc::testing::EchoTestService::NewStub(channel_);
 
-     generic_stub_.reset(new GenericStub(channel_));
 
-     DummyInterceptor::Reset();
 
-   }
 
-   void TearDown() override {
 
-     if (is_server_started_) {
 
-       // Although we would normally do an explicit shutdown, the server
 
-       // should also work correctly with just a destructor call. The regular
 
-       // end2end test uses explicit shutdown, so let this one just do reset.
 
-       server_.reset();
 
-     }
 
-     if (picked_port_ > 0) {
 
-       grpc_recycle_unused_port(picked_port_);
 
-     }
 
-   }
 
-   void SendRpcs(int num_rpcs, bool with_binary_metadata) {
 
-     grpc::string test_string("");
 
-     for (int i = 0; i < num_rpcs; i++) {
 
-       EchoRequest request;
 
-       EchoResponse response;
 
-       ClientContext cli_ctx;
 
-       test_string += "Hello world. ";
 
-       request.set_message(test_string);
 
-       grpc::string val;
 
-       if (with_binary_metadata) {
 
-         request.mutable_param()->set_echo_metadata(true);
 
-         char bytes[8] = {'\0', '\1', '\2', '\3',
 
-                          '\4', '\5', '\6', static_cast<char>(i)};
 
-         val = grpc::string(bytes, 8);
 
-         cli_ctx.AddMetadata("custom-bin", val);
 
-       }
 
-       cli_ctx.set_compression_algorithm(GRPC_COMPRESS_GZIP);
 
-       std::mutex mu;
 
-       std::condition_variable cv;
 
-       bool done = false;
 
-       stub_->experimental_async()->Echo(
 
-           &cli_ctx, &request, &response,
 
-           [&cli_ctx, &request, &response, &done, &mu, &cv, val,
 
-            with_binary_metadata](Status s) {
 
-             GPR_ASSERT(s.ok());
 
-             EXPECT_EQ(request.message(), response.message());
 
-             if (with_binary_metadata) {
 
-               EXPECT_EQ(
 
-                   1u, cli_ctx.GetServerTrailingMetadata().count("custom-bin"));
 
-               EXPECT_EQ(val, ToString(cli_ctx.GetServerTrailingMetadata()
 
-                                           .find("custom-bin")
 
-                                           ->second));
 
-             }
 
-             std::lock_guard<std::mutex> l(mu);
 
-             done = true;
 
-             cv.notify_one();
 
-           });
 
-       std::unique_lock<std::mutex> l(mu);
 
-       while (!done) {
 
-         cv.wait(l);
 
-       }
 
-     }
 
-   }
 
-   void SendRpcsRawReq(int num_rpcs) {
 
-     grpc::string test_string("Hello raw world.");
 
-     EchoRequest request;
 
-     request.set_message(test_string);
 
-     std::unique_ptr<ByteBuffer> send_buf = SerializeToByteBuffer(&request);
 
-     for (int i = 0; i < num_rpcs; i++) {
 
-       EchoResponse response;
 
-       ClientContext cli_ctx;
 
-       std::mutex mu;
 
-       std::condition_variable cv;
 
-       bool done = false;
 
-       stub_->experimental_async()->Echo(
 
-           &cli_ctx, send_buf.get(), &response,
 
-           [&request, &response, &done, &mu, &cv](Status s) {
 
-             GPR_ASSERT(s.ok());
 
-             EXPECT_EQ(request.message(), response.message());
 
-             std::lock_guard<std::mutex> l(mu);
 
-             done = true;
 
-             cv.notify_one();
 
-           });
 
-       std::unique_lock<std::mutex> l(mu);
 
-       while (!done) {
 
-         cv.wait(l);
 
-       }
 
-     }
 
-   }
 
-   void SendRpcsGeneric(int num_rpcs, bool maybe_except) {
 
-     const grpc::string kMethodName("/grpc.testing.EchoTestService/Echo");
 
-     grpc::string test_string("");
 
-     for (int i = 0; i < num_rpcs; i++) {
 
-       EchoRequest request;
 
-       std::unique_ptr<ByteBuffer> send_buf;
 
-       ByteBuffer recv_buf;
 
-       ClientContext cli_ctx;
 
-       test_string += "Hello world. ";
 
-       request.set_message(test_string);
 
-       send_buf = SerializeToByteBuffer(&request);
 
-       std::mutex mu;
 
-       std::condition_variable cv;
 
-       bool done = false;
 
-       generic_stub_->experimental().UnaryCall(
 
-           &cli_ctx, kMethodName, send_buf.get(), &recv_buf,
 
-           [&request, &recv_buf, &done, &mu, &cv, maybe_except](Status s) {
 
-             GPR_ASSERT(s.ok());
 
-             EchoResponse response;
 
-             EXPECT_TRUE(ParseFromByteBuffer(&recv_buf, &response));
 
-             EXPECT_EQ(request.message(), response.message());
 
-             std::lock_guard<std::mutex> l(mu);
 
-             done = true;
 
-             cv.notify_one();
 
- #if GRPC_ALLOW_EXCEPTIONS
 
-             if (maybe_except) {
 
-               throw - 1;
 
-             }
 
- #else
 
-             GPR_ASSERT(!maybe_except);
 
- #endif
 
-           });
 
-       std::unique_lock<std::mutex> l(mu);
 
-       while (!done) {
 
-         cv.wait(l);
 
-       }
 
-     }
 
-   }
 
-   void SendGenericEchoAsBidi(int num_rpcs, int reuses, bool do_writes_done) {
 
-     const grpc::string kMethodName("/grpc.testing.EchoTestService/Echo");
 
-     grpc::string test_string("");
 
-     for (int i = 0; i < num_rpcs; i++) {
 
-       test_string += "Hello world. ";
 
-       class Client : public grpc::experimental::ClientBidiReactor<ByteBuffer,
 
-                                                                   ByteBuffer> {
 
-        public:
 
-         Client(ClientCallbackEnd2endTest* test, const grpc::string& method_name,
 
-                const grpc::string& test_str, int reuses, bool do_writes_done)
 
-             : reuses_remaining_(reuses), do_writes_done_(do_writes_done) {
 
-           activate_ = [this, test, method_name, test_str] {
 
-             if (reuses_remaining_ > 0) {
 
-               cli_ctx_.reset(new ClientContext);
 
-               reuses_remaining_--;
 
-               test->generic_stub_->experimental().PrepareBidiStreamingCall(
 
-                   cli_ctx_.get(), method_name, this);
 
-               request_.set_message(test_str);
 
-               send_buf_ = SerializeToByteBuffer(&request_);
 
-               StartWrite(send_buf_.get());
 
-               StartRead(&recv_buf_);
 
-               StartCall();
 
-             } else {
 
-               std::unique_lock<std::mutex> l(mu_);
 
-               done_ = true;
 
-               cv_.notify_one();
 
-             }
 
-           };
 
-           activate_();
 
-         }
 
-         void OnWriteDone(bool /*ok*/) override {
 
-           if (do_writes_done_) {
 
-             StartWritesDone();
 
-           }
 
-         }
 
-         void OnReadDone(bool /*ok*/) override {
 
-           EchoResponse response;
 
-           EXPECT_TRUE(ParseFromByteBuffer(&recv_buf_, &response));
 
-           EXPECT_EQ(request_.message(), response.message());
 
-         };
 
-         void OnDone(const Status& s) override {
 
-           EXPECT_TRUE(s.ok());
 
-           activate_();
 
-         }
 
-         void Await() {
 
-           std::unique_lock<std::mutex> l(mu_);
 
-           while (!done_) {
 
-             cv_.wait(l);
 
-           }
 
-         }
 
-         EchoRequest request_;
 
-         std::unique_ptr<ByteBuffer> send_buf_;
 
-         ByteBuffer recv_buf_;
 
-         std::unique_ptr<ClientContext> cli_ctx_;
 
-         int reuses_remaining_;
 
-         std::function<void()> activate_;
 
-         std::mutex mu_;
 
-         std::condition_variable cv_;
 
-         bool done_ = false;
 
-         const bool do_writes_done_;
 
-       };
 
-       Client rpc(this, kMethodName, test_string, reuses, do_writes_done);
 
-       rpc.Await();
 
-     }
 
-   }
 
-   bool do_not_test_{false};
 
-   bool is_server_started_{false};
 
-   int picked_port_{0};
 
-   std::shared_ptr<Channel> channel_;
 
-   std::unique_ptr<grpc::testing::EchoTestService::Stub> stub_;
 
-   std::unique_ptr<grpc::GenericStub> generic_stub_;
 
-   TestServiceImpl service_;
 
-   CallbackTestServiceImpl callback_service_;
 
-   std::unique_ptr<Server> server_;
 
-   std::ostringstream server_address_;
 
- };
 
- TEST_P(ClientCallbackEnd2endTest, SimpleRpc) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   SendRpcs(1, false);
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, SimpleRpcUnderLockNested) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   std::mutex mu1, mu2, mu3;
 
-   std::condition_variable cv;
 
-   bool done = false;
 
-   EchoRequest request1, request2, request3;
 
-   request1.set_message("Hello locked world1.");
 
-   request2.set_message("Hello locked world2.");
 
-   request3.set_message("Hello locked world3.");
 
-   EchoResponse response1, response2, response3;
 
-   ClientContext cli_ctx1, cli_ctx2, cli_ctx3;
 
-   {
 
-     std::lock_guard<std::mutex> l(mu1);
 
-     stub_->experimental_async()->Echo(
 
-         &cli_ctx1, &request1, &response1,
 
-         [this, &mu1, &mu2, &mu3, &cv, &done, &request1, &request2, &request3,
 
-          &response1, &response2, &response3, &cli_ctx2, &cli_ctx3](Status s1) {
 
-           std::lock_guard<std::mutex> l1(mu1);
 
-           EXPECT_TRUE(s1.ok());
 
-           EXPECT_EQ(request1.message(), response1.message());
 
-           // start the second level of nesting
 
-           std::unique_lock<std::mutex> l2(mu2);
 
-           this->stub_->experimental_async()->Echo(
 
-               &cli_ctx2, &request2, &response2,
 
-               [this, &mu2, &mu3, &cv, &done, &request2, &request3, &response2,
 
-                &response3, &cli_ctx3](Status s2) {
 
-                 std::lock_guard<std::mutex> l2(mu2);
 
-                 EXPECT_TRUE(s2.ok());
 
-                 EXPECT_EQ(request2.message(), response2.message());
 
-                 // start the third level of nesting
 
-                 std::lock_guard<std::mutex> l3(mu3);
 
-                 stub_->experimental_async()->Echo(
 
-                     &cli_ctx3, &request3, &response3,
 
-                     [&mu3, &cv, &done, &request3, &response3](Status s3) {
 
-                       std::lock_guard<std::mutex> l(mu3);
 
-                       EXPECT_TRUE(s3.ok());
 
-                       EXPECT_EQ(request3.message(), response3.message());
 
-                       done = true;
 
-                       cv.notify_all();
 
-                     });
 
-               });
 
-         });
 
-   }
 
-   std::unique_lock<std::mutex> l(mu3);
 
-   while (!done) {
 
-     cv.wait(l);
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, SimpleRpcUnderLock) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   std::mutex mu;
 
-   std::condition_variable cv;
 
-   bool done = false;
 
-   EchoRequest request;
 
-   request.set_message("Hello locked world.");
 
-   EchoResponse response;
 
-   ClientContext cli_ctx;
 
-   {
 
-     std::lock_guard<std::mutex> l(mu);
 
-     stub_->experimental_async()->Echo(
 
-         &cli_ctx, &request, &response,
 
-         [&mu, &cv, &done, &request, &response](Status s) {
 
-           std::lock_guard<std::mutex> l(mu);
 
-           EXPECT_TRUE(s.ok());
 
-           EXPECT_EQ(request.message(), response.message());
 
-           done = true;
 
-           cv.notify_one();
 
-         });
 
-   }
 
-   std::unique_lock<std::mutex> l(mu);
 
-   while (!done) {
 
-     cv.wait(l);
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, SequentialRpcs) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   SendRpcs(10, false);
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, SequentialRpcsRawReq) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   SendRpcsRawReq(10);
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, SendClientInitialMetadata) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   SimpleRequest request;
 
-   SimpleResponse response;
 
-   ClientContext cli_ctx;
 
-   cli_ctx.AddMetadata(kCheckClientInitialMetadataKey,
 
-                       kCheckClientInitialMetadataVal);
 
-   std::mutex mu;
 
-   std::condition_variable cv;
 
-   bool done = false;
 
-   stub_->experimental_async()->CheckClientInitialMetadata(
 
-       &cli_ctx, &request, &response, [&done, &mu, &cv](Status s) {
 
-         GPR_ASSERT(s.ok());
 
-         std::lock_guard<std::mutex> l(mu);
 
-         done = true;
 
-         cv.notify_one();
 
-       });
 
-   std::unique_lock<std::mutex> l(mu);
 
-   while (!done) {
 
-     cv.wait(l);
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, SimpleRpcWithBinaryMetadata) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   SendRpcs(1, true);
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, SequentialRpcsWithVariedBinaryMetadataValue) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   SendRpcs(10, true);
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcs) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   SendRpcsGeneric(10, false);
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcsAsBidi) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   SendGenericEchoAsBidi(10, 1, /*do_writes_done=*/true);
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcsAsBidiWithReactorReuse) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   SendGenericEchoAsBidi(10, 10, /*do_writes_done=*/true);
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, GenericRpcNoWritesDone) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   SendGenericEchoAsBidi(1, 1, /*do_writes_done=*/false);
 
- }
 
- #if GRPC_ALLOW_EXCEPTIONS
 
- TEST_P(ClientCallbackEnd2endTest, ExceptingRpc) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   SendRpcsGeneric(10, true);
 
- }
 
- #endif
 
- TEST_P(ClientCallbackEnd2endTest, MultipleRpcsWithVariedBinaryMetadataValue) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   std::vector<std::thread> threads;
 
-   threads.reserve(10);
 
-   for (int i = 0; i < 10; ++i) {
 
-     threads.emplace_back([this] { SendRpcs(10, true); });
 
-   }
 
-   for (int i = 0; i < 10; ++i) {
 
-     threads[i].join();
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, MultipleRpcs) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   std::vector<std::thread> threads;
 
-   threads.reserve(10);
 
-   for (int i = 0; i < 10; ++i) {
 
-     threads.emplace_back([this] { SendRpcs(10, false); });
 
-   }
 
-   for (int i = 0; i < 10; ++i) {
 
-     threads[i].join();
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, CancelRpcBeforeStart) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   EchoRequest request;
 
-   EchoResponse response;
 
-   ClientContext context;
 
-   request.set_message("hello");
 
-   context.TryCancel();
 
-   std::mutex mu;
 
-   std::condition_variable cv;
 
-   bool done = false;
 
-   stub_->experimental_async()->Echo(
 
-       &context, &request, &response, [&response, &done, &mu, &cv](Status s) {
 
-         EXPECT_EQ("", response.message());
 
-         EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
 
-         std::lock_guard<std::mutex> l(mu);
 
-         done = true;
 
-         cv.notify_one();
 
-       });
 
-   std::unique_lock<std::mutex> l(mu);
 
-   while (!done) {
 
-     cv.wait(l);
 
-   }
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, RequestEchoServerCancel) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   EchoRequest request;
 
-   EchoResponse response;
 
-   ClientContext context;
 
-   request.set_message("hello");
 
-   context.AddMetadata(kServerTryCancelRequest,
 
-                       grpc::to_string(CANCEL_BEFORE_PROCESSING));
 
-   std::mutex mu;
 
-   std::condition_variable cv;
 
-   bool done = false;
 
-   stub_->experimental_async()->Echo(
 
-       &context, &request, &response, [&done, &mu, &cv](Status s) {
 
-         EXPECT_FALSE(s.ok());
 
-         EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
 
-         std::lock_guard<std::mutex> l(mu);
 
-         done = true;
 
-         cv.notify_one();
 
-       });
 
-   std::unique_lock<std::mutex> l(mu);
 
-   while (!done) {
 
-     cv.wait(l);
 
-   }
 
- }
 
- struct ClientCancelInfo {
 
-   bool cancel{false};
 
-   int ops_before_cancel;
 
-   ClientCancelInfo() : cancel{false} {}
 
-   explicit ClientCancelInfo(int ops) : cancel{true}, ops_before_cancel{ops} {}
 
- };
 
- class WriteClient : public grpc::experimental::ClientWriteReactor<EchoRequest> {
 
-  public:
 
-   WriteClient(grpc::testing::EchoTestService::Stub* stub,
 
-               ServerTryCancelRequestPhase server_try_cancel,
 
-               int num_msgs_to_send, ClientCancelInfo client_cancel = {})
 
-       : server_try_cancel_(server_try_cancel),
 
-         num_msgs_to_send_(num_msgs_to_send),
 
-         client_cancel_{client_cancel} {
 
-     grpc::string msg{"Hello server."};
 
-     for (int i = 0; i < num_msgs_to_send; i++) {
 
-       desired_ += msg;
 
-     }
 
-     if (server_try_cancel != DO_NOT_CANCEL) {
 
-       // Send server_try_cancel value in the client metadata
 
-       context_.AddMetadata(kServerTryCancelRequest,
 
-                            grpc::to_string(server_try_cancel));
 
-     }
 
-     context_.set_initial_metadata_corked(true);
 
-     stub->experimental_async()->RequestStream(&context_, &response_, this);
 
-     StartCall();
 
-     request_.set_message(msg);
 
-     MaybeWrite();
 
-   }
 
-   void OnWriteDone(bool ok) override {
 
-     if (ok) {
 
-       num_msgs_sent_++;
 
-       MaybeWrite();
 
-     }
 
-   }
 
-   void OnDone(const Status& s) override {
 
-     gpr_log(GPR_INFO, "Sent %d messages", num_msgs_sent_);
 
-     int num_to_send =
 
-         (client_cancel_.cancel)
 
-             ? std::min(num_msgs_to_send_, client_cancel_.ops_before_cancel)
 
-             : num_msgs_to_send_;
 
-     switch (server_try_cancel_) {
 
-       case CANCEL_BEFORE_PROCESSING:
 
-       case CANCEL_DURING_PROCESSING:
 
-         // If the RPC is canceled by server before / during messages from the
 
-         // client, it means that the client most likely did not get a chance to
 
-         // send all the messages it wanted to send. i.e num_msgs_sent <=
 
-         // num_msgs_to_send
 
-         EXPECT_LE(num_msgs_sent_, num_to_send);
 
-         break;
 
-       case DO_NOT_CANCEL:
 
-       case CANCEL_AFTER_PROCESSING:
 
-         // If the RPC was not canceled or canceled after all messages were read
 
-         // by the server, the client did get a chance to send all its messages
 
-         EXPECT_EQ(num_msgs_sent_, num_to_send);
 
-         break;
 
-       default:
 
-         assert(false);
 
-         break;
 
-     }
 
-     if ((server_try_cancel_ == DO_NOT_CANCEL) && !client_cancel_.cancel) {
 
-       EXPECT_TRUE(s.ok());
 
-       EXPECT_EQ(response_.message(), desired_);
 
-     } else {
 
-       EXPECT_FALSE(s.ok());
 
-       EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
 
-     }
 
-     std::unique_lock<std::mutex> l(mu_);
 
-     done_ = true;
 
-     cv_.notify_one();
 
-   }
 
-   void Await() {
 
-     std::unique_lock<std::mutex> l(mu_);
 
-     while (!done_) {
 
-       cv_.wait(l);
 
-     }
 
-   }
 
-  private:
 
-   void MaybeWrite() {
 
-     if (client_cancel_.cancel &&
 
-         num_msgs_sent_ == client_cancel_.ops_before_cancel) {
 
-       context_.TryCancel();
 
-     } else if (num_msgs_to_send_ > num_msgs_sent_ + 1) {
 
-       StartWrite(&request_);
 
-     } else if (num_msgs_to_send_ == num_msgs_sent_ + 1) {
 
-       StartWriteLast(&request_, WriteOptions());
 
-     }
 
-   }
 
-   EchoRequest request_;
 
-   EchoResponse response_;
 
-   ClientContext context_;
 
-   const ServerTryCancelRequestPhase server_try_cancel_;
 
-   int num_msgs_sent_{0};
 
-   const int num_msgs_to_send_;
 
-   grpc::string desired_;
 
-   const ClientCancelInfo client_cancel_;
 
-   std::mutex mu_;
 
-   std::condition_variable cv_;
 
-   bool done_ = false;
 
- };
 
- TEST_P(ClientCallbackEnd2endTest, RequestStream) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   WriteClient test{stub_.get(), DO_NOT_CANCEL, 3};
 
-   test.Await();
 
-   // Make sure that the server interceptors were not notified to cancel
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, ClientCancelsRequestStream) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   WriteClient test{stub_.get(), DO_NOT_CANCEL, 3, ClientCancelInfo{2}};
 
-   test.Await();
 
-   // Make sure that the server interceptors got the cancel
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- // Server to cancel before doing reading the request
 
- TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelBeforeReads) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   WriteClient test{stub_.get(), CANCEL_BEFORE_PROCESSING, 1};
 
-   test.Await();
 
-   // Make sure that the server interceptors were notified
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- // Server to cancel while reading a request from the stream in parallel
 
- TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelDuringRead) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   WriteClient test{stub_.get(), CANCEL_DURING_PROCESSING, 10};
 
-   test.Await();
 
-   // Make sure that the server interceptors were notified
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- // Server to cancel after reading all the requests but before returning to the
 
- // client
 
- TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelAfterReads) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   WriteClient test{stub_.get(), CANCEL_AFTER_PROCESSING, 4};
 
-   test.Await();
 
-   // Make sure that the server interceptors were notified
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, UnaryReactor) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   class UnaryClient : public grpc::experimental::ClientUnaryReactor {
 
-    public:
 
-     UnaryClient(grpc::testing::EchoTestService::Stub* stub) {
 
-       cli_ctx_.AddMetadata("key1", "val1");
 
-       cli_ctx_.AddMetadata("key2", "val2");
 
-       request_.mutable_param()->set_echo_metadata_initially(true);
 
-       request_.set_message("Hello metadata");
 
-       stub->experimental_async()->Echo(&cli_ctx_, &request_, &response_, this);
 
-       StartCall();
 
-     }
 
-     void OnReadInitialMetadataDone(bool ok) override {
 
-       EXPECT_TRUE(ok);
 
-       EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key1"));
 
-       EXPECT_EQ(
 
-           "val1",
 
-           ToString(cli_ctx_.GetServerInitialMetadata().find("key1")->second));
 
-       EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key2"));
 
-       EXPECT_EQ(
 
-           "val2",
 
-           ToString(cli_ctx_.GetServerInitialMetadata().find("key2")->second));
 
-       initial_metadata_done_ = true;
 
-     }
 
-     void OnDone(const Status& s) override {
 
-       EXPECT_TRUE(initial_metadata_done_);
 
-       EXPECT_EQ(0u, cli_ctx_.GetServerTrailingMetadata().size());
 
-       EXPECT_TRUE(s.ok());
 
-       EXPECT_EQ(request_.message(), response_.message());
 
-       std::unique_lock<std::mutex> l(mu_);
 
-       done_ = true;
 
-       cv_.notify_one();
 
-     }
 
-     void Await() {
 
-       std::unique_lock<std::mutex> l(mu_);
 
-       while (!done_) {
 
-         cv_.wait(l);
 
-       }
 
-     }
 
-    private:
 
-     EchoRequest request_;
 
-     EchoResponse response_;
 
-     ClientContext cli_ctx_;
 
-     std::mutex mu_;
 
-     std::condition_variable cv_;
 
-     bool done_{false};
 
-     bool initial_metadata_done_{false};
 
-   };
 
-   UnaryClient test{stub_.get()};
 
-   test.Await();
 
-   // Make sure that the server interceptors were not notified of a cancel
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, GenericUnaryReactor) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   const grpc::string kMethodName("/grpc.testing.EchoTestService/Echo");
 
-   class UnaryClient : public grpc::experimental::ClientUnaryReactor {
 
-    public:
 
-     UnaryClient(grpc::GenericStub* stub, const grpc::string& method_name) {
 
-       cli_ctx_.AddMetadata("key1", "val1");
 
-       cli_ctx_.AddMetadata("key2", "val2");
 
-       request_.mutable_param()->set_echo_metadata_initially(true);
 
-       request_.set_message("Hello metadata");
 
-       send_buf_ = SerializeToByteBuffer(&request_);
 
-       stub->experimental().PrepareUnaryCall(&cli_ctx_, method_name,
 
-                                             send_buf_.get(), &recv_buf_, this);
 
-       StartCall();
 
-     }
 
-     void OnReadInitialMetadataDone(bool ok) override {
 
-       EXPECT_TRUE(ok);
 
-       EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key1"));
 
-       EXPECT_EQ(
 
-           "val1",
 
-           ToString(cli_ctx_.GetServerInitialMetadata().find("key1")->second));
 
-       EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key2"));
 
-       EXPECT_EQ(
 
-           "val2",
 
-           ToString(cli_ctx_.GetServerInitialMetadata().find("key2")->second));
 
-       initial_metadata_done_ = true;
 
-     }
 
-     void OnDone(const Status& s) override {
 
-       EXPECT_TRUE(initial_metadata_done_);
 
-       EXPECT_EQ(0u, cli_ctx_.GetServerTrailingMetadata().size());
 
-       EXPECT_TRUE(s.ok());
 
-       EchoResponse response;
 
-       EXPECT_TRUE(ParseFromByteBuffer(&recv_buf_, &response));
 
-       EXPECT_EQ(request_.message(), response.message());
 
-       std::unique_lock<std::mutex> l(mu_);
 
-       done_ = true;
 
-       cv_.notify_one();
 
-     }
 
-     void Await() {
 
-       std::unique_lock<std::mutex> l(mu_);
 
-       while (!done_) {
 
-         cv_.wait(l);
 
-       }
 
-     }
 
-    private:
 
-     EchoRequest request_;
 
-     std::unique_ptr<ByteBuffer> send_buf_;
 
-     ByteBuffer recv_buf_;
 
-     ClientContext cli_ctx_;
 
-     std::mutex mu_;
 
-     std::condition_variable cv_;
 
-     bool done_{false};
 
-     bool initial_metadata_done_{false};
 
-   };
 
-   UnaryClient test{generic_stub_.get(), kMethodName};
 
-   test.Await();
 
-   // Make sure that the server interceptors were not notified of a cancel
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- class ReadClient : public grpc::experimental::ClientReadReactor<EchoResponse> {
 
-  public:
 
-   ReadClient(grpc::testing::EchoTestService::Stub* stub,
 
-              ServerTryCancelRequestPhase server_try_cancel,
 
-              ClientCancelInfo client_cancel = {})
 
-       : server_try_cancel_(server_try_cancel), client_cancel_{client_cancel} {
 
-     if (server_try_cancel_ != DO_NOT_CANCEL) {
 
-       // Send server_try_cancel value in the client metadata
 
-       context_.AddMetadata(kServerTryCancelRequest,
 
-                            grpc::to_string(server_try_cancel));
 
-     }
 
-     request_.set_message("Hello client ");
 
-     stub->experimental_async()->ResponseStream(&context_, &request_, this);
 
-     if (client_cancel_.cancel &&
 
-         reads_complete_ == client_cancel_.ops_before_cancel) {
 
-       context_.TryCancel();
 
-     }
 
-     // Even if we cancel, read until failure because there might be responses
 
-     // pending
 
-     StartRead(&response_);
 
-     StartCall();
 
-   }
 
-   void OnReadDone(bool ok) override {
 
-     if (!ok) {
 
-       if (server_try_cancel_ == DO_NOT_CANCEL && !client_cancel_.cancel) {
 
-         EXPECT_EQ(reads_complete_, kServerDefaultResponseStreamsToSend);
 
-       }
 
-     } else {
 
-       EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend);
 
-       EXPECT_EQ(response_.message(),
 
-                 request_.message() + grpc::to_string(reads_complete_));
 
-       reads_complete_++;
 
-       if (client_cancel_.cancel &&
 
-           reads_complete_ == client_cancel_.ops_before_cancel) {
 
-         context_.TryCancel();
 
-       }
 
-       // Even if we cancel, read until failure because there might be responses
 
-       // pending
 
-       StartRead(&response_);
 
-     }
 
-   }
 
-   void OnDone(const Status& s) override {
 
-     gpr_log(GPR_INFO, "Read %d messages", reads_complete_);
 
-     switch (server_try_cancel_) {
 
-       case DO_NOT_CANCEL:
 
-         if (!client_cancel_.cancel || client_cancel_.ops_before_cancel >
 
-                                           kServerDefaultResponseStreamsToSend) {
 
-           EXPECT_TRUE(s.ok());
 
-           EXPECT_EQ(reads_complete_, kServerDefaultResponseStreamsToSend);
 
-         } else {
 
-           EXPECT_GE(reads_complete_, client_cancel_.ops_before_cancel);
 
-           EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend);
 
-           // Status might be ok or cancelled depending on whether server
 
-           // sent status before client cancel went through
 
-           if (!s.ok()) {
 
-             EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
 
-           }
 
-         }
 
-         break;
 
-       case CANCEL_BEFORE_PROCESSING:
 
-         EXPECT_FALSE(s.ok());
 
-         EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
 
-         EXPECT_EQ(reads_complete_, 0);
 
-         break;
 
-       case CANCEL_DURING_PROCESSING:
 
-       case CANCEL_AFTER_PROCESSING:
 
-         // If server canceled while writing messages, client must have read
 
-         // less than or equal to the expected number of messages. Even if the
 
-         // server canceled after writing all messages, the RPC may be canceled
 
-         // before the Client got a chance to read all the messages.
 
-         EXPECT_FALSE(s.ok());
 
-         EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
 
-         EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend);
 
-         break;
 
-       default:
 
-         assert(false);
 
-     }
 
-     std::unique_lock<std::mutex> l(mu_);
 
-     done_ = true;
 
-     cv_.notify_one();
 
-   }
 
-   void Await() {
 
-     std::unique_lock<std::mutex> l(mu_);
 
-     while (!done_) {
 
-       cv_.wait(l);
 
-     }
 
-   }
 
-  private:
 
-   EchoRequest request_;
 
-   EchoResponse response_;
 
-   ClientContext context_;
 
-   const ServerTryCancelRequestPhase server_try_cancel_;
 
-   int reads_complete_{0};
 
-   const ClientCancelInfo client_cancel_;
 
-   std::mutex mu_;
 
-   std::condition_variable cv_;
 
-   bool done_ = false;
 
- };
 
- TEST_P(ClientCallbackEnd2endTest, ResponseStream) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   ReadClient test{stub_.get(), DO_NOT_CANCEL};
 
-   test.Await();
 
-   // Make sure that the server interceptors were not notified of a cancel
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, ClientCancelsResponseStream) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   ReadClient test{stub_.get(), DO_NOT_CANCEL, ClientCancelInfo{2}};
 
-   test.Await();
 
-   // Because cancel in this case races with server finish, we can't be sure that
 
-   // server interceptors even see cancellation
 
- }
 
- // Server to cancel before sending any response messages
 
- TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelBefore) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   ReadClient test{stub_.get(), CANCEL_BEFORE_PROCESSING};
 
-   test.Await();
 
-   // Make sure that the server interceptors were notified
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- // Server to cancel while writing a response to the stream in parallel
 
- TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelDuring) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   ReadClient test{stub_.get(), CANCEL_DURING_PROCESSING};
 
-   test.Await();
 
-   // Make sure that the server interceptors were notified
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- // Server to cancel after writing all the respones to the stream but before
 
- // returning to the client
 
- TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelAfter) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   ReadClient test{stub_.get(), CANCEL_AFTER_PROCESSING};
 
-   test.Await();
 
-   // Make sure that the server interceptors were notified
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- class BidiClient
 
-     : public grpc::experimental::ClientBidiReactor<EchoRequest, EchoResponse> {
 
-  public:
 
-   BidiClient(grpc::testing::EchoTestService::Stub* stub,
 
-              ServerTryCancelRequestPhase server_try_cancel,
 
-              int num_msgs_to_send, bool cork_metadata, bool first_write_async,
 
-              ClientCancelInfo client_cancel = {})
 
-       : server_try_cancel_(server_try_cancel),
 
-         msgs_to_send_{num_msgs_to_send},
 
-         client_cancel_{client_cancel} {
 
-     if (server_try_cancel_ != DO_NOT_CANCEL) {
 
-       // Send server_try_cancel value in the client metadata
 
-       context_.AddMetadata(kServerTryCancelRequest,
 
-                            grpc::to_string(server_try_cancel));
 
-     }
 
-     request_.set_message("Hello fren ");
 
-     context_.set_initial_metadata_corked(cork_metadata);
 
-     stub->experimental_async()->BidiStream(&context_, this);
 
-     MaybeAsyncWrite(first_write_async);
 
-     StartRead(&response_);
 
-     StartCall();
 
-   }
 
-   void OnReadDone(bool ok) override {
 
-     if (!ok) {
 
-       if (server_try_cancel_ == DO_NOT_CANCEL) {
 
-         if (!client_cancel_.cancel) {
 
-           EXPECT_EQ(reads_complete_, msgs_to_send_);
 
-         } else {
 
-           EXPECT_LE(reads_complete_, writes_complete_);
 
-         }
 
-       }
 
-     } else {
 
-       EXPECT_LE(reads_complete_, msgs_to_send_);
 
-       EXPECT_EQ(response_.message(), request_.message());
 
-       reads_complete_++;
 
-       StartRead(&response_);
 
-     }
 
-   }
 
-   void OnWriteDone(bool ok) override {
 
-     if (async_write_thread_.joinable()) {
 
-       async_write_thread_.join();
 
-       RemoveHold();
 
-     }
 
-     if (server_try_cancel_ == DO_NOT_CANCEL) {
 
-       EXPECT_TRUE(ok);
 
-     } else if (!ok) {
 
-       return;
 
-     }
 
-     writes_complete_++;
 
-     MaybeWrite();
 
-   }
 
-   void OnDone(const Status& s) override {
 
-     gpr_log(GPR_INFO, "Sent %d messages", writes_complete_);
 
-     gpr_log(GPR_INFO, "Read %d messages", reads_complete_);
 
-     switch (server_try_cancel_) {
 
-       case DO_NOT_CANCEL:
 
-         if (!client_cancel_.cancel ||
 
-             client_cancel_.ops_before_cancel > msgs_to_send_) {
 
-           EXPECT_TRUE(s.ok());
 
-           EXPECT_EQ(writes_complete_, msgs_to_send_);
 
-           EXPECT_EQ(reads_complete_, writes_complete_);
 
-         } else {
 
-           EXPECT_FALSE(s.ok());
 
-           EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
 
-           EXPECT_EQ(writes_complete_, client_cancel_.ops_before_cancel);
 
-           EXPECT_LE(reads_complete_, writes_complete_);
 
-         }
 
-         break;
 
-       case CANCEL_BEFORE_PROCESSING:
 
-         EXPECT_FALSE(s.ok());
 
-         EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
 
-         // The RPC is canceled before the server did any work or returned any
 
-         // reads, but it's possible that some writes took place first from the
 
-         // client
 
-         EXPECT_LE(writes_complete_, msgs_to_send_);
 
-         EXPECT_EQ(reads_complete_, 0);
 
-         break;
 
-       case CANCEL_DURING_PROCESSING:
 
-         EXPECT_FALSE(s.ok());
 
-         EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
 
-         EXPECT_LE(writes_complete_, msgs_to_send_);
 
-         EXPECT_LE(reads_complete_, writes_complete_);
 
-         break;
 
-       case CANCEL_AFTER_PROCESSING:
 
-         EXPECT_FALSE(s.ok());
 
-         EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
 
-         EXPECT_EQ(writes_complete_, msgs_to_send_);
 
-         // The Server canceled after reading the last message and after writing
 
-         // the message to the client. However, the RPC cancellation might have
 
-         // taken effect before the client actually read the response.
 
-         EXPECT_LE(reads_complete_, writes_complete_);
 
-         break;
 
-       default:
 
-         assert(false);
 
-     }
 
-     std::unique_lock<std::mutex> l(mu_);
 
-     done_ = true;
 
-     cv_.notify_one();
 
-   }
 
-   void Await() {
 
-     std::unique_lock<std::mutex> l(mu_);
 
-     while (!done_) {
 
-       cv_.wait(l);
 
-     }
 
-   }
 
-  private:
 
-   void MaybeAsyncWrite(bool first_write_async) {
 
-     if (first_write_async) {
 
-       // Make sure that we have a write to issue.
 
-       // TODO(vjpai): Make this work with 0 writes case as well.
 
-       assert(msgs_to_send_ >= 1);
 
-       AddHold();
 
-       async_write_thread_ = std::thread([this] {
 
-         std::unique_lock<std::mutex> lock(async_write_thread_mu_);
 
-         async_write_thread_cv_.wait(
 
-             lock, [this] { return async_write_thread_start_; });
 
-         MaybeWrite();
 
-       });
 
-       std::lock_guard<std::mutex> lock(async_write_thread_mu_);
 
-       async_write_thread_start_ = true;
 
-       async_write_thread_cv_.notify_one();
 
-       return;
 
-     }
 
-     MaybeWrite();
 
-   }
 
-   void MaybeWrite() {
 
-     if (client_cancel_.cancel &&
 
-         writes_complete_ == client_cancel_.ops_before_cancel) {
 
-       context_.TryCancel();
 
-     } else if (writes_complete_ == msgs_to_send_) {
 
-       StartWritesDone();
 
-     } else {
 
-       StartWrite(&request_);
 
-     }
 
-   }
 
-   EchoRequest request_;
 
-   EchoResponse response_;
 
-   ClientContext context_;
 
-   const ServerTryCancelRequestPhase server_try_cancel_;
 
-   int reads_complete_{0};
 
-   int writes_complete_{0};
 
-   const int msgs_to_send_;
 
-   const ClientCancelInfo client_cancel_;
 
-   std::mutex mu_;
 
-   std::condition_variable cv_;
 
-   bool done_ = false;
 
-   std::thread async_write_thread_;
 
-   bool async_write_thread_start_ = false;
 
-   std::mutex async_write_thread_mu_;
 
-   std::condition_variable async_write_thread_cv_;
 
- };
 
- TEST_P(ClientCallbackEnd2endTest, BidiStream) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   BidiClient test(stub_.get(), DO_NOT_CANCEL,
 
-                   kServerDefaultResponseStreamsToSend,
 
-                   /*cork_metadata=*/false, /*first_write_async=*/false);
 
-   test.Await();
 
-   // Make sure that the server interceptors were not notified of a cancel
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, BidiStreamFirstWriteAsync) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   BidiClient test(stub_.get(), DO_NOT_CANCEL,
 
-                   kServerDefaultResponseStreamsToSend,
 
-                   /*cork_metadata=*/false, /*first_write_async=*/true);
 
-   test.Await();
 
-   // Make sure that the server interceptors were not notified of a cancel
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, BidiStreamCorked) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   BidiClient test(stub_.get(), DO_NOT_CANCEL,
 
-                   kServerDefaultResponseStreamsToSend,
 
-                   /*cork_metadata=*/true, /*first_write_async=*/false);
 
-   test.Await();
 
-   // Make sure that the server interceptors were not notified of a cancel
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, BidiStreamCorkedFirstWriteAsync) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   BidiClient test(stub_.get(), DO_NOT_CANCEL,
 
-                   kServerDefaultResponseStreamsToSend,
 
-                   /*cork_metadata=*/true, /*first_write_async=*/true);
 
-   test.Await();
 
-   // Make sure that the server interceptors were not notified of a cancel
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, ClientCancelsBidiStream) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   BidiClient test(stub_.get(), DO_NOT_CANCEL,
 
-                   kServerDefaultResponseStreamsToSend,
 
-                   /*cork_metadata=*/false, /*first_write_async=*/false,
 
-                   ClientCancelInfo(2));
 
-   test.Await();
 
-   // Make sure that the server interceptors were notified of a cancel
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- // Server to cancel before reading/writing any requests/responses on the stream
 
- TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelBefore) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   BidiClient test(stub_.get(), CANCEL_BEFORE_PROCESSING, /*num_msgs_to_send=*/2,
 
-                   /*cork_metadata=*/false, /*first_write_async=*/false);
 
-   test.Await();
 
-   // Make sure that the server interceptors were notified
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- // Server to cancel while reading/writing requests/responses on the stream in
 
- // parallel
 
- TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelDuring) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   BidiClient test(stub_.get(), CANCEL_DURING_PROCESSING,
 
-                   /*num_msgs_to_send=*/10, /*cork_metadata=*/false,
 
-                   /*first_write_async=*/false);
 
-   test.Await();
 
-   // Make sure that the server interceptors were notified
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- // Server to cancel after reading/writing all requests/responses on the stream
 
- // but before returning to the client
 
- TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelAfter) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   BidiClient test(stub_.get(), CANCEL_AFTER_PROCESSING, /*num_msgs_to_send=*/5,
 
-                   /*cork_metadata=*/false, /*first_write_async=*/false);
 
-   test.Await();
 
-   // Make sure that the server interceptors were notified
 
-   if (GetParam().use_interceptors) {
 
-     EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, SimultaneousReadAndWritesDone) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   class Client : public grpc::experimental::ClientBidiReactor<EchoRequest,
 
-                                                               EchoResponse> {
 
-    public:
 
-     Client(grpc::testing::EchoTestService::Stub* stub) {
 
-       request_.set_message("Hello bidi ");
 
-       stub->experimental_async()->BidiStream(&context_, this);
 
-       StartWrite(&request_);
 
-       StartCall();
 
-     }
 
-     void OnReadDone(bool ok) override {
 
-       EXPECT_TRUE(ok);
 
-       EXPECT_EQ(response_.message(), request_.message());
 
-     }
 
-     void OnWriteDone(bool ok) override {
 
-       EXPECT_TRUE(ok);
 
-       // Now send out the simultaneous Read and WritesDone
 
-       StartWritesDone();
 
-       StartRead(&response_);
 
-     }
 
-     void OnDone(const Status& s) override {
 
-       EXPECT_TRUE(s.ok());
 
-       EXPECT_EQ(response_.message(), request_.message());
 
-       std::unique_lock<std::mutex> l(mu_);
 
-       done_ = true;
 
-       cv_.notify_one();
 
-     }
 
-     void Await() {
 
-       std::unique_lock<std::mutex> l(mu_);
 
-       while (!done_) {
 
-         cv_.wait(l);
 
-       }
 
-     }
 
-    private:
 
-     EchoRequest request_;
 
-     EchoResponse response_;
 
-     ClientContext context_;
 
-     std::mutex mu_;
 
-     std::condition_variable cv_;
 
-     bool done_ = false;
 
-   } test{stub_.get()};
 
-   test.Await();
 
- }
 
- TEST_P(ClientCallbackEnd2endTest, UnimplementedRpc) {
 
-   MAYBE_SKIP_TEST;
 
-   ChannelArguments args;
 
-   const auto& channel_creds = GetCredentialsProvider()->GetChannelCredentials(
 
-       GetParam().credentials_type, &args);
 
-   std::shared_ptr<Channel> channel =
 
-       (GetParam().protocol == Protocol::TCP)
 
-           ? ::grpc::CreateCustomChannel(server_address_.str(), channel_creds,
 
-                                         args)
 
-           : server_->InProcessChannel(args);
 
-   std::unique_ptr<grpc::testing::UnimplementedEchoService::Stub> stub;
 
-   stub = grpc::testing::UnimplementedEchoService::NewStub(channel);
 
-   EchoRequest request;
 
-   EchoResponse response;
 
-   ClientContext cli_ctx;
 
-   request.set_message("Hello world.");
 
-   std::mutex mu;
 
-   std::condition_variable cv;
 
-   bool done = false;
 
-   stub->experimental_async()->Unimplemented(
 
-       &cli_ctx, &request, &response, [&done, &mu, &cv](Status s) {
 
-         EXPECT_EQ(StatusCode::UNIMPLEMENTED, s.error_code());
 
-         EXPECT_EQ("", s.error_message());
 
-         std::lock_guard<std::mutex> l(mu);
 
-         done = true;
 
-         cv.notify_one();
 
-       });
 
-   std::unique_lock<std::mutex> l(mu);
 
-   while (!done) {
 
-     cv.wait(l);
 
-   }
 
- }
 
- TEST_P(ClientCallbackEnd2endTest,
 
-        ResponseStreamExtraReactionFlowReadsUntilDone) {
 
-   MAYBE_SKIP_TEST;
 
-   ResetStub();
 
-   class ReadAllIncomingDataClient
 
-       : public grpc::experimental::ClientReadReactor<EchoResponse> {
 
-    public:
 
-     ReadAllIncomingDataClient(grpc::testing::EchoTestService::Stub* stub) {
 
-       request_.set_message("Hello client ");
 
-       stub->experimental_async()->ResponseStream(&context_, &request_, this);
 
-     }
 
-     bool WaitForReadDone() {
 
-       std::unique_lock<std::mutex> l(mu_);
 
-       while (!read_done_) {
 
-         read_cv_.wait(l);
 
-       }
 
-       read_done_ = false;
 
-       return read_ok_;
 
-     }
 
-     void Await() {
 
-       std::unique_lock<std::mutex> l(mu_);
 
-       while (!done_) {
 
-         done_cv_.wait(l);
 
-       }
 
-     }
 
-     const Status& status() {
 
-       std::unique_lock<std::mutex> l(mu_);
 
-       return status_;
 
-     }
 
-    private:
 
-     void OnReadDone(bool ok) override {
 
-       std::unique_lock<std::mutex> l(mu_);
 
-       read_ok_ = ok;
 
-       read_done_ = true;
 
-       read_cv_.notify_one();
 
-     }
 
-     void OnDone(const Status& s) override {
 
-       std::unique_lock<std::mutex> l(mu_);
 
-       done_ = true;
 
-       status_ = s;
 
-       done_cv_.notify_one();
 
-     }
 
-     EchoRequest request_;
 
-     EchoResponse response_;
 
-     ClientContext context_;
 
-     bool read_ok_ = false;
 
-     bool read_done_ = false;
 
-     std::mutex mu_;
 
-     std::condition_variable read_cv_;
 
-     std::condition_variable done_cv_;
 
-     bool done_ = false;
 
-     Status status_;
 
-   } client{stub_.get()};
 
-   int reads_complete = 0;
 
-   client.AddHold();
 
-   client.StartCall();
 
-   EchoResponse response;
 
-   bool read_ok = true;
 
-   while (read_ok) {
 
-     client.StartRead(&response);
 
-     read_ok = client.WaitForReadDone();
 
-     if (read_ok) {
 
-       ++reads_complete;
 
-     }
 
-   }
 
-   client.RemoveHold();
 
-   client.Await();
 
-   EXPECT_EQ(kServerDefaultResponseStreamsToSend, reads_complete);
 
-   EXPECT_EQ(client.status().error_code(), grpc::StatusCode::OK);
 
- }
 
- std::vector<TestScenario> CreateTestScenarios(bool test_insecure) {
 
- #if TARGET_OS_IPHONE
 
-   // Workaround Apple CFStream bug
 
-   gpr_setenv("grpc_cfstream", "0");
 
- #endif
 
-   std::vector<TestScenario> scenarios;
 
-   std::vector<grpc::string> credentials_types{
 
-       GetCredentialsProvider()->GetSecureCredentialsTypeList()};
 
-   auto insec_ok = [] {
 
-     // Only allow insecure credentials type when it is registered with the
 
-     // provider. User may create providers that do not have insecure.
 
-     return GetCredentialsProvider()->GetChannelCredentials(
 
-                kInsecureCredentialsType, nullptr) != nullptr;
 
-   };
 
-   if (test_insecure && insec_ok()) {
 
-     credentials_types.push_back(kInsecureCredentialsType);
 
-   }
 
-   GPR_ASSERT(!credentials_types.empty());
 
-   bool barr[]{false, true};
 
-   Protocol parr[]{Protocol::INPROC, Protocol::TCP};
 
-   for (Protocol p : parr) {
 
-     for (const auto& cred : credentials_types) {
 
-       // TODO(vjpai): Test inproc with secure credentials when feasible
 
-       if (p == Protocol::INPROC &&
 
-           (cred != kInsecureCredentialsType || !insec_ok())) {
 
-         continue;
 
-       }
 
-       for (bool callback_server : barr) {
 
-         for (bool use_interceptors : barr) {
 
-           scenarios.emplace_back(callback_server, p, use_interceptors, cred);
 
-         }
 
-       }
 
-     }
 
-   }
 
-   return scenarios;
 
- }
 
- INSTANTIATE_TEST_SUITE_P(ClientCallbackEnd2endTest, ClientCallbackEnd2endTest,
 
-                          ::testing::ValuesIn(CreateTestScenarios(true)));
 
- }  // namespace
 
- }  // namespace testing
 
- }  // namespace grpc
 
- int main(int argc, char** argv) {
 
-   ::testing::InitGoogleTest(&argc, argv);
 
-   grpc::testing::TestEnvironment env(argc, argv);
 
-   grpc_init();
 
-   int ret = RUN_ALL_TESTS();
 
-   grpc_shutdown();
 
-   return ret;
 
- }
 
 
  |