| /* |
| * |
| * Copyright 2015, Google Inc. |
| * All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions are |
| * met: |
| * |
| * * Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * * Redistributions in binary form must reproduce the above |
| * copyright notice, this list of conditions and the following disclaimer |
| * in the documentation and/or other materials provided with the |
| * distribution. |
| * * Neither the name of Google Inc. nor the names of its |
| * contributors may be used to endorse or promote products derived from |
| * this software without specific prior written permission. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
| * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| * |
| */ |
| |
| #include <mutex> |
| #include <thread> |
| |
| #include <grpc++/channel.h> |
| #include <grpc++/client_context.h> |
| #include <grpc++/create_channel.h> |
| #include <grpc++/server.h> |
| #include <grpc++/server_builder.h> |
| #include <grpc++/server_context.h> |
| #include <grpc/grpc.h> |
| #include <grpc/support/thd.h> |
| #include <grpc/support/time.h> |
| #include <gtest/gtest.h> |
| |
| #include "src/core/lib/surface/api_trace.h" |
| #include "src/proto/grpc/testing/duplicate/echo_duplicate.grpc.pb.h" |
| #include "src/proto/grpc/testing/echo.grpc.pb.h" |
| #include "test/core/util/port.h" |
| #include "test/core/util/test_config.h" |
| |
| using grpc::testing::EchoRequest; |
| using grpc::testing::EchoResponse; |
| using std::chrono::system_clock; |
| |
| const int kNumThreads = 100; // Number of threads |
| const int kNumAsyncSendThreads = 2; |
| const int kNumAsyncReceiveThreads = 50; |
| const int kNumAsyncServerThreads = 50; |
| const int kNumRpcs = 1000; // Number of RPCs per thread |
| |
| namespace grpc { |
| namespace testing { |
| |
| namespace { |
| |
| // When echo_deadline is requested, deadline seen in the ServerContext is set in |
| // the response in seconds. |
| void MaybeEchoDeadline(ServerContext* context, const EchoRequest* request, |
| EchoResponse* response) { |
| if (request->has_param() && request->param().echo_deadline()) { |
| gpr_timespec deadline = gpr_inf_future(GPR_CLOCK_REALTIME); |
| if (context->deadline() != system_clock::time_point::max()) { |
| Timepoint2Timespec(context->deadline(), &deadline); |
| } |
| response->mutable_param()->set_request_deadline(deadline.tv_sec); |
| } |
| } |
| |
| } // namespace |
| |
| class TestServiceImpl : public ::grpc::testing::EchoTestService::Service { |
| public: |
| TestServiceImpl() : signal_client_(false) {} |
| |
| Status Echo(ServerContext* context, const EchoRequest* request, |
| EchoResponse* response) GRPC_OVERRIDE { |
| response->set_message(request->message()); |
| MaybeEchoDeadline(context, request, response); |
| if (request->has_param() && request->param().client_cancel_after_us()) { |
| { |
| unique_lock<mutex> lock(mu_); |
| signal_client_ = true; |
| } |
| while (!context->IsCancelled()) { |
| gpr_sleep_until(gpr_time_add( |
| gpr_now(GPR_CLOCK_REALTIME), |
| gpr_time_from_micros(request->param().client_cancel_after_us(), |
| GPR_TIMESPAN))); |
| } |
| return Status::CANCELLED; |
| } else if (request->has_param() && |
| request->param().server_cancel_after_us()) { |
| gpr_sleep_until(gpr_time_add( |
| gpr_now(GPR_CLOCK_REALTIME), |
| gpr_time_from_micros(request->param().server_cancel_after_us(), |
| GPR_TIMESPAN))); |
| return Status::CANCELLED; |
| } else { |
| EXPECT_FALSE(context->IsCancelled()); |
| } |
| return Status::OK; |
| } |
| |
| // Unimplemented is left unimplemented to test the returned error. |
| |
| Status RequestStream(ServerContext* context, |
| ServerReader<EchoRequest>* reader, |
| EchoResponse* response) GRPC_OVERRIDE { |
| EchoRequest request; |
| response->set_message(""); |
| while (reader->Read(&request)) { |
| response->mutable_message()->append(request.message()); |
| } |
| return Status::OK; |
| } |
| |
| // Return 3 messages. |
| // TODO(yangg) make it generic by adding a parameter into EchoRequest |
| Status ResponseStream(ServerContext* context, const EchoRequest* request, |
| ServerWriter<EchoResponse>* writer) GRPC_OVERRIDE { |
| EchoResponse response; |
| response.set_message(request->message() + "0"); |
| writer->Write(response); |
| response.set_message(request->message() + "1"); |
| writer->Write(response); |
| response.set_message(request->message() + "2"); |
| writer->Write(response); |
| |
| return Status::OK; |
| } |
| |
| Status BidiStream(ServerContext* context, |
| ServerReaderWriter<EchoResponse, EchoRequest>* stream) |
| GRPC_OVERRIDE { |
| EchoRequest request; |
| EchoResponse response; |
| while (stream->Read(&request)) { |
| gpr_log(GPR_INFO, "recv msg %s", request.message().c_str()); |
| response.set_message(request.message()); |
| stream->Write(response); |
| } |
| return Status::OK; |
| } |
| |
| bool signal_client() { |
| unique_lock<mutex> lock(mu_); |
| return signal_client_; |
| } |
| |
| private: |
| bool signal_client_; |
| mutex mu_; |
| }; |
| |
| class TestServiceImplDupPkg |
| : public ::grpc::testing::duplicate::EchoTestService::Service { |
| public: |
| Status Echo(ServerContext* context, const EchoRequest* request, |
| EchoResponse* response) GRPC_OVERRIDE { |
| response->set_message("no package"); |
| return Status::OK; |
| } |
| }; |
| |
| template <class Service> |
| class CommonStressTest { |
| public: |
| CommonStressTest() : kMaxMessageSize_(8192) {} |
| virtual ~CommonStressTest() {} |
| virtual void SetUp() = 0; |
| virtual void TearDown() = 0; |
| void ResetStub() { |
| std::shared_ptr<Channel> channel = |
| CreateChannel(server_address_.str(), InsecureChannelCredentials()); |
| stub_ = grpc::testing::EchoTestService::NewStub(channel); |
| } |
| grpc::testing::EchoTestService::Stub* GetStub() { return stub_.get(); } |
| |
| protected: |
| void SetUpStart(ServerBuilder* builder, Service* service) { |
| int port = grpc_pick_unused_port_or_die(); |
| server_address_ << "localhost:" << port; |
| // Setup server |
| builder->AddListeningPort(server_address_.str(), |
| InsecureServerCredentials()); |
| builder->RegisterService(service); |
| builder->SetMaxMessageSize( |
| kMaxMessageSize_); // For testing max message size. |
| builder->RegisterService(&dup_pkg_service_); |
| } |
| void SetUpEnd(ServerBuilder* builder) { server_ = builder->BuildAndStart(); } |
| void TearDownStart() { server_->Shutdown(); } |
| void TearDownEnd() {} |
| |
| private: |
| std::unique_ptr<grpc::testing::EchoTestService::Stub> stub_; |
| std::unique_ptr<Server> server_; |
| std::ostringstream server_address_; |
| const int kMaxMessageSize_; |
| TestServiceImplDupPkg dup_pkg_service_; |
| }; |
| |
| class CommonStressTestSyncServer : public CommonStressTest<TestServiceImpl> { |
| public: |
| void SetUp() GRPC_OVERRIDE { |
| ServerBuilder builder; |
| SetUpStart(&builder, &service_); |
| SetUpEnd(&builder); |
| } |
| void TearDown() GRPC_OVERRIDE { |
| TearDownStart(); |
| TearDownEnd(); |
| } |
| |
| private: |
| TestServiceImpl service_; |
| }; |
| |
| class CommonStressTestAsyncServer |
| : public CommonStressTest<grpc::testing::EchoTestService::AsyncService> { |
| public: |
| void SetUp() GRPC_OVERRIDE { |
| shutting_down_ = false; |
| ServerBuilder builder; |
| SetUpStart(&builder, &service_); |
| cq_ = builder.AddCompletionQueue(); |
| SetUpEnd(&builder); |
| contexts_ = new Context[kNumAsyncServerThreads * 100]; |
| for (int i = 0; i < kNumAsyncServerThreads * 100; i++) { |
| RefreshContext(i); |
| } |
| for (int i = 0; i < kNumAsyncServerThreads; i++) { |
| server_threads_.push_back( |
| new std::thread(&CommonStressTestAsyncServer::ProcessRpcs, this)); |
| } |
| } |
| void TearDown() GRPC_OVERRIDE { |
| { |
| unique_lock<mutex> l(mu_); |
| TearDownStart(); |
| shutting_down_ = true; |
| cq_->Shutdown(); |
| } |
| |
| for (int i = 0; i < kNumAsyncServerThreads; i++) { |
| server_threads_[i]->join(); |
| delete server_threads_[i]; |
| } |
| |
| void* ignored_tag; |
| bool ignored_ok; |
| while (cq_->Next(&ignored_tag, &ignored_ok)) |
| ; |
| TearDownEnd(); |
| delete[] contexts_; |
| } |
| |
| private: |
| void ProcessRpcs() { |
| void* tag; |
| bool ok; |
| while (cq_->Next(&tag, &ok)) { |
| if (ok) { |
| int i = static_cast<int>(reinterpret_cast<intptr_t>(tag)); |
| switch (contexts_[i].state) { |
| case Context::READY: { |
| contexts_[i].state = Context::DONE; |
| EchoResponse send_response; |
| send_response.set_message(contexts_[i].recv_request.message()); |
| contexts_[i].response_writer->Finish(send_response, Status::OK, |
| tag); |
| break; |
| } |
| case Context::DONE: |
| RefreshContext(i); |
| break; |
| } |
| } |
| } |
| } |
| void RefreshContext(int i) { |
| unique_lock<mutex> l(mu_); |
| if (!shutting_down_) { |
| contexts_[i].state = Context::READY; |
| contexts_[i].srv_ctx.reset(new ServerContext); |
| contexts_[i].response_writer.reset( |
| new grpc::ServerAsyncResponseWriter<EchoResponse>( |
| contexts_[i].srv_ctx.get())); |
| service_.RequestEcho(contexts_[i].srv_ctx.get(), |
| &contexts_[i].recv_request, |
| contexts_[i].response_writer.get(), cq_.get(), |
| cq_.get(), (void*)(intptr_t)i); |
| } |
| } |
| struct Context { |
| std::unique_ptr<ServerContext> srv_ctx; |
| std::unique_ptr<grpc::ServerAsyncResponseWriter<EchoResponse>> |
| response_writer; |
| EchoRequest recv_request; |
| enum { READY, DONE } state; |
| } * contexts_; |
| ::grpc::testing::EchoTestService::AsyncService service_; |
| std::unique_ptr<ServerCompletionQueue> cq_; |
| bool shutting_down_; |
| mutex mu_; |
| std::vector<std::thread*> server_threads_; |
| }; |
| |
| template <class Common> |
| class End2endTest : public ::testing::Test { |
| protected: |
| End2endTest() {} |
| void SetUp() GRPC_OVERRIDE { common_.SetUp(); } |
| void TearDown() GRPC_OVERRIDE { common_.TearDown(); } |
| void ResetStub() { common_.ResetStub(); } |
| |
| Common common_; |
| }; |
| |
| static void SendRpc(grpc::testing::EchoTestService::Stub* stub, int num_rpcs) { |
| EchoRequest request; |
| EchoResponse response; |
| request.set_message("Hello"); |
| |
| for (int i = 0; i < num_rpcs; ++i) { |
| ClientContext context; |
| Status s = stub->Echo(&context, request, &response); |
| EXPECT_EQ(response.message(), request.message()); |
| EXPECT_TRUE(s.ok()); |
| } |
| } |
| |
| typedef ::testing::Types<CommonStressTestSyncServer, |
| CommonStressTestAsyncServer> |
| CommonTypes; |
| TYPED_TEST_CASE(End2endTest, CommonTypes); |
| TYPED_TEST(End2endTest, ThreadStress) { |
| this->common_.ResetStub(); |
| std::vector<std::thread*> threads; |
| for (int i = 0; i < kNumThreads; ++i) { |
| threads.push_back( |
| new std::thread(SendRpc, this->common_.GetStub(), kNumRpcs)); |
| } |
| for (int i = 0; i < kNumThreads; ++i) { |
| threads[i]->join(); |
| delete threads[i]; |
| } |
| } |
| |
| template <class Common> |
| class AsyncClientEnd2endTest : public ::testing::Test { |
| protected: |
| AsyncClientEnd2endTest() : rpcs_outstanding_(0) {} |
| |
| void SetUp() GRPC_OVERRIDE { common_.SetUp(); } |
| void TearDown() GRPC_OVERRIDE { |
| void* ignored_tag; |
| bool ignored_ok; |
| while (cq_.Next(&ignored_tag, &ignored_ok)) |
| ; |
| common_.TearDown(); |
| } |
| |
| void Wait() { |
| unique_lock<mutex> l(mu_); |
| while (rpcs_outstanding_ != 0) { |
| cv_.wait(l); |
| } |
| |
| cq_.Shutdown(); |
| } |
| |
| struct AsyncClientCall { |
| EchoResponse response; |
| ClientContext context; |
| Status status; |
| std::unique_ptr<ClientAsyncResponseReader<EchoResponse>> response_reader; |
| }; |
| |
| void AsyncSendRpc(int num_rpcs) { |
| for (int i = 0; i < num_rpcs; ++i) { |
| AsyncClientCall* call = new AsyncClientCall; |
| EchoRequest request; |
| request.set_message("Hello: " + grpc::to_string(i)); |
| call->response_reader = |
| common_.GetStub()->AsyncEcho(&call->context, request, &cq_); |
| call->response_reader->Finish(&call->response, &call->status, |
| (void*)call); |
| |
| unique_lock<mutex> l(mu_); |
| rpcs_outstanding_++; |
| } |
| } |
| |
| void AsyncCompleteRpc() { |
| while (true) { |
| void* got_tag; |
| bool ok = false; |
| if (!cq_.Next(&got_tag, &ok)) break; |
| AsyncClientCall* call = static_cast<AsyncClientCall*>(got_tag); |
| if (!ok) { |
| gpr_log(GPR_DEBUG, "Error: %d", call->status.error_code()); |
| } |
| delete call; |
| |
| bool notify; |
| { |
| unique_lock<mutex> l(mu_); |
| rpcs_outstanding_--; |
| notify = (rpcs_outstanding_ == 0); |
| } |
| if (notify) { |
| cv_.notify_all(); |
| } |
| } |
| } |
| |
| Common common_; |
| CompletionQueue cq_; |
| mutex mu_; |
| condition_variable cv_; |
| int rpcs_outstanding_; |
| }; |
| |
| TYPED_TEST_CASE(AsyncClientEnd2endTest, CommonTypes); |
| TYPED_TEST(AsyncClientEnd2endTest, ThreadStress) { |
| this->common_.ResetStub(); |
| std::vector<std::thread *> send_threads, completion_threads; |
| for (int i = 0; i < kNumAsyncReceiveThreads; ++i) { |
| completion_threads.push_back(new std::thread( |
| &AsyncClientEnd2endTest_ThreadStress_Test<TypeParam>::AsyncCompleteRpc, |
| this)); |
| } |
| for (int i = 0; i < kNumAsyncSendThreads; ++i) { |
| send_threads.push_back(new std::thread( |
| &AsyncClientEnd2endTest_ThreadStress_Test<TypeParam>::AsyncSendRpc, |
| this, kNumRpcs)); |
| } |
| for (int i = 0; i < kNumAsyncSendThreads; ++i) { |
| send_threads[i]->join(); |
| delete send_threads[i]; |
| } |
| |
| this->Wait(); |
| for (int i = 0; i < kNumAsyncReceiveThreads; ++i) { |
| completion_threads[i]->join(); |
| delete completion_threads[i]; |
| } |
| } |
| |
| } // namespace testing |
| } // namespace grpc |
| |
| int main(int argc, char** argv) { |
| grpc_test_init(argc, argv); |
| ::testing::InitGoogleTest(&argc, argv); |
| return RUN_ALL_TESTS(); |
| } |