blob: 34d15dfc3b9e20cdb54882692e894890da8f95e8 [file] [log] [blame]
Craig Tillerc67cc992017-04-27 10:15:51 -07001/*
2 *
Jan Tattermusch7897ae92017-06-07 22:57:36 +02003 * Copyright 2017 gRPC authors.
Craig Tillerc67cc992017-04-27 10:15:51 -07004 *
Jan Tattermusch7897ae92017-06-07 22:57:36 +02005 * Licensed under the Apache License, Version 2.0 (the "License");
6 * you may not use this file except in compliance with the License.
7 * You may obtain a copy of the License at
Craig Tillerc67cc992017-04-27 10:15:51 -07008 *
Jan Tattermusch7897ae92017-06-07 22:57:36 +02009 * http://www.apache.org/licenses/LICENSE-2.0
Craig Tillerc67cc992017-04-27 10:15:51 -070010 *
Jan Tattermusch7897ae92017-06-07 22:57:36 +020011 * Unless required by applicable law or agreed to in writing, software
12 * distributed under the License is distributed on an "AS IS" BASIS,
13 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14 * See the License for the specific language governing permissions and
15 * limitations under the License.
Craig Tillerc67cc992017-04-27 10:15:51 -070016 *
17 */
18
19#include "src/core/lib/iomgr/port.h"
20
21/* This polling engine is only relevant on linux kernels supporting epoll() */
22#ifdef GRPC_LINUX_EPOLL
23
Craig Tiller4509c472017-04-27 19:05:13 +000024#include "src/core/lib/iomgr/ev_epoll1_linux.h"
Craig Tillerc67cc992017-04-27 10:15:51 -070025
26#include <assert.h>
27#include <errno.h>
28#include <poll.h>
29#include <pthread.h>
30#include <string.h>
31#include <sys/epoll.h>
32#include <sys/socket.h>
33#include <unistd.h>
34
35#include <grpc/support/alloc.h>
Craig Tiller6de05932017-04-28 09:17:38 -070036#include <grpc/support/cpu.h>
Craig Tillerc67cc992017-04-27 10:15:51 -070037#include <grpc/support/log.h>
38#include <grpc/support/string_util.h>
39#include <grpc/support/tls.h>
40#include <grpc/support/useful.h>
41
42#include "src/core/lib/iomgr/ev_posix.h"
43#include "src/core/lib/iomgr/iomgr_internal.h"
44#include "src/core/lib/iomgr/lockfree_event.h"
Craig Tillerc67cc992017-04-27 10:15:51 -070045#include "src/core/lib/iomgr/wakeup_fd_posix.h"
Craig Tillerc67cc992017-04-27 10:15:51 -070046#include "src/core/lib/profiling/timers.h"
47#include "src/core/lib/support/block_annotate.h"
Craig Tillerb89bac02017-05-26 15:20:32 +000048#include "src/core/lib/support/string.h"
Craig Tillerc67cc992017-04-27 10:15:51 -070049
Craig Tillerc67cc992017-04-27 10:15:51 -070050static grpc_wakeup_fd global_wakeup_fd;
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -070051
52/*******************************************************************************
53 * Singleton epoll set related fields
54 */
55
56#define MAX_EPOLL_EVENTS 100
57#define MAX_EPOLL_EVENTS_HANDLED_PER_ITERATION 5
58
59/* Note: Since fields in this struct are only modified by the designated poller,
60 we do not need any locks to protect the struct */
61typedef struct epoll_set {
62 int epfd;
63
64 /* The epoll_events after the last call to epoll_wait() */
65 struct epoll_event events[MAX_EPOLL_EVENTS];
66
67 /* The number of epoll_events after the last call to epoll_wait() */
68 int num_events;
69
70 /* Index of the first event in epoll_events that has to be processed. This
71 * field is only valid if num_events > 0 */
72 int cursor;
73} epoll_set;
74
75/* The global singleton epoll set */
76static epoll_set g_epoll_set;
77
78/* Must be called *only* once */
79static bool epoll_set_init() {
80 g_epoll_set.epfd = epoll_create1(EPOLL_CLOEXEC);
81 if (g_epoll_set.epfd < 0) {
82 gpr_log(GPR_ERROR, "epoll unavailable");
83 return false;
84 }
85
86 gpr_log(GPR_ERROR, "grpc epoll fd: %d", g_epoll_set.epfd);
87 g_epoll_set.num_events = 0;
88 g_epoll_set.cursor = 0;
89 return true;
90}
91
92/* epoll_set_init() MUST be called before calling this. */
93static void epoll_set_shutdown() {
94 if (g_epoll_set.epfd >= 0) {
95 close(g_epoll_set.epfd);
96 g_epoll_set.epfd = -1;
97 }
98}
Craig Tillerc67cc992017-04-27 10:15:51 -070099
100/*******************************************************************************
101 * Fd Declarations
102 */
103
104struct grpc_fd {
105 int fd;
106
Craig Tillerc67cc992017-04-27 10:15:51 -0700107 gpr_atm read_closure;
108 gpr_atm write_closure;
109
110 struct grpc_fd *freelist_next;
Craig Tillerc67cc992017-04-27 10:15:51 -0700111
112 /* The pollset that last noticed that the fd is readable. The actual type
113 * stored in this is (grpc_pollset *) */
114 gpr_atm read_notifier_pollset;
115
116 grpc_iomgr_object iomgr_object;
117};
118
119static void fd_global_init(void);
120static void fd_global_shutdown(void);
121
122/*******************************************************************************
123 * Pollset Declarations
124 */
125
Craig Tiller43bf2592017-04-28 23:21:01 +0000126typedef enum { UNKICKED, KICKED, DESIGNATED_POLLER } kick_state;
Craig Tillerc67cc992017-04-27 10:15:51 -0700127
Craig Tiller830e82a2017-05-31 16:26:27 -0700128static const char *kick_state_string(kick_state st) {
129 switch (st) {
130 case UNKICKED:
131 return "UNKICKED";
132 case KICKED:
133 return "KICKED";
134 case DESIGNATED_POLLER:
135 return "DESIGNATED_POLLER";
136 }
137 GPR_UNREACHABLE_CODE(return "UNKNOWN");
138}
139
Craig Tillerc67cc992017-04-27 10:15:51 -0700140struct grpc_pollset_worker {
Craig Tiller32f90ee2017-04-28 12:46:41 -0700141 kick_state kick_state;
Craig Tiller55624a32017-05-26 08:14:44 -0700142 int kick_state_mutator; // which line of code last changed kick state
Craig Tillerc67cc992017-04-27 10:15:51 -0700143 bool initialized_cv;
Craig Tiller32f90ee2017-04-28 12:46:41 -0700144 grpc_pollset_worker *next;
145 grpc_pollset_worker *prev;
Craig Tillerc67cc992017-04-27 10:15:51 -0700146 gpr_cv cv;
Craig Tiller50da5ec2017-05-01 13:51:14 -0700147 grpc_closure_list schedule_on_end_work;
Craig Tillerc67cc992017-04-27 10:15:51 -0700148};
149
Craig Tiller55624a32017-05-26 08:14:44 -0700150#define SET_KICK_STATE(worker, state) \
151 do { \
152 (worker)->kick_state = (state); \
153 (worker)->kick_state_mutator = __LINE__; \
154 } while (false)
155
Craig Tillerba550da2017-05-01 14:26:31 +0000156#define MAX_NEIGHBOURHOODS 1024
157
Craig Tiller6de05932017-04-28 09:17:38 -0700158typedef struct pollset_neighbourhood {
159 gpr_mu mu;
160 grpc_pollset *active_root;
Craig Tiller6de05932017-04-28 09:17:38 -0700161 char pad[GPR_CACHELINE_SIZE];
162} pollset_neighbourhood;
163
Craig Tillerc67cc992017-04-27 10:15:51 -0700164struct grpc_pollset {
Craig Tiller6de05932017-04-28 09:17:38 -0700165 gpr_mu mu;
166 pollset_neighbourhood *neighbourhood;
Craig Tillere00d7332017-05-01 15:43:51 +0000167 bool reassigning_neighbourhood;
Craig Tiller4509c472017-04-27 19:05:13 +0000168 grpc_pollset_worker *root_worker;
169 bool kicked_without_poller;
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700170
171 /* Set to true if the pollset is observed to have no workers available to
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700172 poll */
Craig Tiller6de05932017-04-28 09:17:38 -0700173 bool seen_inactive;
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700174 bool shutting_down; /* Is the pollset shutting down ? */
Craig Tiller4509c472017-04-27 19:05:13 +0000175 grpc_closure *shutdown_closure; /* Called after after shutdown is complete */
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700176
177 /* Number of workers who are *about-to* attach themselves to the pollset
178 * worker list */
Craig Tillerba550da2017-05-01 14:26:31 +0000179 int begin_refs;
Craig Tiller6de05932017-04-28 09:17:38 -0700180
181 grpc_pollset *next;
182 grpc_pollset *prev;
Craig Tillerc67cc992017-04-27 10:15:51 -0700183};
184
185/*******************************************************************************
186 * Pollset-set Declarations
187 */
Craig Tiller6de05932017-04-28 09:17:38 -0700188
Craig Tiller61f96c12017-05-12 13:36:39 -0700189struct grpc_pollset_set {
190 char unused;
191};
Craig Tillerc67cc992017-04-27 10:15:51 -0700192
193/*******************************************************************************
194 * Common helpers
195 */
196
197static bool append_error(grpc_error **composite, grpc_error *error,
198 const char *desc) {
199 if (error == GRPC_ERROR_NONE) return true;
200 if (*composite == GRPC_ERROR_NONE) {
201 *composite = GRPC_ERROR_CREATE_FROM_COPIED_STRING(desc);
202 }
203 *composite = grpc_error_add_child(*composite, error);
204 return false;
205}
206
207/*******************************************************************************
208 * Fd Definitions
209 */
210
211/* We need to keep a freelist not because of any concerns of malloc performance
212 * but instead so that implementations with multiple threads in (for example)
213 * epoll_wait deal with the race between pollset removal and incoming poll
214 * notifications.
215 *
216 * The problem is that the poller ultimately holds a reference to this
217 * object, so it is very difficult to know when is safe to free it, at least
218 * without some expensive synchronization.
219 *
220 * If we keep the object freelisted, in the worst case losing this race just
221 * becomes a spurious read notification on a reused fd.
222 */
223
224/* The alarm system needs to be able to wakeup 'some poller' sometimes
225 * (specifically when a new alarm needs to be triggered earlier than the next
226 * alarm 'epoch'). This wakeup_fd gives us something to alert on when such a
227 * case occurs. */
228
229static grpc_fd *fd_freelist = NULL;
230static gpr_mu fd_freelist_mu;
231
Craig Tillerc67cc992017-04-27 10:15:51 -0700232static void fd_global_init(void) { gpr_mu_init(&fd_freelist_mu); }
233
234static void fd_global_shutdown(void) {
235 gpr_mu_lock(&fd_freelist_mu);
236 gpr_mu_unlock(&fd_freelist_mu);
237 while (fd_freelist != NULL) {
238 grpc_fd *fd = fd_freelist;
239 fd_freelist = fd_freelist->freelist_next;
Craig Tillerc67cc992017-04-27 10:15:51 -0700240 gpr_free(fd);
241 }
242 gpr_mu_destroy(&fd_freelist_mu);
243}
244
245static grpc_fd *fd_create(int fd, const char *name) {
246 grpc_fd *new_fd = NULL;
247
248 gpr_mu_lock(&fd_freelist_mu);
249 if (fd_freelist != NULL) {
250 new_fd = fd_freelist;
251 fd_freelist = fd_freelist->freelist_next;
252 }
253 gpr_mu_unlock(&fd_freelist_mu);
254
255 if (new_fd == NULL) {
256 new_fd = gpr_malloc(sizeof(grpc_fd));
Craig Tillerc67cc992017-04-27 10:15:51 -0700257 }
258
Craig Tillerc67cc992017-04-27 10:15:51 -0700259 new_fd->fd = fd;
Craig Tillerc67cc992017-04-27 10:15:51 -0700260 grpc_lfev_init(&new_fd->read_closure);
261 grpc_lfev_init(&new_fd->write_closure);
262 gpr_atm_no_barrier_store(&new_fd->read_notifier_pollset, (gpr_atm)NULL);
263
264 new_fd->freelist_next = NULL;
Craig Tillerc67cc992017-04-27 10:15:51 -0700265
266 char *fd_name;
267 gpr_asprintf(&fd_name, "%s fd=%d", name, fd);
268 grpc_iomgr_register_object(&new_fd->iomgr_object, fd_name);
Noah Eisen264879f2017-06-20 17:14:47 -0700269#ifndef NDEBUG
270 if (GRPC_TRACER_ON(grpc_trace_fd_refcount)) {
271 gpr_log(GPR_DEBUG, "FD %d %p create %s", fd, new_fd, fd_name);
272 }
Craig Tillerc67cc992017-04-27 10:15:51 -0700273#endif
274 gpr_free(fd_name);
Craig Tiller9ddb3152017-04-27 21:32:56 +0000275
276 struct epoll_event ev = {.events = (uint32_t)(EPOLLIN | EPOLLOUT | EPOLLET),
277 .data.ptr = new_fd};
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700278 if (epoll_ctl(g_epoll_set.epfd, EPOLL_CTL_ADD, fd, &ev) != 0) {
Craig Tiller9ddb3152017-04-27 21:32:56 +0000279 gpr_log(GPR_ERROR, "epoll_ctl failed: %s", strerror(errno));
280 }
281
Craig Tillerc67cc992017-04-27 10:15:51 -0700282 return new_fd;
283}
284
Craig Tiller4509c472017-04-27 19:05:13 +0000285static int fd_wrapped_fd(grpc_fd *fd) { return fd->fd; }
Craig Tillerc67cc992017-04-27 10:15:51 -0700286
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700287/* if 'releasing_fd' is true, it means that we are going to detach the internal
288 * fd from grpc_fd structure (i.e which means we should not be calling
289 * shutdown() syscall on that fd) */
290static void fd_shutdown_internal(grpc_exec_ctx *exec_ctx, grpc_fd *fd,
291 grpc_error *why, bool releasing_fd) {
Craig Tiller9ddb3152017-04-27 21:32:56 +0000292 if (grpc_lfev_set_shutdown(exec_ctx, &fd->read_closure,
293 GRPC_ERROR_REF(why))) {
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700294 if (!releasing_fd) {
295 shutdown(fd->fd, SHUT_RDWR);
296 }
Craig Tiller9ddb3152017-04-27 21:32:56 +0000297 grpc_lfev_set_shutdown(exec_ctx, &fd->write_closure, GRPC_ERROR_REF(why));
298 }
299 GRPC_ERROR_UNREF(why);
300}
301
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700302/* Might be called multiple times */
303static void fd_shutdown(grpc_exec_ctx *exec_ctx, grpc_fd *fd, grpc_error *why) {
304 fd_shutdown_internal(exec_ctx, fd, why, false);
305}
306
Craig Tillerc67cc992017-04-27 10:15:51 -0700307static void fd_orphan(grpc_exec_ctx *exec_ctx, grpc_fd *fd,
308 grpc_closure *on_done, int *release_fd,
Yuchen Zengd40a7ae2017-07-12 15:59:56 -0700309 bool already_closed, const char *reason) {
Craig Tillerc67cc992017-04-27 10:15:51 -0700310 grpc_error *error = GRPC_ERROR_NONE;
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700311 bool is_release_fd = (release_fd != NULL);
Craig Tillerc67cc992017-04-27 10:15:51 -0700312
Craig Tiller9ddb3152017-04-27 21:32:56 +0000313 if (!grpc_lfev_is_shutdown(&fd->read_closure)) {
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700314 fd_shutdown_internal(exec_ctx, fd,
315 GRPC_ERROR_CREATE_FROM_COPIED_STRING(reason),
316 is_release_fd);
Craig Tiller9ddb3152017-04-27 21:32:56 +0000317 }
318
Craig Tillerc67cc992017-04-27 10:15:51 -0700319 /* If release_fd is not NULL, we should be relinquishing control of the file
320 descriptor fd->fd (but we still own the grpc_fd structure). */
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700321 if (is_release_fd) {
Craig Tillerc67cc992017-04-27 10:15:51 -0700322 *release_fd = fd->fd;
Yuchen Zengd40a7ae2017-07-12 15:59:56 -0700323 } else if (!already_closed) {
Craig Tillerc67cc992017-04-27 10:15:51 -0700324 close(fd->fd);
Craig Tillerc67cc992017-04-27 10:15:51 -0700325 }
326
ncteisen274bbbe2017-06-08 14:57:11 -0700327 GRPC_CLOSURE_SCHED(exec_ctx, on_done, GRPC_ERROR_REF(error));
Craig Tillerc67cc992017-04-27 10:15:51 -0700328
Craig Tiller4509c472017-04-27 19:05:13 +0000329 grpc_iomgr_unregister_object(&fd->iomgr_object);
330 grpc_lfev_destroy(&fd->read_closure);
331 grpc_lfev_destroy(&fd->write_closure);
Craig Tillerc67cc992017-04-27 10:15:51 -0700332
Craig Tiller4509c472017-04-27 19:05:13 +0000333 gpr_mu_lock(&fd_freelist_mu);
334 fd->freelist_next = fd_freelist;
335 fd_freelist = fd;
336 gpr_mu_unlock(&fd_freelist_mu);
Craig Tillerc67cc992017-04-27 10:15:51 -0700337}
338
339static grpc_pollset *fd_get_read_notifier_pollset(grpc_exec_ctx *exec_ctx,
340 grpc_fd *fd) {
341 gpr_atm notifier = gpr_atm_acq_load(&fd->read_notifier_pollset);
342 return (grpc_pollset *)notifier;
343}
344
345static bool fd_is_shutdown(grpc_fd *fd) {
346 return grpc_lfev_is_shutdown(&fd->read_closure);
347}
348
Craig Tillerc67cc992017-04-27 10:15:51 -0700349static void fd_notify_on_read(grpc_exec_ctx *exec_ctx, grpc_fd *fd,
350 grpc_closure *closure) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700351 grpc_lfev_notify_on(exec_ctx, &fd->read_closure, closure, "read");
Craig Tillerc67cc992017-04-27 10:15:51 -0700352}
353
354static void fd_notify_on_write(grpc_exec_ctx *exec_ctx, grpc_fd *fd,
355 grpc_closure *closure) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700356 grpc_lfev_notify_on(exec_ctx, &fd->write_closure, closure, "write");
Craig Tillerc67cc992017-04-27 10:15:51 -0700357}
358
Craig Tiller4509c472017-04-27 19:05:13 +0000359static void fd_become_readable(grpc_exec_ctx *exec_ctx, grpc_fd *fd,
360 grpc_pollset *notifier) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700361 grpc_lfev_set_ready(exec_ctx, &fd->read_closure, "read");
Craig Tiller4509c472017-04-27 19:05:13 +0000362 /* Use release store to match with acquire load in fd_get_read_notifier */
363 gpr_atm_rel_store(&fd->read_notifier_pollset, (gpr_atm)notifier);
364}
365
366static void fd_become_writable(grpc_exec_ctx *exec_ctx, grpc_fd *fd) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700367 grpc_lfev_set_ready(exec_ctx, &fd->write_closure, "write");
Craig Tillerc67cc992017-04-27 10:15:51 -0700368}
369
370/*******************************************************************************
371 * Pollset Definitions
372 */
373
Craig Tiller6de05932017-04-28 09:17:38 -0700374GPR_TLS_DECL(g_current_thread_pollset);
375GPR_TLS_DECL(g_current_thread_worker);
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700376
377/* The designated poller */
Craig Tiller6de05932017-04-28 09:17:38 -0700378static gpr_atm g_active_poller;
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700379
Craig Tiller6de05932017-04-28 09:17:38 -0700380static pollset_neighbourhood *g_neighbourhoods;
Craig Tiller32f90ee2017-04-28 12:46:41 -0700381static size_t g_num_neighbourhoods;
Craig Tiller6de05932017-04-28 09:17:38 -0700382
Craig Tillerc67cc992017-04-27 10:15:51 -0700383/* Return true if first in list */
Craig Tiller32f90ee2017-04-28 12:46:41 -0700384static bool worker_insert(grpc_pollset *pollset, grpc_pollset_worker *worker) {
385 if (pollset->root_worker == NULL) {
386 pollset->root_worker = worker;
387 worker->next = worker->prev = worker;
Craig Tillerc67cc992017-04-27 10:15:51 -0700388 return true;
389 } else {
Craig Tiller32f90ee2017-04-28 12:46:41 -0700390 worker->next = pollset->root_worker;
391 worker->prev = worker->next->prev;
392 worker->next->prev = worker;
393 worker->prev->next = worker;
Craig Tillerc67cc992017-04-27 10:15:51 -0700394 return false;
395 }
396}
397
398/* Return true if last in list */
399typedef enum { EMPTIED, NEW_ROOT, REMOVED } worker_remove_result;
400
Craig Tiller32f90ee2017-04-28 12:46:41 -0700401static worker_remove_result worker_remove(grpc_pollset *pollset,
Craig Tillerc67cc992017-04-27 10:15:51 -0700402 grpc_pollset_worker *worker) {
Craig Tiller32f90ee2017-04-28 12:46:41 -0700403 if (worker == pollset->root_worker) {
404 if (worker == worker->next) {
405 pollset->root_worker = NULL;
Craig Tillerc67cc992017-04-27 10:15:51 -0700406 return EMPTIED;
407 } else {
Craig Tiller32f90ee2017-04-28 12:46:41 -0700408 pollset->root_worker = worker->next;
409 worker->prev->next = worker->next;
410 worker->next->prev = worker->prev;
Craig Tillerc67cc992017-04-27 10:15:51 -0700411 return NEW_ROOT;
412 }
413 } else {
Craig Tiller32f90ee2017-04-28 12:46:41 -0700414 worker->prev->next = worker->next;
415 worker->next->prev = worker->prev;
Craig Tillerc67cc992017-04-27 10:15:51 -0700416 return REMOVED;
417 }
418}
419
Craig Tillerba550da2017-05-01 14:26:31 +0000420static size_t choose_neighbourhood(void) {
421 return (size_t)gpr_cpu_current_cpu() % g_num_neighbourhoods;
422}
423
Craig Tiller4509c472017-04-27 19:05:13 +0000424static grpc_error *pollset_global_init(void) {
Craig Tiller4509c472017-04-27 19:05:13 +0000425 gpr_tls_init(&g_current_thread_pollset);
426 gpr_tls_init(&g_current_thread_worker);
Craig Tiller6de05932017-04-28 09:17:38 -0700427 gpr_atm_no_barrier_store(&g_active_poller, 0);
Craig Tiller375eb252017-04-27 23:29:12 +0000428 global_wakeup_fd.read_fd = -1;
429 grpc_error *err = grpc_wakeup_fd_init(&global_wakeup_fd);
430 if (err != GRPC_ERROR_NONE) return err;
Craig Tiller4509c472017-04-27 19:05:13 +0000431 struct epoll_event ev = {.events = (uint32_t)(EPOLLIN | EPOLLET),
432 .data.ptr = &global_wakeup_fd};
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700433 if (epoll_ctl(g_epoll_set.epfd, EPOLL_CTL_ADD, global_wakeup_fd.read_fd,
434 &ev) != 0) {
Craig Tiller4509c472017-04-27 19:05:13 +0000435 return GRPC_OS_ERROR(errno, "epoll_ctl");
436 }
Craig Tillerba550da2017-05-01 14:26:31 +0000437 g_num_neighbourhoods = GPR_CLAMP(gpr_cpu_num_cores(), 1, MAX_NEIGHBOURHOODS);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700438 g_neighbourhoods =
439 gpr_zalloc(sizeof(*g_neighbourhoods) * g_num_neighbourhoods);
440 for (size_t i = 0; i < g_num_neighbourhoods; i++) {
441 gpr_mu_init(&g_neighbourhoods[i].mu);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700442 }
Craig Tiller4509c472017-04-27 19:05:13 +0000443 return GRPC_ERROR_NONE;
444}
445
446static void pollset_global_shutdown(void) {
Craig Tiller4509c472017-04-27 19:05:13 +0000447 gpr_tls_destroy(&g_current_thread_pollset);
448 gpr_tls_destroy(&g_current_thread_worker);
Craig Tiller375eb252017-04-27 23:29:12 +0000449 if (global_wakeup_fd.read_fd != -1) grpc_wakeup_fd_destroy(&global_wakeup_fd);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700450 for (size_t i = 0; i < g_num_neighbourhoods; i++) {
451 gpr_mu_destroy(&g_neighbourhoods[i].mu);
452 }
453 gpr_free(g_neighbourhoods);
Craig Tiller4509c472017-04-27 19:05:13 +0000454}
455
456static void pollset_init(grpc_pollset *pollset, gpr_mu **mu) {
Craig Tiller6de05932017-04-28 09:17:38 -0700457 gpr_mu_init(&pollset->mu);
458 *mu = &pollset->mu;
Craig Tillerba550da2017-05-01 14:26:31 +0000459 pollset->neighbourhood = &g_neighbourhoods[choose_neighbourhood()];
Sree Kuchibhotla30882302017-08-16 13:46:52 -0700460 pollset->reassigning_neighbourhood = false;
461 pollset->root_worker = NULL;
462 pollset->kicked_without_poller = false;
Craig Tiller6de05932017-04-28 09:17:38 -0700463 pollset->seen_inactive = true;
Sree Kuchibhotla30882302017-08-16 13:46:52 -0700464 pollset->shutting_down = false;
465 pollset->shutdown_closure = NULL;
466 pollset->begin_refs = 0;
467 pollset->next = pollset->prev = NULL;
Craig Tiller6de05932017-04-28 09:17:38 -0700468}
469
Craig Tillerc6109852017-05-01 14:26:49 -0700470static void pollset_destroy(grpc_exec_ctx *exec_ctx, grpc_pollset *pollset) {
Craig Tillere00d7332017-05-01 15:43:51 +0000471 gpr_mu_lock(&pollset->mu);
Craig Tillerba550da2017-05-01 14:26:31 +0000472 if (!pollset->seen_inactive) {
Craig Tillere00d7332017-05-01 15:43:51 +0000473 pollset_neighbourhood *neighbourhood = pollset->neighbourhood;
474 gpr_mu_unlock(&pollset->mu);
Craig Tillera95bacf2017-05-01 12:51:24 -0700475 retry_lock_neighbourhood:
Craig Tillere00d7332017-05-01 15:43:51 +0000476 gpr_mu_lock(&neighbourhood->mu);
477 gpr_mu_lock(&pollset->mu);
478 if (!pollset->seen_inactive) {
479 if (pollset->neighbourhood != neighbourhood) {
480 gpr_mu_unlock(&neighbourhood->mu);
481 neighbourhood = pollset->neighbourhood;
482 gpr_mu_unlock(&pollset->mu);
483 goto retry_lock_neighbourhood;
484 }
485 pollset->prev->next = pollset->next;
486 pollset->next->prev = pollset->prev;
487 if (pollset == pollset->neighbourhood->active_root) {
488 pollset->neighbourhood->active_root =
489 pollset->next == pollset ? NULL : pollset->next;
490 }
Craig Tillerba550da2017-05-01 14:26:31 +0000491 }
492 gpr_mu_unlock(&pollset->neighbourhood->mu);
Craig Tiller6de05932017-04-28 09:17:38 -0700493 }
Craig Tillere00d7332017-05-01 15:43:51 +0000494 gpr_mu_unlock(&pollset->mu);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700495 gpr_mu_destroy(&pollset->mu);
Craig Tiller4509c472017-04-27 19:05:13 +0000496}
497
498static grpc_error *pollset_kick_all(grpc_pollset *pollset) {
yang-gdf92a642017-08-21 22:38:45 -0700499 GPR_TIMER_BEGIN("pollset_kick_all", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000500 grpc_error *error = GRPC_ERROR_NONE;
501 if (pollset->root_worker != NULL) {
502 grpc_pollset_worker *worker = pollset->root_worker;
503 do {
Craig Tiller55624a32017-05-26 08:14:44 -0700504 switch (worker->kick_state) {
505 case KICKED:
506 break;
507 case UNKICKED:
508 SET_KICK_STATE(worker, KICKED);
509 if (worker->initialized_cv) {
510 gpr_cv_signal(&worker->cv);
511 }
512 break;
513 case DESIGNATED_POLLER:
514 SET_KICK_STATE(worker, KICKED);
515 append_error(&error, grpc_wakeup_fd_wakeup(&global_wakeup_fd),
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700516 "pollset_kick_all");
Craig Tiller55624a32017-05-26 08:14:44 -0700517 break;
Craig Tiller4509c472017-04-27 19:05:13 +0000518 }
519
Craig Tiller32f90ee2017-04-28 12:46:41 -0700520 worker = worker->next;
Craig Tiller4509c472017-04-27 19:05:13 +0000521 } while (worker != pollset->root_worker);
522 }
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700523 // TODO: sreek. Check if we need to set 'kicked_without_poller' to true here
524 // in the else case
yang-gdf92a642017-08-21 22:38:45 -0700525 GPR_TIMER_END("pollset_kick_all", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000526 return error;
527}
528
529static void pollset_maybe_finish_shutdown(grpc_exec_ctx *exec_ctx,
530 grpc_pollset *pollset) {
Craig Tillerba550da2017-05-01 14:26:31 +0000531 if (pollset->shutdown_closure != NULL && pollset->root_worker == NULL &&
532 pollset->begin_refs == 0) {
yang-gdf92a642017-08-21 22:38:45 -0700533 GPR_TIMER_MARK("pollset_finish_shutdown", 0);
ncteisen274bbbe2017-06-08 14:57:11 -0700534 GRPC_CLOSURE_SCHED(exec_ctx, pollset->shutdown_closure, GRPC_ERROR_NONE);
Craig Tiller4509c472017-04-27 19:05:13 +0000535 pollset->shutdown_closure = NULL;
536 }
537}
538
539static void pollset_shutdown(grpc_exec_ctx *exec_ctx, grpc_pollset *pollset,
540 grpc_closure *closure) {
yang-gdf92a642017-08-21 22:38:45 -0700541 GPR_TIMER_BEGIN("pollset_shutdown", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000542 GPR_ASSERT(pollset->shutdown_closure == NULL);
Craig Tillerc81512a2017-05-26 09:53:58 -0700543 GPR_ASSERT(!pollset->shutting_down);
Craig Tiller4509c472017-04-27 19:05:13 +0000544 pollset->shutdown_closure = closure;
Craig Tillerc81512a2017-05-26 09:53:58 -0700545 pollset->shutting_down = true;
Craig Tiller4509c472017-04-27 19:05:13 +0000546 GRPC_LOG_IF_ERROR("pollset_shutdown", pollset_kick_all(pollset));
547 pollset_maybe_finish_shutdown(exec_ctx, pollset);
yang-gdf92a642017-08-21 22:38:45 -0700548 GPR_TIMER_END("pollset_shutdown", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000549}
550
Craig Tiller4509c472017-04-27 19:05:13 +0000551static int poll_deadline_to_millis_timeout(gpr_timespec deadline,
552 gpr_timespec now) {
553 gpr_timespec timeout;
554 if (gpr_time_cmp(deadline, gpr_inf_future(deadline.clock_type)) == 0) {
555 return -1;
556 }
557
558 if (gpr_time_cmp(deadline, now) <= 0) {
559 return 0;
560 }
561
562 static const gpr_timespec round_up = {
563 .clock_type = GPR_TIMESPAN, .tv_sec = 0, .tv_nsec = GPR_NS_PER_MS - 1};
564 timeout = gpr_time_sub(deadline, now);
565 int millis = gpr_time_to_millis(gpr_time_add(timeout, round_up));
566 return millis >= 1 ? millis : 1;
567}
568
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700569/* Process the epoll events found by do_epoll_wait() function.
570 - g_epoll_set.cursor points to the index of the first event to be processed
571 - This function then processes up-to MAX_EPOLL_EVENTS_PER_ITERATION and
572 updates the g_epoll_set.cursor
Craig Tiller4509c472017-04-27 19:05:13 +0000573
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700574 NOTE ON SYNCRHONIZATION: Similar to do_epoll_wait(), this function is only
575 called by g_active_poller thread. So there is no need for synchronization
576 when accessing fields in g_epoll_set */
577static grpc_error *process_epoll_events(grpc_exec_ctx *exec_ctx,
578 grpc_pollset *pollset) {
579 static const char *err_desc = "process_events";
Craig Tiller4509c472017-04-27 19:05:13 +0000580 grpc_error *error = GRPC_ERROR_NONE;
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700581
Sree Kuchibhotla3d609f12017-08-25 10:00:18 -0700582 GPR_TIMER_BEGIN("process_epoll_events", 0);
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700583 for (int idx = 0; (idx < MAX_EPOLL_EVENTS_HANDLED_PER_ITERATION) &&
584 g_epoll_set.cursor != g_epoll_set.num_events;
585 idx++) {
586 int c = g_epoll_set.cursor++;
587 struct epoll_event *ev = &g_epoll_set.events[c];
588 void *data_ptr = ev->data.ptr;
589
Craig Tiller4509c472017-04-27 19:05:13 +0000590 if (data_ptr == &global_wakeup_fd) {
Craig Tiller4509c472017-04-27 19:05:13 +0000591 append_error(&error, grpc_wakeup_fd_consume_wakeup(&global_wakeup_fd),
592 err_desc);
593 } else {
594 grpc_fd *fd = (grpc_fd *)(data_ptr);
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700595 bool cancel = (ev->events & (EPOLLERR | EPOLLHUP)) != 0;
596 bool read_ev = (ev->events & (EPOLLIN | EPOLLPRI)) != 0;
597 bool write_ev = (ev->events & EPOLLOUT) != 0;
598
Craig Tiller4509c472017-04-27 19:05:13 +0000599 if (read_ev || cancel) {
600 fd_become_readable(exec_ctx, fd, pollset);
601 }
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700602
Craig Tiller4509c472017-04-27 19:05:13 +0000603 if (write_ev || cancel) {
604 fd_become_writable(exec_ctx, fd);
605 }
606 }
607 }
Sree Kuchibhotla3d609f12017-08-25 10:00:18 -0700608 GPR_TIMER_END("process_epoll_events", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000609 return error;
610}
611
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700612/* Do epoll_wait and store the events in g_epoll_set.events field. This does not
613 "process" any of the events yet; that is done in process_epoll_events().
614 *See process_epoll_events() function for more details.
615
616 NOTE ON SYNCHRONIZATION: At any point of time, only the g_active_poller
617 (i.e the designated poller thread) will be calling this function. So there is
618 no need for any synchronization when accesing fields in g_epoll_set */
619static grpc_error *do_epoll_wait(grpc_exec_ctx *exec_ctx, grpc_pollset *ps,
620 gpr_timespec now, gpr_timespec deadline) {
Sree Kuchibhotla3d609f12017-08-25 10:00:18 -0700621 GPR_TIMER_BEGIN("do_epoll_wait", 0);
622
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700623 int r;
624 int timeout = poll_deadline_to_millis_timeout(deadline, now);
625 if (timeout != 0) {
626 GRPC_SCHEDULING_START_BLOCKING_REGION;
627 }
628 do {
629 r = epoll_wait(g_epoll_set.epfd, g_epoll_set.events, MAX_EPOLL_EVENTS,
630 timeout);
631 } while (r < 0 && errno == EINTR);
632 if (timeout != 0) {
633 GRPC_SCHEDULING_END_BLOCKING_REGION;
634 }
635
636 if (r < 0) return GRPC_OS_ERROR(errno, "epoll_wait");
637
638 if (GRPC_TRACER_ON(grpc_polling_trace)) {
639 gpr_log(GPR_DEBUG, "ps: %p poll got %d events", ps, r);
640 }
641
642 g_epoll_set.num_events = r;
643 g_epoll_set.cursor = 0;
644
Sree Kuchibhotla3d609f12017-08-25 10:00:18 -0700645 GPR_TIMER_END("do_epoll_wait", 0);
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700646 return GRPC_ERROR_NONE;
647}
648
Craig Tiller4509c472017-04-27 19:05:13 +0000649static bool begin_worker(grpc_pollset *pollset, grpc_pollset_worker *worker,
650 grpc_pollset_worker **worker_hdl, gpr_timespec *now,
651 gpr_timespec deadline) {
yang-gdf92a642017-08-21 22:38:45 -0700652 GPR_TIMER_BEGIN("begin_worker", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000653 if (worker_hdl != NULL) *worker_hdl = worker;
654 worker->initialized_cv = false;
Craig Tiller55624a32017-05-26 08:14:44 -0700655 SET_KICK_STATE(worker, UNKICKED);
Craig Tiller50da5ec2017-05-01 13:51:14 -0700656 worker->schedule_on_end_work = (grpc_closure_list)GRPC_CLOSURE_LIST_INIT;
Craig Tillerba550da2017-05-01 14:26:31 +0000657 pollset->begin_refs++;
Craig Tiller4509c472017-04-27 19:05:13 +0000658
Craig Tiller830e82a2017-05-31 16:26:27 -0700659 if (GRPC_TRACER_ON(grpc_polling_trace)) {
660 gpr_log(GPR_ERROR, "PS:%p BEGIN_STARTS:%p", pollset, worker);
661 }
662
Craig Tiller32f90ee2017-04-28 12:46:41 -0700663 if (pollset->seen_inactive) {
664 // pollset has been observed to be inactive, we need to move back to the
665 // active list
Craig Tillere00d7332017-05-01 15:43:51 +0000666 bool is_reassigning = false;
667 if (!pollset->reassigning_neighbourhood) {
668 is_reassigning = true;
669 pollset->reassigning_neighbourhood = true;
670 pollset->neighbourhood = &g_neighbourhoods[choose_neighbourhood()];
671 }
672 pollset_neighbourhood *neighbourhood = pollset->neighbourhood;
Craig Tiller32f90ee2017-04-28 12:46:41 -0700673 gpr_mu_unlock(&pollset->mu);
Craig Tillerba550da2017-05-01 14:26:31 +0000674 // pollset unlocked: state may change (even worker->kick_state)
675 retry_lock_neighbourhood:
Craig Tiller32f90ee2017-04-28 12:46:41 -0700676 gpr_mu_lock(&neighbourhood->mu);
677 gpr_mu_lock(&pollset->mu);
Craig Tiller830e82a2017-05-31 16:26:27 -0700678 if (GRPC_TRACER_ON(grpc_polling_trace)) {
679 gpr_log(GPR_ERROR, "PS:%p BEGIN_REORG:%p kick_state=%s is_reassigning=%d",
680 pollset, worker, kick_state_string(worker->kick_state),
681 is_reassigning);
682 }
Craig Tiller32f90ee2017-04-28 12:46:41 -0700683 if (pollset->seen_inactive) {
Craig Tiller2acab6e2017-04-30 23:06:33 +0000684 if (neighbourhood != pollset->neighbourhood) {
685 gpr_mu_unlock(&neighbourhood->mu);
686 neighbourhood = pollset->neighbourhood;
687 gpr_mu_unlock(&pollset->mu);
688 goto retry_lock_neighbourhood;
689 }
Craig Tiller32f90ee2017-04-28 12:46:41 -0700690 pollset->seen_inactive = false;
Craig Tiller2acab6e2017-04-30 23:06:33 +0000691 if (neighbourhood->active_root == NULL) {
692 neighbourhood->active_root = pollset->next = pollset->prev = pollset;
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700693 /* TODO: sreek. Why would this worker state be other than UNKICKED
694 * here ? (since the worker isn't added to the pollset yet, there is no
695 * way it can be "found" by other threads to get kicked). */
696
697 /* If there is no designated poller, make this the designated poller */
Craig Tiller55624a32017-05-26 08:14:44 -0700698 if (worker->kick_state == UNKICKED &&
699 gpr_atm_no_barrier_cas(&g_active_poller, 0, (gpr_atm)worker)) {
700 SET_KICK_STATE(worker, DESIGNATED_POLLER);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700701 }
Craig Tiller2acab6e2017-04-30 23:06:33 +0000702 } else {
703 pollset->next = neighbourhood->active_root;
704 pollset->prev = pollset->next->prev;
705 pollset->next->prev = pollset->prev->next = pollset;
Craig Tiller4509c472017-04-27 19:05:13 +0000706 }
707 }
Craig Tillere00d7332017-05-01 15:43:51 +0000708 if (is_reassigning) {
709 GPR_ASSERT(pollset->reassigning_neighbourhood);
710 pollset->reassigning_neighbourhood = false;
711 }
Craig Tiller32f90ee2017-04-28 12:46:41 -0700712 gpr_mu_unlock(&neighbourhood->mu);
713 }
Sree Kuchibhotlae6506bc2017-07-18 21:43:45 -0700714
Craig Tiller32f90ee2017-04-28 12:46:41 -0700715 worker_insert(pollset, worker);
Craig Tillerba550da2017-05-01 14:26:31 +0000716 pollset->begin_refs--;
Sree Kuchibhotla949d0752017-07-20 23:49:15 -0700717 if (worker->kick_state == UNKICKED && !pollset->kicked_without_poller) {
Craig Tillera4b8eb02017-04-29 00:13:52 +0000718 GPR_ASSERT(gpr_atm_no_barrier_load(&g_active_poller) != (gpr_atm)worker);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700719 worker->initialized_cv = true;
720 gpr_cv_init(&worker->cv);
Craig Tillerc81512a2017-05-26 09:53:58 -0700721 while (worker->kick_state == UNKICKED && !pollset->shutting_down) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700722 if (GRPC_TRACER_ON(grpc_polling_trace)) {
723 gpr_log(GPR_ERROR, "PS:%p BEGIN_WAIT:%p kick_state=%s shutdown=%d",
724 pollset, worker, kick_state_string(worker->kick_state),
725 pollset->shutting_down);
726 }
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700727
Craig Tiller32f90ee2017-04-28 12:46:41 -0700728 if (gpr_cv_wait(&worker->cv, &pollset->mu, deadline) &&
729 worker->kick_state == UNKICKED) {
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700730 /* If gpr_cv_wait returns true (i.e a timeout), pretend that the worker
731 received a kick */
Craig Tiller55624a32017-05-26 08:14:44 -0700732 SET_KICK_STATE(worker, KICKED);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700733 }
Craig Tillerba550da2017-05-01 14:26:31 +0000734 }
Craig Tiller4509c472017-04-27 19:05:13 +0000735 *now = gpr_now(now->clock_type);
736 }
Sree Kuchibhotla949d0752017-07-20 23:49:15 -0700737
Craig Tiller830e82a2017-05-31 16:26:27 -0700738 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Sree Kuchibhotla949d0752017-07-20 23:49:15 -0700739 gpr_log(GPR_ERROR,
740 "PS:%p BEGIN_DONE:%p kick_state=%s shutdown=%d "
741 "kicked_without_poller: %d",
742 pollset, worker, kick_state_string(worker->kick_state),
743 pollset->shutting_down, pollset->kicked_without_poller);
Craig Tiller830e82a2017-05-31 16:26:27 -0700744 }
Craig Tiller4509c472017-04-27 19:05:13 +0000745
Sree Kuchibhotlae6506bc2017-07-18 21:43:45 -0700746 /* We release pollset lock in this function at a couple of places:
Sree Kuchibhotlaa0616ef2017-07-18 23:49:49 -0700747 * 1. Briefly when assigning pollset to a neighbourhood
Sree Kuchibhotlae6506bc2017-07-18 21:43:45 -0700748 * 2. When doing gpr_cv_wait()
749 * It is possible that 'kicked_without_poller' was set to true during (1) and
750 * 'shutting_down' is set to true during (1) or (2). If either of them is
751 * true, this worker cannot do polling */
Sree Kuchibhotlae6506bc2017-07-18 21:43:45 -0700752 /* TODO(sreek): Perhaps there is a better way to handle kicked_without_poller
753 * case; especially when the worker is the DESIGNATED_POLLER */
754
Sree Kuchibhotlaa0616ef2017-07-18 23:49:49 -0700755 if (pollset->kicked_without_poller) {
756 pollset->kicked_without_poller = false;
yang-gdf92a642017-08-21 22:38:45 -0700757 GPR_TIMER_END("begin_worker", 0);
Sree Kuchibhotlaa0616ef2017-07-18 23:49:49 -0700758 return false;
759 }
760
yang-gdf92a642017-08-21 22:38:45 -0700761 GPR_TIMER_END("begin_worker", 0);
Sree Kuchibhotlaa0616ef2017-07-18 23:49:49 -0700762 return worker->kick_state == DESIGNATED_POLLER && !pollset->shutting_down;
Craig Tiller4509c472017-04-27 19:05:13 +0000763}
764
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700765static bool check_neighbourhood_for_available_poller(
Craig Tillera4b8eb02017-04-29 00:13:52 +0000766 pollset_neighbourhood *neighbourhood) {
yang-gdf92a642017-08-21 22:38:45 -0700767 GPR_TIMER_BEGIN("check_neighbourhood_for_available_poller", 0);
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700768 bool found_worker = false;
769 do {
770 grpc_pollset *inspect = neighbourhood->active_root;
771 if (inspect == NULL) {
772 break;
773 }
774 gpr_mu_lock(&inspect->mu);
775 GPR_ASSERT(!inspect->seen_inactive);
776 grpc_pollset_worker *inspect_worker = inspect->root_worker;
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700777 if (inspect_worker != NULL) {
Craig Tillera4b8eb02017-04-29 00:13:52 +0000778 do {
Craig Tillerba550da2017-05-01 14:26:31 +0000779 switch (inspect_worker->kick_state) {
780 case UNKICKED:
781 if (gpr_atm_no_barrier_cas(&g_active_poller, 0,
782 (gpr_atm)inspect_worker)) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700783 if (GRPC_TRACER_ON(grpc_polling_trace)) {
784 gpr_log(GPR_DEBUG, " .. choose next poller to be %p",
785 inspect_worker);
786 }
Craig Tiller55624a32017-05-26 08:14:44 -0700787 SET_KICK_STATE(inspect_worker, DESIGNATED_POLLER);
Craig Tillerba550da2017-05-01 14:26:31 +0000788 if (inspect_worker->initialized_cv) {
yang-gdf92a642017-08-21 22:38:45 -0700789 GPR_TIMER_MARK("signal worker", 0);
Craig Tillerba550da2017-05-01 14:26:31 +0000790 gpr_cv_signal(&inspect_worker->cv);
791 }
Craig Tiller830e82a2017-05-31 16:26:27 -0700792 } else {
793 if (GRPC_TRACER_ON(grpc_polling_trace)) {
794 gpr_log(GPR_DEBUG, " .. beaten to choose next poller");
795 }
Craig Tillera4b8eb02017-04-29 00:13:52 +0000796 }
Craig Tillerba550da2017-05-01 14:26:31 +0000797 // even if we didn't win the cas, there's a worker, we can stop
798 found_worker = true;
799 break;
800 case KICKED:
801 break;
802 case DESIGNATED_POLLER:
803 found_worker = true; // ok, so someone else found the worker, but
804 // we'll accept that
805 break;
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700806 }
Craig Tillera4b8eb02017-04-29 00:13:52 +0000807 inspect_worker = inspect_worker->next;
Craig Tiller830e82a2017-05-31 16:26:27 -0700808 } while (!found_worker && inspect_worker != inspect->root_worker);
Craig Tillera4b8eb02017-04-29 00:13:52 +0000809 }
810 if (!found_worker) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700811 if (GRPC_TRACER_ON(grpc_polling_trace)) {
812 gpr_log(GPR_DEBUG, " .. mark pollset %p inactive", inspect);
813 }
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700814 inspect->seen_inactive = true;
Craig Tiller2acab6e2017-04-30 23:06:33 +0000815 if (inspect == neighbourhood->active_root) {
Craig Tillera95bacf2017-05-01 12:51:24 -0700816 neighbourhood->active_root =
817 inspect->next == inspect ? NULL : inspect->next;
Craig Tiller2acab6e2017-04-30 23:06:33 +0000818 }
819 inspect->next->prev = inspect->prev;
820 inspect->prev->next = inspect->next;
Craig Tillere00d7332017-05-01 15:43:51 +0000821 inspect->next = inspect->prev = NULL;
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700822 }
823 gpr_mu_unlock(&inspect->mu);
824 } while (!found_worker);
yang-gdf92a642017-08-21 22:38:45 -0700825 GPR_TIMER_END("check_neighbourhood_for_available_poller", 0);
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700826 return found_worker;
827}
828
Craig Tiller4509c472017-04-27 19:05:13 +0000829static void end_worker(grpc_exec_ctx *exec_ctx, grpc_pollset *pollset,
830 grpc_pollset_worker *worker,
831 grpc_pollset_worker **worker_hdl) {
yang-gdf92a642017-08-21 22:38:45 -0700832 GPR_TIMER_BEGIN("end_worker", 0);
Craig Tiller830e82a2017-05-31 16:26:27 -0700833 if (GRPC_TRACER_ON(grpc_polling_trace)) {
834 gpr_log(GPR_DEBUG, "PS:%p END_WORKER:%p", pollset, worker);
835 }
Craig Tiller8502ecb2017-04-28 14:22:01 -0700836 if (worker_hdl != NULL) *worker_hdl = NULL;
Craig Tiller830e82a2017-05-31 16:26:27 -0700837 /* Make sure we appear kicked */
Craig Tiller55624a32017-05-26 08:14:44 -0700838 SET_KICK_STATE(worker, KICKED);
Craig Tiller50da5ec2017-05-01 13:51:14 -0700839 grpc_closure_list_move(&worker->schedule_on_end_work,
840 &exec_ctx->closure_list);
Craig Tiller8502ecb2017-04-28 14:22:01 -0700841 if (gpr_atm_no_barrier_load(&g_active_poller) == (gpr_atm)worker) {
Craig Tillera4b8eb02017-04-29 00:13:52 +0000842 if (worker->next != worker && worker->next->kick_state == UNKICKED) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700843 if (GRPC_TRACER_ON(grpc_polling_trace)) {
844 gpr_log(GPR_DEBUG, " .. choose next poller to be peer %p", worker);
845 }
Craig Tiller2acab6e2017-04-30 23:06:33 +0000846 GPR_ASSERT(worker->next->initialized_cv);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700847 gpr_atm_no_barrier_store(&g_active_poller, (gpr_atm)worker->next);
Craig Tiller55624a32017-05-26 08:14:44 -0700848 SET_KICK_STATE(worker->next, DESIGNATED_POLLER);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700849 gpr_cv_signal(&worker->next->cv);
Craig Tiller8502ecb2017-04-28 14:22:01 -0700850 if (grpc_exec_ctx_has_work(exec_ctx)) {
851 gpr_mu_unlock(&pollset->mu);
852 grpc_exec_ctx_flush(exec_ctx);
853 gpr_mu_lock(&pollset->mu);
854 }
Craig Tiller32f90ee2017-04-28 12:46:41 -0700855 } else {
856 gpr_atm_no_barrier_store(&g_active_poller, 0);
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700857 size_t poller_neighbourhood_idx =
858 (size_t)(pollset->neighbourhood - g_neighbourhoods);
Craig Tillerbb742672017-05-17 22:19:05 +0000859 gpr_mu_unlock(&pollset->mu);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700860 bool found_worker = false;
Craig Tillerba550da2017-05-01 14:26:31 +0000861 bool scan_state[MAX_NEIGHBOURHOODS];
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700862 for (size_t i = 0; !found_worker && i < g_num_neighbourhoods; i++) {
863 pollset_neighbourhood *neighbourhood =
864 &g_neighbourhoods[(poller_neighbourhood_idx + i) %
865 g_num_neighbourhoods];
866 if (gpr_mu_trylock(&neighbourhood->mu)) {
867 found_worker =
Craig Tillera4b8eb02017-04-29 00:13:52 +0000868 check_neighbourhood_for_available_poller(neighbourhood);
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700869 gpr_mu_unlock(&neighbourhood->mu);
Craig Tillerba550da2017-05-01 14:26:31 +0000870 scan_state[i] = true;
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700871 } else {
Craig Tillerba550da2017-05-01 14:26:31 +0000872 scan_state[i] = false;
Craig Tiller32f90ee2017-04-28 12:46:41 -0700873 }
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700874 }
Craig Tiller2acab6e2017-04-30 23:06:33 +0000875 for (size_t i = 0; !found_worker && i < g_num_neighbourhoods; i++) {
Craig Tillerba550da2017-05-01 14:26:31 +0000876 if (scan_state[i]) continue;
Craig Tiller2acab6e2017-04-30 23:06:33 +0000877 pollset_neighbourhood *neighbourhood =
878 &g_neighbourhoods[(poller_neighbourhood_idx + i) %
879 g_num_neighbourhoods];
880 gpr_mu_lock(&neighbourhood->mu);
Craig Tillerba550da2017-05-01 14:26:31 +0000881 found_worker = check_neighbourhood_for_available_poller(neighbourhood);
Craig Tiller2acab6e2017-04-30 23:06:33 +0000882 gpr_mu_unlock(&neighbourhood->mu);
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700883 }
Craig Tiller8502ecb2017-04-28 14:22:01 -0700884 grpc_exec_ctx_flush(exec_ctx);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700885 gpr_mu_lock(&pollset->mu);
886 }
Craig Tiller50da5ec2017-05-01 13:51:14 -0700887 } else if (grpc_exec_ctx_has_work(exec_ctx)) {
888 gpr_mu_unlock(&pollset->mu);
889 grpc_exec_ctx_flush(exec_ctx);
890 gpr_mu_lock(&pollset->mu);
Craig Tiller4509c472017-04-27 19:05:13 +0000891 }
892 if (worker->initialized_cv) {
893 gpr_cv_destroy(&worker->cv);
894 }
Craig Tiller830e82a2017-05-31 16:26:27 -0700895 if (GRPC_TRACER_ON(grpc_polling_trace)) {
896 gpr_log(GPR_DEBUG, " .. remove worker");
897 }
Craig Tiller32f90ee2017-04-28 12:46:41 -0700898 if (EMPTIED == worker_remove(pollset, worker)) {
Craig Tiller4509c472017-04-27 19:05:13 +0000899 pollset_maybe_finish_shutdown(exec_ctx, pollset);
900 }
Craig Tillera4b8eb02017-04-29 00:13:52 +0000901 GPR_ASSERT(gpr_atm_no_barrier_load(&g_active_poller) != (gpr_atm)worker);
yang-gdf92a642017-08-21 22:38:45 -0700902 GPR_TIMER_END("end_worker", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000903}
904
905/* pollset->po.mu lock must be held by the caller before calling this.
906 The function pollset_work() may temporarily release the lock (pollset->po.mu)
907 during the course of its execution but it will always re-acquire the lock and
908 ensure that it is held by the time the function returns */
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700909static grpc_error *pollset_work(grpc_exec_ctx *exec_ctx, grpc_pollset *ps,
Craig Tiller4509c472017-04-27 19:05:13 +0000910 grpc_pollset_worker **worker_hdl,
911 gpr_timespec now, gpr_timespec deadline) {
912 grpc_pollset_worker worker;
913 grpc_error *error = GRPC_ERROR_NONE;
914 static const char *err_desc = "pollset_work";
yang-gdf92a642017-08-21 22:38:45 -0700915 GPR_TIMER_BEGIN("pollset_work", 0);
Sree Kuchibhotlab154cd12017-08-25 10:33:41 -0700916 if (ps->kicked_without_poller) {
917 ps->kicked_without_poller = false;
yang-gdf92a642017-08-21 22:38:45 -0700918 GPR_TIMER_END("pollset_work", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000919 return GRPC_ERROR_NONE;
920 }
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700921
922 if (begin_worker(ps, &worker, worker_hdl, &now, deadline)) {
923 gpr_tls_set(&g_current_thread_pollset, (intptr_t)ps);
Craig Tiller4509c472017-04-27 19:05:13 +0000924 gpr_tls_set(&g_current_thread_worker, (intptr_t)&worker);
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700925 GPR_ASSERT(!ps->shutting_down);
926 GPR_ASSERT(!ps->seen_inactive);
927
928 gpr_mu_unlock(&ps->mu); /* unlock */
929
930 /* This is the designated polling thread at this point and should ideally do
931 polling. However, if there are unprocessed events left from a previous
932 call to do_epoll_wait(), skip calling epoll_wait() in this iteration and
933 process the pending epoll events.
934
935 The reason for decoupling do_epoll_wait and process_epoll_events is to
936 better distrubute the work (i.e handling epoll events) across multiple
937 threads
938
939 process_epoll_events() returns very quickly: It just queues the work on
940 exec_ctx but does not execute it (the actual exectution or more
941 accurately grpc_exec_ctx_flush() happens in end_worker() AFTER selecting
942 a designated poller). So we are not waiting long periods without a
943 designated poller */
944 if (g_epoll_set.cursor == g_epoll_set.num_events) {
945 append_error(&error, do_epoll_wait(exec_ctx, ps, now, deadline),
946 err_desc);
947 }
948 append_error(&error, process_epoll_events(exec_ctx, ps), err_desc);
949
950 gpr_mu_lock(&ps->mu); /* lock */
951
Craig Tiller4509c472017-04-27 19:05:13 +0000952 gpr_tls_set(&g_current_thread_worker, 0);
Craig Tiller830e82a2017-05-31 16:26:27 -0700953 } else {
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700954 gpr_tls_set(&g_current_thread_pollset, (intptr_t)ps);
Craig Tiller4509c472017-04-27 19:05:13 +0000955 }
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700956 end_worker(exec_ctx, ps, &worker, worker_hdl);
957
Craig Tiller8502ecb2017-04-28 14:22:01 -0700958 gpr_tls_set(&g_current_thread_pollset, 0);
yang-gdf92a642017-08-21 22:38:45 -0700959 GPR_TIMER_END("pollset_work", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000960 return error;
961}
962
963static grpc_error *pollset_kick(grpc_pollset *pollset,
964 grpc_pollset_worker *specific_worker) {
yang-gdf92a642017-08-21 22:38:45 -0700965 GPR_TIMER_BEGIN("pollset_kick", 0);
966 grpc_error *ret_err = GRPC_ERROR_NONE;
Craig Tillerb89bac02017-05-26 15:20:32 +0000967 if (GRPC_TRACER_ON(grpc_polling_trace)) {
968 gpr_strvec log;
969 gpr_strvec_init(&log);
970 char *tmp;
Craig Tiller75aef7f2017-05-26 08:26:08 -0700971 gpr_asprintf(
972 &tmp, "PS:%p KICK:%p curps=%p curworker=%p root=%p", pollset,
973 specific_worker, (void *)gpr_tls_get(&g_current_thread_pollset),
974 (void *)gpr_tls_get(&g_current_thread_worker), pollset->root_worker);
Craig Tillerb89bac02017-05-26 15:20:32 +0000975 gpr_strvec_add(&log, tmp);
976 if (pollset->root_worker != NULL) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700977 gpr_asprintf(&tmp, " {kick_state=%s next=%p {kick_state=%s}}",
978 kick_state_string(pollset->root_worker->kick_state),
979 pollset->root_worker->next,
980 kick_state_string(pollset->root_worker->next->kick_state));
Craig Tillerb89bac02017-05-26 15:20:32 +0000981 gpr_strvec_add(&log, tmp);
982 }
983 if (specific_worker != NULL) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700984 gpr_asprintf(&tmp, " worker_kick_state=%s",
985 kick_state_string(specific_worker->kick_state));
Craig Tillerb89bac02017-05-26 15:20:32 +0000986 gpr_strvec_add(&log, tmp);
987 }
988 tmp = gpr_strvec_flatten(&log, NULL);
989 gpr_strvec_destroy(&log);
Craig Tiller830e82a2017-05-31 16:26:27 -0700990 gpr_log(GPR_ERROR, "%s", tmp);
Craig Tillerb89bac02017-05-26 15:20:32 +0000991 gpr_free(tmp);
992 }
Craig Tiller4509c472017-04-27 19:05:13 +0000993 if (specific_worker == NULL) {
994 if (gpr_tls_get(&g_current_thread_pollset) != (intptr_t)pollset) {
Craig Tiller375eb252017-04-27 23:29:12 +0000995 grpc_pollset_worker *root_worker = pollset->root_worker;
996 if (root_worker == NULL) {
Craig Tiller4509c472017-04-27 19:05:13 +0000997 pollset->kicked_without_poller = true;
Craig Tiller75aef7f2017-05-26 08:26:08 -0700998 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700999 gpr_log(GPR_ERROR, " .. kicked_without_poller");
Craig Tiller75aef7f2017-05-26 08:26:08 -07001000 }
yang-gdf92a642017-08-21 22:38:45 -07001001 goto done;
Craig Tiller375eb252017-04-27 23:29:12 +00001002 }
Craig Tiller32f90ee2017-04-28 12:46:41 -07001003 grpc_pollset_worker *next_worker = root_worker->next;
Craig Tiller830e82a2017-05-31 16:26:27 -07001004 if (root_worker->kick_state == KICKED) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001005 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001006 gpr_log(GPR_ERROR, " .. already kicked %p", root_worker);
1007 }
1008 SET_KICK_STATE(root_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001009 goto done;
Craig Tiller830e82a2017-05-31 16:26:27 -07001010 } else if (next_worker->kick_state == KICKED) {
1011 if (GRPC_TRACER_ON(grpc_polling_trace)) {
1012 gpr_log(GPR_ERROR, " .. already kicked %p", next_worker);
1013 }
1014 SET_KICK_STATE(next_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001015 goto done;
Craig Tiller830e82a2017-05-31 16:26:27 -07001016 } else if (root_worker ==
1017 next_worker && // only try and wake up a poller if
1018 // there is no next worker
1019 root_worker == (grpc_pollset_worker *)gpr_atm_no_barrier_load(
1020 &g_active_poller)) {
1021 if (GRPC_TRACER_ON(grpc_polling_trace)) {
1022 gpr_log(GPR_ERROR, " .. kicked %p", root_worker);
Craig Tiller75aef7f2017-05-26 08:26:08 -07001023 }
Craig Tiller55624a32017-05-26 08:14:44 -07001024 SET_KICK_STATE(root_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001025 ret_err = grpc_wakeup_fd_wakeup(&global_wakeup_fd);
1026 goto done;
Craig Tiller8502ecb2017-04-28 14:22:01 -07001027 } else if (next_worker->kick_state == UNKICKED) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001028 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001029 gpr_log(GPR_ERROR, " .. kicked %p", next_worker);
Craig Tiller75aef7f2017-05-26 08:26:08 -07001030 }
Craig Tiller8502ecb2017-04-28 14:22:01 -07001031 GPR_ASSERT(next_worker->initialized_cv);
Craig Tiller55624a32017-05-26 08:14:44 -07001032 SET_KICK_STATE(next_worker, KICKED);
Craig Tiller375eb252017-04-27 23:29:12 +00001033 gpr_cv_signal(&next_worker->cv);
yang-gdf92a642017-08-21 22:38:45 -07001034 goto done;
Craig Tiller55624a32017-05-26 08:14:44 -07001035 } else if (next_worker->kick_state == DESIGNATED_POLLER) {
1036 if (root_worker->kick_state != DESIGNATED_POLLER) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001037 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001038 gpr_log(
1039 GPR_ERROR,
1040 " .. kicked root non-poller %p (initialized_cv=%d) (poller=%p)",
1041 root_worker, root_worker->initialized_cv, next_worker);
Craig Tiller75aef7f2017-05-26 08:26:08 -07001042 }
Craig Tiller55624a32017-05-26 08:14:44 -07001043 SET_KICK_STATE(root_worker, KICKED);
1044 if (root_worker->initialized_cv) {
1045 gpr_cv_signal(&root_worker->cv);
1046 }
yang-gdf92a642017-08-21 22:38:45 -07001047 goto done;
Craig Tiller55624a32017-05-26 08:14:44 -07001048 } else {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001049 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001050 gpr_log(GPR_ERROR, " .. non-root poller %p (root=%p)", next_worker,
Craig Tiller75aef7f2017-05-26 08:26:08 -07001051 root_worker);
1052 }
Craig Tiller55624a32017-05-26 08:14:44 -07001053 SET_KICK_STATE(next_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001054 ret_err = grpc_wakeup_fd_wakeup(&global_wakeup_fd);
1055 goto done;
Craig Tiller55624a32017-05-26 08:14:44 -07001056 }
Craig Tiller8502ecb2017-04-28 14:22:01 -07001057 } else {
Craig Tiller55624a32017-05-26 08:14:44 -07001058 GPR_ASSERT(next_worker->kick_state == KICKED);
1059 SET_KICK_STATE(next_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001060 goto done;
Craig Tiller4509c472017-04-27 19:05:13 +00001061 }
1062 } else {
Craig Tiller830e82a2017-05-31 16:26:27 -07001063 if (GRPC_TRACER_ON(grpc_polling_trace)) {
1064 gpr_log(GPR_ERROR, " .. kicked while waking up");
1065 }
yang-gdf92a642017-08-21 22:38:45 -07001066 goto done;
Craig Tiller4509c472017-04-27 19:05:13 +00001067 }
Craig Tiller43bf2592017-04-28 23:21:01 +00001068 } else if (specific_worker->kick_state == KICKED) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001069 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001070 gpr_log(GPR_ERROR, " .. specific worker already kicked");
Craig Tiller75aef7f2017-05-26 08:26:08 -07001071 }
yang-gdf92a642017-08-21 22:38:45 -07001072 goto done;
Craig Tiller4509c472017-04-27 19:05:13 +00001073 } else if (gpr_tls_get(&g_current_thread_worker) ==
1074 (intptr_t)specific_worker) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001075 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001076 gpr_log(GPR_ERROR, " .. mark %p kicked", specific_worker);
Craig Tiller75aef7f2017-05-26 08:26:08 -07001077 }
Craig Tiller55624a32017-05-26 08:14:44 -07001078 SET_KICK_STATE(specific_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001079 goto done;
Craig Tiller32f90ee2017-04-28 12:46:41 -07001080 } else if (specific_worker ==
1081 (grpc_pollset_worker *)gpr_atm_no_barrier_load(&g_active_poller)) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001082 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001083 gpr_log(GPR_ERROR, " .. kick active poller");
Craig Tiller75aef7f2017-05-26 08:26:08 -07001084 }
Craig Tiller55624a32017-05-26 08:14:44 -07001085 SET_KICK_STATE(specific_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001086 ret_err = grpc_wakeup_fd_wakeup(&global_wakeup_fd);
1087 goto done;
Craig Tiller8502ecb2017-04-28 14:22:01 -07001088 } else if (specific_worker->initialized_cv) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001089 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001090 gpr_log(GPR_ERROR, " .. kick waiting worker");
Craig Tiller75aef7f2017-05-26 08:26:08 -07001091 }
Craig Tiller55624a32017-05-26 08:14:44 -07001092 SET_KICK_STATE(specific_worker, KICKED);
Craig Tiller4509c472017-04-27 19:05:13 +00001093 gpr_cv_signal(&specific_worker->cv);
yang-gdf92a642017-08-21 22:38:45 -07001094 goto done;
Craig Tiller8502ecb2017-04-28 14:22:01 -07001095 } else {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001096 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001097 gpr_log(GPR_ERROR, " .. kick non-waiting worker");
Craig Tiller75aef7f2017-05-26 08:26:08 -07001098 }
Craig Tiller55624a32017-05-26 08:14:44 -07001099 SET_KICK_STATE(specific_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001100 goto done;
Craig Tiller4509c472017-04-27 19:05:13 +00001101 }
yang-gdf92a642017-08-21 22:38:45 -07001102done:
1103 GPR_TIMER_END("pollset_kick", 0);
1104 return ret_err;
Craig Tiller4509c472017-04-27 19:05:13 +00001105}
1106
1107static void pollset_add_fd(grpc_exec_ctx *exec_ctx, grpc_pollset *pollset,
1108 grpc_fd *fd) {}
1109
Craig Tiller4509c472017-04-27 19:05:13 +00001110/*******************************************************************************
Craig Tillerc67cc992017-04-27 10:15:51 -07001111 * Pollset-set Definitions
1112 */
1113
1114static grpc_pollset_set *pollset_set_create(void) {
1115 return (grpc_pollset_set *)((intptr_t)0xdeafbeef);
1116}
1117
1118static void pollset_set_destroy(grpc_exec_ctx *exec_ctx,
1119 grpc_pollset_set *pss) {}
1120
1121static void pollset_set_add_fd(grpc_exec_ctx *exec_ctx, grpc_pollset_set *pss,
1122 grpc_fd *fd) {}
1123
1124static void pollset_set_del_fd(grpc_exec_ctx *exec_ctx, grpc_pollset_set *pss,
1125 grpc_fd *fd) {}
1126
1127static void pollset_set_add_pollset(grpc_exec_ctx *exec_ctx,
1128 grpc_pollset_set *pss, grpc_pollset *ps) {}
1129
1130static void pollset_set_del_pollset(grpc_exec_ctx *exec_ctx,
1131 grpc_pollset_set *pss, grpc_pollset *ps) {}
1132
1133static void pollset_set_add_pollset_set(grpc_exec_ctx *exec_ctx,
1134 grpc_pollset_set *bag,
1135 grpc_pollset_set *item) {}
1136
1137static void pollset_set_del_pollset_set(grpc_exec_ctx *exec_ctx,
1138 grpc_pollset_set *bag,
1139 grpc_pollset_set *item) {}
1140
1141/*******************************************************************************
1142 * Event engine binding
1143 */
1144
1145static void shutdown_engine(void) {
1146 fd_global_shutdown();
1147 pollset_global_shutdown();
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -07001148 epoll_set_shutdown();
Craig Tillerc67cc992017-04-27 10:15:51 -07001149}
1150
1151static const grpc_event_engine_vtable vtable = {
1152 .pollset_size = sizeof(grpc_pollset),
1153
1154 .fd_create = fd_create,
1155 .fd_wrapped_fd = fd_wrapped_fd,
1156 .fd_orphan = fd_orphan,
1157 .fd_shutdown = fd_shutdown,
1158 .fd_is_shutdown = fd_is_shutdown,
1159 .fd_notify_on_read = fd_notify_on_read,
1160 .fd_notify_on_write = fd_notify_on_write,
1161 .fd_get_read_notifier_pollset = fd_get_read_notifier_pollset,
Craig Tillerc67cc992017-04-27 10:15:51 -07001162
1163 .pollset_init = pollset_init,
1164 .pollset_shutdown = pollset_shutdown,
1165 .pollset_destroy = pollset_destroy,
1166 .pollset_work = pollset_work,
1167 .pollset_kick = pollset_kick,
1168 .pollset_add_fd = pollset_add_fd,
1169
1170 .pollset_set_create = pollset_set_create,
1171 .pollset_set_destroy = pollset_set_destroy,
1172 .pollset_set_add_pollset = pollset_set_add_pollset,
1173 .pollset_set_del_pollset = pollset_set_del_pollset,
1174 .pollset_set_add_pollset_set = pollset_set_add_pollset_set,
1175 .pollset_set_del_pollset_set = pollset_set_del_pollset_set,
1176 .pollset_set_add_fd = pollset_set_add_fd,
1177 .pollset_set_del_fd = pollset_set_del_fd,
1178
Craig Tillerc67cc992017-04-27 10:15:51 -07001179 .shutdown_engine = shutdown_engine,
1180};
1181
1182/* It is possible that GLIBC has epoll but the underlying kernel doesn't.
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -07001183 * Create epoll_fd (epoll_set_init() takes care of that) to make sure epoll
1184 * support is available */
Craig Tiller6f0af492017-04-27 19:26:16 +00001185const grpc_event_engine_vtable *grpc_init_epoll1_linux(bool explicit_request) {
Sree Kuchibhotla107840f2017-07-24 18:45:09 -07001186 if (!explicit_request) {
1187 return NULL;
1188 }
1189
Craig Tillerc67cc992017-04-27 10:15:51 -07001190 if (!grpc_has_wakeup_fd()) {
1191 return NULL;
1192 }
1193
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -07001194 if (!epoll_set_init()) {
Craig Tillerc67cc992017-04-27 10:15:51 -07001195 return NULL;
1196 }
1197
Craig Tillerc67cc992017-04-27 10:15:51 -07001198 fd_global_init();
1199
1200 if (!GRPC_LOG_IF_ERROR("pollset_global_init", pollset_global_init())) {
Craig Tiller4509c472017-04-27 19:05:13 +00001201 fd_global_shutdown();
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -07001202 epoll_set_shutdown();
Craig Tillerc67cc992017-04-27 10:15:51 -07001203 return NULL;
1204 }
1205
1206 return &vtable;
1207}
1208
1209#else /* defined(GRPC_LINUX_EPOLL) */
1210#if defined(GRPC_POSIX_SOCKET)
1211#include "src/core/lib/iomgr/ev_posix.h"
1212/* If GRPC_LINUX_EPOLL is not defined, it means epoll is not available. Return
1213 * NULL */
Craig Tiller9ddb3152017-04-27 21:32:56 +00001214const grpc_event_engine_vtable *grpc_init_epoll1_linux(bool explicit_request) {
1215 return NULL;
1216}
Craig Tillerc67cc992017-04-27 10:15:51 -07001217#endif /* defined(GRPC_POSIX_SOCKET) */
1218#endif /* !defined(GRPC_LINUX_EPOLL) */