blob: b940d48ba9eab78f8327cd2077faaefc331d5134 [file] [log] [blame]
Craig Tillerc67cc992017-04-27 10:15:51 -07001/*
2 *
Jan Tattermusch7897ae92017-06-07 22:57:36 +02003 * Copyright 2017 gRPC authors.
Craig Tillerc67cc992017-04-27 10:15:51 -07004 *
Jan Tattermusch7897ae92017-06-07 22:57:36 +02005 * Licensed under the Apache License, Version 2.0 (the "License");
6 * you may not use this file except in compliance with the License.
7 * You may obtain a copy of the License at
Craig Tillerc67cc992017-04-27 10:15:51 -07008 *
Jan Tattermusch7897ae92017-06-07 22:57:36 +02009 * http://www.apache.org/licenses/LICENSE-2.0
Craig Tillerc67cc992017-04-27 10:15:51 -070010 *
Jan Tattermusch7897ae92017-06-07 22:57:36 +020011 * Unless required by applicable law or agreed to in writing, software
12 * distributed under the License is distributed on an "AS IS" BASIS,
13 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14 * See the License for the specific language governing permissions and
15 * limitations under the License.
Craig Tillerc67cc992017-04-27 10:15:51 -070016 *
17 */
18
19#include "src/core/lib/iomgr/port.h"
20
21/* This polling engine is only relevant on linux kernels supporting epoll() */
22#ifdef GRPC_LINUX_EPOLL
23
Craig Tiller4509c472017-04-27 19:05:13 +000024#include "src/core/lib/iomgr/ev_epoll1_linux.h"
Craig Tillerc67cc992017-04-27 10:15:51 -070025
26#include <assert.h>
27#include <errno.h>
28#include <poll.h>
29#include <pthread.h>
30#include <string.h>
31#include <sys/epoll.h>
32#include <sys/socket.h>
33#include <unistd.h>
34
35#include <grpc/support/alloc.h>
Craig Tiller6de05932017-04-28 09:17:38 -070036#include <grpc/support/cpu.h>
Craig Tillerc67cc992017-04-27 10:15:51 -070037#include <grpc/support/log.h>
38#include <grpc/support/string_util.h>
39#include <grpc/support/tls.h>
40#include <grpc/support/useful.h>
41
42#include "src/core/lib/iomgr/ev_posix.h"
43#include "src/core/lib/iomgr/iomgr_internal.h"
44#include "src/core/lib/iomgr/lockfree_event.h"
Craig Tillerc67cc992017-04-27 10:15:51 -070045#include "src/core/lib/iomgr/wakeup_fd_posix.h"
Craig Tillerc67cc992017-04-27 10:15:51 -070046#include "src/core/lib/profiling/timers.h"
47#include "src/core/lib/support/block_annotate.h"
Craig Tillerb89bac02017-05-26 15:20:32 +000048#include "src/core/lib/support/string.h"
Craig Tillerc67cc992017-04-27 10:15:51 -070049
Craig Tillerc67cc992017-04-27 10:15:51 -070050static grpc_wakeup_fd global_wakeup_fd;
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -070051
52/*******************************************************************************
53 * Singleton epoll set related fields
54 */
55
56#define MAX_EPOLL_EVENTS 100
Sree Kuchibhotla19614522017-08-25 17:10:10 -070057#define MAX_EPOLL_EVENTS_HANDLED_PER_ITERATION 1
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -070058
Sree Kuchibhotlae01940f2017-08-27 18:10:12 -070059/* NOTE ON SYNCHRONIZATION:
60 * - Fields in this struct are only modified by the designated poller. Hence
61 * there is no need for any locks to protect the struct.
62 * - num_events and cursor fields have to be of atomic type to provide memory
63 * visibility guarantees only. i.e In case of multiple pollers, the designated
64 * polling thread keeps changing; the thread that wrote these values may be
65 * different from the thread reading the values
66 */
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -070067typedef struct epoll_set {
68 int epfd;
69
70 /* The epoll_events after the last call to epoll_wait() */
71 struct epoll_event events[MAX_EPOLL_EVENTS];
72
73 /* The number of epoll_events after the last call to epoll_wait() */
Sree Kuchibhotlaa92a9cc2017-08-27 14:02:15 -070074 gpr_atm num_events;
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -070075
76 /* Index of the first event in epoll_events that has to be processed. This
77 * field is only valid if num_events > 0 */
Sree Kuchibhotlaa92a9cc2017-08-27 14:02:15 -070078 gpr_atm cursor;
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -070079} epoll_set;
80
81/* The global singleton epoll set */
82static epoll_set g_epoll_set;
83
84/* Must be called *only* once */
85static bool epoll_set_init() {
86 g_epoll_set.epfd = epoll_create1(EPOLL_CLOEXEC);
87 if (g_epoll_set.epfd < 0) {
88 gpr_log(GPR_ERROR, "epoll unavailable");
89 return false;
90 }
91
Sree Kuchibhotlaa92a9cc2017-08-27 14:02:15 -070092 gpr_log(GPR_INFO, "grpc epoll fd: %d", g_epoll_set.epfd);
93 gpr_atm_no_barrier_store(&g_epoll_set.num_events, 0);
94 gpr_atm_no_barrier_store(&g_epoll_set.cursor, 0);
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -070095 return true;
96}
97
98/* epoll_set_init() MUST be called before calling this. */
99static void epoll_set_shutdown() {
100 if (g_epoll_set.epfd >= 0) {
101 close(g_epoll_set.epfd);
102 g_epoll_set.epfd = -1;
103 }
104}
Craig Tillerc67cc992017-04-27 10:15:51 -0700105
106/*******************************************************************************
107 * Fd Declarations
108 */
109
110struct grpc_fd {
111 int fd;
112
Craig Tillerc67cc992017-04-27 10:15:51 -0700113 gpr_atm read_closure;
114 gpr_atm write_closure;
115
116 struct grpc_fd *freelist_next;
Craig Tillerc67cc992017-04-27 10:15:51 -0700117
118 /* The pollset that last noticed that the fd is readable. The actual type
119 * stored in this is (grpc_pollset *) */
120 gpr_atm read_notifier_pollset;
121
122 grpc_iomgr_object iomgr_object;
123};
124
125static void fd_global_init(void);
126static void fd_global_shutdown(void);
127
128/*******************************************************************************
129 * Pollset Declarations
130 */
131
Craig Tiller43bf2592017-04-28 23:21:01 +0000132typedef enum { UNKICKED, KICKED, DESIGNATED_POLLER } kick_state;
Craig Tillerc67cc992017-04-27 10:15:51 -0700133
Craig Tiller830e82a2017-05-31 16:26:27 -0700134static const char *kick_state_string(kick_state st) {
135 switch (st) {
136 case UNKICKED:
137 return "UNKICKED";
138 case KICKED:
139 return "KICKED";
140 case DESIGNATED_POLLER:
141 return "DESIGNATED_POLLER";
142 }
143 GPR_UNREACHABLE_CODE(return "UNKNOWN");
144}
145
Craig Tillerc67cc992017-04-27 10:15:51 -0700146struct grpc_pollset_worker {
Craig Tiller32f90ee2017-04-28 12:46:41 -0700147 kick_state kick_state;
Craig Tiller55624a32017-05-26 08:14:44 -0700148 int kick_state_mutator; // which line of code last changed kick state
Craig Tillerc67cc992017-04-27 10:15:51 -0700149 bool initialized_cv;
Craig Tiller32f90ee2017-04-28 12:46:41 -0700150 grpc_pollset_worker *next;
151 grpc_pollset_worker *prev;
Craig Tillerc67cc992017-04-27 10:15:51 -0700152 gpr_cv cv;
Craig Tiller50da5ec2017-05-01 13:51:14 -0700153 grpc_closure_list schedule_on_end_work;
Craig Tillerc67cc992017-04-27 10:15:51 -0700154};
155
Craig Tiller55624a32017-05-26 08:14:44 -0700156#define SET_KICK_STATE(worker, state) \
157 do { \
158 (worker)->kick_state = (state); \
159 (worker)->kick_state_mutator = __LINE__; \
160 } while (false)
161
Craig Tillerba550da2017-05-01 14:26:31 +0000162#define MAX_NEIGHBOURHOODS 1024
163
Craig Tiller6de05932017-04-28 09:17:38 -0700164typedef struct pollset_neighbourhood {
165 gpr_mu mu;
166 grpc_pollset *active_root;
Craig Tiller6de05932017-04-28 09:17:38 -0700167 char pad[GPR_CACHELINE_SIZE];
168} pollset_neighbourhood;
169
Craig Tillerc67cc992017-04-27 10:15:51 -0700170struct grpc_pollset {
Craig Tiller6de05932017-04-28 09:17:38 -0700171 gpr_mu mu;
172 pollset_neighbourhood *neighbourhood;
Craig Tillere00d7332017-05-01 15:43:51 +0000173 bool reassigning_neighbourhood;
Craig Tiller4509c472017-04-27 19:05:13 +0000174 grpc_pollset_worker *root_worker;
175 bool kicked_without_poller;
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700176
177 /* Set to true if the pollset is observed to have no workers available to
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700178 poll */
Craig Tiller6de05932017-04-28 09:17:38 -0700179 bool seen_inactive;
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700180 bool shutting_down; /* Is the pollset shutting down ? */
Craig Tiller4509c472017-04-27 19:05:13 +0000181 grpc_closure *shutdown_closure; /* Called after after shutdown is complete */
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700182
183 /* Number of workers who are *about-to* attach themselves to the pollset
184 * worker list */
Craig Tillerba550da2017-05-01 14:26:31 +0000185 int begin_refs;
Craig Tiller6de05932017-04-28 09:17:38 -0700186
187 grpc_pollset *next;
188 grpc_pollset *prev;
Craig Tillerc67cc992017-04-27 10:15:51 -0700189};
190
191/*******************************************************************************
192 * Pollset-set Declarations
193 */
Craig Tiller6de05932017-04-28 09:17:38 -0700194
Craig Tiller61f96c12017-05-12 13:36:39 -0700195struct grpc_pollset_set {
196 char unused;
197};
Craig Tillerc67cc992017-04-27 10:15:51 -0700198
199/*******************************************************************************
200 * Common helpers
201 */
202
203static bool append_error(grpc_error **composite, grpc_error *error,
204 const char *desc) {
205 if (error == GRPC_ERROR_NONE) return true;
206 if (*composite == GRPC_ERROR_NONE) {
207 *composite = GRPC_ERROR_CREATE_FROM_COPIED_STRING(desc);
208 }
209 *composite = grpc_error_add_child(*composite, error);
210 return false;
211}
212
213/*******************************************************************************
214 * Fd Definitions
215 */
216
217/* We need to keep a freelist not because of any concerns of malloc performance
218 * but instead so that implementations with multiple threads in (for example)
219 * epoll_wait deal with the race between pollset removal and incoming poll
220 * notifications.
221 *
222 * The problem is that the poller ultimately holds a reference to this
223 * object, so it is very difficult to know when is safe to free it, at least
224 * without some expensive synchronization.
225 *
226 * If we keep the object freelisted, in the worst case losing this race just
227 * becomes a spurious read notification on a reused fd.
228 */
229
230/* The alarm system needs to be able to wakeup 'some poller' sometimes
231 * (specifically when a new alarm needs to be triggered earlier than the next
232 * alarm 'epoch'). This wakeup_fd gives us something to alert on when such a
233 * case occurs. */
234
235static grpc_fd *fd_freelist = NULL;
236static gpr_mu fd_freelist_mu;
237
Craig Tillerc67cc992017-04-27 10:15:51 -0700238static void fd_global_init(void) { gpr_mu_init(&fd_freelist_mu); }
239
240static void fd_global_shutdown(void) {
241 gpr_mu_lock(&fd_freelist_mu);
242 gpr_mu_unlock(&fd_freelist_mu);
243 while (fd_freelist != NULL) {
244 grpc_fd *fd = fd_freelist;
245 fd_freelist = fd_freelist->freelist_next;
Craig Tillerc67cc992017-04-27 10:15:51 -0700246 gpr_free(fd);
247 }
248 gpr_mu_destroy(&fd_freelist_mu);
249}
250
251static grpc_fd *fd_create(int fd, const char *name) {
252 grpc_fd *new_fd = NULL;
253
254 gpr_mu_lock(&fd_freelist_mu);
255 if (fd_freelist != NULL) {
256 new_fd = fd_freelist;
257 fd_freelist = fd_freelist->freelist_next;
258 }
259 gpr_mu_unlock(&fd_freelist_mu);
260
261 if (new_fd == NULL) {
262 new_fd = gpr_malloc(sizeof(grpc_fd));
Craig Tillerc67cc992017-04-27 10:15:51 -0700263 }
264
Craig Tillerc67cc992017-04-27 10:15:51 -0700265 new_fd->fd = fd;
Craig Tillerc67cc992017-04-27 10:15:51 -0700266 grpc_lfev_init(&new_fd->read_closure);
267 grpc_lfev_init(&new_fd->write_closure);
268 gpr_atm_no_barrier_store(&new_fd->read_notifier_pollset, (gpr_atm)NULL);
269
270 new_fd->freelist_next = NULL;
Craig Tillerc67cc992017-04-27 10:15:51 -0700271
272 char *fd_name;
273 gpr_asprintf(&fd_name, "%s fd=%d", name, fd);
274 grpc_iomgr_register_object(&new_fd->iomgr_object, fd_name);
Noah Eisen264879f2017-06-20 17:14:47 -0700275#ifndef NDEBUG
276 if (GRPC_TRACER_ON(grpc_trace_fd_refcount)) {
277 gpr_log(GPR_DEBUG, "FD %d %p create %s", fd, new_fd, fd_name);
278 }
Craig Tillerc67cc992017-04-27 10:15:51 -0700279#endif
280 gpr_free(fd_name);
Craig Tiller9ddb3152017-04-27 21:32:56 +0000281
282 struct epoll_event ev = {.events = (uint32_t)(EPOLLIN | EPOLLOUT | EPOLLET),
283 .data.ptr = new_fd};
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700284 if (epoll_ctl(g_epoll_set.epfd, EPOLL_CTL_ADD, fd, &ev) != 0) {
Craig Tiller9ddb3152017-04-27 21:32:56 +0000285 gpr_log(GPR_ERROR, "epoll_ctl failed: %s", strerror(errno));
286 }
287
Craig Tillerc67cc992017-04-27 10:15:51 -0700288 return new_fd;
289}
290
Craig Tiller4509c472017-04-27 19:05:13 +0000291static int fd_wrapped_fd(grpc_fd *fd) { return fd->fd; }
Craig Tillerc67cc992017-04-27 10:15:51 -0700292
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700293/* if 'releasing_fd' is true, it means that we are going to detach the internal
294 * fd from grpc_fd structure (i.e which means we should not be calling
295 * shutdown() syscall on that fd) */
296static void fd_shutdown_internal(grpc_exec_ctx *exec_ctx, grpc_fd *fd,
297 grpc_error *why, bool releasing_fd) {
Craig Tiller9ddb3152017-04-27 21:32:56 +0000298 if (grpc_lfev_set_shutdown(exec_ctx, &fd->read_closure,
299 GRPC_ERROR_REF(why))) {
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700300 if (!releasing_fd) {
301 shutdown(fd->fd, SHUT_RDWR);
302 }
Craig Tiller9ddb3152017-04-27 21:32:56 +0000303 grpc_lfev_set_shutdown(exec_ctx, &fd->write_closure, GRPC_ERROR_REF(why));
304 }
305 GRPC_ERROR_UNREF(why);
306}
307
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700308/* Might be called multiple times */
309static void fd_shutdown(grpc_exec_ctx *exec_ctx, grpc_fd *fd, grpc_error *why) {
310 fd_shutdown_internal(exec_ctx, fd, why, false);
311}
312
Craig Tillerc67cc992017-04-27 10:15:51 -0700313static void fd_orphan(grpc_exec_ctx *exec_ctx, grpc_fd *fd,
314 grpc_closure *on_done, int *release_fd,
Yuchen Zengd40a7ae2017-07-12 15:59:56 -0700315 bool already_closed, const char *reason) {
Craig Tillerc67cc992017-04-27 10:15:51 -0700316 grpc_error *error = GRPC_ERROR_NONE;
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700317 bool is_release_fd = (release_fd != NULL);
Craig Tillerc67cc992017-04-27 10:15:51 -0700318
Craig Tiller9ddb3152017-04-27 21:32:56 +0000319 if (!grpc_lfev_is_shutdown(&fd->read_closure)) {
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700320 fd_shutdown_internal(exec_ctx, fd,
321 GRPC_ERROR_CREATE_FROM_COPIED_STRING(reason),
322 is_release_fd);
Craig Tiller9ddb3152017-04-27 21:32:56 +0000323 }
324
Craig Tillerc67cc992017-04-27 10:15:51 -0700325 /* If release_fd is not NULL, we should be relinquishing control of the file
326 descriptor fd->fd (but we still own the grpc_fd structure). */
Sree Kuchibhotlaf2641472017-08-02 23:46:40 -0700327 if (is_release_fd) {
Craig Tillerc67cc992017-04-27 10:15:51 -0700328 *release_fd = fd->fd;
Yuchen Zengd40a7ae2017-07-12 15:59:56 -0700329 } else if (!already_closed) {
Craig Tillerc67cc992017-04-27 10:15:51 -0700330 close(fd->fd);
Craig Tillerc67cc992017-04-27 10:15:51 -0700331 }
332
ncteisen274bbbe2017-06-08 14:57:11 -0700333 GRPC_CLOSURE_SCHED(exec_ctx, on_done, GRPC_ERROR_REF(error));
Craig Tillerc67cc992017-04-27 10:15:51 -0700334
Craig Tiller4509c472017-04-27 19:05:13 +0000335 grpc_iomgr_unregister_object(&fd->iomgr_object);
336 grpc_lfev_destroy(&fd->read_closure);
337 grpc_lfev_destroy(&fd->write_closure);
Craig Tillerc67cc992017-04-27 10:15:51 -0700338
Craig Tiller4509c472017-04-27 19:05:13 +0000339 gpr_mu_lock(&fd_freelist_mu);
340 fd->freelist_next = fd_freelist;
341 fd_freelist = fd;
342 gpr_mu_unlock(&fd_freelist_mu);
Craig Tillerc67cc992017-04-27 10:15:51 -0700343}
344
345static grpc_pollset *fd_get_read_notifier_pollset(grpc_exec_ctx *exec_ctx,
346 grpc_fd *fd) {
347 gpr_atm notifier = gpr_atm_acq_load(&fd->read_notifier_pollset);
348 return (grpc_pollset *)notifier;
349}
350
351static bool fd_is_shutdown(grpc_fd *fd) {
352 return grpc_lfev_is_shutdown(&fd->read_closure);
353}
354
Craig Tillerc67cc992017-04-27 10:15:51 -0700355static void fd_notify_on_read(grpc_exec_ctx *exec_ctx, grpc_fd *fd,
356 grpc_closure *closure) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700357 grpc_lfev_notify_on(exec_ctx, &fd->read_closure, closure, "read");
Craig Tillerc67cc992017-04-27 10:15:51 -0700358}
359
360static void fd_notify_on_write(grpc_exec_ctx *exec_ctx, grpc_fd *fd,
361 grpc_closure *closure) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700362 grpc_lfev_notify_on(exec_ctx, &fd->write_closure, closure, "write");
Craig Tillerc67cc992017-04-27 10:15:51 -0700363}
364
Craig Tiller4509c472017-04-27 19:05:13 +0000365static void fd_become_readable(grpc_exec_ctx *exec_ctx, grpc_fd *fd,
366 grpc_pollset *notifier) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700367 grpc_lfev_set_ready(exec_ctx, &fd->read_closure, "read");
Craig Tiller4509c472017-04-27 19:05:13 +0000368 /* Use release store to match with acquire load in fd_get_read_notifier */
369 gpr_atm_rel_store(&fd->read_notifier_pollset, (gpr_atm)notifier);
370}
371
372static void fd_become_writable(grpc_exec_ctx *exec_ctx, grpc_fd *fd) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700373 grpc_lfev_set_ready(exec_ctx, &fd->write_closure, "write");
Craig Tillerc67cc992017-04-27 10:15:51 -0700374}
375
376/*******************************************************************************
377 * Pollset Definitions
378 */
379
Craig Tiller6de05932017-04-28 09:17:38 -0700380GPR_TLS_DECL(g_current_thread_pollset);
381GPR_TLS_DECL(g_current_thread_worker);
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700382
383/* The designated poller */
Craig Tiller6de05932017-04-28 09:17:38 -0700384static gpr_atm g_active_poller;
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700385
Craig Tiller6de05932017-04-28 09:17:38 -0700386static pollset_neighbourhood *g_neighbourhoods;
Craig Tiller32f90ee2017-04-28 12:46:41 -0700387static size_t g_num_neighbourhoods;
Craig Tiller6de05932017-04-28 09:17:38 -0700388
Craig Tillerc67cc992017-04-27 10:15:51 -0700389/* Return true if first in list */
Craig Tiller32f90ee2017-04-28 12:46:41 -0700390static bool worker_insert(grpc_pollset *pollset, grpc_pollset_worker *worker) {
391 if (pollset->root_worker == NULL) {
392 pollset->root_worker = worker;
393 worker->next = worker->prev = worker;
Craig Tillerc67cc992017-04-27 10:15:51 -0700394 return true;
395 } else {
Craig Tiller32f90ee2017-04-28 12:46:41 -0700396 worker->next = pollset->root_worker;
397 worker->prev = worker->next->prev;
398 worker->next->prev = worker;
399 worker->prev->next = worker;
Craig Tillerc67cc992017-04-27 10:15:51 -0700400 return false;
401 }
402}
403
404/* Return true if last in list */
405typedef enum { EMPTIED, NEW_ROOT, REMOVED } worker_remove_result;
406
Craig Tiller32f90ee2017-04-28 12:46:41 -0700407static worker_remove_result worker_remove(grpc_pollset *pollset,
Craig Tillerc67cc992017-04-27 10:15:51 -0700408 grpc_pollset_worker *worker) {
Craig Tiller32f90ee2017-04-28 12:46:41 -0700409 if (worker == pollset->root_worker) {
410 if (worker == worker->next) {
411 pollset->root_worker = NULL;
Craig Tillerc67cc992017-04-27 10:15:51 -0700412 return EMPTIED;
413 } else {
Craig Tiller32f90ee2017-04-28 12:46:41 -0700414 pollset->root_worker = worker->next;
415 worker->prev->next = worker->next;
416 worker->next->prev = worker->prev;
Craig Tillerc67cc992017-04-27 10:15:51 -0700417 return NEW_ROOT;
418 }
419 } else {
Craig Tiller32f90ee2017-04-28 12:46:41 -0700420 worker->prev->next = worker->next;
421 worker->next->prev = worker->prev;
Craig Tillerc67cc992017-04-27 10:15:51 -0700422 return REMOVED;
423 }
424}
425
Craig Tillerba550da2017-05-01 14:26:31 +0000426static size_t choose_neighbourhood(void) {
427 return (size_t)gpr_cpu_current_cpu() % g_num_neighbourhoods;
428}
429
Craig Tiller4509c472017-04-27 19:05:13 +0000430static grpc_error *pollset_global_init(void) {
Craig Tiller4509c472017-04-27 19:05:13 +0000431 gpr_tls_init(&g_current_thread_pollset);
432 gpr_tls_init(&g_current_thread_worker);
Craig Tiller6de05932017-04-28 09:17:38 -0700433 gpr_atm_no_barrier_store(&g_active_poller, 0);
Craig Tiller375eb252017-04-27 23:29:12 +0000434 global_wakeup_fd.read_fd = -1;
435 grpc_error *err = grpc_wakeup_fd_init(&global_wakeup_fd);
436 if (err != GRPC_ERROR_NONE) return err;
Craig Tiller4509c472017-04-27 19:05:13 +0000437 struct epoll_event ev = {.events = (uint32_t)(EPOLLIN | EPOLLET),
438 .data.ptr = &global_wakeup_fd};
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700439 if (epoll_ctl(g_epoll_set.epfd, EPOLL_CTL_ADD, global_wakeup_fd.read_fd,
440 &ev) != 0) {
Craig Tiller4509c472017-04-27 19:05:13 +0000441 return GRPC_OS_ERROR(errno, "epoll_ctl");
442 }
Craig Tillerba550da2017-05-01 14:26:31 +0000443 g_num_neighbourhoods = GPR_CLAMP(gpr_cpu_num_cores(), 1, MAX_NEIGHBOURHOODS);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700444 g_neighbourhoods =
445 gpr_zalloc(sizeof(*g_neighbourhoods) * g_num_neighbourhoods);
446 for (size_t i = 0; i < g_num_neighbourhoods; i++) {
447 gpr_mu_init(&g_neighbourhoods[i].mu);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700448 }
Craig Tiller4509c472017-04-27 19:05:13 +0000449 return GRPC_ERROR_NONE;
450}
451
452static void pollset_global_shutdown(void) {
Craig Tiller4509c472017-04-27 19:05:13 +0000453 gpr_tls_destroy(&g_current_thread_pollset);
454 gpr_tls_destroy(&g_current_thread_worker);
Craig Tiller375eb252017-04-27 23:29:12 +0000455 if (global_wakeup_fd.read_fd != -1) grpc_wakeup_fd_destroy(&global_wakeup_fd);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700456 for (size_t i = 0; i < g_num_neighbourhoods; i++) {
457 gpr_mu_destroy(&g_neighbourhoods[i].mu);
458 }
459 gpr_free(g_neighbourhoods);
Craig Tiller4509c472017-04-27 19:05:13 +0000460}
461
462static void pollset_init(grpc_pollset *pollset, gpr_mu **mu) {
Craig Tiller6de05932017-04-28 09:17:38 -0700463 gpr_mu_init(&pollset->mu);
464 *mu = &pollset->mu;
Craig Tillerba550da2017-05-01 14:26:31 +0000465 pollset->neighbourhood = &g_neighbourhoods[choose_neighbourhood()];
Sree Kuchibhotla30882302017-08-16 13:46:52 -0700466 pollset->reassigning_neighbourhood = false;
467 pollset->root_worker = NULL;
468 pollset->kicked_without_poller = false;
Craig Tiller6de05932017-04-28 09:17:38 -0700469 pollset->seen_inactive = true;
Sree Kuchibhotla30882302017-08-16 13:46:52 -0700470 pollset->shutting_down = false;
471 pollset->shutdown_closure = NULL;
472 pollset->begin_refs = 0;
473 pollset->next = pollset->prev = NULL;
Craig Tiller6de05932017-04-28 09:17:38 -0700474}
475
Craig Tillerc6109852017-05-01 14:26:49 -0700476static void pollset_destroy(grpc_exec_ctx *exec_ctx, grpc_pollset *pollset) {
Craig Tillere00d7332017-05-01 15:43:51 +0000477 gpr_mu_lock(&pollset->mu);
Craig Tillerba550da2017-05-01 14:26:31 +0000478 if (!pollset->seen_inactive) {
Craig Tillere00d7332017-05-01 15:43:51 +0000479 pollset_neighbourhood *neighbourhood = pollset->neighbourhood;
480 gpr_mu_unlock(&pollset->mu);
Craig Tillera95bacf2017-05-01 12:51:24 -0700481 retry_lock_neighbourhood:
Craig Tillere00d7332017-05-01 15:43:51 +0000482 gpr_mu_lock(&neighbourhood->mu);
483 gpr_mu_lock(&pollset->mu);
484 if (!pollset->seen_inactive) {
485 if (pollset->neighbourhood != neighbourhood) {
486 gpr_mu_unlock(&neighbourhood->mu);
487 neighbourhood = pollset->neighbourhood;
488 gpr_mu_unlock(&pollset->mu);
489 goto retry_lock_neighbourhood;
490 }
491 pollset->prev->next = pollset->next;
492 pollset->next->prev = pollset->prev;
493 if (pollset == pollset->neighbourhood->active_root) {
494 pollset->neighbourhood->active_root =
495 pollset->next == pollset ? NULL : pollset->next;
496 }
Craig Tillerba550da2017-05-01 14:26:31 +0000497 }
498 gpr_mu_unlock(&pollset->neighbourhood->mu);
Craig Tiller6de05932017-04-28 09:17:38 -0700499 }
Craig Tillere00d7332017-05-01 15:43:51 +0000500 gpr_mu_unlock(&pollset->mu);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700501 gpr_mu_destroy(&pollset->mu);
Craig Tiller4509c472017-04-27 19:05:13 +0000502}
503
504static grpc_error *pollset_kick_all(grpc_pollset *pollset) {
yang-gdf92a642017-08-21 22:38:45 -0700505 GPR_TIMER_BEGIN("pollset_kick_all", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000506 grpc_error *error = GRPC_ERROR_NONE;
507 if (pollset->root_worker != NULL) {
508 grpc_pollset_worker *worker = pollset->root_worker;
509 do {
Craig Tiller55624a32017-05-26 08:14:44 -0700510 switch (worker->kick_state) {
511 case KICKED:
512 break;
513 case UNKICKED:
514 SET_KICK_STATE(worker, KICKED);
515 if (worker->initialized_cv) {
516 gpr_cv_signal(&worker->cv);
517 }
518 break;
519 case DESIGNATED_POLLER:
520 SET_KICK_STATE(worker, KICKED);
521 append_error(&error, grpc_wakeup_fd_wakeup(&global_wakeup_fd),
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700522 "pollset_kick_all");
Craig Tiller55624a32017-05-26 08:14:44 -0700523 break;
Craig Tiller4509c472017-04-27 19:05:13 +0000524 }
525
Craig Tiller32f90ee2017-04-28 12:46:41 -0700526 worker = worker->next;
Craig Tiller4509c472017-04-27 19:05:13 +0000527 } while (worker != pollset->root_worker);
528 }
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700529 // TODO: sreek. Check if we need to set 'kicked_without_poller' to true here
530 // in the else case
yang-gdf92a642017-08-21 22:38:45 -0700531 GPR_TIMER_END("pollset_kick_all", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000532 return error;
533}
534
535static void pollset_maybe_finish_shutdown(grpc_exec_ctx *exec_ctx,
536 grpc_pollset *pollset) {
Craig Tillerba550da2017-05-01 14:26:31 +0000537 if (pollset->shutdown_closure != NULL && pollset->root_worker == NULL &&
538 pollset->begin_refs == 0) {
yang-gdf92a642017-08-21 22:38:45 -0700539 GPR_TIMER_MARK("pollset_finish_shutdown", 0);
ncteisen274bbbe2017-06-08 14:57:11 -0700540 GRPC_CLOSURE_SCHED(exec_ctx, pollset->shutdown_closure, GRPC_ERROR_NONE);
Craig Tiller4509c472017-04-27 19:05:13 +0000541 pollset->shutdown_closure = NULL;
542 }
543}
544
545static void pollset_shutdown(grpc_exec_ctx *exec_ctx, grpc_pollset *pollset,
546 grpc_closure *closure) {
yang-gdf92a642017-08-21 22:38:45 -0700547 GPR_TIMER_BEGIN("pollset_shutdown", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000548 GPR_ASSERT(pollset->shutdown_closure == NULL);
Craig Tillerc81512a2017-05-26 09:53:58 -0700549 GPR_ASSERT(!pollset->shutting_down);
Craig Tiller4509c472017-04-27 19:05:13 +0000550 pollset->shutdown_closure = closure;
Craig Tillerc81512a2017-05-26 09:53:58 -0700551 pollset->shutting_down = true;
Craig Tiller4509c472017-04-27 19:05:13 +0000552 GRPC_LOG_IF_ERROR("pollset_shutdown", pollset_kick_all(pollset));
553 pollset_maybe_finish_shutdown(exec_ctx, pollset);
yang-gdf92a642017-08-21 22:38:45 -0700554 GPR_TIMER_END("pollset_shutdown", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000555}
556
Craig Tiller4509c472017-04-27 19:05:13 +0000557static int poll_deadline_to_millis_timeout(gpr_timespec deadline,
558 gpr_timespec now) {
559 gpr_timespec timeout;
560 if (gpr_time_cmp(deadline, gpr_inf_future(deadline.clock_type)) == 0) {
561 return -1;
562 }
563
564 if (gpr_time_cmp(deadline, now) <= 0) {
565 return 0;
566 }
567
568 static const gpr_timespec round_up = {
569 .clock_type = GPR_TIMESPAN, .tv_sec = 0, .tv_nsec = GPR_NS_PER_MS - 1};
570 timeout = gpr_time_sub(deadline, now);
571 int millis = gpr_time_to_millis(gpr_time_add(timeout, round_up));
572 return millis >= 1 ? millis : 1;
573}
574
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700575/* Process the epoll events found by do_epoll_wait() function.
576 - g_epoll_set.cursor points to the index of the first event to be processed
577 - This function then processes up-to MAX_EPOLL_EVENTS_PER_ITERATION and
578 updates the g_epoll_set.cursor
Craig Tiller4509c472017-04-27 19:05:13 +0000579
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700580 NOTE ON SYNCRHONIZATION: Similar to do_epoll_wait(), this function is only
581 called by g_active_poller thread. So there is no need for synchronization
582 when accessing fields in g_epoll_set */
583static grpc_error *process_epoll_events(grpc_exec_ctx *exec_ctx,
584 grpc_pollset *pollset) {
585 static const char *err_desc = "process_events";
Craig Tiller4509c472017-04-27 19:05:13 +0000586 grpc_error *error = GRPC_ERROR_NONE;
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700587
Sree Kuchibhotla3d609f12017-08-25 10:00:18 -0700588 GPR_TIMER_BEGIN("process_epoll_events", 0);
Sree Kuchibhotlaa92a9cc2017-08-27 14:02:15 -0700589 long num_events = gpr_atm_acq_load(&g_epoll_set.num_events);
590 long cursor = gpr_atm_acq_load(&g_epoll_set.cursor);
591 for (int idx = 0;
592 (idx < MAX_EPOLL_EVENTS_HANDLED_PER_ITERATION) && cursor != num_events;
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700593 idx++) {
Sree Kuchibhotlaa92a9cc2017-08-27 14:02:15 -0700594 long c = cursor++;
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700595 struct epoll_event *ev = &g_epoll_set.events[c];
596 void *data_ptr = ev->data.ptr;
597
Craig Tiller4509c472017-04-27 19:05:13 +0000598 if (data_ptr == &global_wakeup_fd) {
Craig Tiller4509c472017-04-27 19:05:13 +0000599 append_error(&error, grpc_wakeup_fd_consume_wakeup(&global_wakeup_fd),
600 err_desc);
601 } else {
602 grpc_fd *fd = (grpc_fd *)(data_ptr);
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700603 bool cancel = (ev->events & (EPOLLERR | EPOLLHUP)) != 0;
604 bool read_ev = (ev->events & (EPOLLIN | EPOLLPRI)) != 0;
605 bool write_ev = (ev->events & EPOLLOUT) != 0;
606
Craig Tiller4509c472017-04-27 19:05:13 +0000607 if (read_ev || cancel) {
608 fd_become_readable(exec_ctx, fd, pollset);
609 }
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700610
Craig Tiller4509c472017-04-27 19:05:13 +0000611 if (write_ev || cancel) {
612 fd_become_writable(exec_ctx, fd);
613 }
614 }
615 }
Sree Kuchibhotlaa92a9cc2017-08-27 14:02:15 -0700616 gpr_atm_rel_store(&g_epoll_set.cursor, cursor);
Sree Kuchibhotla3d609f12017-08-25 10:00:18 -0700617 GPR_TIMER_END("process_epoll_events", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000618 return error;
619}
620
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700621/* Do epoll_wait and store the events in g_epoll_set.events field. This does not
622 "process" any of the events yet; that is done in process_epoll_events().
623 *See process_epoll_events() function for more details.
624
625 NOTE ON SYNCHRONIZATION: At any point of time, only the g_active_poller
626 (i.e the designated poller thread) will be calling this function. So there is
627 no need for any synchronization when accesing fields in g_epoll_set */
628static grpc_error *do_epoll_wait(grpc_exec_ctx *exec_ctx, grpc_pollset *ps,
629 gpr_timespec now, gpr_timespec deadline) {
Sree Kuchibhotla3d609f12017-08-25 10:00:18 -0700630 GPR_TIMER_BEGIN("do_epoll_wait", 0);
631
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700632 int r;
633 int timeout = poll_deadline_to_millis_timeout(deadline, now);
634 if (timeout != 0) {
635 GRPC_SCHEDULING_START_BLOCKING_REGION;
636 }
637 do {
638 r = epoll_wait(g_epoll_set.epfd, g_epoll_set.events, MAX_EPOLL_EVENTS,
639 timeout);
640 } while (r < 0 && errno == EINTR);
641 if (timeout != 0) {
642 GRPC_SCHEDULING_END_BLOCKING_REGION;
643 }
644
645 if (r < 0) return GRPC_OS_ERROR(errno, "epoll_wait");
646
647 if (GRPC_TRACER_ON(grpc_polling_trace)) {
648 gpr_log(GPR_DEBUG, "ps: %p poll got %d events", ps, r);
649 }
650
Sree Kuchibhotlaa92a9cc2017-08-27 14:02:15 -0700651 gpr_atm_rel_store(&g_epoll_set.num_events, r);
652 gpr_atm_rel_store(&g_epoll_set.cursor, 0);
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700653
Sree Kuchibhotla3d609f12017-08-25 10:00:18 -0700654 GPR_TIMER_END("do_epoll_wait", 0);
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700655 return GRPC_ERROR_NONE;
656}
657
Craig Tiller4509c472017-04-27 19:05:13 +0000658static bool begin_worker(grpc_pollset *pollset, grpc_pollset_worker *worker,
659 grpc_pollset_worker **worker_hdl, gpr_timespec *now,
660 gpr_timespec deadline) {
yang-gdf92a642017-08-21 22:38:45 -0700661 GPR_TIMER_BEGIN("begin_worker", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000662 if (worker_hdl != NULL) *worker_hdl = worker;
663 worker->initialized_cv = false;
Craig Tiller55624a32017-05-26 08:14:44 -0700664 SET_KICK_STATE(worker, UNKICKED);
Craig Tiller50da5ec2017-05-01 13:51:14 -0700665 worker->schedule_on_end_work = (grpc_closure_list)GRPC_CLOSURE_LIST_INIT;
Craig Tillerba550da2017-05-01 14:26:31 +0000666 pollset->begin_refs++;
Craig Tiller4509c472017-04-27 19:05:13 +0000667
Craig Tiller830e82a2017-05-31 16:26:27 -0700668 if (GRPC_TRACER_ON(grpc_polling_trace)) {
669 gpr_log(GPR_ERROR, "PS:%p BEGIN_STARTS:%p", pollset, worker);
670 }
671
Craig Tiller32f90ee2017-04-28 12:46:41 -0700672 if (pollset->seen_inactive) {
673 // pollset has been observed to be inactive, we need to move back to the
674 // active list
Craig Tillere00d7332017-05-01 15:43:51 +0000675 bool is_reassigning = false;
676 if (!pollset->reassigning_neighbourhood) {
677 is_reassigning = true;
678 pollset->reassigning_neighbourhood = true;
679 pollset->neighbourhood = &g_neighbourhoods[choose_neighbourhood()];
680 }
681 pollset_neighbourhood *neighbourhood = pollset->neighbourhood;
Craig Tiller32f90ee2017-04-28 12:46:41 -0700682 gpr_mu_unlock(&pollset->mu);
Craig Tillerba550da2017-05-01 14:26:31 +0000683 // pollset unlocked: state may change (even worker->kick_state)
684 retry_lock_neighbourhood:
Craig Tiller32f90ee2017-04-28 12:46:41 -0700685 gpr_mu_lock(&neighbourhood->mu);
686 gpr_mu_lock(&pollset->mu);
Craig Tiller830e82a2017-05-31 16:26:27 -0700687 if (GRPC_TRACER_ON(grpc_polling_trace)) {
688 gpr_log(GPR_ERROR, "PS:%p BEGIN_REORG:%p kick_state=%s is_reassigning=%d",
689 pollset, worker, kick_state_string(worker->kick_state),
690 is_reassigning);
691 }
Craig Tiller32f90ee2017-04-28 12:46:41 -0700692 if (pollset->seen_inactive) {
Craig Tiller2acab6e2017-04-30 23:06:33 +0000693 if (neighbourhood != pollset->neighbourhood) {
694 gpr_mu_unlock(&neighbourhood->mu);
695 neighbourhood = pollset->neighbourhood;
696 gpr_mu_unlock(&pollset->mu);
697 goto retry_lock_neighbourhood;
698 }
Craig Tiller32f90ee2017-04-28 12:46:41 -0700699 pollset->seen_inactive = false;
Craig Tiller2acab6e2017-04-30 23:06:33 +0000700 if (neighbourhood->active_root == NULL) {
701 neighbourhood->active_root = pollset->next = pollset->prev = pollset;
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700702 /* TODO: sreek. Why would this worker state be other than UNKICKED
703 * here ? (since the worker isn't added to the pollset yet, there is no
704 * way it can be "found" by other threads to get kicked). */
705
706 /* If there is no designated poller, make this the designated poller */
Craig Tiller55624a32017-05-26 08:14:44 -0700707 if (worker->kick_state == UNKICKED &&
708 gpr_atm_no_barrier_cas(&g_active_poller, 0, (gpr_atm)worker)) {
709 SET_KICK_STATE(worker, DESIGNATED_POLLER);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700710 }
Craig Tiller2acab6e2017-04-30 23:06:33 +0000711 } else {
712 pollset->next = neighbourhood->active_root;
713 pollset->prev = pollset->next->prev;
714 pollset->next->prev = pollset->prev->next = pollset;
Craig Tiller4509c472017-04-27 19:05:13 +0000715 }
716 }
Craig Tillere00d7332017-05-01 15:43:51 +0000717 if (is_reassigning) {
718 GPR_ASSERT(pollset->reassigning_neighbourhood);
719 pollset->reassigning_neighbourhood = false;
720 }
Craig Tiller32f90ee2017-04-28 12:46:41 -0700721 gpr_mu_unlock(&neighbourhood->mu);
722 }
Sree Kuchibhotlae6506bc2017-07-18 21:43:45 -0700723
Craig Tiller32f90ee2017-04-28 12:46:41 -0700724 worker_insert(pollset, worker);
Craig Tillerba550da2017-05-01 14:26:31 +0000725 pollset->begin_refs--;
Sree Kuchibhotla949d0752017-07-20 23:49:15 -0700726 if (worker->kick_state == UNKICKED && !pollset->kicked_without_poller) {
Craig Tillera4b8eb02017-04-29 00:13:52 +0000727 GPR_ASSERT(gpr_atm_no_barrier_load(&g_active_poller) != (gpr_atm)worker);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700728 worker->initialized_cv = true;
729 gpr_cv_init(&worker->cv);
Craig Tillerc81512a2017-05-26 09:53:58 -0700730 while (worker->kick_state == UNKICKED && !pollset->shutting_down) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700731 if (GRPC_TRACER_ON(grpc_polling_trace)) {
732 gpr_log(GPR_ERROR, "PS:%p BEGIN_WAIT:%p kick_state=%s shutdown=%d",
733 pollset, worker, kick_state_string(worker->kick_state),
734 pollset->shutting_down);
735 }
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700736
Craig Tiller32f90ee2017-04-28 12:46:41 -0700737 if (gpr_cv_wait(&worker->cv, &pollset->mu, deadline) &&
738 worker->kick_state == UNKICKED) {
Sree Kuchibhotla0d8431a2017-07-18 16:21:54 -0700739 /* If gpr_cv_wait returns true (i.e a timeout), pretend that the worker
740 received a kick */
Craig Tiller55624a32017-05-26 08:14:44 -0700741 SET_KICK_STATE(worker, KICKED);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700742 }
Craig Tillerba550da2017-05-01 14:26:31 +0000743 }
Craig Tiller4509c472017-04-27 19:05:13 +0000744 *now = gpr_now(now->clock_type);
745 }
Sree Kuchibhotla949d0752017-07-20 23:49:15 -0700746
Craig Tiller830e82a2017-05-31 16:26:27 -0700747 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Sree Kuchibhotla949d0752017-07-20 23:49:15 -0700748 gpr_log(GPR_ERROR,
749 "PS:%p BEGIN_DONE:%p kick_state=%s shutdown=%d "
750 "kicked_without_poller: %d",
751 pollset, worker, kick_state_string(worker->kick_state),
752 pollset->shutting_down, pollset->kicked_without_poller);
Craig Tiller830e82a2017-05-31 16:26:27 -0700753 }
Craig Tiller4509c472017-04-27 19:05:13 +0000754
Sree Kuchibhotlae6506bc2017-07-18 21:43:45 -0700755 /* We release pollset lock in this function at a couple of places:
Sree Kuchibhotlaa0616ef2017-07-18 23:49:49 -0700756 * 1. Briefly when assigning pollset to a neighbourhood
Sree Kuchibhotlae6506bc2017-07-18 21:43:45 -0700757 * 2. When doing gpr_cv_wait()
758 * It is possible that 'kicked_without_poller' was set to true during (1) and
759 * 'shutting_down' is set to true during (1) or (2). If either of them is
760 * true, this worker cannot do polling */
Sree Kuchibhotlae6506bc2017-07-18 21:43:45 -0700761 /* TODO(sreek): Perhaps there is a better way to handle kicked_without_poller
762 * case; especially when the worker is the DESIGNATED_POLLER */
763
Sree Kuchibhotlaa0616ef2017-07-18 23:49:49 -0700764 if (pollset->kicked_without_poller) {
765 pollset->kicked_without_poller = false;
yang-gdf92a642017-08-21 22:38:45 -0700766 GPR_TIMER_END("begin_worker", 0);
Sree Kuchibhotlaa0616ef2017-07-18 23:49:49 -0700767 return false;
768 }
769
yang-gdf92a642017-08-21 22:38:45 -0700770 GPR_TIMER_END("begin_worker", 0);
Sree Kuchibhotlaa0616ef2017-07-18 23:49:49 -0700771 return worker->kick_state == DESIGNATED_POLLER && !pollset->shutting_down;
Craig Tiller4509c472017-04-27 19:05:13 +0000772}
773
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700774static bool check_neighbourhood_for_available_poller(
Craig Tillera4b8eb02017-04-29 00:13:52 +0000775 pollset_neighbourhood *neighbourhood) {
yang-gdf92a642017-08-21 22:38:45 -0700776 GPR_TIMER_BEGIN("check_neighbourhood_for_available_poller", 0);
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700777 bool found_worker = false;
778 do {
779 grpc_pollset *inspect = neighbourhood->active_root;
780 if (inspect == NULL) {
781 break;
782 }
783 gpr_mu_lock(&inspect->mu);
784 GPR_ASSERT(!inspect->seen_inactive);
785 grpc_pollset_worker *inspect_worker = inspect->root_worker;
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700786 if (inspect_worker != NULL) {
Craig Tillera4b8eb02017-04-29 00:13:52 +0000787 do {
Craig Tillerba550da2017-05-01 14:26:31 +0000788 switch (inspect_worker->kick_state) {
789 case UNKICKED:
790 if (gpr_atm_no_barrier_cas(&g_active_poller, 0,
791 (gpr_atm)inspect_worker)) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700792 if (GRPC_TRACER_ON(grpc_polling_trace)) {
793 gpr_log(GPR_DEBUG, " .. choose next poller to be %p",
794 inspect_worker);
795 }
Craig Tiller55624a32017-05-26 08:14:44 -0700796 SET_KICK_STATE(inspect_worker, DESIGNATED_POLLER);
Craig Tillerba550da2017-05-01 14:26:31 +0000797 if (inspect_worker->initialized_cv) {
yang-gdf92a642017-08-21 22:38:45 -0700798 GPR_TIMER_MARK("signal worker", 0);
Craig Tillerba550da2017-05-01 14:26:31 +0000799 gpr_cv_signal(&inspect_worker->cv);
800 }
Craig Tiller830e82a2017-05-31 16:26:27 -0700801 } else {
802 if (GRPC_TRACER_ON(grpc_polling_trace)) {
803 gpr_log(GPR_DEBUG, " .. beaten to choose next poller");
804 }
Craig Tillera4b8eb02017-04-29 00:13:52 +0000805 }
Craig Tillerba550da2017-05-01 14:26:31 +0000806 // even if we didn't win the cas, there's a worker, we can stop
807 found_worker = true;
808 break;
809 case KICKED:
810 break;
811 case DESIGNATED_POLLER:
812 found_worker = true; // ok, so someone else found the worker, but
813 // we'll accept that
814 break;
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700815 }
Craig Tillera4b8eb02017-04-29 00:13:52 +0000816 inspect_worker = inspect_worker->next;
Craig Tiller830e82a2017-05-31 16:26:27 -0700817 } while (!found_worker && inspect_worker != inspect->root_worker);
Craig Tillera4b8eb02017-04-29 00:13:52 +0000818 }
819 if (!found_worker) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700820 if (GRPC_TRACER_ON(grpc_polling_trace)) {
821 gpr_log(GPR_DEBUG, " .. mark pollset %p inactive", inspect);
822 }
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700823 inspect->seen_inactive = true;
Craig Tiller2acab6e2017-04-30 23:06:33 +0000824 if (inspect == neighbourhood->active_root) {
Craig Tillera95bacf2017-05-01 12:51:24 -0700825 neighbourhood->active_root =
826 inspect->next == inspect ? NULL : inspect->next;
Craig Tiller2acab6e2017-04-30 23:06:33 +0000827 }
828 inspect->next->prev = inspect->prev;
829 inspect->prev->next = inspect->next;
Craig Tillere00d7332017-05-01 15:43:51 +0000830 inspect->next = inspect->prev = NULL;
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700831 }
832 gpr_mu_unlock(&inspect->mu);
833 } while (!found_worker);
yang-gdf92a642017-08-21 22:38:45 -0700834 GPR_TIMER_END("check_neighbourhood_for_available_poller", 0);
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700835 return found_worker;
836}
837
Craig Tiller4509c472017-04-27 19:05:13 +0000838static void end_worker(grpc_exec_ctx *exec_ctx, grpc_pollset *pollset,
839 grpc_pollset_worker *worker,
840 grpc_pollset_worker **worker_hdl) {
yang-gdf92a642017-08-21 22:38:45 -0700841 GPR_TIMER_BEGIN("end_worker", 0);
Craig Tiller830e82a2017-05-31 16:26:27 -0700842 if (GRPC_TRACER_ON(grpc_polling_trace)) {
843 gpr_log(GPR_DEBUG, "PS:%p END_WORKER:%p", pollset, worker);
844 }
Craig Tiller8502ecb2017-04-28 14:22:01 -0700845 if (worker_hdl != NULL) *worker_hdl = NULL;
Craig Tiller830e82a2017-05-31 16:26:27 -0700846 /* Make sure we appear kicked */
Craig Tiller55624a32017-05-26 08:14:44 -0700847 SET_KICK_STATE(worker, KICKED);
Craig Tiller50da5ec2017-05-01 13:51:14 -0700848 grpc_closure_list_move(&worker->schedule_on_end_work,
849 &exec_ctx->closure_list);
Craig Tiller8502ecb2017-04-28 14:22:01 -0700850 if (gpr_atm_no_barrier_load(&g_active_poller) == (gpr_atm)worker) {
Craig Tillera4b8eb02017-04-29 00:13:52 +0000851 if (worker->next != worker && worker->next->kick_state == UNKICKED) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700852 if (GRPC_TRACER_ON(grpc_polling_trace)) {
853 gpr_log(GPR_DEBUG, " .. choose next poller to be peer %p", worker);
854 }
Craig Tiller2acab6e2017-04-30 23:06:33 +0000855 GPR_ASSERT(worker->next->initialized_cv);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700856 gpr_atm_no_barrier_store(&g_active_poller, (gpr_atm)worker->next);
Craig Tiller55624a32017-05-26 08:14:44 -0700857 SET_KICK_STATE(worker->next, DESIGNATED_POLLER);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700858 gpr_cv_signal(&worker->next->cv);
Craig Tiller8502ecb2017-04-28 14:22:01 -0700859 if (grpc_exec_ctx_has_work(exec_ctx)) {
860 gpr_mu_unlock(&pollset->mu);
861 grpc_exec_ctx_flush(exec_ctx);
862 gpr_mu_lock(&pollset->mu);
863 }
Craig Tiller32f90ee2017-04-28 12:46:41 -0700864 } else {
865 gpr_atm_no_barrier_store(&g_active_poller, 0);
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700866 size_t poller_neighbourhood_idx =
867 (size_t)(pollset->neighbourhood - g_neighbourhoods);
Craig Tillerbb742672017-05-17 22:19:05 +0000868 gpr_mu_unlock(&pollset->mu);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700869 bool found_worker = false;
Craig Tillerba550da2017-05-01 14:26:31 +0000870 bool scan_state[MAX_NEIGHBOURHOODS];
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700871 for (size_t i = 0; !found_worker && i < g_num_neighbourhoods; i++) {
872 pollset_neighbourhood *neighbourhood =
873 &g_neighbourhoods[(poller_neighbourhood_idx + i) %
874 g_num_neighbourhoods];
875 if (gpr_mu_trylock(&neighbourhood->mu)) {
876 found_worker =
Craig Tillera4b8eb02017-04-29 00:13:52 +0000877 check_neighbourhood_for_available_poller(neighbourhood);
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700878 gpr_mu_unlock(&neighbourhood->mu);
Craig Tillerba550da2017-05-01 14:26:31 +0000879 scan_state[i] = true;
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700880 } else {
Craig Tillerba550da2017-05-01 14:26:31 +0000881 scan_state[i] = false;
Craig Tiller32f90ee2017-04-28 12:46:41 -0700882 }
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700883 }
Craig Tiller2acab6e2017-04-30 23:06:33 +0000884 for (size_t i = 0; !found_worker && i < g_num_neighbourhoods; i++) {
Craig Tillerba550da2017-05-01 14:26:31 +0000885 if (scan_state[i]) continue;
Craig Tiller2acab6e2017-04-30 23:06:33 +0000886 pollset_neighbourhood *neighbourhood =
887 &g_neighbourhoods[(poller_neighbourhood_idx + i) %
888 g_num_neighbourhoods];
889 gpr_mu_lock(&neighbourhood->mu);
Craig Tillerba550da2017-05-01 14:26:31 +0000890 found_worker = check_neighbourhood_for_available_poller(neighbourhood);
Craig Tiller2acab6e2017-04-30 23:06:33 +0000891 gpr_mu_unlock(&neighbourhood->mu);
Craig Tillerbbf4c7a2017-04-28 15:12:10 -0700892 }
Craig Tiller8502ecb2017-04-28 14:22:01 -0700893 grpc_exec_ctx_flush(exec_ctx);
Craig Tiller32f90ee2017-04-28 12:46:41 -0700894 gpr_mu_lock(&pollset->mu);
895 }
Craig Tiller50da5ec2017-05-01 13:51:14 -0700896 } else if (grpc_exec_ctx_has_work(exec_ctx)) {
897 gpr_mu_unlock(&pollset->mu);
898 grpc_exec_ctx_flush(exec_ctx);
899 gpr_mu_lock(&pollset->mu);
Craig Tiller4509c472017-04-27 19:05:13 +0000900 }
901 if (worker->initialized_cv) {
902 gpr_cv_destroy(&worker->cv);
903 }
Craig Tiller830e82a2017-05-31 16:26:27 -0700904 if (GRPC_TRACER_ON(grpc_polling_trace)) {
905 gpr_log(GPR_DEBUG, " .. remove worker");
906 }
Craig Tiller32f90ee2017-04-28 12:46:41 -0700907 if (EMPTIED == worker_remove(pollset, worker)) {
Craig Tiller4509c472017-04-27 19:05:13 +0000908 pollset_maybe_finish_shutdown(exec_ctx, pollset);
909 }
Craig Tillera4b8eb02017-04-29 00:13:52 +0000910 GPR_ASSERT(gpr_atm_no_barrier_load(&g_active_poller) != (gpr_atm)worker);
yang-gdf92a642017-08-21 22:38:45 -0700911 GPR_TIMER_END("end_worker", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000912}
913
914/* pollset->po.mu lock must be held by the caller before calling this.
915 The function pollset_work() may temporarily release the lock (pollset->po.mu)
916 during the course of its execution but it will always re-acquire the lock and
917 ensure that it is held by the time the function returns */
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700918static grpc_error *pollset_work(grpc_exec_ctx *exec_ctx, grpc_pollset *ps,
Craig Tiller4509c472017-04-27 19:05:13 +0000919 grpc_pollset_worker **worker_hdl,
920 gpr_timespec now, gpr_timespec deadline) {
921 grpc_pollset_worker worker;
922 grpc_error *error = GRPC_ERROR_NONE;
923 static const char *err_desc = "pollset_work";
yang-gdf92a642017-08-21 22:38:45 -0700924 GPR_TIMER_BEGIN("pollset_work", 0);
Sree Kuchibhotlab154cd12017-08-25 10:33:41 -0700925 if (ps->kicked_without_poller) {
926 ps->kicked_without_poller = false;
yang-gdf92a642017-08-21 22:38:45 -0700927 GPR_TIMER_END("pollset_work", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000928 return GRPC_ERROR_NONE;
929 }
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700930
931 if (begin_worker(ps, &worker, worker_hdl, &now, deadline)) {
932 gpr_tls_set(&g_current_thread_pollset, (intptr_t)ps);
Craig Tiller4509c472017-04-27 19:05:13 +0000933 gpr_tls_set(&g_current_thread_worker, (intptr_t)&worker);
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700934 GPR_ASSERT(!ps->shutting_down);
935 GPR_ASSERT(!ps->seen_inactive);
936
937 gpr_mu_unlock(&ps->mu); /* unlock */
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700938 /* This is the designated polling thread at this point and should ideally do
939 polling. However, if there are unprocessed events left from a previous
940 call to do_epoll_wait(), skip calling epoll_wait() in this iteration and
941 process the pending epoll events.
942
943 The reason for decoupling do_epoll_wait and process_epoll_events is to
944 better distrubute the work (i.e handling epoll events) across multiple
945 threads
946
947 process_epoll_events() returns very quickly: It just queues the work on
948 exec_ctx but does not execute it (the actual exectution or more
949 accurately grpc_exec_ctx_flush() happens in end_worker() AFTER selecting
950 a designated poller). So we are not waiting long periods without a
951 designated poller */
Sree Kuchibhotlaa92a9cc2017-08-27 14:02:15 -0700952 if (gpr_atm_acq_load(&g_epoll_set.cursor) ==
953 gpr_atm_acq_load(&g_epoll_set.num_events)) {
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700954 append_error(&error, do_epoll_wait(exec_ctx, ps, now, deadline),
955 err_desc);
956 }
957 append_error(&error, process_epoll_events(exec_ctx, ps), err_desc);
958
959 gpr_mu_lock(&ps->mu); /* lock */
960
Craig Tiller4509c472017-04-27 19:05:13 +0000961 gpr_tls_set(&g_current_thread_worker, 0);
Craig Tiller830e82a2017-05-31 16:26:27 -0700962 } else {
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700963 gpr_tls_set(&g_current_thread_pollset, (intptr_t)ps);
Craig Tiller4509c472017-04-27 19:05:13 +0000964 }
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -0700965 end_worker(exec_ctx, ps, &worker, worker_hdl);
966
Craig Tiller8502ecb2017-04-28 14:22:01 -0700967 gpr_tls_set(&g_current_thread_pollset, 0);
yang-gdf92a642017-08-21 22:38:45 -0700968 GPR_TIMER_END("pollset_work", 0);
Craig Tiller4509c472017-04-27 19:05:13 +0000969 return error;
970}
971
972static grpc_error *pollset_kick(grpc_pollset *pollset,
973 grpc_pollset_worker *specific_worker) {
yang-gdf92a642017-08-21 22:38:45 -0700974 GPR_TIMER_BEGIN("pollset_kick", 0);
975 grpc_error *ret_err = GRPC_ERROR_NONE;
Craig Tillerb89bac02017-05-26 15:20:32 +0000976 if (GRPC_TRACER_ON(grpc_polling_trace)) {
977 gpr_strvec log;
978 gpr_strvec_init(&log);
979 char *tmp;
Craig Tiller75aef7f2017-05-26 08:26:08 -0700980 gpr_asprintf(
981 &tmp, "PS:%p KICK:%p curps=%p curworker=%p root=%p", pollset,
982 specific_worker, (void *)gpr_tls_get(&g_current_thread_pollset),
983 (void *)gpr_tls_get(&g_current_thread_worker), pollset->root_worker);
Craig Tillerb89bac02017-05-26 15:20:32 +0000984 gpr_strvec_add(&log, tmp);
985 if (pollset->root_worker != NULL) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700986 gpr_asprintf(&tmp, " {kick_state=%s next=%p {kick_state=%s}}",
987 kick_state_string(pollset->root_worker->kick_state),
988 pollset->root_worker->next,
989 kick_state_string(pollset->root_worker->next->kick_state));
Craig Tillerb89bac02017-05-26 15:20:32 +0000990 gpr_strvec_add(&log, tmp);
991 }
992 if (specific_worker != NULL) {
Craig Tiller830e82a2017-05-31 16:26:27 -0700993 gpr_asprintf(&tmp, " worker_kick_state=%s",
994 kick_state_string(specific_worker->kick_state));
Craig Tillerb89bac02017-05-26 15:20:32 +0000995 gpr_strvec_add(&log, tmp);
996 }
997 tmp = gpr_strvec_flatten(&log, NULL);
998 gpr_strvec_destroy(&log);
Craig Tiller830e82a2017-05-31 16:26:27 -0700999 gpr_log(GPR_ERROR, "%s", tmp);
Craig Tillerb89bac02017-05-26 15:20:32 +00001000 gpr_free(tmp);
1001 }
Craig Tiller4509c472017-04-27 19:05:13 +00001002 if (specific_worker == NULL) {
1003 if (gpr_tls_get(&g_current_thread_pollset) != (intptr_t)pollset) {
Craig Tiller375eb252017-04-27 23:29:12 +00001004 grpc_pollset_worker *root_worker = pollset->root_worker;
1005 if (root_worker == NULL) {
Craig Tiller4509c472017-04-27 19:05:13 +00001006 pollset->kicked_without_poller = true;
Craig Tiller75aef7f2017-05-26 08:26:08 -07001007 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001008 gpr_log(GPR_ERROR, " .. kicked_without_poller");
Craig Tiller75aef7f2017-05-26 08:26:08 -07001009 }
yang-gdf92a642017-08-21 22:38:45 -07001010 goto done;
Craig Tiller375eb252017-04-27 23:29:12 +00001011 }
Craig Tiller32f90ee2017-04-28 12:46:41 -07001012 grpc_pollset_worker *next_worker = root_worker->next;
Craig Tiller830e82a2017-05-31 16:26:27 -07001013 if (root_worker->kick_state == KICKED) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001014 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001015 gpr_log(GPR_ERROR, " .. already kicked %p", root_worker);
1016 }
1017 SET_KICK_STATE(root_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001018 goto done;
Craig Tiller830e82a2017-05-31 16:26:27 -07001019 } else if (next_worker->kick_state == KICKED) {
1020 if (GRPC_TRACER_ON(grpc_polling_trace)) {
1021 gpr_log(GPR_ERROR, " .. already kicked %p", next_worker);
1022 }
1023 SET_KICK_STATE(next_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001024 goto done;
Craig Tiller830e82a2017-05-31 16:26:27 -07001025 } else if (root_worker ==
1026 next_worker && // only try and wake up a poller if
1027 // there is no next worker
1028 root_worker == (grpc_pollset_worker *)gpr_atm_no_barrier_load(
1029 &g_active_poller)) {
1030 if (GRPC_TRACER_ON(grpc_polling_trace)) {
1031 gpr_log(GPR_ERROR, " .. kicked %p", root_worker);
Craig Tiller75aef7f2017-05-26 08:26:08 -07001032 }
Craig Tiller55624a32017-05-26 08:14:44 -07001033 SET_KICK_STATE(root_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001034 ret_err = grpc_wakeup_fd_wakeup(&global_wakeup_fd);
1035 goto done;
Craig Tiller8502ecb2017-04-28 14:22:01 -07001036 } else if (next_worker->kick_state == UNKICKED) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001037 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001038 gpr_log(GPR_ERROR, " .. kicked %p", next_worker);
Craig Tiller75aef7f2017-05-26 08:26:08 -07001039 }
Craig Tiller8502ecb2017-04-28 14:22:01 -07001040 GPR_ASSERT(next_worker->initialized_cv);
Craig Tiller55624a32017-05-26 08:14:44 -07001041 SET_KICK_STATE(next_worker, KICKED);
Craig Tiller375eb252017-04-27 23:29:12 +00001042 gpr_cv_signal(&next_worker->cv);
yang-gdf92a642017-08-21 22:38:45 -07001043 goto done;
Craig Tiller55624a32017-05-26 08:14:44 -07001044 } else if (next_worker->kick_state == DESIGNATED_POLLER) {
1045 if (root_worker->kick_state != DESIGNATED_POLLER) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001046 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001047 gpr_log(
1048 GPR_ERROR,
1049 " .. kicked root non-poller %p (initialized_cv=%d) (poller=%p)",
1050 root_worker, root_worker->initialized_cv, next_worker);
Craig Tiller75aef7f2017-05-26 08:26:08 -07001051 }
Craig Tiller55624a32017-05-26 08:14:44 -07001052 SET_KICK_STATE(root_worker, KICKED);
1053 if (root_worker->initialized_cv) {
1054 gpr_cv_signal(&root_worker->cv);
1055 }
yang-gdf92a642017-08-21 22:38:45 -07001056 goto done;
Craig Tiller55624a32017-05-26 08:14:44 -07001057 } else {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001058 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001059 gpr_log(GPR_ERROR, " .. non-root poller %p (root=%p)", next_worker,
Craig Tiller75aef7f2017-05-26 08:26:08 -07001060 root_worker);
1061 }
Craig Tiller55624a32017-05-26 08:14:44 -07001062 SET_KICK_STATE(next_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001063 ret_err = grpc_wakeup_fd_wakeup(&global_wakeup_fd);
1064 goto done;
Craig Tiller55624a32017-05-26 08:14:44 -07001065 }
Craig Tiller8502ecb2017-04-28 14:22:01 -07001066 } else {
Craig Tiller55624a32017-05-26 08:14:44 -07001067 GPR_ASSERT(next_worker->kick_state == KICKED);
1068 SET_KICK_STATE(next_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001069 goto done;
Craig Tiller4509c472017-04-27 19:05:13 +00001070 }
1071 } else {
Craig Tiller830e82a2017-05-31 16:26:27 -07001072 if (GRPC_TRACER_ON(grpc_polling_trace)) {
1073 gpr_log(GPR_ERROR, " .. kicked while waking up");
1074 }
yang-gdf92a642017-08-21 22:38:45 -07001075 goto done;
Craig Tiller4509c472017-04-27 19:05:13 +00001076 }
Craig Tiller43bf2592017-04-28 23:21:01 +00001077 } else if (specific_worker->kick_state == KICKED) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001078 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001079 gpr_log(GPR_ERROR, " .. specific worker already kicked");
Craig Tiller75aef7f2017-05-26 08:26:08 -07001080 }
yang-gdf92a642017-08-21 22:38:45 -07001081 goto done;
Craig Tiller4509c472017-04-27 19:05:13 +00001082 } else if (gpr_tls_get(&g_current_thread_worker) ==
1083 (intptr_t)specific_worker) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001084 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001085 gpr_log(GPR_ERROR, " .. mark %p kicked", specific_worker);
Craig Tiller75aef7f2017-05-26 08:26:08 -07001086 }
Craig Tiller55624a32017-05-26 08:14:44 -07001087 SET_KICK_STATE(specific_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001088 goto done;
Craig Tiller32f90ee2017-04-28 12:46:41 -07001089 } else if (specific_worker ==
1090 (grpc_pollset_worker *)gpr_atm_no_barrier_load(&g_active_poller)) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001091 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001092 gpr_log(GPR_ERROR, " .. kick active poller");
Craig Tiller75aef7f2017-05-26 08:26:08 -07001093 }
Craig Tiller55624a32017-05-26 08:14:44 -07001094 SET_KICK_STATE(specific_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001095 ret_err = grpc_wakeup_fd_wakeup(&global_wakeup_fd);
1096 goto done;
Craig Tiller8502ecb2017-04-28 14:22:01 -07001097 } else if (specific_worker->initialized_cv) {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001098 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001099 gpr_log(GPR_ERROR, " .. kick waiting worker");
Craig Tiller75aef7f2017-05-26 08:26:08 -07001100 }
Craig Tiller55624a32017-05-26 08:14:44 -07001101 SET_KICK_STATE(specific_worker, KICKED);
Craig Tiller4509c472017-04-27 19:05:13 +00001102 gpr_cv_signal(&specific_worker->cv);
yang-gdf92a642017-08-21 22:38:45 -07001103 goto done;
Craig Tiller8502ecb2017-04-28 14:22:01 -07001104 } else {
Craig Tiller75aef7f2017-05-26 08:26:08 -07001105 if (GRPC_TRACER_ON(grpc_polling_trace)) {
Craig Tiller830e82a2017-05-31 16:26:27 -07001106 gpr_log(GPR_ERROR, " .. kick non-waiting worker");
Craig Tiller75aef7f2017-05-26 08:26:08 -07001107 }
Craig Tiller55624a32017-05-26 08:14:44 -07001108 SET_KICK_STATE(specific_worker, KICKED);
yang-gdf92a642017-08-21 22:38:45 -07001109 goto done;
Craig Tiller4509c472017-04-27 19:05:13 +00001110 }
yang-gdf92a642017-08-21 22:38:45 -07001111done:
1112 GPR_TIMER_END("pollset_kick", 0);
1113 return ret_err;
Craig Tiller4509c472017-04-27 19:05:13 +00001114}
1115
1116static void pollset_add_fd(grpc_exec_ctx *exec_ctx, grpc_pollset *pollset,
1117 grpc_fd *fd) {}
1118
Craig Tiller4509c472017-04-27 19:05:13 +00001119/*******************************************************************************
Craig Tillerc67cc992017-04-27 10:15:51 -07001120 * Pollset-set Definitions
1121 */
1122
1123static grpc_pollset_set *pollset_set_create(void) {
1124 return (grpc_pollset_set *)((intptr_t)0xdeafbeef);
1125}
1126
1127static void pollset_set_destroy(grpc_exec_ctx *exec_ctx,
1128 grpc_pollset_set *pss) {}
1129
1130static void pollset_set_add_fd(grpc_exec_ctx *exec_ctx, grpc_pollset_set *pss,
1131 grpc_fd *fd) {}
1132
1133static void pollset_set_del_fd(grpc_exec_ctx *exec_ctx, grpc_pollset_set *pss,
1134 grpc_fd *fd) {}
1135
1136static void pollset_set_add_pollset(grpc_exec_ctx *exec_ctx,
1137 grpc_pollset_set *pss, grpc_pollset *ps) {}
1138
1139static void pollset_set_del_pollset(grpc_exec_ctx *exec_ctx,
1140 grpc_pollset_set *pss, grpc_pollset *ps) {}
1141
1142static void pollset_set_add_pollset_set(grpc_exec_ctx *exec_ctx,
1143 grpc_pollset_set *bag,
1144 grpc_pollset_set *item) {}
1145
1146static void pollset_set_del_pollset_set(grpc_exec_ctx *exec_ctx,
1147 grpc_pollset_set *bag,
1148 grpc_pollset_set *item) {}
1149
1150/*******************************************************************************
1151 * Event engine binding
1152 */
1153
1154static void shutdown_engine(void) {
1155 fd_global_shutdown();
1156 pollset_global_shutdown();
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -07001157 epoll_set_shutdown();
Craig Tillerc67cc992017-04-27 10:15:51 -07001158}
1159
1160static const grpc_event_engine_vtable vtable = {
1161 .pollset_size = sizeof(grpc_pollset),
1162
1163 .fd_create = fd_create,
1164 .fd_wrapped_fd = fd_wrapped_fd,
1165 .fd_orphan = fd_orphan,
1166 .fd_shutdown = fd_shutdown,
1167 .fd_is_shutdown = fd_is_shutdown,
1168 .fd_notify_on_read = fd_notify_on_read,
1169 .fd_notify_on_write = fd_notify_on_write,
1170 .fd_get_read_notifier_pollset = fd_get_read_notifier_pollset,
Craig Tillerc67cc992017-04-27 10:15:51 -07001171
1172 .pollset_init = pollset_init,
1173 .pollset_shutdown = pollset_shutdown,
1174 .pollset_destroy = pollset_destroy,
1175 .pollset_work = pollset_work,
1176 .pollset_kick = pollset_kick,
1177 .pollset_add_fd = pollset_add_fd,
1178
1179 .pollset_set_create = pollset_set_create,
1180 .pollset_set_destroy = pollset_set_destroy,
1181 .pollset_set_add_pollset = pollset_set_add_pollset,
1182 .pollset_set_del_pollset = pollset_set_del_pollset,
1183 .pollset_set_add_pollset_set = pollset_set_add_pollset_set,
1184 .pollset_set_del_pollset_set = pollset_set_del_pollset_set,
1185 .pollset_set_add_fd = pollset_set_add_fd,
1186 .pollset_set_del_fd = pollset_set_del_fd,
1187
Craig Tillerc67cc992017-04-27 10:15:51 -07001188 .shutdown_engine = shutdown_engine,
1189};
1190
1191/* It is possible that GLIBC has epoll but the underlying kernel doesn't.
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -07001192 * Create epoll_fd (epoll_set_init() takes care of that) to make sure epoll
1193 * support is available */
Craig Tiller6f0af492017-04-27 19:26:16 +00001194const grpc_event_engine_vtable *grpc_init_epoll1_linux(bool explicit_request) {
Sree Kuchibhotla107840f2017-07-24 18:45:09 -07001195 if (!explicit_request) {
1196 return NULL;
1197 }
1198
Craig Tillerc67cc992017-04-27 10:15:51 -07001199 if (!grpc_has_wakeup_fd()) {
1200 return NULL;
1201 }
1202
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -07001203 if (!epoll_set_init()) {
Craig Tillerc67cc992017-04-27 10:15:51 -07001204 return NULL;
1205 }
1206
Craig Tillerc67cc992017-04-27 10:15:51 -07001207 fd_global_init();
1208
1209 if (!GRPC_LOG_IF_ERROR("pollset_global_init", pollset_global_init())) {
Craig Tiller4509c472017-04-27 19:05:13 +00001210 fd_global_shutdown();
Sree Kuchibhotla5efc9132017-08-17 14:10:38 -07001211 epoll_set_shutdown();
Craig Tillerc67cc992017-04-27 10:15:51 -07001212 return NULL;
1213 }
1214
1215 return &vtable;
1216}
1217
1218#else /* defined(GRPC_LINUX_EPOLL) */
1219#if defined(GRPC_POSIX_SOCKET)
1220#include "src/core/lib/iomgr/ev_posix.h"
1221/* If GRPC_LINUX_EPOLL is not defined, it means epoll is not available. Return
1222 * NULL */
Craig Tiller9ddb3152017-04-27 21:32:56 +00001223const grpc_event_engine_vtable *grpc_init_epoll1_linux(bool explicit_request) {
1224 return NULL;
1225}
Craig Tillerc67cc992017-04-27 10:15:51 -07001226#endif /* defined(GRPC_POSIX_SOCKET) */
1227#endif /* !defined(GRPC_LINUX_EPOLL) */