blob: 199862ddddee4e5c0726fb2799cbcc7317731680 [file] [log] [blame]
/*
*
* Copyright 2015, Google Inc.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are
* met:
*
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above
* copyright notice, this list of conditions and the following disclaimer
* in the documentation and/or other materials provided with the
* distribution.
* * Neither the name of Google Inc. nor the names of its
* contributors may be used to endorse or promote products derived from
* this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*
*/
#include "src/core/channel/channel_stack.h"
#include <grpc/support/log.h>
#include <stdlib.h>
#include <string.h>
int grpc_trace_channel = 0;
/* Memory layouts.
Channel stack is laid out as: {
grpc_channel_stack stk;
padding to GPR_MAX_ALIGNMENT
grpc_channel_element[stk.count];
per-filter memory, aligned to GPR_MAX_ALIGNMENT
}
Call stack is laid out as: {
grpc_call_stack stk;
padding to GPR_MAX_ALIGNMENT
grpc_call_element[stk.count];
per-filter memory, aligned to GPR_MAX_ALIGNMENT
} */
/* Given a size, round up to the next multiple of sizeof(void*) */
#define ROUND_UP_TO_ALIGNMENT_SIZE(x) \
(((x) + GPR_MAX_ALIGNMENT - 1u) & ~(GPR_MAX_ALIGNMENT - 1u))
size_t
grpc_channel_stack_size (const grpc_channel_filter ** filters, size_t filter_count)
{
/* always need the header, and size for the channel elements */
size_t size = ROUND_UP_TO_ALIGNMENT_SIZE (sizeof (grpc_channel_stack)) + ROUND_UP_TO_ALIGNMENT_SIZE (filter_count * sizeof (grpc_channel_element));
size_t i;
GPR_ASSERT ((GPR_MAX_ALIGNMENT & (GPR_MAX_ALIGNMENT - 1)) == 0 && "GPR_MAX_ALIGNMENT must be a power of two");
/* add the size for each filter */
for (i = 0; i < filter_count; i++)
{
size += ROUND_UP_TO_ALIGNMENT_SIZE (filters[i]->sizeof_channel_data);
}
return size;
}
#define CHANNEL_ELEMS_FROM_STACK(stk) \
((grpc_channel_element *)((char *)(stk) + ROUND_UP_TO_ALIGNMENT_SIZE( \
sizeof(grpc_channel_stack))))
#define CALL_ELEMS_FROM_STACK(stk) \
((grpc_call_element *)((char *)(stk) + \
ROUND_UP_TO_ALIGNMENT_SIZE(sizeof(grpc_call_stack))))
grpc_channel_element *
grpc_channel_stack_element (grpc_channel_stack * channel_stack, size_t index)
{
return CHANNEL_ELEMS_FROM_STACK (channel_stack) + index;
}
grpc_channel_element *
grpc_channel_stack_last_element (grpc_channel_stack * channel_stack)
{
return grpc_channel_stack_element (channel_stack, channel_stack->count - 1);
}
grpc_call_element *
grpc_call_stack_element (grpc_call_stack * call_stack, size_t index)
{
return CALL_ELEMS_FROM_STACK (call_stack) + index;
}
void
grpc_channel_stack_init (grpc_exec_ctx * exec_ctx, const grpc_channel_filter ** filters, size_t filter_count, grpc_channel * master, const grpc_channel_args * args, grpc_mdctx * metadata_context, grpc_channel_stack * stack)
{
size_t call_size = ROUND_UP_TO_ALIGNMENT_SIZE (sizeof (grpc_call_stack)) + ROUND_UP_TO_ALIGNMENT_SIZE (filter_count * sizeof (grpc_call_element));
grpc_channel_element *elems;
char *user_data;
size_t i;
stack->count = filter_count;
elems = CHANNEL_ELEMS_FROM_STACK (stack);
user_data = ((char *) elems) + ROUND_UP_TO_ALIGNMENT_SIZE (filter_count * sizeof (grpc_channel_element));
/* init per-filter data */
for (i = 0; i < filter_count; i++)
{
elems[i].filter = filters[i];
elems[i].channel_data = user_data;
elems[i].filter->init_channel_elem (&elems[i], master, args, metadata_context, i == 0, i == (exec_ctx, filter_count - 1));
user_data += ROUND_UP_TO_ALIGNMENT_SIZE (filters[i]->sizeof_channel_data);
call_size += ROUND_UP_TO_ALIGNMENT_SIZE (filters[i]->sizeof_call_data);
}
GPR_ASSERT (user_data > (char *) stack);
GPR_ASSERT ((gpr_uintptr) (user_data - (char *) stack) == grpc_channel_stack_size (filters, filter_count));
stack->call_stack_size = call_size;
}
void
grpc_channel_stack_destroy (grpc_exec_ctx * exec_ctx, grpc_channel_stack * stack)
{
grpc_channel_element *channel_elems = CHANNEL_ELEMS_FROM_STACK (stack);
size_t count = stack->count;
size_t i;
/* destroy per-filter data */
for (i = 0; i < count; i++)
{
channel_elems[i].filter->destroy_channel_elem (exec_ctx, &channel_elems[i]);
}
}
void
grpc_call_stack_init (grpc_exec_ctx * exec_ctx, grpc_channel_stack * channel_stack, const void *transport_server_data, grpc_transport_stream_op * initial_op, grpc_call_stack * call_stack)
{
grpc_channel_element *channel_elems = CHANNEL_ELEMS_FROM_STACK (channel_stack);
size_t count = channel_stack->count;
grpc_call_element *call_elems;
char *user_data;
size_t i;
call_stack->count = count;
call_elems = CALL_ELEMS_FROM_STACK (call_stack);
user_data = ((char *) call_elems) + ROUND_UP_TO_ALIGNMENT_SIZE (count * sizeof (grpc_call_element));
/* init per-filter data */
for (i = 0; i < count; i++)
{
call_elems[i].filter = channel_elems[i].filter;
call_elems[i].channel_data = channel_elems[i].channel_data;
call_elems[i].call_data = user_data;
call_elems[i].filter->init_call_elem (exec_ctx, &call_elems[i], transport_server_data, initial_op);
user_data += ROUND_UP_TO_ALIGNMENT_SIZE (call_elems[i].filter->sizeof_call_data);
}
}
void
grpc_call_stack_destroy (grpc_exec_ctx * exec_ctx, grpc_call_stack * stack)
{
grpc_call_element *elems = CALL_ELEMS_FROM_STACK (stack);
size_t count = stack->count;
size_t i;
/* destroy per-filter data */
for (i = 0; i < count; i++)
{
elems[i].filter->destroy_call_elem (exec_ctx, &elems[i]);
}
}
void
grpc_call_next_op (grpc_exec_ctx * exec_ctx, grpc_call_element * elem, grpc_transport_stream_op * op)
{
grpc_call_element *next_elem = elem + 1;
next_elem->filter->start_transport_stream_op (exec_ctx, next_elem, op);
}
char *
grpc_call_next_get_peer (grpc_exec_ctx * exec_ctx, grpc_call_element * elem)
{
grpc_call_element *next_elem = elem + 1;
return next_elem->filter->get_peer (exec_ctx, next_elem);
}
void
grpc_channel_next_op (grpc_exec_ctx * exec_ctx, grpc_channel_element * elem, grpc_transport_op * op)
{
grpc_channel_element *next_elem = elem + 1;
next_elem->filter->start_transport_op (exec_ctx, next_elem, op);
}
grpc_channel_stack *
grpc_channel_stack_from_top_element (grpc_channel_element * elem)
{
return (grpc_channel_stack *) ((char *) (elem) - ROUND_UP_TO_ALIGNMENT_SIZE (sizeof (grpc_channel_stack)));
}
grpc_call_stack *
grpc_call_stack_from_top_element (grpc_call_element * elem)
{
return (grpc_call_stack *) ((char *) (elem) - ROUND_UP_TO_ALIGNMENT_SIZE (sizeof (grpc_call_stack)));
}
void
grpc_call_element_send_cancel (grpc_exec_ctx * exec_ctx, grpc_call_element * cur_elem)
{
grpc_transport_stream_op op;
memset (&op, 0, sizeof (op));
op.cancel_with_status = GRPC_STATUS_CANCELLED;
grpc_call_next_op (exec_ctx, cur_elem, &op);
}