summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorMaria Matejka <mq@ucw.cz>2022-09-12 11:09:43 +0200
committerMaria Matejka <mq@ucw.cz>2022-09-18 16:33:51 +0200
commit0861d3a8a3af3935d474a12c6227ea2403375b4c (patch)
tree199d958d60c4cdb69b7dc5a1253d32dff70e4a6b
parente3e15f36fa54414e497fab1a054c89c6b1cdd1df (diff)
Fixing several race-conditions in event code.
After a suggestion by Santiago, I added the direct list pointer into events and the events are now using this value to check whether the route is active or not. Also the whole trick with sentinel node unioned with event list is now gone. For debugging, there is also an internal circular buffer to store what has been recently happening in event code before e.g. a crash happened. By default, this debug is off and must be manually enabled in lib/event.c as it eats quite some time and space.
-rw-r--r--lib/event.c141
-rw-r--r--lib/event.h30
2 files changed, 123 insertions, 48 deletions
diff --git a/lib/event.c b/lib/event.c
index 07d7dc53..68ee4c06 100644
--- a/lib/event.c
+++ b/lib/event.c
@@ -28,7 +28,50 @@
event_list global_event_list;
event_list global_work_list;
-STATIC_ASSERT(OFFSETOF(event_list, _sentinel.next) >= OFFSETOF(event_list, _end[0]));
+//#ifdef DEBUGGING
+#if 0
+#define EDL_MAX 16384
+enum edl_caller {
+ EDL_REMOVE_FROM = 1,
+ EDL_POSTPONE = 2,
+ EDL_RUN = 3,
+ EDL_SEND = 4,
+ EDL_RUN_LIST = 5,
+} caller;
+static struct event_debug_log {
+ event_list *target_list;
+ event *event;
+ event *receiver;
+ uint pos;
+ uint prev_edl_pos;
+ uint thread;
+ enum edl_caller caller;
+} edl[EDL_MAX];
+static _Atomic uint edl_cnt;
+_Thread_local static uint edl_thread;
+_Thread_local static uint prev_edl_pos = ~0;
+static inline void edlog(event_list *list, event *e, event *receiver, uint pos, enum edl_caller caller)
+{
+ uint edl_pos = atomic_fetch_add_explicit(&edl_cnt, 1, memory_order_acq_rel);
+ if (!edl_thread)
+ edl_thread = edl_pos;
+
+ edl[edl_pos % EDL_MAX] = (struct event_debug_log) {
+ .target_list = list,
+ .event = e,
+ .receiver = receiver,
+ .pos = pos,
+ .prev_edl_pos = prev_edl_pos,
+ .thread = edl_thread,
+ .caller = caller,
+ };
+
+ prev_edl_pos = edl_pos;
+}
+#else
+#define edlog(...)
+#endif
+
void
ev_init_list(event_list *el, struct birdloop *loop, const char *name)
@@ -36,9 +79,8 @@ ev_init_list(event_list *el, struct birdloop *loop, const char *name)
el->name = name;
el->loop = loop;
- atomic_store_explicit(&el->receiver, &el->_sentinel, memory_order_relaxed);
- atomic_store_explicit(&el->_executor, &el->_sentinel, memory_order_relaxed);
- atomic_store_explicit(&el->_sentinel.next, NULL, memory_order_relaxed);
+ atomic_store_explicit(&el->receiver, NULL, memory_order_release);
+ atomic_store_explicit(&el->_executor, NULL, memory_order_release);
}
/*
@@ -61,13 +103,20 @@ ev_remove_from(event *e, event * _Atomic * head)
/* The current event in queue to check */
event *cur = atomic_load_explicit(prev, memory_order_acquire);
- /* Pre-loaded next pointer; if NULL, this is sentinel */
- event *next = atomic_load_explicit(&cur->next, memory_order_acquire);
+ /* This part of queue is empty! */
+ if (!cur)
+ return 0;
- while (next)
+ edlog(NULL, e, cur, 1, EDL_REMOVE_FROM);
+ while (cur)
{
+ /* Pre-loaded next pointer */
+ event *next = atomic_load_explicit(&cur->next, memory_order_acquire);
+
if (e == cur)
{
+ edlog(NULL, e, next, 3, EDL_REMOVE_FROM);
+
/* Check whether we have collided with somebody else
* adding an item to the queue. */
if (!atomic_compare_exchange_strong_explicit(
@@ -86,12 +135,15 @@ ev_remove_from(event *e, event * _Atomic * head)
return 1;
}
+ edlog(NULL, e, next, 2, EDL_REMOVE_FROM);
+
/* Go to the next event. */
prev = &cur->next;
cur = next;
- next = atomic_load_explicit(&cur->next, memory_order_acquire);
}
+ edlog(NULL, e, cur, 4, EDL_REMOVE_FROM);
+
return 0;
}
@@ -100,6 +152,7 @@ ev_postpone(event *e)
{
/* Find the list to remove the event from */
event_list *sl = ev_get_list(e);
+ edlog(sl, e, NULL, 1, EDL_POSTPONE);
if (!sl)
return;
@@ -108,6 +161,10 @@ ev_postpone(event *e)
/* Remove from one of these lists. */
ASSERT(ev_remove_from(e, &sl->_executor) || ev_remove_from(e, &sl->receiver));
+
+ /* Mark as inactive */
+ ASSERT_DIE(sl == atomic_exchange_explicit(&e->list, NULL, memory_order_acq_rel));
+ edlog(sl, e, NULL, 2, EDL_POSTPONE);
}
static void
@@ -157,8 +214,10 @@ ev_new(pool *p)
inline void
ev_run(event *e)
{
+ edlog(NULL, e, NULL, 1, EDL_RUN);
ev_postpone(e);
e->hook(e->data);
+ edlog(NULL, e, NULL, 2, EDL_RUN);
}
/**
@@ -172,18 +231,36 @@ ev_run(event *e)
inline void
ev_send(event_list *l, event *e)
{
- event_list *sl = ev_get_list(e);
- if (sl == l)
- return;
- if (sl)
- bug("Queuing an already queued event to another queue is not supported.");
-
+ edlog(l, e, NULL, 1, EDL_SEND);
+ /* Set the target list */
+ event_list *ol = NULL;
+ if (!atomic_compare_exchange_strong_explicit(
+ &e->list, &ol, l,
+ memory_order_acq_rel, memory_order_acquire))
+ if (ol == l)
+ return;
+ else
+ bug("Queuing an already queued event to another queue is not supported.");
+
+ /* Here should be no concurrent senders */
event *next = atomic_load_explicit(&l->receiver, memory_order_acquire);
- do atomic_store_explicit(&e->next, next, memory_order_relaxed);
+ edlog(l, e, next, 2, EDL_SEND);
+ event *old_next = NULL;
+ do
+ if (!atomic_compare_exchange_strong_explicit(
+ &e->next, &old_next, next,
+ memory_order_acq_rel, memory_order_acquire))
+ bug("Event %p in inconsistent state");
+ else
+ {
+ old_next = next;
+ edlog(l, old_next, next, 3, EDL_SEND);
+ }
while (!atomic_compare_exchange_strong_explicit(
&l->receiver, &next, e,
memory_order_acq_rel, memory_order_acquire));
+ edlog(l, e, next, 4, EDL_SEND);
birdloop_ping(l->loop);
}
@@ -199,37 +276,41 @@ int
ev_run_list_limited(event_list *l, uint limit)
{
event * _Atomic *ep = &l->_executor;
+ edlog(l, NULL, NULL, 1, EDL_RUN_LIST);
/* No pending events, refill the queue. */
- if (atomic_load_explicit(ep, memory_order_relaxed) == &l->_sentinel)
+ if (!atomic_load_explicit(ep, memory_order_acquire))
{
/* Move the current event list aside and create a new one. */
- event *received = atomic_exchange_explicit(
- &l->receiver, &l->_sentinel, memory_order_acq_rel);
+ event *received = atomic_exchange_explicit(&l->receiver, NULL, memory_order_acq_rel);
+ edlog(l, NULL, received, 2, EDL_RUN_LIST);
/* No event to run. */
- if (received == &l->_sentinel)
+ if (!received)
return 0;
/* Setup the executor queue */
- event *head = &l->_sentinel;
+ event *head = NULL;
/* Flip the order of the events by relinking them one by one (push-pop) */
- while (received != &l->_sentinel)
+ while (received)
{
event *cur = received;
- received = atomic_exchange_explicit(&cur->next, head, memory_order_relaxed);
+ received = atomic_exchange_explicit(&cur->next, head, memory_order_acq_rel);
+ edlog(l, head, received, 3, EDL_RUN_LIST);
head = cur;
}
/* Store the executor queue to its designated place */
- atomic_store_explicit(ep, head, memory_order_relaxed);
+ ASSERT_DIE(atomic_exchange_explicit(ep, head, memory_order_acq_rel) == NULL);
+ edlog(l, NULL, head, 4, EDL_RUN_LIST);
}
/* Run the events in order. */
event *e;
- while ((e = atomic_load_explicit(ep, memory_order_relaxed)) != &l->_sentinel)
+ while (e = atomic_load_explicit(ep, memory_order_acquire))
{
+ edlog(l, e, NULL, 5, EDL_RUN_LIST);
/* Check limit */
if (!--limit)
return 1;
@@ -238,14 +319,20 @@ ev_run_list_limited(event_list *l, uint limit)
if ((l == &global_event_list) || (l == &global_work_list))
io_log_event(e->hook, e->data);
+ edlog(l, e, NULL, 6, EDL_RUN_LIST);
/* Inactivate the event */
- atomic_store_explicit(ep, atomic_load_explicit(&e->next, memory_order_relaxed), memory_order_relaxed);
- atomic_store_explicit(&e->next, NULL, memory_order_relaxed);
+ event *next = atomic_load_explicit(&e->next, memory_order_relaxed);
+ ASSERT_DIE(e == atomic_exchange_explicit(ep, next, memory_order_acq_rel));
+ ASSERT_DIE(next == atomic_exchange_explicit(&e->next, NULL, memory_order_acq_rel));
+ ASSERT_DIE(l == atomic_exchange_explicit(&e->list, NULL, memory_order_acq_rel));
+ edlog(l, e, next, 7, EDL_RUN_LIST);
/* Run the event */
e->hook(e->data);
tmp_flush();
+
+ edlog(l, e, next, 8, EDL_RUN_LIST);
}
- return atomic_load_explicit(&l->receiver, memory_order_relaxed) != &l->_sentinel;
+ return !!atomic_load_explicit(&l->receiver, memory_order_acquire);
}
diff --git a/lib/event.h b/lib/event.h
index 9773c3a9..0bef737a 100644
--- a/lib/event.h
+++ b/lib/event.h
@@ -11,6 +11,7 @@
#include "lib/resource.h"
#include "lib/locking.h"
+#include "lib/rcu.h"
#include <stdatomic.h>
@@ -21,17 +22,14 @@ typedef struct event {
void (*hook)(void *);
void *data;
struct event * _Atomic next;
+ struct event_list * _Atomic list;
} event;
-typedef union event_list {
- struct {
- event * _Atomic receiver; /* Event receive list */
- event * _Atomic _executor; /* Event execute list */
- const char *name;
- struct birdloop *loop; /* The executor loop */
- char _end[0];
- };
- event _sentinel; /* Sentinel node to actively detect list end */
+typedef struct event_list {
+ event * _Atomic receiver; /* Event receive list */
+ event * _Atomic _executor; /* Event execute list */
+ const char *name;
+ struct birdloop *loop; /* The executor loop */
} event_list;
extern event_list global_event_list;
@@ -56,23 +54,13 @@ int ev_run_list_limited(event_list *, uint);
static inline int
ev_active(event *e)
{
- return atomic_load_explicit(&e->next, memory_order_relaxed) != NULL;
+ return atomic_load_explicit(&e->list, memory_order_acquire) != NULL;
}
static inline event_list *
ev_get_list(event *e)
{
- /* We are looking for the sentinel node at the list end.
- * After this, we have s->next == NULL */
- event *s = e;
- for (event *sn; sn = atomic_load_explicit(&s->next, memory_order_acquire); s = sn)
- ;
-
- /* No sentinel, no list. */
- if (s == e)
- return NULL;
- else
- return SKIP_BACK(event_list, _sentinel, s);
+ return atomic_load_explicit(&e->list, memory_order_acquire);
}
static inline event*