summaryrefslogtreecommitdiff
path: root/src/ipcpd/unicast/routing
diff options
context:
space:
mode:
Diffstat (limited to 'src/ipcpd/unicast/routing')
-rw-r--r--src/ipcpd/unicast/routing/graph.c833
-rw-r--r--src/ipcpd/unicast/routing/graph.h69
-rw-r--r--src/ipcpd/unicast/routing/link-state.c1127
-rw-r--r--src/ipcpd/unicast/routing/link-state.h46
-rw-r--r--src/ipcpd/unicast/routing/ops.h43
-rw-r--r--src/ipcpd/unicast/routing/pol.h23
-rw-r--r--src/ipcpd/unicast/routing/tests/CMakeLists.txt34
-rw-r--r--src/ipcpd/unicast/routing/tests/graph_test.c351
8 files changed, 2526 insertions, 0 deletions
diff --git a/src/ipcpd/unicast/routing/graph.c b/src/ipcpd/unicast/routing/graph.c
new file mode 100644
index 00000000..0226c762
--- /dev/null
+++ b/src/ipcpd/unicast/routing/graph.c
@@ -0,0 +1,833 @@
+/*
+ * Ouroboros - Copyright (C) 2016 - 2026
+ *
+ * Undirected graph structure
+ *
+ * Dimitri Staessens <dimitri@ouroboros.rocks>
+ * Sander Vrijders <sander@ouroboros.rocks>
+ * Nick Aerts <nick.aerts@ugent.be>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License version 2 as
+ * published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., http://www.fsf.org/about/contact/.
+ */
+
+#if defined(__linux__) || defined(__CYGWIN__)
+#define _DEFAULT_SOURCE
+#else
+#define _POSIX_C_SOURCE 200112L
+#endif
+
+#define OUROBOROS_PREFIX "graph"
+
+#include <ouroboros/logs.h>
+#include <ouroboros/errno.h>
+#include <ouroboros/list.h>
+
+#include "graph.h"
+#include "ipcp.h"
+
+#include <assert.h>
+#include <pthread.h>
+#include <stdlib.h>
+#include <limits.h>
+#include <string.h>
+
+struct vertex {
+ struct list_head next;
+ uint64_t addr;
+ struct list_head edges;
+ int index;
+};
+
+struct edge {
+ struct list_head next;
+ struct vertex * nb;
+ qosspec_t qs;
+ int announced;
+};
+
+struct graph {
+ struct llist vertices;
+
+ pthread_mutex_t lock;
+};
+
+static struct edge * find_edge_by_addr(struct vertex * vertex,
+ uint64_t dst_addr)
+{
+ struct list_head * p;
+
+ assert(vertex != NULL);
+
+ list_for_each(p, &vertex->edges) {
+ struct edge * e = list_entry(p, struct edge, next);
+ if (e->nb->addr == dst_addr)
+ return e;
+ }
+
+ return NULL;
+}
+
+static struct vertex * find_vertex_by_addr(struct graph * graph,
+ uint64_t addr)
+{
+ struct list_head * p;
+
+ assert(graph);
+
+ llist_for_each(p, &graph->vertices) {
+ struct vertex * e = list_entry(p, struct vertex, next);
+ if (e->addr == addr)
+ return e;
+ }
+
+ return NULL;
+}
+
+static struct edge * add_edge(struct vertex * vertex,
+ struct vertex * nb)
+{
+ struct edge * edge;
+
+ assert(vertex != NULL);
+ assert(nb != NULL);
+
+ edge = malloc(sizeof(*edge));
+ if (edge == NULL)
+ return NULL;
+
+ edge->nb = nb;
+ edge->announced = 0;
+
+ list_add(&edge->next, &vertex->edges);
+
+ return edge;
+}
+
+static void del_edge(struct edge * edge)
+{
+ assert(edge);
+
+ list_del(&edge->next);
+ free(edge);
+}
+
+static struct vertex * add_vertex(struct graph * graph,
+ uint64_t addr)
+{
+ struct vertex * vertex;
+ struct list_head * p;
+ int i = 0;
+
+ assert(graph);
+
+ vertex = malloc(sizeof(*vertex));
+ if (vertex == NULL)
+ return NULL;
+
+ list_head_init(&vertex->edges);
+ vertex->addr = addr;
+
+ /* Keep them ordered on address. */
+ llist_for_each(p, &graph->vertices) {
+ struct vertex * v = list_entry(p, struct vertex, next);
+ if (v->addr > addr)
+ break;
+ i++;
+ }
+
+ vertex->index = i;
+
+ llist_add_tail_at(&vertex->next, p, &graph->vertices);
+
+ /* Increase the index of the vertices to the right. */
+ list_for_each(p, &vertex->next) {
+ struct vertex * v = list_entry(p, struct vertex, next);
+ if (v->addr > addr)
+ v->index++;
+ }
+
+ return vertex;
+}
+
+static void free_edges(struct list_head * edges)
+{
+ struct list_head * p;
+ struct list_head * h;
+
+ list_for_each_safe(p, h, edges) {
+ struct edge * e = list_entry(p, struct edge, next);
+ list_del(&e->next);
+ free(e);
+ }
+}
+
+static void del_vertex(struct graph * graph,
+ struct vertex * vertex)
+{
+ struct list_head * p;
+
+ assert(graph != NULL);
+ assert(vertex != NULL);
+
+ llist_del(&vertex->next, &graph->vertices);
+
+ /* Decrease the index of the vertices to the right. */
+ llist_for_each(p, &graph->vertices) {
+ struct vertex * v = list_entry(p, struct vertex, next);
+ if (v->addr > vertex->addr)
+ v->index--;
+ }
+
+ free_edges(&vertex->edges);
+
+ free(vertex);
+}
+
+struct graph * graph_create(void)
+{
+ struct graph * graph;
+
+ graph = malloc(sizeof(*graph));
+ if (graph == NULL)
+ return NULL;
+
+ if (pthread_mutex_init(&graph->lock, NULL)) {
+ free(graph);
+ return NULL;
+ }
+
+ llist_init(&graph->vertices);
+
+ return graph;
+}
+
+void graph_destroy(struct graph * graph)
+{
+ struct list_head * p = NULL;
+ struct list_head * n = NULL;
+
+ assert(graph);
+
+ pthread_mutex_lock(&graph->lock);
+
+ llist_for_each_safe(p, n, &graph->vertices) {
+ struct vertex * e = list_entry(p, struct vertex, next);
+ del_vertex(graph, e);
+ }
+
+ pthread_mutex_unlock(&graph->lock);
+
+ pthread_mutex_destroy(&graph->lock);
+
+ assert(llist_is_empty(&graph->vertices));
+
+ free(graph);
+}
+
+int graph_update_edge(struct graph * graph,
+ uint64_t s_addr,
+ uint64_t d_addr,
+ qosspec_t qs)
+{
+ struct vertex * v;
+ struct edge * e;
+ struct vertex * nb;
+ struct edge * nb_e;
+
+ assert(graph != NULL);
+
+ pthread_mutex_lock(&graph->lock);
+
+ v = find_vertex_by_addr(graph, s_addr);
+ if (v == NULL && ((v = add_vertex(graph, s_addr)) == NULL)) {;
+ log_err("Failed to add src vertex.");
+ goto fail_add_s;
+ }
+
+ nb = find_vertex_by_addr(graph, d_addr);
+ if (nb == NULL && ((nb = add_vertex(graph, d_addr)) == NULL)) {
+ log_err("Failed to add dst vertex.");
+ goto fail_add_d;
+ }
+
+ e = find_edge_by_addr(v, d_addr);
+ if (e == NULL && ((e = add_edge(v, nb)) == NULL)) {
+ log_err("Failed to add edge to dst.");
+ goto fail_add_edge_d;
+ }
+
+ e->announced++;
+ e->qs = qs;
+
+ nb_e = find_edge_by_addr(nb, s_addr);
+ if (nb_e == NULL && ((nb_e = add_edge(nb, v)) == NULL)) {;
+ log_err("Failed to add edge to src.");
+ goto fail_add_edge_s;
+ }
+
+ nb_e->announced++;
+ nb_e->qs = qs;
+
+ pthread_mutex_unlock(&graph->lock);
+
+ return 0;
+ fail_add_edge_s:
+ if (--e->announced == 0)
+ del_edge(e);
+ fail_add_edge_d:
+ if (list_is_empty(&nb->edges))
+ del_vertex(graph, nb);
+ fail_add_d:
+ if (list_is_empty(&v->edges))
+ del_vertex(graph, v);
+ fail_add_s:
+ pthread_mutex_unlock(&graph->lock);
+ return -ENOMEM;
+
+}
+
+int graph_del_edge(struct graph * graph,
+ uint64_t s_addr,
+ uint64_t d_addr)
+{
+ struct vertex * v;
+ struct edge * e;
+ struct vertex * nb;
+ struct edge * nb_e;
+
+ assert(graph);
+
+ pthread_mutex_lock(&graph->lock);
+
+ v = find_vertex_by_addr(graph, s_addr);
+ if (v == NULL) {
+ log_err("Failed to find src vertex.");
+ goto fail;
+ }
+
+ nb = find_vertex_by_addr(graph, d_addr);
+ if (nb == NULL) {
+ log_err("No such destination vertex.");
+ goto fail;
+ }
+
+ e = find_edge_by_addr(v, d_addr);
+ if (e == NULL) {
+ log_err("No such source edge.");
+ goto fail;
+ }
+
+ nb_e = find_edge_by_addr(nb, s_addr);
+ if (nb_e == NULL) {
+ log_err("No such destination edge.");
+ goto fail;
+ }
+
+ if (--e->announced == 0)
+ del_edge(e);
+ if (--nb_e->announced == 0)
+ del_edge(nb_e);
+
+ /* Removing vertex if it was the last edge */
+ if (list_is_empty(&v->edges))
+ del_vertex(graph, v);
+ if (list_is_empty(&nb->edges))
+ del_vertex(graph, nb);
+
+ pthread_mutex_unlock(&graph->lock);
+
+ return 0;
+
+ fail:
+ pthread_mutex_unlock(&graph->lock);
+ return -1;
+}
+
+static int get_min_vertex(struct graph * graph,
+ int * dist,
+ bool * used,
+ struct vertex ** v)
+{
+ int min = INT_MAX;
+ int index = -1;
+ int i = 0;
+ struct list_head * p;
+
+ assert(v);
+ assert(graph);
+ assert(dist);
+ assert(used);
+
+ *v = NULL;
+
+ llist_for_each(p, &graph->vertices) {
+ if (!used[i] && dist[i] < min) {
+ min = dist[i];
+ index = i;
+ *v = list_entry(p, struct vertex, next);
+ }
+
+ i++;
+ }
+
+ if (index != -1)
+ used[index] = true;
+
+ return index;
+}
+
+static int dijkstra(struct graph * graph,
+ uint64_t src,
+ struct vertex *** nhops,
+ int ** dist)
+{
+ bool * used;
+ struct list_head * p = NULL;
+ int i = 0;
+ struct vertex * v = NULL;
+ struct edge * e = NULL;
+ int alt;
+
+ assert(graph);
+ assert(nhops);
+ assert(dist);
+
+ *nhops = malloc(sizeof(**nhops) * graph->vertices.len);
+ if (*nhops == NULL)
+ goto fail_pnhops;
+
+ *dist = malloc(sizeof(**dist) * graph->vertices.len);
+ if (*dist == NULL)
+ goto fail_pdist;
+
+ used = malloc(sizeof(*used) * graph->vertices.len);
+ if (used == NULL)
+ goto fail_used;
+
+ /* Init the data structures */
+ memset(used, 0, sizeof(*used) * graph->vertices.len);
+ memset(*nhops, 0, sizeof(**nhops) * graph->vertices.len);
+ memset(*dist, 0, sizeof(**dist) * graph->vertices.len);
+
+ llist_for_each(p, &graph->vertices) {
+ v = list_entry(p, struct vertex, next);
+ (*dist)[i++] = (v->addr == src) ? 0 : INT_MAX;
+ }
+
+ /* Perform actual Dijkstra */
+ i = get_min_vertex(graph, *dist, used, &v);
+ while (v != NULL) {
+ list_for_each(p, &v->edges) {
+ e = list_entry(p, struct edge, next);
+
+ /* Only include it if both sides announced it. */
+ if (e->announced != 2)
+ continue;
+
+ /*
+ * NOTE: Current weight is just hop count.
+ * Method could be extended to use a different
+ * weight for a different QoS cube.
+ */
+ alt = (*dist)[i] + 1;
+ if (alt < (*dist)[e->nb->index]) {
+ (*dist)[e->nb->index] = alt;
+ if (v->addr == src)
+ (*nhops)[e->nb->index] = e->nb;
+ else
+ (*nhops)[e->nb->index] = (*nhops)[i];
+ }
+ }
+ i = get_min_vertex(graph, *dist, used, &v);
+ }
+
+ free(used);
+
+ return 0;
+
+ fail_used:
+ free(*dist);
+ fail_pdist:
+ free(*nhops);
+ fail_pnhops:
+ return -1;
+
+}
+
+static void free_routing_table(struct list_head * table)
+{
+ struct list_head * h;
+ struct list_head * p;
+ struct list_head * q;
+ struct list_head * i;
+
+ assert(table);
+
+ list_for_each_safe(p, h, table) {
+ struct routing_table * t =
+ list_entry(p, struct routing_table, next);
+ list_for_each_safe(q, i, &t->nhops) {
+ struct nhop * n =
+ list_entry(q, struct nhop, next);
+ list_del(&n->next);
+ free(n);
+ }
+ list_del(&t->next);
+ free(t);
+ }
+}
+
+void graph_free_routing_table(struct graph * graph,
+ struct list_head * table)
+{
+ assert(table);
+
+ pthread_mutex_lock(&graph->lock);
+
+ free_routing_table(table);
+
+ pthread_mutex_unlock(&graph->lock);
+}
+
+static int graph_routing_table_simple(struct graph * graph,
+ uint64_t s_addr,
+ struct list_head * table,
+ int ** dist)
+{
+ struct vertex ** nhops;
+ struct list_head * p;
+ int i = 0;
+ struct vertex * v;
+ struct routing_table * t;
+ struct nhop * n;
+
+ assert(graph);
+ assert(table);
+ assert(dist);
+
+ /* We need at least 2 vertices for a table */
+ if (graph->vertices.len < 2)
+ goto fail_vertices;
+
+ if (dijkstra(graph, s_addr, &nhops, dist))
+ goto fail_vertices;
+
+ list_head_init(table);
+
+ /* Now construct the routing table from the nhops. */
+ llist_for_each(p, &graph->vertices) {
+ v = list_entry(p, struct vertex, next);
+
+ /* This is the src */
+ if (nhops[i] == NULL) {
+ i++;
+ continue;
+ }
+
+ t = malloc(sizeof(*t));
+ if (t == NULL)
+ goto fail_t;
+
+ list_head_init(&t->nhops);
+
+ n = malloc(sizeof(*n));
+ if (n == NULL)
+ goto fail_n;
+
+ t->dst = v->addr;
+ n->nhop = nhops[i]->addr;
+
+ list_add(&n->next, &t->nhops);
+ list_add(&t->next, table);
+
+ i++;
+ }
+
+ free(nhops);
+
+ return 0;
+
+ fail_n:
+ free(t);
+ fail_t:
+ free_routing_table(table);
+ free(nhops);
+ free(*dist);
+ fail_vertices:
+ *dist = NULL;
+ return -1;
+}
+
+static int add_lfa_to_table(struct list_head * table,
+ uint64_t addr,
+ uint64_t lfa)
+{
+ struct list_head * p;
+ struct nhop * n;
+
+ assert(table);
+
+ n = malloc(sizeof(*n));
+ if (n == NULL)
+ return -1;
+
+ n->nhop = lfa;
+
+ list_for_each(p, table) {
+ struct routing_table * t =
+ list_entry(p, struct routing_table, next);
+ if (t->dst == addr) {
+ list_add_tail(&n->next, &t->nhops);
+ return 0;
+ }
+ }
+
+ free(n);
+
+ return -1;
+}
+
+static int graph_routing_table_lfa(struct graph * graph,
+ uint64_t s_addr,
+ struct list_head * table,
+ int ** dist)
+{
+ int * n_dist[PROG_MAX_FLOWS];
+ uint64_t addrs[PROG_MAX_FLOWS];
+ int n_index[PROG_MAX_FLOWS];
+ struct list_head * p;
+ struct list_head * q;
+ struct vertex * v;
+ struct edge * e;
+ struct vertex ** nhops;
+ int i = 0;
+ int j;
+ int k;
+
+ if (graph_routing_table_simple(graph, s_addr, table, dist))
+ goto fail_table;
+
+ for (j = 0; j < PROG_MAX_FLOWS; j++) {
+ n_dist[j] = NULL;
+ n_index[j] = -1;
+ addrs[j] = -1;
+ }
+
+ llist_for_each(p, &graph->vertices) {
+ v = list_entry(p, struct vertex, next);
+
+ if (v->addr != s_addr)
+ continue;
+
+ /*
+ * Get the distances for every neighbor
+ * of the source.
+ */
+ list_for_each(q, &v->edges) {
+ e = list_entry(q, struct edge, next);
+
+ addrs[i] = e->nb->addr;
+ n_index[i] = e->nb->index;
+ if (dijkstra(graph, e->nb->addr,
+ &nhops, &(n_dist[i++])))
+ goto fail_dijkstra;
+
+ free(nhops);
+ }
+
+ break;
+ }
+
+ /* Loop though all nodes to see if we have a LFA for them. */
+ llist_for_each(p, &graph->vertices) {
+ v = list_entry(p, struct vertex, next);
+
+ if (v->addr == s_addr)
+ continue;
+
+ /*
+ * Check for every neighbor if
+ * dist(neighbor, destination) <
+ * dist(neighbor, source) + dist(source, destination).
+ */
+ for (j = 0; j < i; j++) {
+ /* Exclude ourselves. */
+ if (addrs[j] == v->addr)
+ continue;
+
+ if (n_dist[j][v->index] <
+ (*dist)[n_index[j]] + (*dist)[v->index])
+ if (add_lfa_to_table(table, v->addr,
+ addrs[j]))
+ goto fail_add_lfa;
+ }
+ }
+
+ for (j = 0; j < i; j++)
+ free(n_dist[j]);
+
+ return 0;
+
+ fail_add_lfa:
+ for (k = j; k < i; k++)
+ free(n_dist[k]);
+ fail_dijkstra:
+ free_routing_table(table);
+ fail_table:
+ return -1;
+}
+
+static int graph_routing_table_ecmp(struct graph * graph,
+ uint64_t s_addr,
+ struct list_head * table,
+ int ** dist)
+{
+ struct vertex ** nhops;
+ struct list_head * p;
+ size_t i;
+ struct vertex * v;
+ struct vertex * src_v;
+ struct edge * e;
+ struct routing_table * t;
+ struct nhop * n;
+ struct list_head * forwarding;
+
+ assert(graph);
+ assert(dist);
+
+ if (graph->vertices.len < 2)
+ goto fail_vertices;
+
+ forwarding = malloc(sizeof(*forwarding) * graph->vertices.len);
+ if (forwarding == NULL)
+ goto fail_vertices;
+
+ for (i = 0; i < graph->vertices.len; ++i)
+ list_head_init(&forwarding[i]);
+
+ if (dijkstra(graph, s_addr, &nhops, dist))
+ goto fail_dijkstra;
+
+ free(nhops);
+
+ src_v = find_vertex_by_addr(graph, s_addr);
+ if (src_v == NULL)
+ goto fail_src_v;
+
+ list_for_each(p, &src_v->edges) {
+ int * tmp_dist;
+
+ e = list_entry(p, struct edge, next);
+ if (dijkstra(graph, e->nb->addr, &nhops, &tmp_dist))
+ goto fail_src_v;
+
+ free(nhops);
+
+ for (i = 0; i < graph->vertices.len; ++i) {
+ if (tmp_dist[i] + 1 == (*dist)[i]) {
+ n = malloc(sizeof(*n));
+ if (n == NULL) {
+ free(tmp_dist);
+ goto fail_src_v;
+ }
+ n->nhop = e->nb->addr;
+ list_add_tail(&n->next, &forwarding[i]);
+ }
+ }
+
+ free(tmp_dist);
+ }
+
+ list_head_init(table);
+ i = 0;
+ llist_for_each(p, &graph->vertices) {
+ v = list_entry(p, struct vertex, next);
+ if (v->addr == s_addr || list_is_empty(&forwarding[i])) {
+ ++i;
+ continue;
+ }
+
+ t = malloc(sizeof(*t));
+ if (t == NULL)
+ goto fail_malloc;
+
+ t->dst = v->addr;
+
+ list_head_init(&t->nhops);
+ t->nhops.nxt = forwarding[i].nxt;
+ t->nhops.prv = forwarding[i].prv;
+ forwarding[i].prv->nxt = &t->nhops;
+ forwarding[i].nxt->prv = &t->nhops;
+
+ list_add(&t->next, table);
+ ++i;
+ }
+
+ free(forwarding);
+
+ return 0;
+
+ fail_malloc:
+ free_routing_table(table);
+ fail_src_v:
+ free(*dist);
+ fail_dijkstra:
+ free(forwarding);
+ fail_vertices:
+ *dist = NULL;
+ return -1;
+}
+
+int graph_routing_table(struct graph * graph,
+ enum routing_algo algo,
+ uint64_t s_addr,
+ struct list_head * table)
+{
+ int * s_dist;
+
+ assert(graph);
+ assert(table);
+
+ pthread_mutex_lock(&graph->lock);
+
+ switch (algo) {
+ case ROUTING_SIMPLE:
+ /* LFA uses the s_dist this returns. */
+ if (graph_routing_table_simple(graph, s_addr, table, &s_dist))
+ goto fail_table;
+ break;
+ case ROUTING_LFA:
+ if (graph_routing_table_lfa(graph, s_addr, table, &s_dist))
+ goto fail_table;
+ break;
+
+ case ROUTING_ECMP:
+ if (graph_routing_table_ecmp(graph, s_addr, table, &s_dist))
+ goto fail_table;
+ break;
+ default:
+ log_err("Unsupported algorithm.");
+ goto fail_table;
+ }
+
+ pthread_mutex_unlock(&graph->lock);
+
+ free(s_dist);
+
+ return 0;
+
+ fail_table:
+ pthread_mutex_unlock(&graph->lock);
+ return -1;
+}
diff --git a/src/ipcpd/unicast/routing/graph.h b/src/ipcpd/unicast/routing/graph.h
new file mode 100644
index 00000000..f3766771
--- /dev/null
+++ b/src/ipcpd/unicast/routing/graph.h
@@ -0,0 +1,69 @@
+/*
+ * Ouroboros - Copyright (C) 2016 - 2026
+ *
+ * Undirected graph structure
+ *
+ * Dimitri Staessens <dimitri@ouroboros.rocks>
+ * Sander Vrijders <sander@ouroboros.rocks>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License version 2 as
+ * published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., http://www.fsf.org/about/contact/.
+ */
+
+#ifndef OUROBOROS_IPCPD_UNICAST_GRAPH_H
+#define OUROBOROS_IPCPD_UNICAST_GRAPH_H
+
+#include <ouroboros/list.h>
+#include <ouroboros/qos.h>
+
+#include <inttypes.h>
+
+enum routing_algo {
+ ROUTING_SIMPLE = 0,
+ ROUTING_LFA,
+ ROUTING_ECMP
+};
+
+struct nhop {
+ struct list_head next;
+ uint64_t nhop;
+};
+
+struct routing_table {
+ struct list_head next;
+ uint64_t dst;
+ struct list_head nhops;
+};
+
+struct graph * graph_create(void);
+
+void graph_destroy(struct graph * graph);
+
+int graph_update_edge(struct graph * graph,
+ uint64_t s_addr,
+ uint64_t d_addr,
+ qosspec_t qs);
+
+int graph_del_edge(struct graph * graph,
+ uint64_t s_addr,
+ uint64_t d_addr);
+
+int graph_routing_table(struct graph * graph,
+ enum routing_algo algo,
+ uint64_t s_addr,
+ struct list_head * table);
+
+void graph_free_routing_table(struct graph * graph,
+ struct list_head * table);
+
+#endif /* OUROBOROS_IPCPD_UNICAST_GRAPH_H */
diff --git a/src/ipcpd/unicast/routing/link-state.c b/src/ipcpd/unicast/routing/link-state.c
new file mode 100644
index 00000000..051dd98d
--- /dev/null
+++ b/src/ipcpd/unicast/routing/link-state.c
@@ -0,0 +1,1127 @@
+/*
+ * Ouroboros - Copyright (C) 2016 - 2026
+ *
+ * Link state routing policy
+ *
+ * Dimitri Staessens <dimitri@ouroboros.rocks>
+ * Sander Vrijders <sander@ouroboros.rocks>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License version 2 as
+ * published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., http://www.fsf.org/about/contact/.
+ */
+
+#if defined(__linux__) || defined(__CYGWIN__)
+#define _DEFAULT_SOURCE
+#else
+#define _POSIX_C_SOURCE 200112L
+#endif
+
+#include "config.h"
+
+#define OUROBOROS_PREFIX "link-state-routing"
+
+#include <ouroboros/endian.h>
+#include <ouroboros/dev.h>
+#include <ouroboros/errno.h>
+#include <ouroboros/fccntl.h>
+#include <ouroboros/fqueue.h>
+#include <ouroboros/list.h>
+#include <ouroboros/logs.h>
+#include <ouroboros/notifier.h>
+#include <ouroboros/pthread.h>
+#include <ouroboros/rib.h>
+#include <ouroboros/utils.h>
+
+#include "addr-auth.h"
+#include "common/comp.h"
+#include "common/connmgr.h"
+#include "graph.h"
+#include "ipcp.h"
+#include "link-state.h"
+#include "pff.h"
+
+#include <assert.h>
+#include <stdlib.h>
+#include <inttypes.h>
+#include <string.h>
+
+#define LS_ENTRY_SIZE 104
+#define lsdb "lsdb"
+
+#ifndef CLOCK_REALTIME_COARSE
+#define CLOCK_REALTIME_COARSE CLOCK_REALTIME
+#endif
+
+#define LINK_FMT ADDR_FMT32 "--" ADDR_FMT32
+#define LINK_VAL(src, dst) ADDR_VAL32(&src), ADDR_VAL32(&dst)
+
+#define LSU_FMT "LSU ["ADDR_FMT32 " -- " ADDR_FMT32 " seq: %09" PRIu64 "]"
+#define LSU_VAL(src, dst, seqno) ADDR_VAL32(&src), ADDR_VAL32(&dst), seqno
+
+struct lsa {
+ uint64_t d_addr;
+ uint64_t s_addr;
+ uint64_t seqno;
+} __attribute__((packed));
+
+struct routing_i {
+ struct list_head next;
+
+ struct pff * pff;
+ pthread_t calculator;
+
+ bool modified;
+ pthread_mutex_t lock;
+};
+
+/* TODO: link weight support. */
+struct adjacency {
+ struct list_head next;
+
+ uint64_t dst;
+ uint64_t src;
+
+ uint64_t seqno;
+
+ time_t stamp;
+};
+
+enum nb_type {
+ NB_DT = 0,
+ NB_MGMT
+};
+
+struct nb {
+ struct list_head next;
+
+ uint64_t addr;
+ int fd;
+ enum nb_type type;
+};
+
+struct {
+ uint64_t addr;
+
+ enum routing_algo routing_algo;
+
+ struct ls_config conf;
+
+ fset_t * mgmt_set;
+
+ struct graph * graph;
+
+ struct {
+ struct llist nbs;
+ struct llist db;
+ pthread_rwlock_t lock;
+ };
+
+ struct {
+ struct list_head list;
+ pthread_mutex_t mtx;
+ } instances;
+
+ pthread_t lsupdate;
+ pthread_t lsreader;
+ pthread_t listener;
+} ls;
+
+struct routing_ops link_state_ops = {
+ .init = (int (*)(void *, enum pol_pff *)) link_state_init,
+ .fini = link_state_fini,
+ .start = link_state_start,
+ .stop = link_state_stop,
+ .routing_i_create = link_state_routing_i_create,
+ .routing_i_destroy = link_state_routing_i_destroy
+};
+
+static int str_adj(struct adjacency * adj,
+ char * buf,
+ size_t len)
+{
+ char tmstr[RIB_TM_STRLEN];
+ char srcbuf[64];
+ char dstbuf[64];
+ char seqnobuf[64];
+ struct tm * tm;
+
+ assert(adj);
+
+ if (len < LS_ENTRY_SIZE)
+ return -1;
+
+ tm = gmtime(&adj->stamp);
+ strftime(tmstr, sizeof(tmstr), RIB_TM_FORMAT, tm);
+
+ sprintf(srcbuf, ADDR_FMT32, ADDR_VAL32(&adj->src));
+ sprintf(dstbuf, ADDR_FMT32, ADDR_VAL32(&adj->dst));
+ sprintf(seqnobuf, "%" PRIu64, adj->seqno);
+
+ sprintf(buf, "src: %20s\ndst: %20s\nseqno: %18s\n"
+ "upd: %s\n",
+ srcbuf, dstbuf, seqnobuf, tmstr);
+
+ return LS_ENTRY_SIZE;
+}
+
+static struct adjacency * get_adj(const char * path)
+{
+ struct list_head * p;
+ char entry[RIB_PATH_LEN + 1];
+
+ assert(path);
+
+ llist_for_each(p, &ls.db) {
+ struct adjacency * a = list_entry(p, struct adjacency, next);
+ sprintf(entry, LINK_FMT, LINK_VAL(a->src, a->dst));
+ if (strcmp(entry, path) == 0)
+ return a;
+ }
+
+ return NULL;
+}
+
+static int lsdb_rib_getattr(const char * path,
+ struct rib_attr * attr)
+{
+ struct adjacency * adj;
+ struct timespec now;
+ char * entry;
+
+ assert(path);
+ assert(attr);
+
+ entry = strstr(path, RIB_SEPARATOR) + 1;
+ assert(entry);
+
+ clock_gettime(CLOCK_REALTIME_COARSE, &now);
+
+ pthread_rwlock_rdlock(&ls.lock);
+
+ adj = get_adj(entry);
+ if (adj != NULL) {
+ attr->mtime = adj->stamp;
+ attr->size = LS_ENTRY_SIZE;
+ } else {
+ attr->mtime = now.tv_sec;
+ attr->size = 0;
+ }
+
+ pthread_rwlock_unlock(&ls.lock);
+
+ return 0;
+}
+
+static int lsdb_rib_read(const char * path,
+ char * buf,
+ size_t len)
+{
+ struct adjacency * a;
+ char * entry;
+ int size;
+
+ assert(path);
+
+ entry = strstr(path, RIB_SEPARATOR) + 1;
+ assert(entry);
+
+ pthread_rwlock_rdlock(&ls.lock);
+
+ if (llist_is_empty(&ls.db) && llist_is_empty(&ls.nbs))
+ goto fail;
+
+ a = get_adj(entry);
+ if (a == NULL)
+ goto fail;
+
+ size = str_adj(a, buf, len);
+ if (size < 0)
+ goto fail;
+
+ pthread_rwlock_unlock(&ls.lock);
+ return size;
+
+ fail:
+ pthread_rwlock_unlock(&ls.lock);
+ return -1;
+}
+
+static int lsdb_rib_readdir(char *** buf)
+{
+ struct list_head * p;
+ char entry[RIB_PATH_LEN + 1];
+ ssize_t idx = 0;
+
+ assert(buf != NULL);
+
+ pthread_rwlock_rdlock(&ls.lock);
+
+ if (llist_is_empty(&ls.db) && llist_is_empty(&ls.nbs)) {
+ *buf = NULL;
+ goto no_entries;
+ }
+
+
+ *buf = malloc(sizeof(**buf) * (ls.db.len + ls.nbs.len));
+ if (*buf == NULL)
+ goto fail_entries;
+
+ llist_for_each(p, &ls.nbs) {
+ struct nb * nb = list_entry(p, struct nb, next);
+ char * str = (nb->type == NB_DT ? ".dt " : ".mgmt ");
+ sprintf(entry, "%s" ADDR_FMT32 , str, ADDR_VAL32(&nb->addr));
+ (*buf)[idx] = malloc(strlen(entry) + 1);
+ if ((*buf)[idx] == NULL)
+ goto fail_entry;
+
+ strcpy((*buf)[idx++], entry);
+ }
+
+ llist_for_each(p, &ls.db) {
+ struct adjacency * a = list_entry(p, struct adjacency, next);
+ sprintf(entry, LINK_FMT, LINK_VAL(a->src, a->dst));
+ (*buf)[idx] = malloc(strlen(entry) + 1);
+ if ((*buf)[idx] == NULL)
+ goto fail_entry;
+
+ strcpy((*buf)[idx++], entry);
+ }
+ no_entries:
+ pthread_rwlock_unlock(&ls.lock);
+
+ return idx;
+
+ fail_entry:
+ while (idx-- > 0)
+ free((*buf)[idx]);
+ free(*buf);
+ fail_entries:
+ pthread_rwlock_unlock(&ls.lock);
+ return -ENOMEM;
+}
+
+static struct rib_ops r_ops = {
+ .read = lsdb_rib_read,
+ .readdir = lsdb_rib_readdir,
+ .getattr = lsdb_rib_getattr
+};
+
+static int lsdb_add_nb(uint64_t addr,
+ int fd,
+ enum nb_type type)
+{
+ struct list_head * p;
+ struct nb * nb;
+
+ pthread_rwlock_wrlock(&ls.lock);
+
+ llist_for_each(p, &ls.nbs) {
+ struct nb * el = list_entry(p, struct nb, next);
+ if (addr > el->addr)
+ break;
+ if (el->addr != addr || el->type != type)
+ continue;
+
+ log_dbg("Already know %s neighbor " ADDR_FMT32 ".",
+ type == NB_DT ? "dt" : "mgmt", ADDR_VAL32(&addr));
+ if (el->fd != fd) {
+ log_warn("Existing neighbor assigned new fd.");
+ el->fd = fd;
+ }
+ pthread_rwlock_unlock(&ls.lock);
+ return -EPERM;
+ }
+
+ nb = malloc(sizeof(*nb));
+ if (nb == NULL) {
+ pthread_rwlock_unlock(&ls.lock);
+ return -ENOMEM;
+ }
+
+ nb->addr = addr;
+ nb->fd = fd;
+ nb->type = type;
+
+ llist_add_tail_at(&nb->next, p, &ls.nbs);
+
+ log_dbg("Type %s neighbor " ADDR_FMT32 " added.",
+ nb->type == NB_DT ? "dt" : "mgmt", ADDR_VAL32(&addr));
+
+ pthread_rwlock_unlock(&ls.lock);
+
+ return 0;
+}
+
+static int lsdb_del_nb(uint64_t addr,
+ int fd)
+{
+ struct list_head * p;
+ struct list_head * h;
+
+ pthread_rwlock_wrlock(&ls.lock);
+
+ llist_for_each_safe(p, h, &ls.nbs) {
+ struct nb * nb = list_entry(p, struct nb, next);
+ if (nb->addr != addr || nb->fd != fd)
+ continue;
+
+ llist_del(&nb->next, &ls.nbs);
+ pthread_rwlock_unlock(&ls.lock);
+ log_dbg("Type %s neighbor " ADDR_FMT32 " deleted.",
+ nb->type == NB_DT ? "dt" : "mgmt", ADDR_VAL32(&addr));
+ free(nb);
+ return 0;
+ }
+
+ pthread_rwlock_unlock(&ls.lock);
+
+ return -EPERM;
+}
+
+static int nbr_to_fd(uint64_t addr)
+{
+ struct list_head * p;
+ int fd;
+
+ pthread_rwlock_rdlock(&ls.lock);
+
+ llist_for_each(p, &ls.nbs) {
+ struct nb * nb = list_entry(p, struct nb, next);
+ if (nb->addr == addr && nb->type == NB_DT) {
+ fd = nb->fd;
+ pthread_rwlock_unlock(&ls.lock);
+ return fd;
+ }
+ }
+
+ pthread_rwlock_unlock(&ls.lock);
+
+ return -1;
+}
+
+static void calculate_pff(struct routing_i * instance)
+{
+ int fd;
+ struct list_head table;
+ struct list_head * p;
+ struct list_head * q;
+ int fds[PROG_MAX_FLOWS];
+
+ assert(instance);
+
+ if (graph_routing_table(ls.graph, ls.routing_algo, ls.addr, &table))
+ return;
+
+ pff_lock(instance->pff);
+
+ pff_flush(instance->pff);
+
+ /* Calculate forwarding table from routing table. */
+ list_for_each(p, &table) {
+ int i = 0;
+ struct routing_table * t =
+ list_entry(p, struct routing_table, next);
+
+ list_for_each(q, &t->nhops) {
+ struct nhop * n = list_entry(q, struct nhop, next);
+
+ fd = nbr_to_fd(n->nhop);
+ if (fd == -1)
+ continue;
+
+ fds[i++] = fd;
+ }
+ if (i > 0)
+ pff_add(instance->pff, t->dst, fds, i);
+ }
+
+ pff_unlock(instance->pff);
+
+ graph_free_routing_table(ls.graph, &table);
+}
+
+static void set_pff_modified(bool calc)
+{
+ struct list_head * p;
+
+ pthread_mutex_lock(&ls.instances.mtx);
+ list_for_each(p, &ls.instances.list) {
+ struct routing_i * inst =
+ list_entry(p, struct routing_i, next);
+ pthread_mutex_lock(&inst->lock);
+ inst->modified = true;
+ pthread_mutex_unlock(&inst->lock);
+ if (calc)
+ calculate_pff(inst);
+ }
+ pthread_mutex_unlock(&ls.instances.mtx);
+}
+
+static int lsdb_add_link(uint64_t src,
+ uint64_t dst,
+ uint64_t seqno,
+ qosspec_t * qs)
+{
+ struct list_head * p;
+ struct adjacency * adj;
+ struct timespec now;
+ int ret = -1;
+
+ assert(qs);
+
+ clock_gettime(CLOCK_REALTIME_COARSE, &now);
+
+ pthread_rwlock_wrlock(&ls.lock);
+
+ llist_for_each(p, &ls.db) {
+ struct adjacency * a = list_entry(p, struct adjacency, next);
+ if (a->dst == dst && a->src == src) {
+ if (a->seqno < seqno) {
+ a->stamp = now.tv_sec;
+ a->seqno = seqno;
+ ret = 0;
+ }
+ pthread_rwlock_unlock(&ls.lock);
+ return ret;
+ }
+
+ if (a->dst > dst || (a->dst == dst && a->src > src))
+ break;
+ }
+
+ adj = malloc(sizeof(*adj));
+ if (adj == NULL) {
+ pthread_rwlock_unlock(&ls.lock);
+ return -ENOMEM;
+ }
+
+ adj->dst = dst;
+ adj->src = src;
+ adj->seqno = seqno;
+ adj->stamp = now.tv_sec;
+
+ llist_add_tail_at(&adj->next, p, &ls.db);
+
+ if (graph_update_edge(ls.graph, src, dst, *qs))
+ log_warn("Failed to add edge to graph.");
+
+ pthread_rwlock_unlock(&ls.lock);
+
+ set_pff_modified(true);
+
+ return 0;
+}
+
+static int lsdb_del_link(uint64_t src,
+ uint64_t dst)
+{
+ struct list_head * p;
+ struct list_head * h;
+
+ pthread_rwlock_wrlock(&ls.lock);
+
+ llist_for_each_safe(p, h, &ls.db) {
+ struct adjacency * a = list_entry(p, struct adjacency, next);
+ if (a->dst == dst && a->src == src) {
+ llist_del(&a->next, &ls.db);
+ if (graph_del_edge(ls.graph, src, dst))
+ log_warn("Failed to delete edge from graph.");
+
+ pthread_rwlock_unlock(&ls.lock);
+ set_pff_modified(false);
+ free(a);
+ return 0;
+ }
+ }
+
+ pthread_rwlock_unlock(&ls.lock);
+
+ return -EPERM;
+}
+
+static void * periodic_recalc_pff(void * o)
+{
+ bool modified;
+ struct routing_i * inst;
+
+ assert(o);
+
+ inst = (struct routing_i *) o;
+
+ while (true) {
+ pthread_mutex_lock(&inst->lock);
+ modified = inst->modified;
+ inst->modified = false;
+ pthread_mutex_unlock(&inst->lock);
+
+ if (modified)
+ calculate_pff(inst);
+
+ sleep(ls.conf.t_recalc);
+ }
+
+ return (void *) 0;
+}
+
+static void send_lsm(uint64_t src,
+ uint64_t dst,
+ uint64_t seqno)
+{
+ struct lsa lsm;
+ struct list_head * p;
+
+ lsm.d_addr = hton64(dst);
+ lsm.s_addr = hton64(src);
+ lsm.seqno = hton64(seqno);
+
+ llist_for_each(p, &ls.nbs) {
+ struct nb * nb = list_entry(p, struct nb, next);
+ if (nb->type != NB_MGMT)
+ continue;
+
+ if (flow_write(nb->fd, &lsm, sizeof(lsm)) < 0)
+ log_err("Failed to send LSM to " ADDR_FMT32,
+ ADDR_VAL32(&nb->addr));
+#ifdef DEBUG_PROTO_LS
+ else
+ log_proto(LSU_FMT " --> " ADDR_FMT32,
+ LSU_VAL(src, dst, seqno),
+ ADDR_VAL32(&nb->addr));
+#endif
+ }
+}
+
+/* replicate the lsdb to a mgmt neighbor */
+static void lsdb_replicate(int fd)
+{
+ struct list_head * p;
+ struct list_head * h;
+ struct list_head copy;
+
+ list_head_init(&copy);
+
+ /* Lock the lsdb, copy the lsms and send outside of lock. */
+ pthread_rwlock_rdlock(&ls.lock);
+
+ llist_for_each(p, &ls.db) {
+ struct adjacency * adj;
+ struct adjacency * cpy;
+ adj = list_entry(p, struct adjacency, next);
+ cpy = malloc(sizeof(*cpy));
+ if (cpy == NULL) {
+ log_warn("Failed to replicate full lsdb.");
+ break;
+ }
+
+ cpy->dst = adj->dst;
+ cpy->src = adj->src;
+ cpy->seqno = adj->seqno;
+
+ list_add_tail(&cpy->next, &copy);
+ }
+
+ pthread_rwlock_unlock(&ls.lock);
+
+ list_for_each_safe(p, h, &copy) {
+ struct lsa lsm;
+ struct adjacency * adj;
+ adj = list_entry(p, struct adjacency, next);
+ lsm.d_addr = hton64(adj->dst);
+ lsm.s_addr = hton64(adj->src);
+ lsm.seqno = hton64(adj->seqno);
+ list_del(&adj->next);
+ free(adj);
+ flow_write(fd, &lsm, sizeof(lsm));
+ }
+}
+
+static void * lsupdate(void * o)
+{
+ struct list_head * p;
+ struct list_head * h;
+ struct timespec now;
+
+ (void) o;
+
+ while (true) {
+ clock_gettime(CLOCK_REALTIME_COARSE, &now);
+
+ pthread_rwlock_wrlock(&ls.lock);
+
+ pthread_cleanup_push(__cleanup_rwlock_unlock, &ls.lock);
+
+ llist_for_each_safe(p, h, &ls.db) {
+ struct adjacency * adj;
+ adj = list_entry(p, struct adjacency, next);
+ if (now.tv_sec > adj->stamp + ls.conf.t_timeo) {
+ llist_del(&adj->next, &ls.db);
+ log_dbg(LINK_FMT " timed out.",
+ LINK_VAL(adj->src, adj->dst));
+ if (graph_del_edge(ls.graph, adj->src,
+ adj->dst))
+ log_err("Failed to del edge.");
+ free(adj);
+ continue;
+ }
+
+ if (adj->src == ls.addr) {
+ adj->seqno++;
+ send_lsm(adj->src, adj->dst, adj->seqno);
+ adj->stamp = now.tv_sec;
+ }
+ }
+
+ pthread_cleanup_pop(true);
+
+ sleep(ls.conf.t_update);
+ }
+
+ return (void *) 0;
+}
+
+static void * ls_conn_handle(void * o)
+{
+ struct conn conn;
+
+ (void) o;
+
+ while (true) {
+ if (connmgr_wait(COMPID_MGMT, &conn)) {
+ log_err("Failed to get next MGMT connection.");
+ continue;
+ }
+
+ /* NOTE: connection acceptance policy could be here. */
+
+ notifier_event(NOTIFY_MGMT_CONN_ADD, &conn);
+ }
+
+ return 0;
+}
+
+
+static void forward_lsm(uint8_t * buf,
+ size_t len,
+ int in_fd)
+{
+ struct list_head * p;
+#ifdef DEBUG_PROTO_LS
+ struct lsa lsm;
+
+ assert(buf);
+ assert(len >= sizeof(struct lsa));
+
+ memcpy(&lsm, buf, sizeof(lsm));
+
+ lsm.s_addr = ntoh64(lsm.s_addr);
+ lsm.d_addr = ntoh64(lsm.d_addr);
+ lsm.seqno = ntoh64(lsm.seqno);
+#endif
+ pthread_rwlock_rdlock(&ls.lock);
+
+ pthread_cleanup_push(__cleanup_rwlock_unlock, &ls.lock);
+
+ llist_for_each(p, &ls.nbs) {
+ struct nb * nb = list_entry(p, struct nb, next);
+ if (nb->type != NB_MGMT || nb->fd == in_fd)
+ continue;
+
+ if (flow_write(nb->fd, buf, len) < 0)
+ log_err("Failed to forward LSM to " ADDR_FMT32,
+ ADDR_VAL32(&nb->addr));
+#ifdef DEBUG_PROTO_LS
+ else
+ log_proto(LSU_FMT " --> " ADDR_FMT32 " [forwarded]",
+ LSU_VAL(lsm.s_addr, lsm.d_addr, lsm.seqno),
+ ADDR_VAL32(&nb->addr));
+#endif
+ }
+
+ pthread_cleanup_pop(true);
+}
+
+static void cleanup_fqueue(void * fq)
+{
+ fqueue_destroy((fqueue_t *) fq);
+}
+
+static void * lsreader(void * o)
+{
+ fqueue_t * fq;
+ int ret;
+ uint8_t buf[sizeof(struct lsa)];
+ int fd;
+ qosspec_t qs;
+ struct lsa msg;
+ size_t len;
+
+ (void) o;
+
+ memset(&qs, 0, sizeof(qs));
+
+ fq = fqueue_create();
+ if (fq == NULL)
+ return (void *) -1;
+
+ pthread_cleanup_push(cleanup_fqueue, fq);
+
+ while (true) {
+ ret = fevent(ls.mgmt_set, fq, NULL);
+ if (ret < 0) {
+ log_warn("Event error: %d.", ret);
+ continue;
+ }
+
+ while ((fd = fqueue_next(fq)) >= 0) {
+ if (fqueue_type(fq) != FLOW_PKT)
+ continue;
+
+ len = flow_read(fd, buf, sizeof(msg));
+ if (len <= 0 || len != sizeof(msg))
+ continue;
+
+ memcpy(&msg, buf, sizeof(msg));
+ msg.s_addr = ntoh64(msg.s_addr);
+ msg.d_addr = ntoh64(msg.d_addr);
+ msg.seqno = ntoh64(msg.seqno);
+#ifdef DEBUG_PROTO_LS
+ log_proto(LSU_FMT " <-- " ADDR_FMT32,
+ LSU_VAL(msg.s_addr, msg.d_addr, msg.seqno),
+ ADDR_VAL32(&ls.addr));
+#endif
+ if (lsdb_add_link(msg.s_addr,
+ msg.d_addr,
+ msg.seqno,
+ &qs))
+ continue;
+
+ forward_lsm(buf, len, fd);
+ }
+ }
+
+ pthread_cleanup_pop(true);
+
+ return (void *) 0;
+}
+
+static void flow_event(int fd,
+ bool up)
+{
+
+ struct list_head * p;
+
+ log_dbg("Notifying routing instances of flow event.");
+
+ pthread_mutex_lock(&ls.instances.mtx);
+
+ list_for_each(p, &ls.instances.list) {
+ struct routing_i * ri = list_entry(p, struct routing_i, next);
+ pff_flow_state_change(ri->pff, fd, up);
+ }
+
+ pthread_mutex_unlock(&ls.instances.mtx);
+}
+
+static void handle_event(void * self,
+ int event,
+ const void * o)
+{
+ /* FIXME: Apply correct QoS on graph */
+ struct conn * c;
+ qosspec_t qs;
+ int flags;
+
+ (void) self;
+
+ assert(o);
+
+ c = (struct conn *) o;
+
+ memset(&qs, 0, sizeof(qs));
+
+ switch (event) {
+ case NOTIFY_DT_CONN_ADD:
+ pthread_rwlock_rdlock(&ls.lock);
+
+ pthread_cleanup_push(__cleanup_rwlock_unlock, &ls.lock);
+
+ send_lsm(ls.addr, c->conn_info.addr, 0);
+ pthread_cleanup_pop(true);
+
+ if (lsdb_add_nb(c->conn_info.addr, c->flow_info.fd, NB_DT))
+ log_dbg("Failed to add neighbor to lsdb.");
+
+ if (lsdb_add_link(ls.addr, c->conn_info.addr, 0, &qs))
+ log_dbg("Failed to add new adjacency to lsdb.");
+ break;
+ case NOTIFY_DT_CONN_DEL:
+ flow_event(c->flow_info.fd, false);
+
+ if (lsdb_del_nb(c->conn_info.addr, c->flow_info.fd))
+ log_dbg("Failed to delete neighbor from lsdb.");
+
+ if (lsdb_del_link(ls.addr, c->conn_info.addr))
+ log_dbg("Local link was not in lsdb.");
+ break;
+ case NOTIFY_DT_CONN_QOS:
+ log_dbg("QoS changes currently unsupported.");
+ break;
+ case NOTIFY_DT_CONN_UP:
+ flow_event(c->flow_info.fd, true);
+ break;
+ case NOTIFY_DT_CONN_DOWN:
+ flow_event(c->flow_info.fd, false);
+ break;
+ case NOTIFY_MGMT_CONN_ADD:
+ fccntl(c->flow_info.fd, FLOWGFLAGS, &flags);
+ fccntl(c->flow_info.fd, FLOWSFLAGS, flags | FLOWFRNOPART);
+ fset_add(ls.mgmt_set, c->flow_info.fd);
+ if (lsdb_add_nb(c->conn_info.addr, c->flow_info.fd, NB_MGMT))
+ log_warn("Failed to add mgmt neighbor to lsdb.");
+ /* replicate the entire lsdb */
+ lsdb_replicate(c->flow_info.fd);
+ break;
+ case NOTIFY_MGMT_CONN_DEL:
+ fset_del(ls.mgmt_set, c->flow_info.fd);
+ if (lsdb_del_nb(c->conn_info.addr, c->flow_info.fd))
+ log_warn("Failed to delete mgmt neighbor from lsdb.");
+ break;
+ default:
+ break;
+ }
+}
+
+struct routing_i * link_state_routing_i_create(struct pff * pff)
+{
+ struct routing_i * tmp;
+
+ assert(pff);
+
+ tmp = malloc(sizeof(*tmp));
+ if (tmp == NULL)
+ goto fail_tmp;
+
+ tmp->pff = pff;
+ tmp->modified = false;
+
+ if (pthread_mutex_init(&tmp->lock, NULL))
+ goto fail_instance_lock_init;
+
+ if (pthread_create(&tmp->calculator, NULL,
+ periodic_recalc_pff, tmp))
+ goto fail_pthread_create_lsupdate;
+
+ pthread_mutex_lock(&ls.instances.mtx);
+
+ list_add(&tmp->next, &ls.instances.list);
+
+ pthread_mutex_unlock(&ls.instances.mtx);
+
+ return tmp;
+
+ fail_pthread_create_lsupdate:
+ pthread_mutex_destroy(&tmp->lock);
+ fail_instance_lock_init:
+ free(tmp);
+ fail_tmp:
+ return NULL;
+}
+
+void link_state_routing_i_destroy(struct routing_i * instance)
+{
+ assert(instance);
+
+ pthread_mutex_lock(&ls.instances.mtx);
+
+ list_del(&instance->next);
+
+ pthread_mutex_unlock(&ls.instances.mtx);
+
+ pthread_cancel(instance->calculator);
+
+ pthread_join(instance->calculator, NULL);
+
+ pthread_mutex_destroy(&instance->lock);
+
+ free(instance);
+}
+
+int link_state_start(void)
+{
+ if (notifier_reg(handle_event, NULL)) {
+ log_err("Failed to register link-state with notifier.");
+ goto fail_notifier_reg;
+ }
+
+ if (pthread_create(&ls.lsupdate, NULL, lsupdate, NULL)) {
+ log_err("Failed to create lsupdate thread.");
+ goto fail_pthread_create_lsupdate;
+ }
+
+ if (pthread_create(&ls.lsreader, NULL, lsreader, NULL)) {
+ log_err("Failed to create lsreader thread.");
+ goto fail_pthread_create_lsreader;
+ }
+
+ if (pthread_create(&ls.listener, NULL, ls_conn_handle, NULL)) {
+ log_err("Failed to create listener thread.");
+ goto fail_pthread_create_listener;
+ }
+
+ return 0;
+
+ fail_pthread_create_listener:
+ pthread_cancel(ls.lsreader);
+ pthread_join(ls.lsreader, NULL);
+ fail_pthread_create_lsreader:
+ pthread_cancel(ls.lsupdate);
+ pthread_join(ls.lsupdate, NULL);
+ fail_pthread_create_lsupdate:
+ notifier_unreg(handle_event);
+ fail_notifier_reg:
+ return -1;
+}
+
+void link_state_stop(void)
+{
+ pthread_cancel(ls.listener);
+ pthread_cancel(ls.lsreader);
+ pthread_cancel(ls.lsupdate);
+
+ pthread_join(ls.listener, NULL);
+ pthread_join(ls.lsreader, NULL);
+ pthread_join(ls.lsupdate, NULL);
+
+ notifier_unreg(handle_event);
+}
+
+
+int link_state_init(struct ls_config * conf,
+ enum pol_pff * pff_type)
+{
+ struct conn_info info;
+
+ assert(conf != NULL);
+ assert(pff_type != NULL);
+
+ memset(&info, 0, sizeof(info));
+
+ ls.addr = addr_auth_address();
+
+ strcpy(info.comp_name, LS_COMP);
+ strcpy(info.protocol, LS_PROTO);
+ info.pref_version = 1;
+ info.pref_syntax = PROTO_GPB;
+ info.addr = ls.addr;
+
+ ls.conf = *conf;
+
+ switch (conf->pol) {
+ case LS_SIMPLE:
+ *pff_type = PFF_SIMPLE;
+ ls.routing_algo = ROUTING_SIMPLE;
+ log_dbg("Using Link State Routing policy.");
+ break;
+ case LS_LFA:
+ ls.routing_algo = ROUTING_LFA;
+ *pff_type = PFF_ALTERNATE;
+ log_dbg("Using Loop-Free Alternates policy.");
+ break;
+ case LS_ECMP:
+ ls.routing_algo = ROUTING_ECMP;
+ *pff_type = PFF_MULTIPATH;
+ log_dbg("Using Equal-Cost Multipath policy.");
+ break;
+ default:
+ goto fail_graph;
+ }
+
+ log_dbg("LS update interval: %ld seconds.", ls.conf.t_update);
+ log_dbg("LS link timeout : %ld seconds.", ls.conf.t_timeo);
+ log_dbg("LS recalc interval: %ld seconds.", ls.conf.t_recalc);
+
+ ls.graph = graph_create();
+ if (ls.graph == NULL)
+ goto fail_graph;
+
+ if (pthread_rwlock_init(&ls.lock, NULL)) {
+ log_err("Failed to init lock.");
+ goto fail_lock_init;
+ }
+
+ if (pthread_mutex_init(&ls.instances.mtx, NULL)) {
+ log_err("Failed to init instances mutex.");
+ goto fail_routing_i_lock_init;
+ }
+
+ if (connmgr_comp_init(COMPID_MGMT, &info)) {
+ log_err("Failed to init connmgr.");
+ goto fail_connmgr_comp_init;
+ }
+
+ ls.mgmt_set = fset_create();
+ if (ls.mgmt_set == NULL) {
+ log_err("Failed to create fset.");
+ goto fail_fset_create;
+ }
+
+ llist_init(&ls.db);
+ llist_init(&ls.nbs);
+ list_head_init(&ls.instances.list);
+
+ if (rib_reg(lsdb, &r_ops))
+ goto fail_rib_reg;
+
+ return 0;
+
+ fail_rib_reg:
+ fset_destroy(ls.mgmt_set);
+ fail_fset_create:
+ connmgr_comp_fini(COMPID_MGMT);
+ fail_connmgr_comp_init:
+ pthread_mutex_destroy(&ls.instances.mtx);
+ fail_routing_i_lock_init:
+ pthread_rwlock_destroy(&ls.lock);
+ fail_lock_init:
+ graph_destroy(ls.graph);
+ fail_graph:
+ return -1;
+}
+
+void link_state_fini(void)
+{
+ struct list_head * p;
+ struct list_head * h;
+
+ rib_unreg(lsdb);
+
+ fset_destroy(ls.mgmt_set);
+
+ connmgr_comp_fini(COMPID_MGMT);
+
+ graph_destroy(ls.graph);
+
+ pthread_rwlock_wrlock(&ls.lock);
+
+ llist_for_each_safe(p, h, &ls.db) {
+ struct adjacency * a = list_entry(p, struct adjacency, next);
+ llist_del(&a->next, &ls.db);
+ free(a);
+ }
+
+ pthread_rwlock_unlock(&ls.lock);
+
+ pthread_rwlock_destroy(&ls.lock);
+
+ pthread_mutex_destroy(&ls.instances.mtx);
+}
diff --git a/src/ipcpd/unicast/routing/link-state.h b/src/ipcpd/unicast/routing/link-state.h
new file mode 100644
index 00000000..38e19065
--- /dev/null
+++ b/src/ipcpd/unicast/routing/link-state.h
@@ -0,0 +1,46 @@
+/*
+ * Ouroboros - Copyright (C) 2016 - 2026
+ *
+ * Link state routing policy
+ *
+ * Dimitri Staessens <dimitri@ouroboros.rocks>
+ * Sander Vrijders <sander@ouroboros.rocks>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License version 2 as
+ * published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., http://www.fsf.org/about/contact/.
+ */
+
+#ifndef OUROBOROS_IPCPD_UNICAST_POL_LINK_STATE_H
+#define OUROBOROS_IPCPD_UNICAST_POL_LINK_STATE_H
+
+#define LS_COMP "Management"
+#define LS_PROTO "LSP"
+
+#include "ops.h"
+
+int link_state_init(struct ls_config * ls,
+ enum pol_pff * pff_type);
+
+void link_state_fini(void);
+
+int link_state_start(void);
+
+void link_state_stop(void);
+
+struct routing_i * link_state_routing_i_create(struct pff * pff);
+
+void link_state_routing_i_destroy(struct routing_i * instance);
+
+extern struct routing_ops link_state_ops;
+
+#endif /* OUROBOROS_IPCPD_UNICAST_POL_LINK_STATE_H */
diff --git a/src/ipcpd/unicast/routing/ops.h b/src/ipcpd/unicast/routing/ops.h
new file mode 100644
index 00000000..b19c5176
--- /dev/null
+++ b/src/ipcpd/unicast/routing/ops.h
@@ -0,0 +1,43 @@
+/*
+ * Ouroboros - Copyright (C) 2016 - 2026
+ *
+ * Routing policy ops
+ *
+ * Dimitri Staessens <dimitri@ouroboros.rocks>
+ * Sander Vrijders <sander@ouroboros.rocks>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License version 2 as
+ * published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., http://www.fsf.org/about/contact/.
+ */
+
+#ifndef OUROBOROS_IPCPD_UNICAST_ROUTING_OPS_H
+#define OUROBOROS_IPCPD_UNICAST_ROUTING_OPS_H
+
+#include "pff.h"
+
+struct routing_ops {
+ int (* init)(void * conf,
+ enum pol_pff * pff_type);
+
+ void (* fini)(void);
+
+ int (* start)(void);
+
+ void (* stop)(void);
+
+ struct routing_i * (* routing_i_create)(struct pff * pff);
+
+ void (* routing_i_destroy)(struct routing_i * instance);
+};
+
+#endif /* OUROBOROS_IPCPD_UNICAST_ROUTING_OPS_H */
diff --git a/src/ipcpd/unicast/routing/pol.h b/src/ipcpd/unicast/routing/pol.h
new file mode 100644
index 00000000..545f5df2
--- /dev/null
+++ b/src/ipcpd/unicast/routing/pol.h
@@ -0,0 +1,23 @@
+/*
+ * Ouroboros - Copyright (C) 2016 - 2026
+ *
+ * Routing policies
+ *
+ * Dimitri Staessens <dimitri@ouroboros.rocks>
+ * Sander Vrijders <sander@ouroboros.rocks>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License version 2 as
+ * published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., http://www.fsf.org/about/contact/.
+ */
+
+#include "link-state.h"
diff --git a/src/ipcpd/unicast/routing/tests/CMakeLists.txt b/src/ipcpd/unicast/routing/tests/CMakeLists.txt
new file mode 100644
index 00000000..be2de72c
--- /dev/null
+++ b/src/ipcpd/unicast/routing/tests/CMakeLists.txt
@@ -0,0 +1,34 @@
+get_filename_component(CURRENT_SOURCE_PARENT_DIR
+ ${CMAKE_CURRENT_SOURCE_DIR} DIRECTORY)
+get_filename_component(CURRENT_BINARY_PARENT_DIR
+ ${CMAKE_CURRENT_BINARY_DIR} DIRECTORY)
+
+get_filename_component(PARENT_PATH ${CMAKE_CURRENT_SOURCE_DIR} DIRECTORY)
+get_filename_component(PARENT_DIR ${PARENT_PATH} NAME)
+
+compute_test_prefix()
+
+create_test_sourcelist(${PARENT_DIR}_tests test_suite.c
+ # Add new tests here
+ graph_test.c
+ )
+
+add_executable(${PARENT_DIR}_test ${${PARENT_DIR}_tests})
+
+target_include_directories(${PARENT_DIR}_test PRIVATE
+ ${CMAKE_CURRENT_SOURCE_DIR}
+ ${CMAKE_CURRENT_BINARY_DIR}
+ ${CURRENT_SOURCE_PARENT_DIR}
+ ${CURRENT_BINARY_PARENT_DIR}
+ ${CMAKE_SOURCE_DIR}/include
+ ${CMAKE_BINARY_DIR}/include
+ ${CMAKE_SOURCE_DIR}/src/ipcpd
+ ${CMAKE_BINARY_DIR}/src/ipcpd
+)
+
+disable_test_logging_for_target(${PARENT_DIR}_test)
+target_link_libraries(${PARENT_DIR}_test PRIVATE ouroboros-common)
+
+add_dependencies(build_tests ${PARENT_DIR}_test)
+
+ouroboros_register_tests(TARGET ${PARENT_DIR}_test TESTS ${${PARENT_DIR}_tests})
diff --git a/src/ipcpd/unicast/routing/tests/graph_test.c b/src/ipcpd/unicast/routing/tests/graph_test.c
new file mode 100644
index 00000000..40a744ff
--- /dev/null
+++ b/src/ipcpd/unicast/routing/tests/graph_test.c
@@ -0,0 +1,351 @@
+/*
+ * Ouroboros - Copyright (C) 2016 - 2026
+ *
+ * Test of the graph structure
+ *
+ * Dimitri Staessens <dimitri@ouroboros.rocks>
+ * Sander Vrijders <sander@ouroboros.rocks>
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License version 2 as
+ * published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, Inc., http://www.fsf.org/about/contact/.
+ */
+
+#define _POSIX_C_SOURCE 200112L
+
+#include <ouroboros/utils.h>
+
+#include <stdio.h>
+#include <stdlib.h>
+#include <string.h>
+
+#include "graph.c"
+
+struct graph * graph;
+struct list_head table;
+qosspec_t qs;
+
+int graph_test_entries(int entries)
+{
+ struct list_head * p;
+ int i = 0;
+
+ if (graph_routing_table(graph, ROUTING_SIMPLE, 1, &table)) {
+ printf("Failed to get routing table.\n");
+ return -1;
+ }
+
+ list_for_each(p, &table)
+ i++;
+
+ if (i != entries) {
+ printf("Wrong number of entries.\n");
+ graph_free_routing_table(graph, &table);
+ return -1;
+ }
+
+ graph_free_routing_table(graph, &table);
+
+ return 0;
+}
+
+int graph_test_double_link(void)
+{
+ struct list_head * p;
+ int i = 0;
+
+ if (graph_routing_table(graph, ROUTING_SIMPLE, 1, &table)) {
+ printf("Failed to get routing table.\n");
+ return -1;
+ }
+
+ list_for_each(p, &table)
+ i++;
+
+ if (i != 2) {
+ printf("Wrong number of entries.\n");
+ graph_free_routing_table(graph, &table);
+ return -1;
+ }
+
+ list_for_each(p, &table) {
+ struct routing_table * t =
+ list_entry(p, struct routing_table, next);
+ struct nhop * n =
+ list_first_entry(&t->nhops, struct nhop, next);
+
+ if ((t->dst != 2 && n->nhop != 2) ||
+ (t->dst != 3 && n->nhop != 2)) {
+ printf("Wrong routing entry.\n");
+ graph_free_routing_table(graph, &table);
+ return -1;
+ }
+ }
+
+ graph_free_routing_table(graph, &table);
+
+ return 0;
+}
+
+int graph_test_single_link(void)
+{
+ struct list_head * p;
+ int i = 0;
+
+ if (graph_routing_table(graph, ROUTING_SIMPLE, 1, &table)) {
+ printf("Failed to get routing table.\n");
+ return -1;
+ }
+
+ list_for_each(p, &table)
+ i++;
+
+ if (i != 1) {
+ printf("Wrong number of entries.\n");
+ graph_free_routing_table(graph, &table);
+ return -1;
+ }
+
+ list_for_each(p, &table) {
+ struct routing_table * t =
+ list_entry(p, struct routing_table, next);
+ struct nhop * n =
+ list_first_entry(&t->nhops, struct nhop, next);
+
+ if (t->dst != 2 && n->nhop != 2) {
+ printf("Wrong routing entry.\n");
+ graph_free_routing_table(graph, &table);
+ return -1;
+ }
+ }
+
+ graph_free_routing_table(graph, &table);
+
+ return 0;
+}
+
+static int distance_check(int dst,
+ int nhop,
+ enum routing_algo algo)
+{
+ (void) algo;
+
+ /* Same distances for all algorithms at the moment. */
+ if (dst == 2 && nhop != 2) {
+ printf("Wrong entry: 2.\n");
+ return -1;
+ }
+
+ if (dst == 3 && nhop != 3) {
+ printf("Wrong entry: 3.\n");
+ return -1;
+ }
+
+ if (dst == 4 && nhop != 3) {
+ printf("Wrong entry: 4.\n");
+ return -1;
+ }
+
+ if (dst == 5 && nhop != 3) {
+ printf("Wrong entry: 5.\n");
+ return -1;
+ }
+
+ if (dst == 6 && nhop != 2) {
+ printf("Wrong entry: 6.\n");
+ return -1;
+ }
+
+ if (dst == 7 && nhop != 3) {
+ printf("Wrong entry: 7.\n");
+ return -1;
+ }
+
+ return 0;
+}
+
+int graph_test(int argc,
+ char ** argv)
+{
+ int nhop;
+ int dst;
+ struct list_head * p;
+
+ (void) argc;
+ (void) argv;
+
+ memset(&qs, 0, sizeof(qs));
+
+ graph = graph_create();
+ if (graph == NULL) {
+ printf("Failed to create graph.\n");
+ return -1;
+ }
+
+ graph_destroy(graph);
+
+ graph = graph_create();
+ if (graph == NULL) {
+ printf("Failed to create graph.\n");
+ return -1;
+ }
+
+ if (graph_update_edge(graph, 1, 2, qs)) {
+ printf("Failed to add edge.\n");
+ graph_destroy(graph);
+ return -1;
+ }
+
+ if (graph_update_edge(graph, 2, 1, qs)) {
+ printf("Failed to add edge.\n");
+ graph_destroy(graph);
+ return -1;
+ }
+
+ if (graph_test_single_link()) {
+ graph_destroy(graph);
+ return -1;
+ }
+
+ if (graph_update_edge(graph, 2, 3, qs)) {
+ printf("Failed to add edge.\n");
+ graph_destroy(graph);
+ return -1;
+ }
+
+ if (graph_update_edge(graph, 3, 2, qs)) {
+ printf("Failed to add edge.\n");
+ graph_destroy(graph);
+ return -1;
+ }
+
+ if (graph_test_double_link()) {
+ graph_destroy(graph);
+ return -1;
+ }
+
+ if (graph_del_edge(graph, 2, 3)) {
+ printf("Failed to delete edge.\n");
+ goto fail_graph;
+ }
+
+ if (graph_del_edge(graph, 3, 2)) {
+ printf("Failed to delete edge.\n");
+ goto fail_graph;
+ }
+
+ if (graph_test_single_link())
+ goto fail_graph;
+
+ graph_update_edge(graph, 2, 3, qs);
+ graph_update_edge(graph, 3, 2, qs);
+ graph_update_edge(graph, 1, 3, qs);
+ graph_update_edge(graph, 3, 1, qs);
+
+ if (graph_test_entries(2))
+ goto fail_graph;
+
+ graph_update_edge(graph, 3, 4, qs);
+ graph_update_edge(graph, 4, 3, qs);
+ graph_update_edge(graph, 4, 5, qs);
+ graph_update_edge(graph, 5, 4, qs);
+
+ if (graph_test_entries(4))
+ goto fail_graph;
+
+ graph_update_edge(graph, 2, 6, qs);
+ graph_update_edge(graph, 6, 2, qs);
+ graph_update_edge(graph, 6, 7, qs);
+ graph_update_edge(graph, 7, 6, qs);
+ graph_update_edge(graph, 3, 7, qs);
+ graph_update_edge(graph, 7, 3, qs);
+
+ if (graph_test_entries(6))
+ goto fail_graph;
+
+
+ if (graph_routing_table(graph, ROUTING_SIMPLE, 1, &table)) {
+ printf("Failed to get routing table.\n");
+ goto fail_graph;
+ }
+
+ list_for_each(p, &table) {
+ struct routing_table * t =
+ list_entry(p, struct routing_table, next);
+ struct nhop * n =
+ list_first_entry(&t->nhops, struct nhop, next);
+
+ dst = t->dst;
+ nhop = n->nhop;
+
+ if (distance_check(dst, nhop, ROUTING_SIMPLE)) {
+ printf("Simple distance check failed.\n");
+ goto fail_routing;
+ }
+ }
+
+ graph_free_routing_table(graph, &table);
+
+ if (graph_routing_table(graph, ROUTING_LFA, 1, &table)) {
+ printf("Failed to get routing table for LFA.\n");
+ goto fail_graph;
+ }
+
+ list_for_each(p, &table) {
+ struct routing_table * t =
+ list_entry(p, struct routing_table, next);
+ struct nhop * n =
+ list_first_entry(&t->nhops, struct nhop, next);
+
+ dst = t->dst;
+ nhop = n->nhop;
+
+ if (distance_check(dst, nhop, ROUTING_LFA)) {
+ printf("LFA distance check failed.\n");
+ goto fail_routing;
+ }
+ }
+
+ graph_free_routing_table(graph, &table);
+
+ if (graph_routing_table(graph, ROUTING_ECMP, 1, &table)) {
+ printf("Failed to get routing table for ECMP.\n");
+ goto fail_graph;
+ }
+
+ list_for_each(p, &table) {
+ struct routing_table * t =
+ list_entry(p, struct routing_table, next);
+ struct nhop * n =
+ list_first_entry(&t->nhops, struct nhop, next);
+
+ dst = t->dst;
+ nhop = n->nhop;
+
+ if (distance_check(dst, nhop, ROUTING_LFA)) {
+ printf("LFA distance check failed.\n");
+ goto fail_routing;
+ }
+ }
+
+ graph_free_routing_table(graph, &table);
+
+ graph_destroy(graph);
+
+ return 0;
+
+ fail_routing:
+ graph_free_routing_table(graph, &table);
+ fail_graph:
+ graph_destroy(graph);
+ return -1;
+}