summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--source4/dsdb/kcc/kcc_topology.c1692
1 files changed, 1692 insertions, 0 deletions
diff --git a/source4/dsdb/kcc/kcc_topology.c b/source4/dsdb/kcc/kcc_topology.c
index 50a1fee49b..ea303b9076 100644
--- a/source4/dsdb/kcc/kcc_topology.c
+++ b/source4/dsdb/kcc/kcc_topology.c
@@ -27,10 +27,16 @@
#define FLAG_CR_NTDS_DOMAIN 0x00000002
+#define NTDSDSA_OPT_IS_GC 0x00000001
+
+#define NTDSSETTINGS_OPT_IS_TOPL_DETECT_STALE_DISABLED 0x00000008
+#define NTDSSETTINGS_OPT_IS_RAND_BH_SELECTION_DISABLED 0x00000100
#define NTDSSETTINGS_OPT_W2K3_BRIDGES_REQUIRED 0x00001000
#define NTDSTRANSPORT_OPT_BRIDGES_REQUIRED 0x00000002
+#define DS_BEHAVIOR_WIN2008 3
+
/** replication parameters of a graph edge */
struct kcctpl_repl_info {
uint32_t cost;
@@ -130,6 +136,160 @@ struct message_list {
};
/**
+ * sort internal edges based on:
+ * - descending red_red,
+ * - ascending repl_info.cost,
+ * - descending available time in repl_info.schedule,
+ * - ascending v1id,
+ * - ascending v2id,
+ * - ascending type.
+ *
+ * this function is used in 'kcctpl_kruskal'.
+ */
+static int kcctpl_sort_internal_edges(const void *internal_edge1,
+ const void *internal_edge2)
+{
+ const struct kcctpl_internal_edge *ie1, *ie2;
+ int cmp_red_red;
+
+ ie1 = (const struct kcctpl_internal_edge *) internal_edge1;
+ ie2 = (const struct kcctpl_internal_edge *) internal_edge2;
+
+ cmp_red_red = ie2->red_red - ie1->red_red;
+ if (cmp_red_red == 0) {
+ int cmp_cost = ie1->repl_info.cost - ie2->repl_info.cost;
+
+ if (cmp_cost == 0) {
+ uint32_t available1, available2, i;
+ int cmp_schedule;
+
+ available1 = available2 = 0;
+ for (i = 0; i < 84; i++) {
+ if (ie1->repl_info.schedule[i] == 0) {
+ available1++;
+ }
+
+ if (ie2->repl_info.schedule[i] == 0) {
+ available2++;
+ }
+ }
+ cmp_schedule = available2 - available1;
+
+ if (cmp_schedule == 0) {
+ int cmp_v1id = GUID_compare(&ie1->v1id,
+ &ie2->v1id);
+
+ if (cmp_v1id == 0) {
+ int cmp_v2id = GUID_compare(&ie1->v2id,
+ &ie2->v2id);
+
+ if (cmp_v2id == 0) {
+ return GUID_compare(&ie1->type,
+ &ie2->type);
+ } else {
+ return cmp_v2id;
+ }
+ } else {
+ return cmp_v1id;
+ }
+ } else {
+ return cmp_schedule;
+ }
+ } else {
+ return cmp_cost;
+ }
+ } else {
+ return cmp_red_red;
+ }
+}
+
+/**
+ * sort vertices based on the following criteria:
+ * - ascending color (RED < BLACK),
+ * - ascending repl_info.cost,
+ * - ascending id.
+ *
+ * this function is used in 'kcctpl_process_edge'.
+ */
+static int kcctpl_sort_vertices(const void *vertex1, const void *vertex2)
+{
+ const struct kcctpl_vertex *v1, *v2;
+ int cmp_color;
+
+ v1 = (const struct kcctpl_vertex *) vertex1;
+ v2 = (const struct kcctpl_vertex *) vertex2;
+
+ cmp_color = v1->color - v2->color;
+ if (cmp_color == 0) {
+ int cmp_cost = v1->repl_info.cost - v2->repl_info.cost;
+ if (cmp_cost == 0) {
+ return GUID_compare(&v1->id, &v2->id);
+ } else {
+ return cmp_cost;
+ }
+ } else {
+ return cmp_color;
+ }
+}
+
+/**
+ * sort bridgehead elements (nTDSDSA) based on the following criteria:
+ * - GC servers precede non-GC servers
+ * - ascending objectGUID
+ *
+ * this function is used in 'kcctpl_get_all_bridgehead_dcs'.
+ */
+static int kcctpl_sort_bridgeheads(const void *bridgehead1,
+ const void *bridgehead2)
+{
+ const struct ldb_message *bh1, *bh2;
+ uint64_t bh1_opts, bh2_opts, cmp_gc;
+
+ bh1 = (const struct ldb_message *) bridgehead1;
+ bh2 = (const struct ldb_message *) bridgehead2;
+
+ bh1_opts = samdb_result_int64(bh1, "options", 0);
+ bh2_opts = samdb_result_int64(bh2, "options", 0);
+
+ cmp_gc = (bh1_opts & NTDSDSA_OPT_IS_GC) -
+ (bh2_opts & NTDSDSA_OPT_IS_GC);
+
+ if (cmp_gc == 0) {
+ struct GUID bh1_id, bh2_id;
+
+ bh1_id = samdb_result_guid(bh1, "objectGUID");
+ bh2_id = samdb_result_guid(bh2, "objectGUID");
+
+ return GUID_compare(&bh1_id, &bh2_id);
+ } else {
+ return cmp_gc;
+ }
+}
+
+/**
+ * sort bridgehead elements (nTDSDSA) in a random order.
+ *
+ * this function is used in 'kcctpl_get_all_bridgehead_dcs'.
+ */
+static void kcctpl_shuffle_bridgeheads(struct message_list bridgeheads)
+{
+ uint32_t i;
+
+ srandom(time(NULL));
+
+ for (i = bridgeheads.count; i > 1; i--) {
+ uint32_t r;
+ struct ldb_message tmp;
+
+ r = random() % i;
+
+ tmp = bridgeheads.data[i - 1];
+ bridgeheads.data[i - 1] = bridgeheads.data[r];
+ bridgeheads.data[r] = tmp;
+ }
+}
+
+/**
* find a graph vertex based on its GUID.
*/
static struct kcctpl_vertex *kcctpl_find_vertex_by_guid(struct kcctpl_graph *graph,
@@ -193,6 +353,22 @@ static struct kcctpl_multi_edge *kcctpl_find_edge_by_vertex_guid(struct kcctpl_g
}
/**
+ * search for an occurrence of a GUID inside a list of GUIDs.
+ */
+static bool kcctpl_guid_list_contains(struct GUID_list list, struct GUID guid)
+{
+ uint32_t i;
+
+ for (i = 0; i < list.count; i++) {
+ if (GUID_equal(&list.data[i], &guid)) {
+ return true;
+ }
+ }
+
+ return false;
+}
+
+/**
* get the Transports DN
* (CN=Inter-Site Transports,CN=Sites,CN=Configuration,DC=<domain>).
*/
@@ -248,6 +424,44 @@ static struct ldb_message *kcctpl_local_site(struct ldb_context *ldb,
return res->msgs[0];
}
+/*
+ * compare two internal edges for equality. every field of the structure will be
+ * compared.
+ */
+static bool kcctpl_internal_edge_equal(struct kcctpl_internal_edge *edge1,
+ struct kcctpl_internal_edge *edge2)
+{
+ if (!edge1 || !edge2) {
+ return false;
+ }
+
+ if (!GUID_equal(&edge1->v1id, &edge2->v1id)) {
+ return false;
+ }
+
+ if (!GUID_equal(&edge1->v2id, &edge2->v2id)) {
+ return false;
+ }
+
+ if (edge1->red_red != edge2->red_red) {
+ return false;
+ }
+
+ if (edge1->repl_info.cost != edge2->repl_info.cost ||
+ edge1->repl_info.interval != edge2->repl_info.interval ||
+ edge1->repl_info.options != edge2->repl_info.options ||
+ memcmp(&edge1->repl_info.schedule,
+ &edge2->repl_info.schedule, 84) != 0) {
+ return false;
+ }
+
+ if (!GUID_equal(&edge1->type, &edge2->type)) {
+ return false;
+ }
+
+ return true;
+}
+
/**
* create a kcctpl_graph instance.
*/
@@ -747,6 +961,324 @@ static NTSTATUS kcctpl_setup_graph(struct ldb_context *ldb, TALLOC_CTX *mem_ctx,
}
/**
+ * determine whether a given DC is known to be in a failed state.
+ */
+static NTSTATUS kcctpl_bridgehead_dc_failed(struct ldb_context *ldb,
+ struct GUID guid,
+ bool detect_failed_dcs,
+ bool *_failed)
+{
+ TALLOC_CTX *tmp_ctx;
+ struct ldb_dn *settings_dn;
+ struct ldb_result *res;
+ const char * const attrs[] = { "options", NULL };
+ int ret;
+ struct ldb_message *settings;
+ uint64_t settings_opts;
+ bool failed;
+
+ tmp_ctx = talloc_new(ldb);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ settings_dn = samdb_ntds_settings_dn(ldb);
+ if (!settings_dn) {
+ DEBUG(1, (__location__ ": failed to find our own NTDS Settings "
+ "DN\n"));
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ ret = ldb_search(ldb, tmp_ctx, &res, settings_dn, LDB_SCOPE_BASE, attrs,
+ "objectClass=nTDSSiteSettings");
+ if (ret != LDB_SUCCESS) {
+ DEBUG(1, (__location__ ": failed to find site settings object "
+ "%s: %s\n", ldb_dn_get_linearized(settings_dn),
+ ldb_strerror(ret)));
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+ if (res->count == 0) {
+ DEBUG(1, ("failed to find site settings object %s\n",
+ ldb_dn_get_linearized(settings_dn)));
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ settings = res->msgs[0];
+
+ settings_opts = samdb_result_int64(settings, "options", 0);
+ if (settings_opts & NTDSSETTINGS_OPT_IS_TOPL_DETECT_STALE_DISABLED) {
+ failed = false;
+ } else if (true) { /* TODO: how to get kCCFailedLinks and
+ kCCFailedConnections? */
+ failed = true;
+ } else {
+ failed = detect_failed_dcs;
+ }
+
+ *_failed = failed;
+ talloc_free(tmp_ctx);
+ return NT_STATUS_OK;
+}
+
+/**
+ * get all bridgehead DCs satisfying the given criteria.
+ */
+static NTSTATUS kcctpl_get_all_bridgehead_dcs(struct ldb_context *ldb,
+ TALLOC_CTX *mem_ctx,
+ struct GUID site_guid,
+ struct ldb_message *cross_ref,
+ struct ldb_message *transport,
+ bool partial_replica_okay,
+ bool detect_failed_dcs,
+ struct message_list *_bridgeheads)
+{
+ struct message_list bridgeheads, all_dcs_in_site;
+ TALLOC_CTX *tmp_ctx;
+ struct ldb_result *res;
+ struct ldb_dn *sites_dn, *schemas_dn;
+ const char * const attrs[] = { "options", NULL };
+ int ret;
+ struct ldb_message *site, *schema;
+ const char * const dc_attrs[] = { "objectGUID", "options", NULL };
+ struct ldb_message_element *el;
+ uint32_t i;
+ bool rodc;
+ const char *transport_name, *transport_address_attr;
+ uint64_t site_opts;
+
+ ZERO_STRUCT(bridgeheads);
+
+ tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ sites_dn = samdb_sites_dn(ldb, tmp_ctx);
+ if (!sites_dn) {
+ DEBUG(1, (__location__ ": failed to find our own Sites DN\n"));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ ret = ldb_search(ldb, tmp_ctx, &res, sites_dn, LDB_SCOPE_ONELEVEL,
+ attrs, "(&(objectClass=site)(objectGUID=%s))",
+ GUID_string(tmp_ctx, &site_guid));
+ if (ret != LDB_SUCCESS) {
+ DEBUG(1, (__location__ ": failed to find site object %s: %s\n",
+ GUID_string(tmp_ctx, &site_guid),
+ ldb_strerror(ret)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+ if (res->count == 0) {
+ DEBUG(1, (__location__ ": failed to find site object %s\n",
+ GUID_string(tmp_ctx, &site_guid)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+ site = res->msgs[0];
+
+ schemas_dn = samdb_schema_dn(ldb);
+ if (!schemas_dn) {
+ DEBUG(1, (__location__ ": failed to find our own Schemas DN\n"));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ ret = ldb_search(ldb, tmp_ctx, &res, schemas_dn, LDB_SCOPE_SUBTREE,
+ NULL,
+ "(&(lDAPDisplayName=nTDSDSA)(objectClass=classSchema))");
+ if (ret != LDB_SUCCESS) {
+ DEBUG(1, (__location__ ": failed to find classSchema object :"
+ "%s\n", ldb_strerror(ret)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+ if (res->count == 0) {
+ DEBUG(1, (__location__ ": failed to find classSchema "
+ "object\n"));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+ schema = res->msgs[0];
+
+ ZERO_STRUCT(all_dcs_in_site);
+
+ ret = ldb_search(ldb, tmp_ctx, &res, site->dn, LDB_SCOPE_SUBTREE,
+ dc_attrs, "objectCategory=%s",
+ ldb_dn_get_linearized(schema->dn));
+ if (ret != LDB_SUCCESS) {
+ DEBUG(1, (__location__ ": failed to find DCs objects :%s\n",
+ ldb_strerror(ret)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ el = ldb_msg_find_element(transport, "bridgeheadServerListBL");
+
+ rodc = samdb_rodc(ldb);
+
+ transport_name = samdb_result_string(transport, "name", NULL);
+ if (!transport_name) {
+ DEBUG(1, (__location__ ": failed to find name attribute of "
+ "object %s\n", ldb_dn_get_linearized(transport->dn)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ transport_address_attr = samdb_result_string(transport,
+ "transportAddressAttribute",
+ NULL);
+ if (!transport_address_attr) {
+ DEBUG(1, (__location__ ": failed to find "
+ "transportAddressAttribute attribute of object %s\n",
+ ldb_dn_get_linearized(transport->dn)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ site_opts = samdb_result_int64(site, "options", 0);
+
+ for (i = 0; i < res->count; i++) {
+ struct ldb_message *dc, *new_data;
+ uint32_t j;
+ struct ldb_dn *parent_dn;
+ uint64_t behavior_version;
+ const char *dc_transport_address;
+ struct ldb_result *parent_res;
+ const char *parent_attrs[] = { transport_address_attr, NULL };
+ NTSTATUS status;
+ struct GUID dc_guid;
+ bool failed;
+
+ dc = res->msgs[i];
+
+ parent_dn = ldb_dn_get_parent(tmp_ctx, dc->dn);
+ if (!parent_dn) {
+ DEBUG(1, (__location__ ": failed to get parent DN of "
+ "%s\n", ldb_dn_get_linearized(dc->dn)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ if (el && (el->num_values >= 1)) {
+ bool contains = false;
+
+ for (j = 0; j < el->num_values; j++) {
+ struct ldb_val val;
+ struct ldb_dn *dn;
+
+ val = el->values[j];
+
+ dn = ldb_dn_from_ldb_val(tmp_ctx, ldb, &val);
+ if (!dn) {
+ DEBUG(1, (__location__ ": failed to read a DN "
+ "from bridgeheadServerListBL "
+ "attribute of %s\n",
+ ldb_dn_get_linearized(transport->dn)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ if (ldb_dn_compare(dn, parent_dn) == 0) {
+ contains = true;
+ break;
+ }
+ }
+
+ if (!contains) {
+ continue;
+ }
+ }
+
+ /* TODO: if dc is in the same site as the local DC */
+ if (true) {
+ /* TODO: if a replica of cr!nCName is not in the set of
+ * NC replicas that "should be present" on 'dc' */
+ /* TODO: a partial replica of the NC "should be
+ present" */
+ if (true || (true && !partial_replica_okay)) {
+ continue;
+ }
+ } else {
+ /* TODO: if an NC replica of cr!nCName is not in the set
+ * of NC replicas that "are present" on 'dc' */
+ /* TODO: a partial replica of the NC "is present" */
+ if (true || (true && !partial_replica_okay)) {
+ continue;
+ }
+ }
+
+ behavior_version = samdb_result_int64(dc,
+ "msDS-Behavior-Version", 0);
+ /* TODO: cr!nCName corresponds to default NC */
+ if (rodc && true && behavior_version < DS_BEHAVIOR_WIN2008) {
+ continue;
+ }
+
+ ret = ldb_search(ldb, tmp_ctx, &parent_res, parent_dn,
+ LDB_SCOPE_BASE, parent_attrs , NULL);
+
+ dc_transport_address = samdb_result_string(parent_res->msgs[0],
+ transport_address_attr,
+ NULL);
+
+ if (strncmp(transport_name, "IP", 2) != 0 &&
+ dc_transport_address == NULL) {
+ continue;
+ }
+
+ dc_guid = samdb_result_guid(dc, "objectGUID");
+
+ status = kcctpl_bridgehead_dc_failed(ldb, dc_guid,
+ detect_failed_dcs,
+ &failed);
+ if (NT_STATUS_IS_ERR(status)) {
+ DEBUG(1, (__location__ ": failed to check if "
+ "bridgehead DC has failed: %s\n",
+ nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+
+ if (failed) {
+ continue;
+ }
+
+ new_data = talloc_realloc(tmp_ctx, bridgeheads.data,
+ struct ldb_message,
+ bridgeheads.count + 1);
+ NT_STATUS_HAVE_NO_MEMORY_AND_FREE(new_data, tmp_ctx);
+ new_data[bridgeheads.count + 1] = *dc;
+ bridgeheads.data = new_data;
+ bridgeheads.count++;
+ }
+
+ if (site_opts & NTDSSETTINGS_OPT_IS_RAND_BH_SELECTION_DISABLED) {
+ qsort(bridgeheads.data, bridgeheads.count,
+ sizeof(struct ldb_message), kcctpl_sort_bridgeheads);
+ } else {
+ kcctpl_shuffle_bridgeheads(bridgeheads);
+ }
+
+ talloc_steal(mem_ctx, bridgeheads.data);
+ *_bridgeheads = bridgeheads;
+ talloc_free(tmp_ctx);
+ return NT_STATUS_OK;
+}
+
+/**
* get a bridgehead DC.
*/
static NTSTATUS kcctpl_get_bridgehead_dc(struct ldb_context *ldb,
@@ -758,6 +1290,21 @@ static NTSTATUS kcctpl_get_bridgehead_dc(struct ldb_context *ldb,
bool detect_failed_dcs,
struct ldb_message **_dsa)
{
+ struct message_list dsa_list;
+ NTSTATUS status;
+
+ status = kcctpl_get_all_bridgehead_dcs(ldb, mem_ctx,
+ site_guid, cross_ref, transport,
+ partial_replica_okay,
+ detect_failed_dcs, &dsa_list);
+ if (NT_STATUS_IS_ERR(status)) {
+ DEBUG(1, (__location__ ": failed to get all bridgehead DCs: "
+ "%s\n", nt_errstr(status)));
+ return status;
+ }
+
+ *_dsa = (dsa_list.count == 0) ? NULL : &dsa_list.data[0];
+
return NT_STATUS_OK;
}
@@ -974,3 +1521,1148 @@ static NTSTATUS kcctpl_color_vertices(struct ldb_context *ldb,
talloc_free(tmp_ctx);
return NT_STATUS_OK;
}
+
+/**
+ * setup the fields of the vertices that are relevant to Phase I (Dijkstra's
+ * Algorithm). for each vertex, set up its cost, root vertex and component. this
+ * defines the shortest-path forest structures.
+ */
+static void kcctpl_setup_vertices(struct kcctpl_graph *graph)
+{
+ uint32_t i;
+
+ for (i = 0; i < graph->vertices.count; i++) {
+ struct kcctpl_vertex *vertex = &graph->vertices.data[i];
+
+ if (vertex->color == WHITE) {
+ vertex->repl_info.cost = UINT32_MAX;
+ vertex->root_id = vertex->component_id = GUID_zero();
+ } else {
+ vertex->repl_info.cost = 0;
+ vertex->root_id = vertex->component_id = vertex->id;
+ }
+
+ vertex->repl_info.interval = 0;
+ vertex->repl_info.options = 0xFFFFFFFF;
+ ZERO_STRUCT(vertex->repl_info.schedule);
+ vertex->demoted = false;
+ }
+}
+
+/**
+ * demote one vertex if necessary.
+ */
+static void kcctpl_check_demote_one_vertex(struct kcctpl_vertex *vertex,
+ struct GUID type)
+{
+ if (vertex->color == WHITE) {
+ return;
+ }
+
+ if (!kcctpl_guid_list_contains(vertex->accept_black, type) &&
+ !kcctpl_guid_list_contains(vertex->accept_red_red, type)) {
+ vertex->repl_info.cost = UINT32_MAX;
+ vertex->root_id = GUID_zero();
+ vertex->demoted = true;
+ }
+}
+
+/**
+ * clear the demoted state of a vertex.
+ */
+static void kcctpl_undemote_one_vertex(struct kcctpl_vertex *vertex)
+{
+ if (vertex->color == WHITE) {
+ return;
+ }
+
+ vertex->repl_info.cost = 0;
+ vertex->root_id = vertex->id;
+ vertex->demoted = false;
+}
+
+/**
+ * returns the id of the component containing 'vertex' by traversing the up-tree
+ * implied by the component pointers.
+ */
+static struct GUID kcctpl_get_component_id(struct kcctpl_graph *graph,
+ struct kcctpl_vertex *vertex)
+{
+ struct kcctpl_vertex *u;
+ struct GUID root;
+
+ u = vertex;
+ while (!GUID_equal(&u->component_id, &u->id)) {
+ u = kcctpl_find_vertex_by_guid(graph, u->component_id);
+ }
+
+ root = u->id;
+
+ u = vertex;
+ while (!GUID_equal(&u->component_id, &u->id)) {
+ struct kcctpl_vertex *w;
+
+ w = kcctpl_find_vertex_by_guid(graph, u->component_id);
+ u->component_id = root;
+ u = w;
+ }
+
+ return root;
+}
+
+/**
+ * copy all spanning tree edges from 'output_edges' that contain the vertex for
+ * DCs in the local DC's site.
+ */
+static NTSTATUS kcctpl_copy_output_edges(struct ldb_context *ldb,
+ TALLOC_CTX *mem_ctx,
+ struct kcctpl_graph *graph,
+ struct kcctpl_multi_edge_list output_edges,
+ struct kcctpl_multi_edge_list *_copy)
+{
+ struct kcctpl_multi_edge_list copy;
+ TALLOC_CTX *tmp_ctx;
+ struct ldb_message *site;
+ struct GUID site_guid;
+ uint32_t i;
+
+ ZERO_STRUCT(copy);
+
+ tmp_ctx = talloc_new(ldb);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ site = kcctpl_local_site(ldb, tmp_ctx);
+ if (!site) {
+ DEBUG(1, (__location__ ": failed to find our own local DC's "
+ "site\n"));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+ site_guid = samdb_result_guid(site, "objectGUID");
+
+ for (i = 0; i < output_edges.count; i++) {
+ struct kcctpl_multi_edge *edge;
+ struct kcctpl_vertex *vertex1, *vertex2;
+
+ edge = &output_edges.data[i];
+
+ vertex1 = kcctpl_find_vertex_by_guid(graph,
+ edge->vertex_ids.data[0]);
+ if (!vertex1) {
+ DEBUG(1, (__location__ ": failed to find vertex %s\n",
+ GUID_string(tmp_ctx,
+ &edge->vertex_ids.data[0])));
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ vertex2 = kcctpl_find_vertex_by_guid(graph,
+ edge->vertex_ids.data[1]);
+ if (!vertex2) {
+ DEBUG(1, (__location__ ": failed to find vertex %s\n",
+ GUID_string(tmp_ctx,
+ &edge->vertex_ids.data[1])));
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ if (GUID_equal(&vertex1->id, &site_guid) ||
+ GUID_equal(&vertex2->id, &site_guid)) {
+ struct kcctpl_multi_edge *new_data;
+
+ if ((vertex1->color == BLACK ||
+ vertex2->color == BLACK) &&
+ vertex1->dist_to_red != UINT32_MAX) {
+
+ edge->directed = true;
+
+ if (vertex2->dist_to_red <
+ vertex1->dist_to_red) {
+ struct GUID tmp;
+
+ tmp = edge->vertex_ids.data[0];
+ edge->vertex_ids.data[0] = edge->vertex_ids.data[1];
+ edge->vertex_ids.data[1] = tmp;
+ }
+ }
+
+ new_data = talloc_realloc(tmp_ctx, copy.data,
+ struct kcctpl_multi_edge,
+ copy.count + 1);
+ NT_STATUS_HAVE_NO_MEMORY_AND_FREE(new_data, tmp_ctx);
+ new_data[copy.count + 1] = *edge;
+ copy.data = new_data;
+ copy.count++;
+ }
+ }
+
+ talloc_steal(mem_ctx, copy.data);
+ *_copy = copy;
+ return NT_STATUS_OK;
+}
+
+/**
+ * build the initial sequence for use with Dijkstra's algorithm. it will contain
+ * the red and black vertices as root vertices, unless these vertices accept no
+ * edges of the current 'type', or unless black vertices are not being
+ * including.
+ */
+static NTSTATUS kcctpl_setup_dijkstra(TALLOC_CTX *mem_ctx,
+ struct kcctpl_graph *graph,
+ struct GUID type, bool include_black,
+ struct kcctpl_vertex_list *_vertices)
+{
+ struct kcctpl_vertex_list vertices;
+ uint32_t i;
+
+ kcctpl_setup_vertices(graph);
+
+ ZERO_STRUCT(vertices);
+
+ for (i = 0; i < graph->vertices.count; i++) {
+ struct kcctpl_vertex *vertex = &graph->vertices.data[i];
+
+ if (vertex->color == WHITE) {
+ continue;
+ }
+
+ if ((vertex->color == BLACK && !include_black) ||
+ !kcctpl_guid_list_contains(vertex->accept_black, type) ||
+ !kcctpl_guid_list_contains(vertex->accept_red_red, type)) {
+ vertex->repl_info.cost = UINT32_MAX;
+ vertex->root_id = GUID_zero();
+ vertex->demoted = true;
+ } else {
+ struct kcctpl_vertex *new_data;
+
+ new_data = talloc_realloc(mem_ctx, vertices.data,
+ struct kcctpl_vertex,
+ vertices.count + 1);
+ NT_STATUS_HAVE_NO_MEMORY(new_data);
+ new_data[vertices.count] = *vertex;
+ vertices.data = new_data;
+ vertices.count++;
+ }
+ }
+
+ *_vertices = vertices;
+ return NT_STATUS_OK;
+}
+
+/**
+ * merge schedules, replication intervals, options and costs.
+ */
+static bool kcctpl_combine_repl_info(struct kcctpl_graph *graph,
+ struct kcctpl_repl_info *ria,
+ struct kcctpl_repl_info *rib,
+ struct kcctpl_repl_info *ric)
+{
+ uint8_t schedule[84];
+ bool is_available;
+ uint32_t i;
+ int32_t ric_cost;
+
+ is_available = false;
+ for (i = 0; i < 84; i++) {
+ schedule[i] = ria->schedule[i] & rib->schedule[i];
+
+ if (schedule[i] == 1) {
+ is_available = true;
+ }
+ }
+ if (!is_available) {
+ return false;
+ }
+
+ ric_cost = ria->cost + rib->cost;
+ ric->cost = (ric_cost < 0) ? UINT32_MAX : ric_cost;
+
+ ric->interval = MAX(ria->interval, rib->interval);
+ ric->options = ria->options & rib->options;
+ memcpy(&ric->schedule, &schedule, 84);
+
+ return true;
+}
+
+/**
+ * helper function for Dijkstra's algorithm. a new path has been found from a
+ * root vertex to vertex 'vertex2'. this path is ('vertex1->root, ..., vertex1,
+ * vertex2'). 'edge' is the edge connecting 'vertex1' and 'vertex2'. if this new
+ * path is better (in this case cheaper, or has a longer schedule), update
+ * 'vertex2' to use the new path.
+ */
+static NTSTATUS kcctpl_try_new_path(TALLOC_CTX *mem_ctx,
+ struct kcctpl_graph *graph,
+ struct kcctpl_vertex_list vertices,
+ struct kcctpl_vertex *vertex1,
+ struct kcctpl_multi_edge *edge,
+ struct kcctpl_vertex *vertex2)
+{
+ struct kcctpl_repl_info new_repl_info;
+ bool intersect;
+ uint32_t i, new_duration, old_duration;
+
+ ZERO_STRUCT(new_repl_info);
+
+ intersect = kcctpl_combine_repl_info(graph, &vertex1->repl_info,
+ &edge->repl_info, &new_repl_info);
+
+ if (new_repl_info.cost > vertex2->repl_info.cost) {
+ return NT_STATUS_OK;
+ }
+
+ if (new_repl_info.cost < vertex2->repl_info.cost && !intersect) {
+ return NT_STATUS_OK;
+ }
+
+ new_duration = old_duration = 0;
+ for (i = 0; i < 84; i++) {
+ if (new_repl_info.schedule[i] == 1) {
+ new_duration++;
+ }
+
+ if (vertex2->repl_info.schedule[i] == 1) {
+ old_duration++;
+ }
+ }
+
+ if (new_repl_info.cost < vertex2->repl_info.cost ||
+ new_duration > old_duration) {
+ struct kcctpl_vertex *new_data;
+
+ vertex2->root_id = vertex1->root_id;
+ vertex2->component_id = vertex1->component_id;
+ vertex2->repl_info = new_repl_info;
+
+ new_data = talloc_realloc(mem_ctx, vertices.data,
+ struct kcctpl_vertex,
+ vertices.count + 1);
+ NT_STATUS_HAVE_NO_MEMORY(new_data);
+ new_data[vertices.count + 1] = *vertex2;
+ vertices.data = new_data;
+ vertices.count++;
+ }
+
+ return NT_STATUS_OK;
+}
+
+/**
+ * run Dijkstra's algorithm with the red (and possibly black) vertices as the
+ * root vertices, and build up a shortest-path forest.
+ */
+static NTSTATUS kcctpl_dijkstra(struct kcctpl_graph *graph, struct GUID type,
+ bool include_black)
+{
+ TALLOC_CTX *tmp_ctx;
+ struct kcctpl_vertex_list vertices;
+ NTSTATUS status;
+
+ tmp_ctx = talloc_new(graph);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ status = kcctpl_setup_dijkstra(tmp_ctx, graph, type, include_black,
+ &vertices);
+ if (NT_STATUS_IS_ERR(status)) {
+ DEBUG(1, (__location__ ": failed to build the initial sequence "
+ "for Dijkstra's algorithm: %s\n", nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+
+ while (vertices.count > 0) {
+ uint32_t minimum_cost, minimum_index, i;
+ struct kcctpl_vertex *minimum_vertex, *new_data;
+
+ minimum_cost = UINT32_MAX;
+ minimum_index = -1;
+ minimum_vertex = NULL;
+ for (i = 0; i < vertices.count; i++) {
+ struct kcctpl_vertex *vertex = &vertices.data[i];
+
+ if (vertex->repl_info.cost < minimum_cost) {
+ minimum_cost = vertex->repl_info.cost;
+ minimum_vertex = vertex;
+ minimum_index = i;
+ } else if (vertex->repl_info.cost == minimum_cost &&
+ GUID_compare(&vertex->id,
+ &minimum_vertex->id) < 0) {
+ minimum_vertex = vertex;
+ minimum_index = i;
+ }
+ }
+
+ if (minimum_index < vertices.count - 1) {
+ memcpy(&vertices.data[minimum_index + 1],
+ &vertices.data[minimum_index],
+ vertices.count - minimum_index - 1);
+ }
+ new_data = talloc_realloc(tmp_ctx, vertices.data,
+ struct kcctpl_vertex,
+ vertices.count - 1);
+ NT_STATUS_HAVE_NO_MEMORY_AND_FREE(new_data, tmp_ctx);
+ talloc_free(vertices.data);
+ vertices.data = new_data;
+ vertices.count--;
+
+ for (i = 0; i < graph->edges.count; i++) {
+ struct kcctpl_multi_edge *edge = &graph->edges.data[i];
+
+ if (kcctpl_guid_list_contains(minimum_vertex->edge_ids,
+ edge->id)) {
+ uint32_t j;
+
+ for (j = 0; j < edge->vertex_ids.count; j++) {
+ struct GUID vertex_id;
+ struct kcctpl_vertex *vertex;
+
+ vertex_id = edge->vertex_ids.data[j];
+ vertex = kcctpl_find_vertex_by_guid(graph,
+ vertex_id);
+ if (!vertex) {
+ DEBUG(1, (__location__
+ ": failed to find "
+ "vertex %s\n",
+ GUID_string(tmp_ctx,
+ &vertex_id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ kcctpl_try_new_path(tmp_ctx, graph,
+ vertices,
+ minimum_vertex,
+ edge, vertex);
+ }
+ }
+ }
+ }
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_OK;
+}
+
+/**
+ * add an edge to the list of edges that will be processed with Kruskal's. the
+ * endpoints are in fact the root of the vertices to pass in, so the endpoints
+ * are always colored vertices.
+ */
+static NTSTATUS kcctpl_add_int_edge(TALLOC_CTX *mem_ctx,
+ struct kcctpl_graph *graph,
+ struct kcctpl_internal_edge_list internal_edges,
+ struct kcctpl_multi_edge *edge,
+ struct kcctpl_vertex *vertex1,
+ struct kcctpl_vertex *vertex2)
+{
+ struct kcctpl_vertex *root1, *root2;
+ bool red_red, found;
+ struct kcctpl_repl_info repl_info1, repl_info2;
+ struct kcctpl_internal_edge new_internal_edge, *new_data;
+ uint32_t i;
+
+ root1 = kcctpl_find_vertex_by_guid(graph, vertex1->root_id);
+ if (!root1) {
+ TALLOC_CTX *tmp_ctx = talloc_new(graph);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ DEBUG(1, (__location__ ": failed to find vertex %s\n",
+ GUID_string(tmp_ctx, &vertex1->root_id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ root2 = kcctpl_find_vertex_by_guid(graph, vertex2->root_id);
+ if (!root2) {
+ TALLOC_CTX *tmp_ctx = talloc_new(graph);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ DEBUG(1, (__location__ ": failed to find vertex %s\n",
+ GUID_string(tmp_ctx, &vertex2->root_id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ red_red = (root1->color == RED && root2->color == RED);
+
+ if (red_red) {
+ if (!kcctpl_guid_list_contains(root1->accept_red_red,
+ edge->type) ||
+ !kcctpl_guid_list_contains(root2->accept_red_red,
+ edge->type)) {
+ return NT_STATUS_OK;
+ }
+ } else if (!kcctpl_guid_list_contains(root1->accept_black,
+ edge->type) ||
+ !kcctpl_guid_list_contains(root2->accept_black,
+ edge->type)) {
+ return NT_STATUS_OK;
+ }
+
+ if (!kcctpl_combine_repl_info(graph, &vertex1->repl_info,
+ &vertex2->repl_info, &repl_info1) ||
+ !kcctpl_combine_repl_info(graph, &repl_info1, &edge->repl_info,
+ &repl_info2)) {
+ return NT_STATUS_OK;
+ }
+
+ new_internal_edge.v1id = root1->id;
+ new_internal_edge.v2id = root2->id;
+ new_internal_edge.red_red = red_red;
+ new_internal_edge.repl_info = repl_info2;
+ new_internal_edge.type = edge->type;
+
+ if (GUID_compare(&new_internal_edge.v1id,
+ &new_internal_edge.v2id) > 0) {
+ struct GUID tmp_guid = new_internal_edge.v1id;
+
+ new_internal_edge.v1id = new_internal_edge.v2id;
+ new_internal_edge.v2id = tmp_guid;
+ }
+
+ found = false;
+ for (i = 0; i < internal_edges.count; i++) {
+ struct kcctpl_internal_edge *ie = &internal_edges.data[i];
+
+ if (kcctpl_internal_edge_equal(ie, &new_internal_edge)) {
+ found = true;
+ }
+ }
+ if (found) {
+ return NT_STATUS_OK;
+ }
+
+ new_data = talloc_realloc(mem_ctx, internal_edges.data,
+ struct kcctpl_internal_edge,
+ internal_edges.count + 1);
+ NT_STATUS_HAVE_NO_MEMORY(new_data);
+ new_data[internal_edges.count + 1] = new_internal_edge;
+ internal_edges.data = new_data;
+ internal_edges.count++;
+
+ return NT_STATUS_OK;
+}
+
+/**
+ * after running Dijkstra's algorithm, this function examines a multi-edge and
+ * adds internal edges between every tree connected by this edge.
+ */
+static NTSTATUS kcctpl_process_edge(TALLOC_CTX *mem_ctx,
+ struct kcctpl_graph *graph,
+ struct kcctpl_multi_edge *edge,
+ struct kcctpl_internal_edge_list internal_edges)
+{
+ TALLOC_CTX *tmp_ctx;
+ struct kcctpl_vertex_list vertices;
+ uint32_t i;
+ struct kcctpl_vertex *best_vertex;
+
+ ZERO_STRUCT(vertices);
+
+ tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ for (i = 0; i < edge->vertex_ids.count; i++) {
+ struct GUID id;
+ struct kcctpl_vertex *vertex, *new_data;
+
+ id = edge->vertex_ids.data[i];
+
+ vertex = kcctpl_find_vertex_by_guid(graph, id);
+ if (!vertex) {
+ DEBUG(1, (__location__ ": failed to find vertex %s\n",
+ GUID_string(tmp_ctx, &id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ new_data = talloc_realloc(tmp_ctx, vertices.data,
+ struct kcctpl_vertex,
+ vertices.count + 1);
+ NT_STATUS_HAVE_NO_MEMORY_AND_FREE(new_data, tmp_ctx);
+ new_data[vertices.count] = *vertex;
+ vertices.data = new_data;
+ vertices.count++;
+ }
+
+ qsort(vertices.data, vertices.count, sizeof(struct kcctpl_vertex),
+ kcctpl_sort_vertices);
+
+ best_vertex = &vertices.data[0];
+
+ for (i = 0; i < edge->vertex_ids.count; i++) {
+ struct GUID id, empty_id = GUID_zero();
+ struct kcctpl_vertex *vertex = &graph->vertices.data[i];
+
+ id = edge->vertex_ids.data[i];
+
+ vertex = kcctpl_find_vertex_by_guid(graph, id);
+ if (!vertex) {
+ DEBUG(1, (__location__ ": failed to find vertex %s\n",
+ GUID_string(tmp_ctx, &id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ if (!GUID_equal(&vertex->component_id, &empty_id) &&
+ !GUID_equal(&vertex->root_id, &empty_id)) {
+ continue;
+ }
+
+ if (!GUID_equal(&best_vertex->component_id,
+ &empty_id) &&
+ !GUID_equal(&best_vertex->root_id, &empty_id) &&
+ !GUID_equal(&vertex->component_id, &empty_id) &&
+ !GUID_equal(&vertex->root_id, &empty_id) &&
+ !GUID_equal(&best_vertex->component_id,
+ &vertex->component_id)) {
+ NTSTATUS status;
+
+ status = kcctpl_add_int_edge(mem_ctx, graph,
+ internal_edges,
+ edge, best_vertex,
+ vertex);
+ if (NT_STATUS_IS_ERR(status)) {
+ DEBUG(1, (__location__ ": failed to add an "
+ "internal edge for %s: %s\n",
+ GUID_string(tmp_ctx, &vertex->id),
+ nt_errstr(status)));
+ talloc_free(tmp_ctx);
+ return status;
+ }
+ }
+ }
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_OK;
+}
+
+/**
+ * after running Dijkstra's algorithm to determine the shortest-path forest,
+ * examine all edges in this edge set. find all inter-tree edges, from which to
+ * build the list of 'internal edges', which will later be passed on to
+ * Kruskal's algorithm.
+ */
+static NTSTATUS kcctpl_process_edge_set(TALLOC_CTX *mem_ctx,
+ struct kcctpl_graph *graph,
+ struct kcctpl_multi_edge_set *set,
+ struct kcctpl_internal_edge_list internal_edges)
+{
+ uint32_t i;
+
+ if (!set) {
+ for (i = 0; i < graph->edges.count; i++) {
+ struct kcctpl_multi_edge *edge;
+ uint32_t j;
+ NTSTATUS status;
+
+ edge = &graph->edges.data[i];
+
+ for (j = 0; j < edge->vertex_ids.count; j++) {
+ struct GUID id;
+ struct kcctpl_vertex *vertex;
+
+ id = edge->vertex_ids.data[j];
+
+ vertex = kcctpl_find_vertex_by_guid(graph, id);
+ if (!vertex) {
+ TALLOC_CTX *tmp_ctx;
+
+ tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ DEBUG(1, (__location__ ": failed to "
+ "find vertex %s\n",
+ GUID_string(tmp_ctx, &id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ kcctpl_check_demote_one_vertex(vertex,
+ edge->type);
+ }
+
+ status = kcctpl_process_edge(mem_ctx, graph, edge,
+ internal_edges);
+ if (NT_STATUS_IS_ERR(status)) {
+ TALLOC_CTX *tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ DEBUG(1, (__location__ ": failed to process "
+ "edge %s: %s\n",
+ GUID_string(tmp_ctx, &edge->id),
+ nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+
+ for (j = 0; j < edge->vertex_ids.count; j++) {
+ struct GUID id;
+ struct kcctpl_vertex *vertex;
+
+ id = edge->vertex_ids.data[j];
+
+ vertex = kcctpl_find_vertex_by_guid(graph, id);
+ if (!vertex) {
+ TALLOC_CTX *tmp_ctx;
+
+ tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ DEBUG(1, (__location__ ": failed to "
+ "find vertex %s\n",
+ GUID_string(tmp_ctx, &id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ kcctpl_undemote_one_vertex(vertex);
+ }
+ }
+ } else {
+ for (i = 0; i < graph->edges.count; i++) {
+ struct kcctpl_multi_edge *edge = &graph->edges.data[i];
+
+ if (kcctpl_guid_list_contains(set->edge_ids,
+ edge->id)) {
+ NTSTATUS status;
+
+ status = kcctpl_process_edge(mem_ctx, graph,
+ edge,
+ internal_edges);
+ if (NT_STATUS_IS_ERR(status)) {
+ TALLOC_CTX *tmp_ctx;
+
+ tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ DEBUG(1, (__location__ ": failed to "
+ "process edge %s: %s\n",
+ GUID_string(tmp_ctx,
+ &edge->id),
+ nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+ }
+ }
+ }
+
+ return NT_STATUS_OK;
+}
+
+/**
+ * a new edge, 'internal_edge', has been found for the spanning tree edge. add
+ * this edge to the list of output edges.
+ */
+static NTSTATUS kcctpl_add_out_edge(TALLOC_CTX *mem_ctx,
+ struct kcctpl_graph *graph,
+ struct kcctpl_multi_edge_list output_edges,
+ struct kcctpl_internal_edge *internal_edge)
+{
+ struct kcctpl_vertex *vertex1, *vertex2;
+ TALLOC_CTX *tmp_ctx;
+ struct kcctpl_multi_edge *new_edge, *new_data;
+ struct GUID *new_data_id;
+
+ tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ vertex1 = kcctpl_find_vertex_by_guid(graph, internal_edge->v1id);
+ if (!vertex1) {
+ DEBUG(1, (__location__ ": failed to find vertex %s\n",
+ GUID_string(tmp_ctx, &internal_edge->v1id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ vertex2 = kcctpl_find_vertex_by_guid(graph, internal_edge->v2id);
+ if (!vertex2) {
+ DEBUG(1, (__location__ ": failed to find vertex %s\n",
+ GUID_string(tmp_ctx, &internal_edge->v2id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ new_edge = talloc(tmp_ctx, struct kcctpl_multi_edge);
+ NT_STATUS_HAVE_NO_MEMORY_AND_FREE(new_edge, tmp_ctx);
+
+ new_edge->id = GUID_random(); /* TODO: what should be new_edge->GUID? */
+ new_edge->directed = false;
+
+ new_edge->vertex_ids.data = talloc_array(new_edge, struct GUID, 2);
+ NT_STATUS_HAVE_NO_MEMORY_AND_FREE(new_edge->vertex_ids.data, tmp_ctx);
+
+ new_edge->vertex_ids.data[0] = vertex1->id;
+ new_edge->vertex_ids.data[1] = vertex2->id;
+ new_edge->vertex_ids.count = 2;
+
+ new_edge->type = internal_edge->type;
+ new_edge->repl_info = internal_edge->repl_info;
+
+ new_data = talloc_realloc(tmp_ctx, output_edges.data,
+ struct kcctpl_multi_edge,
+ output_edges.count + 1);
+ NT_STATUS_HAVE_NO_MEMORY_AND_FREE(new_data, tmp_ctx);
+ new_data[output_edges.count + 1] = *new_edge;
+ output_edges.data = new_data;
+ output_edges.count++;
+
+ new_data_id = talloc_realloc(vertex1, vertex1->edge_ids.data,
+ struct GUID, vertex1->edge_ids.count);
+ NT_STATUS_HAVE_NO_MEMORY_AND_FREE(new_data_id, tmp_ctx);
+ new_data_id[vertex1->edge_ids.count] = new_edge->id;
+ talloc_free(vertex1->edge_ids.data);
+ vertex1->edge_ids.data = new_data_id;
+ vertex1->edge_ids.count++;
+
+ new_data_id = talloc_realloc(vertex2, vertex2->edge_ids.data,
+ struct GUID, vertex2->edge_ids.count);
+ NT_STATUS_HAVE_NO_MEMORY_AND_FREE(new_data_id, tmp_ctx);
+ new_data_id[vertex2->edge_ids.count] = new_edge->id;
+ talloc_free(vertex2->edge_ids.data);
+ vertex2->edge_ids.data = new_data_id;
+ vertex2->edge_ids.count++;
+
+ talloc_steal(graph, new_edge);
+ talloc_steal(mem_ctx, output_edges.data);
+ talloc_free(tmp_ctx);
+ return NT_STATUS_OK;
+}
+
+/**
+ * run Kruskal's minimum-cost spanning tree algorithm on the internal edges
+ * (that represent shortest paths in the original graph between colored
+ * vertices).
+ */
+static NTSTATUS kcctpl_kruskal(TALLOC_CTX *mem_ctx, struct kcctpl_graph *graph,
+ struct kcctpl_internal_edge_list internal_edges,
+ struct kcctpl_multi_edge_list *_output_edges)
+{
+ uint32_t i, num_expected_tree_edges, cst_edges;
+ struct kcctpl_multi_edge_list output_edges;
+
+ num_expected_tree_edges = 0;
+ for (i = 0; i < graph->vertices.count; i++) {
+ struct kcctpl_vertex *vertex = &graph->vertices.data[i];
+
+ talloc_free(vertex->edge_ids.data);
+ ZERO_STRUCT(vertex->edge_ids);
+
+ if (vertex->color == RED || vertex->color == WHITE) {
+ num_expected_tree_edges++;
+ }
+ }
+
+ qsort(internal_edges.data, internal_edges.count,
+ sizeof(struct kcctpl_internal_edge), kcctpl_sort_internal_edges);
+
+ cst_edges = 0;
+
+ ZERO_STRUCT(output_edges);
+
+ while (internal_edges.count > 0 &&
+ cst_edges < num_expected_tree_edges) {
+ struct kcctpl_internal_edge *edge, *new_data;
+ struct kcctpl_vertex *vertex1, *vertex2;
+ struct GUID comp1, comp2;
+
+ edge = &internal_edges.data[0];
+
+ vertex1 = kcctpl_find_vertex_by_guid(graph, edge->v1id);
+ if (!vertex1) {
+ TALLOC_CTX *tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ DEBUG(1, (__location__ ": failed to find vertex %s\n",
+ GUID_string(tmp_ctx, &edge->v1id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ vertex2 = kcctpl_find_vertex_by_guid(graph, edge->v2id);
+ if (!vertex2) {
+ TALLOC_CTX *tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ DEBUG(1, (__location__ ": failed to find vertex %s\n",
+ GUID_string(tmp_ctx, &edge->v2id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ comp1 = kcctpl_get_component_id(graph, vertex1);
+ comp2 = kcctpl_get_component_id(graph, vertex2);
+
+ if (!GUID_equal(&comp1, &comp2)) {
+ NTSTATUS status;
+ struct kcctpl_vertex *vertex;
+
+ cst_edges++;
+
+ status = kcctpl_add_out_edge(mem_ctx, graph,
+ output_edges, edge);
+ if (NT_STATUS_IS_ERR(status)) {
+ TALLOC_CTX *tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ DEBUG(1, (__location__ ": failed to add an "
+ "output edge between %s and %s: %s\n",
+ GUID_string(tmp_ctx, &edge->v1id),
+ GUID_string(tmp_ctx, &edge->v2id),
+ nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+
+ vertex = kcctpl_find_vertex_by_guid(graph, comp1);
+ if (!vertex) {
+ TALLOC_CTX *tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ DEBUG(1, (__location__ ": failed to find "
+ "vertex %s\n", GUID_string(tmp_ctx,
+ &comp1)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+ vertex->component_id = comp2;
+ }
+
+ internal_edges.data = internal_edges.data + 1;
+ new_data = talloc_realloc(mem_ctx, internal_edges.data,
+ struct kcctpl_internal_edge,
+ internal_edges.count - 1);
+ NT_STATUS_HAVE_NO_MEMORY(new_data);
+ talloc_free(internal_edges.data);
+ internal_edges.data = new_data;
+ internal_edges.count--;
+ }
+
+ *_output_edges = output_edges;
+ return NT_STATUS_OK;
+}
+
+/**
+ * count the number of components. a component is considered to be a bunch of
+ * colored vertices that are connected by the spanning tree. vertices whose
+ * component ID is the same as their vertex ID are the root of the connected
+ * component.
+ */
+static uint32_t kcctpl_count_components(struct kcctpl_graph *graph)
+{
+ uint32_t num_components = 0, i;
+
+ for (i = 0; i < graph->vertices.count; i++) {
+ struct kcctpl_vertex *vertex;
+ struct GUID component_id;
+
+ vertex = &graph->vertices.data[i];
+
+ if (vertex->color == WHITE) {
+ continue;
+ }
+
+ component_id = kcctpl_get_component_id(graph, vertex);
+ if (GUID_equal(&component_id, &vertex->id)) {
+ vertex->component_index = num_components;
+ num_components++;
+ }
+ }
+
+ return num_components;
+}
+
+/**
+ * calculate the spanning tree and return the edges that include the vertex for
+ * the local site.
+ */
+static NTSTATUS kcctpl_get_spanning_tree_edges(struct ldb_context *ldb,
+ TALLOC_CTX *mem_ctx,
+ struct kcctpl_graph *graph,
+ uint32_t *_component_count,
+ struct kcctpl_multi_edge_list *_st_edge_list)
+{
+ TALLOC_CTX *tmp_ctx;
+ struct kcctpl_internal_edge_list internal_edges;
+ uint32_t i, component_count;
+ NTSTATUS status;
+ struct kcctpl_multi_edge_list output_edges, st_edge_list;
+
+ ZERO_STRUCT(internal_edges);
+
+ tmp_ctx = talloc_new(mem_ctx);
+ NT_STATUS_HAVE_NO_MEMORY(tmp_ctx);
+
+ for (i = 0; i < graph->edge_sets.count; i++) {
+ struct kcctpl_multi_edge_set *set;
+ struct GUID edge_type;
+ uint32_t j;
+
+ set = &graph->edge_sets.data[i];
+
+ edge_type = GUID_zero();
+
+ for (j = 0; j < graph->vertices.count; j++) {
+ struct kcctpl_vertex *vertex = &graph->vertices.data[j];
+
+ talloc_free(vertex->edge_ids.data);
+ ZERO_STRUCT(vertex->edge_ids.data);
+ }
+
+ for (j = 0; j < set->edge_ids.count; j++) {
+ struct GUID edge_id;
+ struct kcctpl_multi_edge *edge;
+ uint32_t k;
+
+ edge_id = set->edge_ids.data[j];
+ edge = kcctpl_find_edge_by_guid(graph, edge_id);
+ if (!edge) {
+ DEBUG(1, (__location__ ": failed to find a "
+ "graph edge with ID=%s\n",
+ GUID_string(tmp_ctx, &edge_id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ edge_type = edge->type;
+
+ for (k = 0; k < edge->vertex_ids.count; k++) {
+ struct GUID vertex_id, *new_data;
+ struct kcctpl_vertex *vertex;
+
+ vertex_id = edge->vertex_ids.data[k];
+ vertex = kcctpl_find_vertex_by_guid(graph,
+ vertex_id);
+ if (!vertex) {
+ DEBUG(1, (__location__ ": failed to "
+ "find a graph vertex with "
+ "ID=%s\n",
+ GUID_string(tmp_ctx,
+ &edge_id)));
+
+ talloc_free(tmp_ctx);
+ return NT_STATUS_INTERNAL_DB_CORRUPTION;
+ }
+
+ new_data = talloc_realloc(tmp_ctx,
+ vertex->edge_ids.data,
+ struct GUID,
+ vertex->edge_ids.count + 1);
+ NT_STATUS_HAVE_NO_MEMORY_AND_FREE(new_data,
+ tmp_ctx);
+ new_data[vertex->edge_ids.count] = edge->id;
+ vertex->edge_ids.data = new_data;
+ vertex->edge_ids.count++;
+ }
+ }
+
+ status = kcctpl_dijkstra(graph, edge_type, false);
+ if (NT_STATUS_IS_ERR(status)) {
+ DEBUG(1, (__location__ ": failed to run Dijkstra's "
+ "algorithm: %s\n", nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+
+ status = kcctpl_process_edge_set(tmp_ctx, graph, set,
+ internal_edges);
+ if (NT_STATUS_IS_ERR(status)) {
+ DEBUG(1, (__location__ ": failed to process edge set "
+ "%s: %s\n", GUID_string(tmp_ctx, &set->id),
+ nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+
+ status = kcctpl_dijkstra(graph, edge_type, true);
+ if (NT_STATUS_IS_ERR(status)) {
+ DEBUG(1, (__location__ ": failed to run Dijkstra's "
+ "algorithm: %s\n", nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+
+ status = kcctpl_process_edge_set(tmp_ctx, graph, set,
+ internal_edges);
+ if (NT_STATUS_IS_ERR(status)) {
+ DEBUG(1, (__location__ ": failed to process edge set "
+ "%s: %s\n", GUID_string(tmp_ctx, &set->id),
+ nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+ }
+
+ kcctpl_setup_vertices(graph);
+
+ status = kcctpl_process_edge_set(tmp_ctx, graph, NULL, internal_edges);
+ if (NT_STATUS_IS_ERR(status)) {
+ DEBUG(1, (__location__ ": failed to process empty edge set: "
+ "%s\n", nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+
+ status = kcctpl_kruskal(tmp_ctx, graph, internal_edges, &output_edges);
+ if (NT_STATUS_IS_ERR(status)) {
+ DEBUG(1, (__location__ ": failed to run Kruskal's algorithm: "
+ "%s\n", nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+
+ for (i = 0; i < graph->vertices.count; i++) {
+ struct kcctpl_vertex *vertex = &graph->vertices.data[i];
+
+ if (vertex->color == RED) {
+ vertex->dist_to_red = 0;
+ } else if (true) { /* TODO: if there exists a path from 'vertex'
+ to a RED vertex */
+ vertex->dist_to_red = -1; /* TODO: the length of the
+ shortest such path */
+ } else {
+ vertex->dist_to_red = UINT32_MAX;
+ }
+ }
+
+ component_count = kcctpl_count_components(graph);
+
+ status = kcctpl_copy_output_edges(ldb, tmp_ctx, graph, output_edges,
+ &st_edge_list);
+ if (NT_STATUS_IS_ERR(status)) {
+ DEBUG(1, (__location__ ": failed to copy edge list: %s\n",
+ nt_errstr(status)));
+
+ talloc_free(tmp_ctx);
+ return status;
+ }
+
+ *_component_count = component_count;
+ talloc_steal(mem_ctx, st_edge_list.data);
+ *_st_edge_list = st_edge_list;
+ talloc_free(tmp_ctx);
+ return NT_STATUS_OK;
+}