/*++ Copyright (c) 2020 Microsoft Corporation Module Name: smt_parallel.cpp Abstract: Parallel SMT, portfolio loop specialized to SMT core. Author: nbjorner 2020-01-31 Ilana Shapiro 2025 --*/ #include "util/scoped_ptr_vector.h" #include "ast/ast_util.h" #include "ast/ast_pp.h" #include "ast/ast_ll_pp.h" #include "ast/ast_translation.h" #include "ast/simplifiers/then_simplifier.h" #include "smt/smt_parallel.h" #include "smt/smt_lookahead.h" #include "solver/solver_preprocess.h" #include "params/smt_parallel_params.hpp" #include #include #include class bounded_pp_exprs { expr_ref_vector const &es; public: bounded_pp_exprs(expr_ref_vector const &es) : es(es) {} std::ostream &display(std::ostream &out) const { for (auto e : es) out << mk_bounded_pp(e, es.get_manager()) << "\n"; return out; } }; inline std::ostream &operator<<(std::ostream &out, bounded_pp_exprs const &pp) { return pp.display(out); } #ifdef SINGLE_THREAD namespace smt { lbool parallel::operator()(expr_ref_vector const &asms) { return l_undef; } } // namespace smt #else #include #define LOG_WORKER(lvl, s) IF_VERBOSE(lvl, verbose_stream() << "Worker " << id << s) #define LOG_BB_WORKER(lvl, s) IF_VERBOSE(lvl, verbose_stream() << "Backbones Worker " << id << s) namespace smt { static bool is_cancellation_exception(char const *msg) { return msg && (strstr(msg, "canceled") != nullptr || strstr(msg, "cancelled") != nullptr); } void parallel::sls_worker::run() { ptr_vector assertions; p.ctx.get_assertions(assertions); for (expr *e : assertions) m_sls->assert_expr(m_g2l(e)); lbool res = l_undef; try { if (!m.inc()) return; res = m_sls->check(); } catch (z3_exception &ex) { // Cancellation is normal in portfolio mode if (m.limit().is_canceled()) { IF_VERBOSE(1, verbose_stream() << "SLS worker canceled\n"); return; } if (strstr(ex.what(), "unsupported for sls") != nullptr) { IF_VERBOSE(1, verbose_stream() << "SLS opted out: " << ex.what() << "\n"); return; } // Anything else is a real error IF_VERBOSE(1, verbose_stream() << "SLS threw exception: " << ex.what() << "\n"); b.set_exception(ex.what()); return; } if (res == l_true) { IF_VERBOSE(2, verbose_stream() << "SLS worker found SAT\n"); model_ref mdl = m_sls->get_model(); b.set_sat(m_l2g, *mdl); } } void parallel::backbones_worker::run() { if (m_use_failed_literal_test) run_failed_literal_mode(); else run_batch_mode(); } lbool parallel::backbones_worker::check_sat(expr_ref_vector const &asms) { lbool r = l_undef; try { r = ctx->check(asms.size(), asms.data()); } catch (z3_error &err) { if (!m.limit().is_canceled()) b.set_exception(err.error_code()); } catch (z3_exception &ex) { if (!m.limit().is_canceled() && !is_cancellation_exception(ex.what())) b.set_exception(ex.what()); } catch (...) { if (!m.limit().is_canceled()) b.set_exception("unknown exception"); } return r; } void parallel::backbones_worker::run_failed_literal_mode() { ctx->get_fparams().m_max_conflicts = 10; auto num_atoms = ctx->get_num_bool_vars(); uint_set units; for (unsigned v = 0; v < num_atoms && m.inc(); ++v) if (ctx->get_assignment(v) != l_undef && ctx->get_assign_level(v) == ctx->m_base_lvl) units.insert(v); while (m.inc()) { for (unsigned v = 0; v < num_atoms && m.inc(); ++v) { if (units.contains(v)) continue; expr_ref e(ctx->bool_var2expr(v), m); if (!e) continue; if (m.is_or(e) || m.is_ite(e)) continue; if (ctx->get_assignment(v) != l_undef && ctx->get_assign_level(v) == ctx->m_base_lvl) { bool is_true = ctx->get_assignment(v) == l_true; IF_VERBOSE(2, verbose_stream() << "backbone on trail " << e << "\n"); units.insert(v); if (!is_true) e = m.mk_not(e); if (b.collect_global_backbone(m_l2g, e)) m_stats.m_backbones_found++; continue; } auto r = probe_literal(v, units, e); if (r != l_undef) break; if (units.contains(v)) continue; e = mk_not(e); r = probe_literal(v, units, e); if (r != l_undef) break; } } } lbool parallel::backbones_worker::probe_literal(bool_var v, uint_set& units, expr *e) { asms.push_back(e); auto r = check_sat(asms); asms.pop_back(); if (r == l_false) { if (!ctx->unsat_core().contains(e)) { b.set_unsat(m_l2g, ctx->unsat_core()); return l_false; } IF_VERBOSE(2, verbose_stream() << "failed literal " << mk_bounded_pp(e, m) << "\n"); expr_ref not_e(mk_not(m, e), m); if (b.collect_global_backbone(m_l2g, not_e)) m_stats.m_backbones_found++; units.insert(v); r = l_undef; } if (r == l_true) { model_ref mdl; ctx->get_model(mdl); b.set_sat(m_l2g, *mdl); } return r; } void parallel::backbones_worker::run_batch_mode() { bb_candidates bb_candidates; while (m.inc()) { if (!b.wait_for_backbone_job(id, m_g2l, bb_candidates, m.limit())) return; if (bb_candidates.empty()) continue; LOG_BB_WORKER(1, " received batch of " << bb_candidates.size() << " candidates\n"); unsigned local_cancel_epoch = b.get_cancel_epoch(); auto canceled = [&] { return local_cancel_epoch != b.get_cancel_epoch(); }; auto fallback_singletons = [&](expr_ref_vector const& chunk_lits, expr_ref_vector& bb_candidate_lits) { m_stats.m_fallback_singleton_checks++; for (expr* c : chunk_lits) { if (!bb_candidate_lits.contains(c)) continue; // already handled in singleton core → backbone case below expr_ref bb_ref(c, m); if (!m.inc() || canceled()) return; if (m_mode == bb_mode::bb_positive) bb_ref = mk_not(bb_ref); // F ∧ U since check_backbone flips it back if (!b.is_global_backbone(m_l2g, bb_ref) && check_backbone(bb_ref)) { m_stats.m_backbones_detected++; LOG_BB_WORKER(1, " fallback found backbone: " << mk_bounded_pp(bb_ref.get(), m, 3) << "\n"); bool is_new_bb = b.collect_global_backbone(m_l2g, bb_ref); if (is_new_bb) { m_stats.m_backbones_found++; ctx->assert_expr(bb_ref.get()); // since bb workers don't collect clauses they themselves shared } } bb_candidate_lits.erase(c); } }; m_stats.m_batches_total++; m_stats.m_candidates_total += bb_candidates.size(); expr_ref_vector bb_candidate_lits(m); for (auto const& c : bb_candidates) bb_candidate_lits.push_back(c.lit); unsigned chunk_delta = 1; // in mode bb_neg this is Algorithm 7 from https://sat.inesc-id.pt/~mikolas/bb-aicom-preprint.pdf while (!bb_candidate_lits.empty() && !canceled() && m.inc()) { // remove candidates that the other backbone thread found to be backbones if (m_num_global_bb_threads > 1) { for (unsigned i = 0; i < bb_candidate_lits.size();) { expr* tmp = bb_candidate_lits.get(i); if (b.is_global_backbone(m_l2g, tmp)) bb_candidate_lits.erase(i); else ++i; } } unsigned chunk_size = std::min(m_bb_chunk_size * chunk_delta, bb_candidate_lits.size()); expr_ref_vector chunk_lits(m); expr_ref_vector negated_chunk_lits(m); for (unsigned i = 0; i < chunk_size; ++i) { expr *e = bb_candidate_lits.get(i); chunk_lits.push_back(e); negated_chunk_lits.push_back(m.mk_not(e)); } expr_ref_vector bb_asms(m); if (m_mode == bb_mode::bb_negated) bb_asms.append(negated_chunk_lits); // F ∧ ¬U else bb_asms.append(chunk_lits); // F ∧ U collect_shared_clauses(); while (true) { if (!m.inc()) return; if (canceled()) break; m_stats.m_core_refinement_rounds++; unsigned base_asms_sz = asms.size(); for (expr* a : bb_asms) asms.push_back(a); lbool r = check_sat(asms); asms.shrink(base_asms_sz); if (!m.inc()) return; if (canceled()) break; if (r == l_undef) { LOG_BB_WORKER(1, " UNDEF at chunk_size=" << chunk_size << "\n"); if (chunk_size < bb_candidate_lits.size()) { chunk_delta++; // try again with a bigger chunk m_stats.m_num_chunk_increases++; break; } LOG_BB_WORKER(1, " UNDEF and max chunk → fallback\n"); fallback_singletons(chunk_lits, bb_candidate_lits); m_stats.m_fallback_reason_undef++; chunk_delta = 1; break; } if (r == l_true) { LOG_BB_WORKER(1, " batch check returned SAT, thus entire formula is SAT\n"); model_ref mdl; ctx->get_model(mdl); b.set_sat(m_l2g, *mdl); bb_candidates.reset(); return; } // ----- UNSAT: inspect core ----- expr_ref_vector bb_asms_in_core(m); auto const& unsat_core = ctx->unsat_core(); for (expr* a : unsat_core) if (bb_asms.contains(a)) bb_asms_in_core.push_back(a); // ---- empty core intersection → formula is UNSAT independent of backbone assumptions ---- if (bb_asms_in_core.empty()) { expr_ref_vector core_vec(m); for (expr* e : unsat_core) core_vec.push_back(e); b.set_unsat(m_l2g, core_vec); return; } // ---- singleton core → backbone ---- if (bb_asms_in_core.size() == 1) { expr* a = bb_asms_in_core.back(); expr_ref backbone_lit(m.mk_not(a), m); m_stats.m_singleton_backbones++; m_stats.m_backbones_detected++; bool is_new_bb = b.collect_global_backbone(m_l2g, backbone_lit); if (is_new_bb) { m_stats.m_backbones_found++; ctx->assert_expr(backbone_lit.get()); // since bb workers don't collect clauses they themselves shared } expr* candidate_to_remove = (m_mode == bb_mode::bb_negated) ? backbone_lit.get() // since core contains ¬candidates in negated mode : a; // since core contains candidates in positive mode bb_candidate_lits.erase(candidate_to_remove); } unsigned sz_before = bb_asms.size(); for (expr* a : bb_asms_in_core) bb_asms.erase(a); m_stats.m_lits_removed_by_core += sz_before - bb_asms.size(); chunk_delta = 1; if (bb_asms.empty()) { LOG_BB_WORKER(1, " no more negated chunk literals, fallback to individual checks\n"); fallback_singletons(chunk_lits, bb_candidate_lits); m_stats.m_fallback_reason_chunk_exhausted++; break; } } } if (!canceled()) { b.cancel_current_backbone_batch(); } bb_candidates.reset(); } } void parallel::backbones_worker::cancel() { LOG_BB_WORKER(1, " BACKBONES WORKER cancelling\n"); m.limit().cancel(); } bool parallel::batch_manager::collect_global_backbone(ast_translation &l2g, expr_ref const &backbone) { IF_VERBOSE(1, verbose_stream() << "collect-global-backbone\n"); std::scoped_lock lock(mux); SASSERT(&m == &l2g.to()); if (is_global_backbone_unlocked(l2g, backbone)) return false; expr_ref g_bb_ref(l2g(backbone.get()), m); m_global_backbones.push_back(g_bb_ref); IF_VERBOSE(1, verbose_stream() << " Found and sharing new global backbone: " << mk_bounded_pp(g_bb_ref, m, 3) << "\n"); collect_clause_unlocked(l2g, /*source_worker_id=*/UINT_MAX, backbone.get()); expr_ref neg_g_bb_ref(mk_not(g_bb_ref), m); ptr_vector matches; m_search_tree.find_nonclosed_nodes_with_literal(neg_g_bb_ref, matches); if (!matches.empty()) { IF_VERBOSE(1, verbose_stream() << " Closing negation of the new global backbone: " << mk_bounded_pp(g_bb_ref, m, 3) << "\n"); expr_ref_vector l_core(l2g.from()); l_core.push_back(mk_not(backbone)); vector targets; for (node* t : matches) { if (!t || m_search_tree.is_lease_canceled(t, t->get_cancel_epoch())) continue; // Keep only highest matching nodes: if an ancestor is already selected, // closing it will also close this descendant subtree. bool is_highest_ancestor = true; for (node* p = t->parent(); p; p = p->parent()) { if (any_of(targets, [&](node_lease const& target) { return target.node == p; })) { is_highest_ancestor = false; break; } } if (!is_highest_ancestor) continue; targets.push_back({ t, t->epoch(), t->get_cancel_epoch() }); } if (!targets.empty()) backtrack_unlocked(l2g, UINT_MAX, l_core, nullptr, &targets); } return true; } void parallel::backbones_worker::collect_statistics(::statistics& st) const { st.update("bb-batches-total", m_stats.m_batches_total); st.update("bb-candidates-total", m_stats.m_candidates_total); st.update("bb-backbones-detected", m_stats.m_backbones_detected); st.update("bb-backbones-found", m_stats.m_backbones_found); st.update("bb-core-refinement-rounds", m_stats.m_core_refinement_rounds); st.update("bb-singleton-backbones", m_stats.m_singleton_backbones); st.update("bb-fallback-singleton-checks", m_stats.m_fallback_singleton_checks); st.update("bb-fallback-chunk-exhausted", m_stats.m_fallback_reason_chunk_exhausted); st.update("bb-fallback-undef", m_stats.m_fallback_reason_undef); st.update("bb-literals-removed-by-core", m_stats.m_lits_removed_by_core); st.update("bb-num-chunk-increases", m_stats.m_num_chunk_increases); double backbone_yield = 0.0; if (m_stats.m_candidates_total > 0) backbone_yield = 100.0 * (double)m_stats.m_backbones_found / (double)m_stats.m_candidates_total; double avg_backbones_per_batch = 0.0; if (m_stats.m_batches_total > 0) avg_backbones_per_batch = (double)m_stats.m_backbones_found / (double)m_stats.m_batches_total; double core_refinement_cost = 0.0; if (m_stats.m_batches_total > 0) core_refinement_cost = (double)m_stats.m_core_refinement_rounds / (double)m_stats.m_batches_total; double core_effectiveness = 0.0; if (m_stats.m_core_refinement_rounds > 0) core_effectiveness = (double)m_stats.m_lits_removed_by_core / (double)m_stats.m_core_refinement_rounds; st.update("bb-backbone-yield-pct", backbone_yield); st.update("bb-avg-backbones-per-batch", avg_backbones_per_batch); st.update("bb-core-refinement-rounds-per-batch", core_refinement_cost); st.update("bb-core-effectiveness-lit-removed-per-round", core_effectiveness); } void parallel::sls_worker::cancel() { IF_VERBOSE(1, verbose_stream() << " SLS WORKER cancelling\n"); m.limit().cancel(); } void parallel::sls_worker::collect_statistics(::statistics &st) const { m_sls->collect_statistics(st); } void parallel::worker::run() { bool is_first_run = true; node_lease lease; expr_ref_vector cube(m); while (true) { if (!b.get_cube(m_g2l, id, cube, is_first_run, lease)) { LOG_WORKER(1, " no more cubes\n"); return; } is_first_run = false; collect_shared_clauses(); check_cube_start: LOG_WORKER(1, " CUBE SIZE IN MAIN LOOP: " << cube.size() << "\n"); if (m_config.m_global_backbones) { bb_candidates local_candidates = find_backbone_candidates(); b.collect_backbone_candidates(m_l2g, local_candidates); if (!m.inc()) return; } lbool r = check_cube(cube); if (b.lease_canceled(lease)) { LOG_WORKER(1, " abandoning canceled lease\n"); lease = {}; m.limit().dec_cancel(); continue; } if (!m.inc()) return; switch (r) { case l_undef: { update_max_thread_conflicts(); LOG_WORKER(1, " found undef cube\n"); if (m_config.m_max_cube_depth <= cube.size()) goto check_cube_start; auto atom = get_split_atom(); if (!atom) goto check_cube_start; b.try_split(m_l2g, id, lease, atom, m_config.m_threads_max_conflicts); lease = {}; simplify(); break; } case l_true: { LOG_WORKER(1, " found sat cube\n"); model_ref mdl; ctx->get_model(mdl); b.set_sat(m_l2g, *mdl); return; } case l_false: { expr_ref_vector const &unsat_core = ctx->unsat_core(); LOG_WORKER(2, " unsat core:\n"; for (auto c : unsat_core) verbose_stream() << mk_bounded_pp(c, m, 3) << "\n"); // If the unsat core only contains external assumptions, // unsatisfiability does not depend on the current cube and the entire problem is unsat. if (all_of(unsat_core, [&](expr *e) { return asms.contains(e); })) { LOG_WORKER(1, " determined formula unsat\n"); b.set_unsat(m_l2g, unsat_core); return; } LOG_WORKER(1, " found unsat cube\n"); b.backtrack(m_l2g, id, unsat_core, lease); lease = {}; if (m_config.m_share_conflicts) b.collect_clause(m_l2g, id, mk_not(mk_and(unsat_core))); break; } } if (m_config.m_share_units) share_units(); } } parallel::worker::worker(unsigned id, parallel &p, expr_ref_vector const &_asms) : id(id), p(p), b(p.m_batch_manager), asms(m), m_smt_params(p.ctx.get_fparams()), m_g2l(p.ctx.m, m), m_l2g(m, p.ctx.m) { for (auto e : _asms) asms.push_back(m_g2l(e)); LOG_WORKER(1, " created with " << asms.size() << " assumptions\n"); m_smt_params.m_random_seed += id; // ensure different random seed for each worker ctx = alloc(context, m, m_smt_params, p.ctx.get_params()); ctx->set_logic(p.ctx.m_setup.get_logic()); context::copy(p.ctx, *ctx, true); // don't share initial units ctx->pop_to_base_lvl(); m_num_shared_units = ctx->assigned_literals().size(); m_num_initial_atoms = ctx->get_num_bool_vars(); ctx->get_fparams().m_preprocess = false; // avoid preprocessing lemmas that are exchanged smt_parallel_params pp(p.ctx.m_params); m_config.m_inprocessing = pp.inprocessing(); m_config.m_global_backbones = pp.num_global_bb_threads() > 0 && !pp.failed_literal_backbones(); } parallel::sls_worker::sls_worker(parallel& p) : p(p), b(p.m_batch_manager), m(), m_g2l(p.ctx.m, m), m_l2g(m, p.ctx.m) { IF_VERBOSE(1, verbose_stream() << "Initialized SLS portfolio thread\n"); m_params.append(p.ctx.m_params); m_sls = alloc(sls::smt_solver, m, m_params); } parallel::backbones_worker::backbones_worker(unsigned id, parallel &p, expr_ref_vector const &_asms) : id(id), b(p.m_batch_manager), m(), asms(m), m_smt_params(p.ctx.get_fparams()), m_g2l(p.ctx.m, m), m_l2g(m, p.ctx.m) { for (auto e : _asms) asms.push_back(m_g2l(e)); IF_VERBOSE(1, verbose_stream() << "Initialized backbones thread " << id << "\n"); m_mode = id == 0 ? bb_mode::bb_negated : bb_mode::bb_positive; ctx = alloc(context, m, m_smt_params, p.ctx.get_params()); ctx->set_logic(p.ctx.m_setup.get_logic()); ctx->get_fparams().m_max_conflicts = m_bb_conflicts_per_chunk; context::copy(p.ctx, *ctx, true); smt_parallel_params pp(p.ctx.m_params); m_num_global_bb_threads = pp.num_global_bb_threads(); m_use_failed_literal_test = pp.failed_literal_backbones(); } parallel::bb_candidates parallel::worker::find_backbone_candidates(unsigned k) { bb_candidates backbone_candidates; expr_ref candidate(m); unsigned curr_time = ctx->m_stats.m_num_assignments; for (bool_var v = 0; v < ctx->get_num_bool_vars(); ++v) { if (ctx->get_assignment(v) != l_undef && ctx->get_assign_level(v) == ctx->m_base_lvl) continue; candidate = ctx->bool_var2expr(v); if (!candidate) continue; auto birth = ctx->m_birthdate[v]; auto age = curr_time - birth; auto const& d = ctx->get_bdata(v); if (d.m_phase_available && !d.m_phase) candidate = m.mk_not(candidate); if (b.is_global_backbone(m_l2g, candidate)) continue; bb_candidate bb_cand(m, candidate, age, 1); backbone_candidates.push_back(bb_cand); } // sort from oldest to youngest std::stable_sort( backbone_candidates.begin(), backbone_candidates.end(), [](bb_candidate const& a, bb_candidate const& b) { return a.age > b.age; } ); // take top-k oldest if (backbone_candidates.size() > k) backbone_candidates.shrink(k); return backbone_candidates; } // // Assume the negation of all candidates (or a batch of them) // run the solver with a low budget of conflicts // if the unsat core contains a single candidate we have found a backbone literal // bool parallel::backbones_worker::check_backbone(expr* bb_candidate) { unsigned sz = asms.size(); asms.push_back(m.mk_not(bb_candidate)); lbool r = l_undef; try { r = ctx->check(asms.size(), asms.data()); } catch (z3_error &err) { if (!m.limit().is_canceled()) b.set_exception(err.error_code()); } catch (z3_exception &ex) { if (!m.limit().is_canceled() && !is_cancellation_exception(ex.what())) b.set_exception(ex.what()); } catch (...) { if (!m.limit().is_canceled()) b.set_exception("unknown exception"); } asms.shrink(sz); LOG_BB_WORKER(1, " RESULT: " << r << " FOR CANDIDATE: " << mk_bounded_pp(bb_candidate, m, 3) << "\n"); if (r == l_false) { auto core = ctx->unsat_core(); if (core.size() == 1) { return true; } } return false; } void parallel::worker::share_units() { // Collect new units learned locally by this worker and send to batch manager unsigned sz = ctx->assigned_literals().size(); for (unsigned j = m_num_shared_units; j < sz; ++j) { // iterate only over new literals since last sync literal lit = ctx->assigned_literals()[j]; // filter by assign level: do not pop to base level as this destroys the current search state if (ctx->get_assign_level(lit) > ctx->m_base_lvl) continue; if (!ctx->is_relevant(lit.var()) && m_config.m_share_units_relevant_only) continue; if (m_config.m_share_units_initial_only && lit.var() >= m_num_initial_atoms) { LOG_WORKER(4, " Skipping non-initial unit: " << lit.var() << "\n"); continue; // skip non-initial atoms if configured to do so } expr_ref e(ctx->bool_var2expr(lit.var()), ctx->m); // turn literal into a Boolean expression if (m.is_and(e) || m.is_or(e)) continue; if (lit.sign()) e = mk_not(e); // negate if literal is negative b.collect_clause(m_l2g, id, e); } m_num_shared_units = sz; } void parallel::worker::simplify() { if (!m.inc()) return; // first attempt: one-shot simplification of the context. // a precise schedule of repeated simplification is TBD. // also, the in-processing simplifier should be applied to // a current set of irredundant clauses that may be reduced by // unit propagation. By including the units we are effectively // repeating unit propagation, but potentially not subsumption or // Boolean simplifications that a solver could perform (smt_context doesnt really) // Integration of inprocssing simplifcation here or in sat/smt solver could // be based on taking the current clause set instead of the asserted formulas. if (!m_config.m_inprocessing) return; if (m_config.m_inprocessing_delay > 0) { --m_config.m_inprocessing_delay; return; } ctx->pop_to_base_lvl(); if (ctx->m_base_lvl > 0) return; // simplification only at base level m_config.m_inprocessing = false; // initial strategy is to immediately disable inprocessing for future calls. dependent_expr_simplifier *s = ctx->m_simplifier.get(); if (!s) { // create a simplifier if none exists // initialize it to a default pre-processing simplifier. ctx->m_fmls = alloc(base_dependent_expr_state, m); auto then_s = alloc(then_simplifier, m, ctx->get_params(), *ctx->m_fmls); s = then_s; ctx->m_simplifier = s; init_preprocess(m, ctx->get_params(), *then_s, *ctx->m_fmls); } dependent_expr_state &fmls = *ctx->m_fmls.get(); // extract assertions from ctx. // it is possible to track proof objects here if wanted. // feed them to the simplifier ptr_vector assertions; expr_ref_vector units(m); ctx->get_assertions(assertions); ctx->get_units(units); for (expr *e : assertions) fmls.add(dependent_expr(m, e, nullptr, nullptr)); for (expr *e : units) fmls.add(dependent_expr(m, e, nullptr, nullptr)); // run in-processing on the assertions s->reduce(); scoped_ptr new_ctx = alloc(context, m, m_smt_params, p.ctx.get_params()); // extract simplified assertions from the simplifier // create a new context with the simplified assertions // update ctx with the new context. for (unsigned i = 0; i < fmls.qtail(); ++i) { auto const &de = fmls[i]; new_ctx->assert_expr(de.fml()); } asserted_formulas &src_af = ctx->m_asserted_formulas; asserted_formulas &dst_af = new_ctx->m_asserted_formulas; src_af.get_macro_manager().copy_to(dst_af.get_macro_manager()); new_ctx->copy_user_propagator(*ctx, true); ctx = new_ctx.detach(); ctx->setup_context(true); ctx->internalize_assertions(); auto old_atoms = m_num_initial_atoms; m_num_shared_units = ctx->assigned_literals().size(); m_num_initial_atoms = ctx->get_num_bool_vars(); LOG_WORKER(1, " inprocess " << old_atoms << " -> " << m_num_initial_atoms << "\n"); } void parallel::worker::collect_statistics(::statistics &st) const { ctx->collect_statistics(st); } void parallel::worker::cancel() { LOG_WORKER(1, " canceling\n"); m.limit().cancel(); } void parallel::worker::cancel_lease() { LOG_WORKER(1, " canceling lease\n"); m.limit().inc_cancel(); } void parallel::batch_manager::release_lease_unlocked(unsigned worker_id, node* n, unsigned epoch) { if (worker_id >= m_worker_leases.size()) return; auto &lease = m_worker_leases[worker_id]; if (!lease.node || lease.node != n || lease.epoch != epoch) return; m_search_tree.dec_active_workers(lease.node); lease = {}; } void parallel::batch_manager::cancel_closed_leases_unlocked(unsigned source_worker_id) { unsigned n = std::min(m_worker_leases.size(), p.m_workers.size()); for (unsigned worker_id = 0; worker_id < n; ++worker_id) { if (worker_id == source_worker_id) continue; auto const& lease = m_worker_leases[worker_id]; // only cancel workers that currently hold a lease, whose lease is canceled, // and haven't already been signaled (prevents multiple inc_cancel() for same lease) if (lease.node && !lease.cancel_signaled && m_search_tree.is_lease_canceled(lease.node, lease.cancel_epoch)) { p.m_workers[worker_id]->cancel_lease(); m_worker_leases[worker_id].cancel_signaled = true; } } } void parallel::batch_manager::backtrack(ast_translation &l2g, unsigned worker_id, expr_ref_vector const &core, node_lease const &lease) { std::scoped_lock lock(mux); backtrack_unlocked(l2g, worker_id, core, &lease, nullptr); } void parallel::batch_manager::backtrack_unlocked(ast_translation& l2g, unsigned worker_id, expr_ref_vector const& core, node_lease const* lease, vector const* targets) { if (m_state != state::is_running) return; vector g_core; for (auto c : core) g_core.push_back(expr_ref(l2g(c), m)); SASSERT((lease != nullptr) != (targets != nullptr)); if (lease) { // we close/backtrack regardless of whether this lease is stale or not, as long as the lease isn't canceled // i.e. worker 1 splits this node, but then worker 2 determines UNSAT --> worker 2 is stale but we still close this node and backtrack if (m_search_tree.is_lease_canceled(lease->node, lease->cancel_epoch)) return; IF_VERBOSE(1, verbose_stream() << "Batch manager backtracking.\n"); release_lease_unlocked(worker_id, lease->node, lease->epoch); m_search_tree.backtrack(lease->node, g_core); } else { bool has_open_targets = false; for (auto const& target : *targets) { if (m_search_tree.is_lease_canceled(target.node, target.cancel_epoch)) continue; has_open_targets = true; IF_VERBOSE(1, verbose_stream() << "Batch manager backtracking external targets.\n"); m_search_tree.backtrack(target.node, g_core); } if (!has_open_targets) return; } // terminate on-demand the workers that are currently exploring the now-closed nodes cancel_closed_leases_unlocked(worker_id); IF_VERBOSE(1, m_search_tree.display(verbose_stream() << bounded_pp_exprs(core) << "\n");); if (m_search_tree.is_closed()) { IF_VERBOSE(1, verbose_stream() << "Search tree closed, setting UNSAT\n"); m_state = state::is_unsat; SASSERT(p.ctx.m_unsat_core.empty()); for (auto e : m_search_tree.get_core_from_root()) p.ctx.m_unsat_core.push_back(e); cancel_background_threads(); } } void parallel::batch_manager::try_split(ast_translation &l2g, unsigned worker_id, node_lease const &lease, expr *atom, unsigned effort) { std::scoped_lock lock(mux); if (m_state != state::is_running) return; if (m_search_tree.is_lease_canceled(lease.node, lease.cancel_epoch)) return; expr_ref lit(m), nlit(m); lit = l2g(atom); nlit = mk_not(m, lit); bool did_split = m_search_tree.try_split(lease.node, lease.epoch, lease.cancel_epoch, lit, nlit, effort); release_lease_unlocked(worker_id, lease.node, lease.epoch); if (did_split) { ++m_stats.m_num_cubes; m_stats.m_max_cube_depth = std::max(m_stats.m_max_cube_depth, lease.node->depth() + 1); IF_VERBOSE(1, verbose_stream() << "Batch manager splitting on literal: " << mk_bounded_pp(lit, m, 3) << "\n"); } } void parallel::batch_manager::release_lease(unsigned worker_id, node_lease const &lease) { std::scoped_lock lock(mux); release_lease_unlocked(worker_id, lease.node, lease.epoch); } bool parallel::batch_manager::lease_canceled(node_lease const &lease) { std::scoped_lock lock(mux); return m_state == state::is_running && m_search_tree.is_lease_canceled(lease.node, lease.cancel_epoch); } void parallel::batch_manager::collect_clause(ast_translation &l2g, unsigned source_worker_id, expr *clause) { std::scoped_lock lock(mux); collect_clause_unlocked(l2g, source_worker_id, clause); } void parallel::batch_manager::collect_clause_unlocked(ast_translation &l2g, unsigned source_worker_id, expr *clause) { expr *g_clause = l2g(clause); if (!shared_clause_set.contains(g_clause)) { shared_clause_set.insert(g_clause); shared_clause sc{source_worker_id, expr_ref(g_clause, m)}; shared_clause_trail.push_back(sc); } } void parallel::worker::collect_shared_clauses() { expr_ref_vector new_clauses = b.return_shared_clauses(m_g2l, m_shared_clause_limit, id); // iterate over new clauses and assert them in the local context for (expr *e : new_clauses) { ctx->assert_expr(e); LOG_WORKER(4, " asserting shared clause: " << mk_bounded_pp(e, m, 3) << "\n"); } } void parallel::backbones_worker::collect_shared_clauses() { expr_ref_vector new_clauses = b.return_shared_clauses(m_g2l, m_shared_clause_limit, UINT_MAX); // iterate over new clauses and assert them in the local context for (expr *e : new_clauses) { ctx->assert_expr(e); LOG_BB_WORKER(4, " asserting shared clause: " << mk_bounded_pp(e, m, 3) << "\n"); } } void parallel::batch_manager::collect_backbone_candidates(ast_translation& l2g, bb_candidates& bb_candidates) { std::scoped_lock lock(mux); auto find_existing_candidate_idx = [&](expr* e) -> int { for (unsigned i = 0; i < m_bb_candidates.size(); ++i) { if (m_bb_candidates[i].lit.get() == e) return i; } return -1; }; auto rank_of = [&](bb_candidate const& c) { return c.age * std::log2(2.0 + c.hits); }; for (auto const& c : bb_candidates) { if (is_global_backbone_unlocked(l2g, c.lit)) continue; expr* worker_lit = c.lit.get(); expr_ref g_lit(l2g(worker_lit), m); double age = c.age; int idx = find_existing_candidate_idx(g_lit.get()); if (idx >= 0) { auto& existing = m_bb_candidates[idx]; existing.age = (existing.age * existing.hits + age) / (existing.hits + 1); existing.hits++; continue; } if (m_bb_candidates.size() < m_max_global_bb_candidates) { m_bb_candidates.push_back(bb_candidate(m, g_lit.get(), age, 1)); continue; } // Find worst candidate to evict unsigned worst_idx = 0; double worst_rank = rank_of(m_bb_candidates[0]); for (unsigned i = 1; i < m_bb_candidates.size(); ++i) { double r = rank_of(m_bb_candidates[i]); if (r < worst_rank) { worst_rank = r; worst_idx = i; } } bb_candidate new_bb_candidate = bb_candidate(m, g_lit.get(), age, 1); if (rank_of(new_bb_candidate) > worst_rank) m_bb_candidates[worst_idx] = new_bb_candidate; } if (!m_bb_candidates.empty()) { std::stable_sort( m_bb_candidates.begin(), m_bb_candidates.end(), [&](bb_candidate const& a, bb_candidate const& b) { return rank_of(a) > rank_of(b); } ); m_bb_cv.notify_all(); } } parallel::bb_candidates parallel::batch_manager::return_global_bb_candidates(ast_translation& g2l) { bb_candidates bb_candidates_local; std::scoped_lock lock(mux); for (auto const& gc : m_bb_candidates) { expr_ref l_lit(g2l(gc.lit.get()), g2l.to()); bb_candidates_local.push_back(bb_candidate(g2l.to(), l_lit, gc.age, gc.hits)); } return bb_candidates_local; } bool parallel::batch_manager::wait_for_backbone_job(unsigned bb_thread_id, ast_translation& g2l, bb_candidates& out, reslimit& lim) { out.reset(); std::unique_lock lock(mux); // ---- WAIT UNTIL: // (a) a new batch is ready that this thread hasn't seen yet, OR // (b) candidates are available AND the previous batch is finished (not in progress) m_bb_cv.wait(lock, [&]() { return lim.is_canceled() || m_state != state::is_running || m_bb_last_batch_processed[bb_thread_id] < m_bb_batch_id || !m_bb_candidates.empty(); }); if (lim.is_canceled()) return false; if (m_state != state::is_running) return false; // ---- NEED NEW BATCH? ---- // Only create a new batch if this thread has already seen the current batch. if (m_bb_last_batch_processed[bb_thread_id] == m_bb_batch_id) { // pop new batch once unsigned n = std::min(m_bb_batch_size, m_bb_candidates.size()); m_bb_current_batch.reset(); for (unsigned i = 0; i < n; ++i) { m_bb_current_batch.push_back(m_bb_candidates.back()); m_bb_candidates.pop_back(); } m_bb_batch_id++; // wake all threads to see new batch m_bb_cv.notify_all(); } for (auto const& gc : m_bb_current_batch) { expr_ref l_lit(g2l(gc.lit.get()), g2l.to()); out.push_back(bb_candidate(g2l.to(), l_lit, gc.age, gc.hits)); } m_bb_last_batch_processed[bb_thread_id] = m_bb_batch_id; return true; } expr_ref_vector parallel::batch_manager::return_shared_clauses(ast_translation &g2l, unsigned &worker_limit, unsigned worker_id) { std::scoped_lock lock(mux); expr_ref_vector result(g2l.to()); for (unsigned i = worker_limit; i < shared_clause_trail.size(); ++i) { if (shared_clause_trail[i].source_worker_id != worker_id) result.push_back(g2l(shared_clause_trail[i].clause.get())); } worker_limit = shared_clause_trail.size(); // update the worker limit to the end of the current trail return result; } lbool parallel::worker::check_cube(expr_ref_vector const &cube) { for (auto &atom : cube) asms.push_back(atom); lbool r = l_undef; ctx->get_fparams().m_max_conflicts = std::min(m_config.m_threads_max_conflicts, m_config.m_max_conflicts); IF_VERBOSE(1, verbose_stream() << " Checking cube\n" << bounded_pp_exprs(cube) << "with max_conflicts: " << ctx->get_fparams().m_max_conflicts << "\n";); try { r = ctx->check(asms.size(), asms.data()); } catch (z3_error &err) { if (!is_cancellation_exception(err.what())) b.set_exception(err.error_code()); } catch (z3_exception &ex) { if (!is_cancellation_exception(ex.what())) b.set_exception(ex.what()); } catch (...) { b.set_exception("unknown exception"); } asms.shrink(asms.size() - cube.size()); LOG_WORKER(1, " DONE checking cube " << r << "\n";); return r; } expr_ref parallel::worker::get_split_atom() { expr_ref result(m); double score = 0; unsigned n = 0; ctx->pop_to_search_lvl(); for (bool_var v = 0; v < ctx->get_num_bool_vars(); ++v) { if (ctx->get_assignment(v) != l_undef) continue; expr *e = ctx->bool_var2expr(v); if (!e) continue; // don't split on a backbone if (m_config.m_global_backbones) { expr_ref e_ref(e, m); expr_ref neg_e_ref(mk_not(e_ref), m); if (b.is_global_backbone(m_l2g, e_ref) || b.is_global_backbone(m_l2g, neg_e_ref)) continue; } double new_score = ctx->m_lit_scores[0][v] * ctx->m_lit_scores[1][v]; ctx->m_lit_scores[0][v] /= 2; ctx->m_lit_scores[1][v] /= 2; if (new_score > score || !result || (new_score == score && m_rand(++n) == 0)) { score = new_score; result = e; } } return result; } void parallel::batch_manager::set_sat(ast_translation &l2g, model &m) { std::scoped_lock lock(mux); IF_VERBOSE(1, verbose_stream() << "Batch manager setting SAT.\n"); if (m_state != state::is_running) return; m_state = state::is_sat; p.ctx.set_model(m.translate(l2g)); cancel_background_threads(); } void parallel::batch_manager::set_unsat(ast_translation &l2g, expr_ref_vector const &unsat_core) { std::scoped_lock lock(mux); IF_VERBOSE(1, verbose_stream() << "Batch manager setting UNSAT.\n"); if (m_state != state::is_running) return; m_state = state::is_unsat; // each call to check_sat needs to have a fresh unsat core SASSERT(p.ctx.m_unsat_core.empty()); for (expr *e : unsat_core) p.ctx.m_unsat_core.push_back(l2g(e)); cancel_background_threads(); } void parallel::batch_manager::set_exception(unsigned error_code) { std::scoped_lock lock(mux); IF_VERBOSE(1, verbose_stream() << "Batch manager setting exception code: " << error_code << ".\n"); if (m_state != state::is_running) return; m_state = state::is_exception_code; m_exception_code = error_code; cancel_background_threads(); } void parallel::batch_manager::set_exception(std::string const &msg) { std::scoped_lock lock(mux); IF_VERBOSE(1, verbose_stream() << "Batch manager setting exception msg: " << msg << ".\n"); if (m_state != state::is_running) return; m_state = state::is_exception_msg; m_exception_msg = msg; cancel_background_threads(); } lbool parallel::batch_manager::get_result() const { if (m.limit().is_canceled()) return l_undef; // the main context was cancelled, so we return undef. switch (m_state) { case state::is_running: // batch manager is still running, but all threads have processed their cubes, which // means all cubes were unsat throw default_exception("inconsistent end state"); case state::is_unsat: return l_false; case state::is_sat: return l_true; case state::is_exception_msg: throw default_exception(m_exception_msg.c_str()); case state::is_exception_code: throw z3_error(m_exception_code); default: UNREACHABLE(); return l_undef; } } bool parallel::batch_manager::get_cube(ast_translation &g2l, unsigned id, expr_ref_vector &cube, bool is_first_run, node_lease &lease) { std::scoped_lock lock(mux); cube.reset(); if (m_search_tree.is_closed()) { IF_VERBOSE(1, verbose_stream() << "all done\n";); return false; } if (m_state != state::is_running) { IF_VERBOSE(1, verbose_stream() << "aborting get_cube\n";); return false; } node *t = is_first_run ? m_search_tree.activate_root() : m_search_tree.activate_best_node(); if (!t) return false; IF_VERBOSE(1, m_search_tree.display(verbose_stream()); verbose_stream() << "\n";); lease.node = t; lease.epoch = t->epoch(); lease.cancel_epoch = t->get_cancel_epoch(); if (id >= m_worker_leases.size()) m_worker_leases.resize(id + 1); m_worker_leases[id] = lease; while (t) { if (cube_config::literal_is_null(t->get_literal())) break; expr_ref lit(g2l.to()); lit = g2l(t->get_literal().get()); cube.push_back(std::move(lit)); t = t->parent(); } return true; } void parallel::batch_manager::initialize(unsigned num_global_bb_threads, unsigned initial_max_thread_conflicts) { m_state = state::is_running; m_num_global_bb_threads = num_global_bb_threads; m_bb_last_batch_processed.reset(); m_bb_last_batch_processed.resize(m_num_global_bb_threads); m_bb_candidates.reset(); m_search_tree.reset(); m_search_tree.set_effort_unit(initial_max_thread_conflicts); m_worker_leases.reset(); m_worker_leases.resize(p.m_workers.size()); } void parallel::batch_manager::collect_statistics(::statistics &st) const { st.update("parallel-num_cubes", m_stats.m_num_cubes); st.update("parallel-max-cube-size", m_stats.m_max_cube_depth); } lbool parallel::operator()(expr_ref_vector const &asms) { smt_parallel_params pp(ctx.m_params); unsigned num_workers = std::min((unsigned)std::thread::hardware_concurrency(), ctx.get_fparams().m_threads); unsigned num_sls_threads = (pp.sls() ? 1 : 0); unsigned num_global_bb_threads = pp.num_global_bb_threads(); unsigned total_threads = num_workers + num_sls_threads + num_global_bb_threads; IF_VERBOSE(1, verbose_stream() << "Parallel SMT with " << total_threads << " threads\n";); ast_manager &m = ctx.m; if (m.has_trace_stream()) throw default_exception("trace streams have to be off in parallel mode"); struct scoped_clear { parallel &p; scoped_clear(parallel &p) : p(p) {} ~scoped_clear() { p.m_workers.reset(); p.m_sls_worker = nullptr; p.m_global_backbones_workers.reset(); } }; scoped_clear clear(*this); m_workers.reset(); scoped_limits sl(m.limit()); flet _nt(ctx.m_fparams.m_threads, 1); SASSERT(num_workers > 1); for (unsigned i = 0; i < num_workers; ++i) m_workers.push_back(alloc(worker, i, *this, asms)); for (auto w : m_workers) sl.push_child(&(w->limit())); if (num_sls_threads == 1) { m_sls_worker = alloc(sls_worker, *this); sl.push_child(&(m_sls_worker->limit())); } for (unsigned i = 0; i < num_global_bb_threads; ++i) { auto *w = alloc(backbones_worker, i, *this, asms); m_global_backbones_workers.push_back(w); sl.push_child(&(w->limit())); } IF_VERBOSE(1, verbose_stream() << "Launched " << m_workers.size() << " CDCL threads, " << (m_sls_worker ? 1 : 0) << " SLS threads, " << m_global_backbones_workers.size() << " global backbone threads.\n";); m_batch_manager.initialize(num_global_bb_threads); // Launch threads vector threads; threads.resize(total_threads); unsigned thread_idx = 0; for (auto* w : m_workers) threads[thread_idx++] = std::thread([&, w]() { w->run(); }); if (m_sls_worker) threads[thread_idx++] = std::thread([&]() { m_sls_worker->run(); }); for (auto* w : m_global_backbones_workers) threads[thread_idx++] = std::thread([&, w]() { w->run(); }); // Wait for all threads to finish for (auto &th : threads) th.join(); for (auto w : m_workers) w->collect_statistics(ctx.m_aux_stats); m_batch_manager.collect_statistics(ctx.m_aux_stats); if (m_sls_worker) m_sls_worker->collect_statistics(ctx.m_aux_stats); for (auto* bb_w : m_global_backbones_workers) bb_w->collect_statistics(ctx.m_aux_stats); return m_batch_manager.get_result(); } } // namespace smt #endif