fuzz coverage

Coverage Report

Created: 2026-05-08 05:52

next uncovered line (L), next uncovered region (R), next uncovered branch (B)
/root/bitcoin/src/txgraph.cpp
Line
Count
Source
1
// Copyright (c) The Bitcoin Core developers
2
// Distributed under the MIT software license, see the accompanying
3
// file COPYING or http://www.opensource.org/licenses/mit-license.php.
4
5
#include <txgraph.h>
6
7
#include <cluster_linearize.h>
8
#include <random.h>
9
#include <util/bitset.h>
10
#include <util/check.h>
11
#include <util/feefrac.h>
12
#include <util/vector.h>
13
14
#include <compare>
15
#include <functional>
16
#include <memory>
17
#include <ranges>
18
#include <set>
19
#include <span>
20
#include <unordered_set>
21
#include <utility>
22
23
namespace {
24
25
using namespace cluster_linearize;
26
27
/** The maximum number of levels a TxGraph can have (0 = main, 1 = staging). */
28
static constexpr int MAX_LEVELS{2};
29
30
// Forward declare the TxGraph implementation class.
31
class TxGraphImpl;
32
33
/** Position of a DepGraphIndex within a Cluster::m_linearization. */
34
using LinearizationIndex = uint32_t;
35
/** Position of a Cluster within TxGraphImpl::ClusterSet::m_clusters. */
36
using ClusterSetIndex = uint32_t;
37
38
/** Quality levels for cached cluster linearizations. */
39
enum class QualityLevel
40
{
41
    /** This is a singleton cluster consisting of a transaction that individually exceeds the
42
     *  cluster size limit. It cannot be merged with anything. */
43
    OVERSIZED_SINGLETON,
44
    /** This cluster may have multiple disconnected components, which are all NEEDS_FIX. */
45
    NEEDS_SPLIT_FIX,
46
    /** This cluster may have multiple disconnected components, which are all NEEDS_RELINEARIZE. */
47
    NEEDS_SPLIT,
48
    /** This cluster may be non-topological. */
49
    NEEDS_FIX,
50
    /** This cluster has undergone changes that warrant re-linearization. */
51
    NEEDS_RELINEARIZE,
52
    /** The minimal level of linearization has been performed, but it is not known to be optimal. */
53
    ACCEPTABLE,
54
    /** The linearization is known to be optimal. */
55
    OPTIMAL,
56
    /** This cluster is not registered in any ClusterSet::m_clusters.
57
     *  This must be the last entry in QualityLevel as ClusterSet::m_clusters is sized using it. */
58
    NONE,
59
};
60
61
/** Information about a transaction inside TxGraphImpl::Trim. */
62
struct TrimTxData
63
{
64
    // Fields populated by Cluster::AppendTrimData(). These are immutable after TrimTxData
65
    // construction.
66
    /** Chunk feerate for this transaction. */
67
    FeePerWeight m_chunk_feerate;
68
    /** GraphIndex of the transaction. */
69
    TxGraph::GraphIndex m_index;
70
    /** Size of the transaction. */
71
    uint32_t m_tx_size;
72
73
    // Fields only used internally by TxGraphImpl::Trim():
74
    /** Number of unmet dependencies this transaction has. -1 if the transaction is included. */
75
    uint32_t m_deps_left;
76
    /** Number of dependencies that apply to this transaction as child. */
77
    uint32_t m_parent_count;
78
    /** Where in deps_by_child those dependencies begin. */
79
    uint32_t m_parent_offset;
80
    /** Number of dependencies that apply to this transaction as parent. */
81
    uint32_t m_children_count;
82
    /** Where in deps_by_parent those dependencies begin. */
83
    uint32_t m_children_offset;
84
85
    // Fields only used internally by TxGraphImpl::Trim()'s union-find implementation, and only for
86
    // transactions that are definitely included or definitely rejected.
87
    //
88
    // As transactions get processed, they get organized into trees which form partitions
89
    // representing the would-be clusters up to that point. The root of each tree is a
90
    // representative for that partition. See
91
    // https://en.wikipedia.org/wiki/Disjoint-set_data_structure.
92
    //
93
    /** Pointer to another TrimTxData, towards the root of the tree. If this is a root, m_uf_parent
94
     *  is equal to this itself. */
95
    TrimTxData* m_uf_parent;
96
    /** If this is a root, the total number of transactions in the partition. */
97
    uint32_t m_uf_count;
98
    /** If this is a root, the total size of transactions in the partition. */
99
    uint64_t m_uf_size;
100
};
101
102
/** A grouping of connected transactions inside a TxGraphImpl::ClusterSet. */
103
class Cluster
104
{
105
    friend class TxGraphImpl;
106
    friend class BlockBuilderImpl;
107
108
protected:
109
    using GraphIndex = TxGraph::GraphIndex;
110
    using SetType = BitSet<MAX_CLUSTER_COUNT_LIMIT>;
111
    /** The quality level of m_linearization. */
112
    QualityLevel m_quality{QualityLevel::NONE};
113
    /** Which position this Cluster has in TxGraphImpl::ClusterSet::m_clusters[m_quality]. */
114
    ClusterSetIndex m_setindex{ClusterSetIndex(-1)};
115
    /** Sequence number for this Cluster (for tie-breaking comparison between equal-chunk-feerate
116
        transactions in distinct clusters). */
117
    uint64_t m_sequence;
118
119
14.7k
    explicit Cluster(uint64_t sequence) noexcept : m_sequence(sequence) {}
120
121
public:
122
    // Provide virtual destructor, for safe polymorphic usage inside std::unique_ptr.
123
14.7k
    virtual ~Cluster() = default;
124
125
    // Cannot move or copy (would invalidate Cluster* in Locator and ClusterSet). */
126
    Cluster(const Cluster&) = delete;
127
    Cluster& operator=(const Cluster&) = delete;
128
    Cluster(Cluster&&) = delete;
129
    Cluster& operator=(Cluster&&) = delete;
130
131
    // Generic helper functions.
132
133
    /** Whether the linearization of this Cluster can be exposed. */
134
    bool IsAcceptable() const noexcept
135
1.18M
    {
136
1.18M
        return m_quality == QualityLevel::ACCEPTABLE || m_quality == QualityLevel::OPTIMAL;
137
1.18M
    }
138
    /** Whether the linearization of this Cluster is topological. */
139
    bool IsTopological() const noexcept
140
117k
    {
141
117k
        return m_quality != QualityLevel::NEEDS_FIX && 
m_quality != QualityLevel::NEEDS_SPLIT_FIX111k
;
142
117k
    }
143
    /** Whether the linearization of this Cluster is optimal. */
144
    bool IsOptimal() const noexcept
145
21.9k
    {
146
21.9k
        return m_quality == QualityLevel::OPTIMAL;
147
21.9k
    }
148
    /** Whether this cluster is oversized. Note that no changes that can cause oversizedness are
149
     *  ever applied, so the only way a materialized Cluster object can be oversized is by being
150
     *  an individually oversized transaction singleton. */
151
84.8k
    bool IsOversized() const noexcept { return m_quality == QualityLevel::OVERSIZED_SINGLETON; }
152
    /** Whether this cluster requires splitting. */
153
    bool NeedsSplitting() const noexcept
154
1.11M
    {
155
1.11M
        return m_quality == QualityLevel::NEEDS_SPLIT || 
m_quality == QualityLevel::NEEDS_SPLIT_FIX1.11M
;
156
1.11M
    }
157
158
    /** Get the smallest number of transactions this Cluster is intended for. */
159
    virtual DepGraphIndex GetMinIntendedTxCount() const noexcept = 0;
160
    /** Get the maximum number of transactions this Cluster supports. */
161
    virtual DepGraphIndex GetMaxTxCount() const noexcept = 0;
162
    /** Total memory usage currently for this Cluster, including all its dynamic memory, plus Cluster
163
     *  structure itself, and ClusterSet::m_clusters entry. */
164
    virtual size_t TotalMemoryUsage() const noexcept = 0;
165
    /** Determine the range of DepGraphIndexes used by this Cluster. */
166
    virtual DepGraphIndex GetDepGraphIndexRange() const noexcept = 0;
167
    /** Get the number of transactions in this Cluster. */
168
    virtual LinearizationIndex GetTxCount() const noexcept = 0;
169
    /** Get the total size of the transactions in this Cluster. */
170
    virtual uint64_t GetTotalTxSize() const noexcept = 0;
171
    /** Given a DepGraphIndex into this Cluster, find the corresponding GraphIndex. */
172
    virtual GraphIndex GetClusterEntry(DepGraphIndex index) const noexcept = 0;
173
    /** Append a transaction with given GraphIndex at the end of this Cluster and its
174
     *  linearization. Return the DepGraphIndex it was placed at. */
175
    virtual DepGraphIndex AppendTransaction(GraphIndex graph_idx, FeePerWeight feerate) noexcept = 0;
176
    /** Add dependencies to a given child in this cluster. */
177
    virtual void AddDependencies(SetType parents, DepGraphIndex child) noexcept = 0;
178
    /** Invoke visit1_fn for each transaction in the cluster, in linearization order, then
179
     *  visit2_fn in the same order, then wipe this Cluster. */
180
    virtual void ExtractTransactions(const std::function<void (DepGraphIndex, GraphIndex, FeePerWeight)>& visit1_fn, const std::function<void (DepGraphIndex, GraphIndex, SetType)>& visit2_fn) noexcept = 0;
181
    /** Figure out what level this Cluster exists at in the graph. In most cases this is known by
182
     *  the caller already (see all "int level" arguments below), but not always. */
183
    virtual int GetLevel(const TxGraphImpl& graph) const noexcept = 0;
184
    /** Only called by TxGraphImpl::SwapIndexes. */
185
    virtual void UpdateMapping(DepGraphIndex cluster_idx, GraphIndex graph_idx) noexcept = 0;
186
    /** Push changes to Cluster and its linearization to the TxGraphImpl Entry objects. Main chunk
187
     *  information is computed if the cluster is acceptable, or when rename is set. Rename is used
188
     *  when called from Compact, to recompute after GraphIndexes may have changed; in this case,
189
     *  no chunk index objects are removed or created either. */
190
    virtual void Updated(TxGraphImpl& graph, int level, bool rename) noexcept = 0;
191
    /** Remove all chunk index entries for this cluster (level=0 only). */
192
    virtual void RemoveChunkData(TxGraphImpl& graph) noexcept = 0;
193
    /** Create a copy of this Cluster in staging, returning a pointer to it (used by PullIn). */
194
    virtual Cluster* CopyToStaging(TxGraphImpl& graph) const noexcept = 0;
195
    /** Get the list of Clusters in main that conflict with this one (which is assumed to be in staging). */
196
    virtual void GetConflicts(const TxGraphImpl& graph, std::vector<Cluster*>& out) const noexcept = 0;
197
    /** Mark all the Entry objects belonging to this staging Cluster as missing. The Cluster must be
198
     *  deleted immediately after. */
199
    virtual void MakeStagingTransactionsMissing(TxGraphImpl& graph) noexcept = 0;
200
    /** Remove all transactions from a (non-empty) Cluster. */
201
    virtual void Clear(TxGraphImpl& graph, int level) noexcept = 0;
202
    /** Change a Cluster's level from 1 (staging) to 0 (main). */
203
    virtual void MoveToMain(TxGraphImpl& graph) noexcept = 0;
204
    /** Minimize this Cluster's memory usage. */
205
    virtual void Compact() noexcept = 0;
206
207
    // Functions that implement the Cluster-specific side of internal TxGraphImpl mutations.
208
209
    /** Apply all removals from the front of to_remove that apply to this Cluster, popping them
210
     *  off. There must be at least one such entry. */
211
    virtual void ApplyRemovals(TxGraphImpl& graph, int level, std::span<GraphIndex>& to_remove) noexcept = 0;
212
    /** Split this cluster (must have a NEEDS_SPLIT* quality). Returns whether to delete this
213
     *  Cluster afterwards. */
214
    [[nodiscard]] virtual bool Split(TxGraphImpl& graph, int level) noexcept = 0;
215
    /** Move all transactions from cluster to *this (as separate components). */
216
    virtual void Merge(TxGraphImpl& graph, int level, Cluster& cluster) noexcept = 0;
217
    /** Given a span of (parent, child) pairs that all belong to this Cluster, apply them. */
218
    virtual void ApplyDependencies(TxGraphImpl& graph, int level, std::span<std::pair<GraphIndex, GraphIndex>> to_apply) noexcept = 0;
219
    /** Improve the linearization of this Cluster. Returns how much work was performed and whether
220
     *  the Cluster's QualityLevel improved as a result. */
221
    virtual std::pair<uint64_t, bool> Relinearize(TxGraphImpl& graph, int level, uint64_t max_cost) noexcept = 0;
222
    /** For every chunk in the cluster, append its FeeFrac to ret. */
223
    virtual void AppendChunkFeerates(std::vector<FeeFrac>& ret) const noexcept = 0;
224
    /** Add a TrimTxData entry (filling m_chunk_feerate, m_index, m_tx_size) for every
225
     *  transaction in the Cluster to ret. Implicit dependencies between consecutive transactions
226
     *  in the linearization are added to deps. Return the Cluster's total transaction size. */
227
    virtual uint64_t AppendTrimData(std::vector<TrimTxData>& ret, std::vector<std::pair<GraphIndex, GraphIndex>>& deps) const noexcept = 0;
228
229
    // Functions that implement the Cluster-specific side of public TxGraph functions.
230
231
    /** Process elements from the front of args that apply to this cluster, and append Refs for the
232
     *  union of their ancestors to output. */
233
    virtual void GetAncestorRefs(const TxGraphImpl& graph, std::span<std::pair<Cluster*, DepGraphIndex>>& args, std::vector<TxGraph::Ref*>& output) noexcept = 0;
234
    /** Process elements from the front of args that apply to this cluster, and append Refs for the
235
     *  union of their descendants to output. */
236
    virtual void GetDescendantRefs(const TxGraphImpl& graph, std::span<std::pair<Cluster*, DepGraphIndex>>& args, std::vector<TxGraph::Ref*>& output) noexcept = 0;
237
    /** Populate range with refs for the transactions in this Cluster's linearization, from
238
     *  position start_pos until start_pos+range.size()-1, inclusive. Returns whether that
239
     *  range includes the last transaction in the linearization. */
240
    virtual bool GetClusterRefs(TxGraphImpl& graph, std::span<TxGraph::Ref*> range, LinearizationIndex start_pos) noexcept = 0;
241
    /** Get the individual transaction feerate of a Cluster element. */
242
    virtual FeePerWeight GetIndividualFeerate(DepGraphIndex idx) noexcept = 0;
243
    /** Modify the fee of a Cluster element. */
244
    virtual void SetFee(TxGraphImpl& graph, int level, DepGraphIndex idx, int64_t fee) noexcept = 0;
245
246
    // Debugging functions.
247
248
    virtual void SanityCheck(const TxGraphImpl& graph, int level) const = 0;
249
};
250
251
/** An implementation of Cluster that uses a DepGraph and vectors, to support arbitrary numbers of
252
 *  transactions up to MAX_CLUSTER_COUNT_LIMIT. */
253
class GenericClusterImpl final : public Cluster
254
{
255
    friend class TxGraphImpl;
256
    /** The DepGraph for this cluster, holding all feerates, and ancestors/descendants. */
257
    DepGraph<SetType> m_depgraph;
258
    /** m_mapping[i] gives the GraphIndex for the position i transaction in m_depgraph. Values for
259
     *  positions i that do not exist in m_depgraph shouldn't ever be accessed and thus don't
260
     *  matter. m_mapping.size() equals m_depgraph.PositionRange(). */
261
    std::vector<GraphIndex> m_mapping;
262
    /** The current linearization of the cluster. m_linearization.size() equals
263
     *  m_depgraph.TxCount(). This is always kept topological. */
264
    std::vector<DepGraphIndex> m_linearization;
265
266
public:
267
    /** The smallest number of transactions this Cluster implementation is intended for. */
268
    static constexpr DepGraphIndex MIN_INTENDED_TX_COUNT{2};
269
    /** The largest number of transactions this Cluster implementation supports. */
270
    static constexpr DepGraphIndex MAX_TX_COUNT{SetType::Size()};
271
272
    GenericClusterImpl() noexcept = delete;
273
    /** Construct an empty GenericClusterImpl. */
274
    explicit GenericClusterImpl(uint64_t sequence) noexcept;
275
276
    size_t TotalMemoryUsage() const noexcept final;
277
23.2k
    constexpr DepGraphIndex GetMinIntendedTxCount() const noexcept final { return MIN_INTENDED_TX_COUNT; }
278
26.8k
    constexpr DepGraphIndex GetMaxTxCount() const noexcept final { return MAX_TX_COUNT; }
279
0
    DepGraphIndex GetDepGraphIndexRange() const noexcept final { return m_depgraph.PositionRange(); }
280
123k
    LinearizationIndex GetTxCount() const noexcept final { return m_linearization.size(); }
281
    uint64_t GetTotalTxSize() const noexcept final;
282
108k
    GraphIndex GetClusterEntry(DepGraphIndex index) const noexcept final { return m_mapping[index]; }
283
    DepGraphIndex AppendTransaction(GraphIndex graph_idx, FeePerWeight feerate) noexcept final;
284
    void AddDependencies(SetType parents, DepGraphIndex child) noexcept final;
285
    void ExtractTransactions(const std::function<void (DepGraphIndex, GraphIndex, FeePerWeight)>& visit1_fn, const std::function<void (DepGraphIndex, GraphIndex, SetType)>& visit2_fn) noexcept final;
286
    int GetLevel(const TxGraphImpl& graph) const noexcept final;
287
191
    void UpdateMapping(DepGraphIndex cluster_idx, GraphIndex graph_idx) noexcept final { m_mapping[cluster_idx] = graph_idx; }
288
    void Updated(TxGraphImpl& graph, int level, bool rename) noexcept final;
289
    void RemoveChunkData(TxGraphImpl& graph) noexcept final;
290
    Cluster* CopyToStaging(TxGraphImpl& graph) const noexcept final;
291
    void GetConflicts(const TxGraphImpl& graph, std::vector<Cluster*>& out) const noexcept final;
292
    void MakeStagingTransactionsMissing(TxGraphImpl& graph) noexcept final;
293
    void Clear(TxGraphImpl& graph, int level) noexcept final;
294
    void MoveToMain(TxGraphImpl& graph) noexcept final;
295
    void Compact() noexcept final;
296
    void ApplyRemovals(TxGraphImpl& graph, int level, std::span<GraphIndex>& to_remove) noexcept final;
297
    [[nodiscard]] bool Split(TxGraphImpl& graph, int level) noexcept final;
298
    void Merge(TxGraphImpl& graph, int level, Cluster& cluster) noexcept final;
299
    void ApplyDependencies(TxGraphImpl& graph, int level, std::span<std::pair<GraphIndex, GraphIndex>> to_apply) noexcept final;
300
    std::pair<uint64_t, bool> Relinearize(TxGraphImpl& graph, int level, uint64_t max_cost) noexcept final;
301
    void AppendChunkFeerates(std::vector<FeeFrac>& ret) const noexcept final;
302
    uint64_t AppendTrimData(std::vector<TrimTxData>& ret, std::vector<std::pair<GraphIndex, GraphIndex>>& deps) const noexcept final;
303
    void GetAncestorRefs(const TxGraphImpl& graph, std::span<std::pair<Cluster*, DepGraphIndex>>& args, std::vector<TxGraph::Ref*>& output) noexcept final;
304
    void GetDescendantRefs(const TxGraphImpl& graph, std::span<std::pair<Cluster*, DepGraphIndex>>& args, std::vector<TxGraph::Ref*>& output) noexcept final;
305
    bool GetClusterRefs(TxGraphImpl& graph, std::span<TxGraph::Ref*> range, LinearizationIndex start_pos) noexcept final;
306
    FeePerWeight GetIndividualFeerate(DepGraphIndex idx) noexcept final;
307
    void SetFee(TxGraphImpl& graph, int level, DepGraphIndex idx, int64_t fee) noexcept final;
308
    void SanityCheck(const TxGraphImpl& graph, int level) const final;
309
};
310
311
/** An implementation of Cluster that only supports 1 transaction. */
312
class SingletonClusterImpl final : public Cluster
313
{
314
    friend class TxGraphImpl;
315
316
    /** The feerate of the (singular) transaction in this Cluster. */
317
    FeePerWeight m_feerate;
318
    /** Constant to indicate that this Cluster is empty. */
319
    static constexpr auto NO_GRAPH_INDEX = GraphIndex(-1);
320
    /** The GraphIndex of the transaction. NO_GRAPH_INDEX if this Cluster is empty. */
321
    GraphIndex m_graph_index = NO_GRAPH_INDEX;
322
323
public:
324
    /** The smallest number of transactions this Cluster implementation is intended for. */
325
    static constexpr DepGraphIndex MIN_INTENDED_TX_COUNT{1};
326
    /** The largest number of transactions this Cluster implementation supports. */
327
    static constexpr DepGraphIndex MAX_TX_COUNT{1};
328
329
    SingletonClusterImpl() noexcept = delete;
330
    /** Construct an empty SingletonClusterImpl. */
331
12.7k
    explicit SingletonClusterImpl(uint64_t sequence) noexcept : Cluster(sequence) {}
332
333
    size_t TotalMemoryUsage() const noexcept final;
334
16.3k
    constexpr DepGraphIndex GetMinIntendedTxCount() const noexcept final { return MIN_INTENDED_TX_COUNT; }
335
18.3k
    constexpr DepGraphIndex GetMaxTxCount() const noexcept final { return MAX_TX_COUNT; }
336
240k
    LinearizationIndex GetTxCount() const noexcept final { return m_graph_index != NO_GRAPH_INDEX; }
337
7.55k
    DepGraphIndex GetDepGraphIndexRange() const noexcept final { return GetTxCount(); }
338
23.8k
    uint64_t GetTotalTxSize() const noexcept final { return GetTxCount() ? m_feerate.size : 
00
; }
339
16.3k
    GraphIndex GetClusterEntry(DepGraphIndex index) const noexcept final { Assume(index == 0); Assume(GetTxCount()); return m_graph_index; }
Line
Count
Source
128
16.3k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
    GraphIndex GetClusterEntry(DepGraphIndex index) const noexcept final { Assume(index == 0); Assume(GetTxCount()); return m_graph_index; }
Line
Count
Source
128
16.3k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
340
    DepGraphIndex AppendTransaction(GraphIndex graph_idx, FeePerWeight feerate) noexcept final;
341
    void AddDependencies(SetType parents, DepGraphIndex child) noexcept final;
342
    void ExtractTransactions(const std::function<void (DepGraphIndex, GraphIndex, FeePerWeight)>& visit1_fn, const std::function<void (DepGraphIndex, GraphIndex, SetType)>& visit2_fn) noexcept final;
343
    int GetLevel(const TxGraphImpl& graph) const noexcept final;
344
615
    void UpdateMapping(DepGraphIndex cluster_idx, GraphIndex graph_idx) noexcept final { Assume(cluster_idx == 0); m_graph_index = graph_idx; }
Line
Count
Source
128
615
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
345
    void Updated(TxGraphImpl& graph, int level, bool rename) noexcept final;
346
    void RemoveChunkData(TxGraphImpl& graph) noexcept final;
347
    Cluster* CopyToStaging(TxGraphImpl& graph) const noexcept final;
348
    void GetConflicts(const TxGraphImpl& graph, std::vector<Cluster*>& out) const noexcept final;
349
    void MakeStagingTransactionsMissing(TxGraphImpl& graph) noexcept final;
350
    void Clear(TxGraphImpl& graph, int level) noexcept final;
351
    void MoveToMain(TxGraphImpl& graph) noexcept final;
352
    void Compact() noexcept final;
353
    void ApplyRemovals(TxGraphImpl& graph, int level, std::span<GraphIndex>& to_remove) noexcept final;
354
    [[nodiscard]] bool Split(TxGraphImpl& graph, int level) noexcept final;
355
    void Merge(TxGraphImpl& graph, int level, Cluster& cluster) noexcept final;
356
    void ApplyDependencies(TxGraphImpl& graph, int level, std::span<std::pair<GraphIndex, GraphIndex>> to_apply) noexcept final;
357
    std::pair<uint64_t, bool> Relinearize(TxGraphImpl& graph, int level, uint64_t max_cost) noexcept final;
358
    void AppendChunkFeerates(std::vector<FeeFrac>& ret) const noexcept final;
359
    uint64_t AppendTrimData(std::vector<TrimTxData>& ret, std::vector<std::pair<GraphIndex, GraphIndex>>& deps) const noexcept final;
360
    void GetAncestorRefs(const TxGraphImpl& graph, std::span<std::pair<Cluster*, DepGraphIndex>>& args, std::vector<TxGraph::Ref*>& output) noexcept final;
361
    void GetDescendantRefs(const TxGraphImpl& graph, std::span<std::pair<Cluster*, DepGraphIndex>>& args, std::vector<TxGraph::Ref*>& output) noexcept final;
362
    bool GetClusterRefs(TxGraphImpl& graph, std::span<TxGraph::Ref*> range, LinearizationIndex start_pos) noexcept final;
363
    FeePerWeight GetIndividualFeerate(DepGraphIndex idx) noexcept final;
364
    void SetFee(TxGraphImpl& graph, int level, DepGraphIndex idx, int64_t fee) noexcept final;
365
    void SanityCheck(const TxGraphImpl& graph, int level) const final;
366
};
367
368
/** The transaction graph, including staged changes.
369
 *
370
 * The overall design of the data structure consists of 3 interlinked representations:
371
 * - The transactions (held as a vector of TxGraphImpl::Entry inside TxGraphImpl).
372
 * - The clusters (Cluster objects in per-quality vectors inside TxGraphImpl::ClusterSet).
373
 * - The Refs (TxGraph::Ref objects, held externally by users of the TxGraph class)
374
 *
375
 * The Clusters are kept in one or two ClusterSet objects, one for the "main" graph, and one for
376
 * the proposed changes ("staging"). If a transaction occurs in both, they share the same Entry,
377
 * but there will be a separate Cluster per graph.
378
 *
379
 * Clusters and Refs contain the index of the Entry objects they refer to, and the Entry objects
380
 * refer back to the Clusters and Refs the corresponding transaction is contained in.
381
 *
382
 * While redundant, this permits moving all of them independently, without invalidating things
383
 * or costly iteration to fix up everything:
384
 * - Entry objects can be moved to fill holes left by removed transactions in the Entry vector
385
 *   (see TxGraphImpl::Compact).
386
 * - Clusters can be rewritten continuously (removals can cause them to split, new dependencies
387
 *   can cause them to be merged).
388
 * - Ref objects can be held outside the class, while permitting them to be moved around, and
389
 *   inherited from.
390
 */
391
class TxGraphImpl final : public TxGraph
392
{
393
    friend class Cluster;
394
    friend class SingletonClusterImpl;
395
    friend class GenericClusterImpl;
396
    friend class BlockBuilderImpl;
397
private:
398
    /** Internal RNG. */
399
    FastRandomContext m_rng;
400
    /** This TxGraphImpl's maximum cluster count limit. */
401
    const DepGraphIndex m_max_cluster_count;
402
    /** This TxGraphImpl's maximum cluster size limit. */
403
    const uint64_t m_max_cluster_size;
404
    /** The amount of linearization work needed per cluster to be considered acceptable. */
405
    const uint64_t m_acceptable_cost;
406
    /** Fallback ordering for transactions. */
407
    const std::function<std::strong_ordering(const TxGraph::Ref&, const TxGraph::Ref&)> m_fallback_order;
408
409
    /** Information about one group of Clusters to be merged. */
410
    struct GroupEntry
411
    {
412
        /** Where the clusters to be merged start in m_group_clusters. */
413
        uint32_t m_cluster_offset;
414
        /** How many clusters to merge. */
415
        uint32_t m_cluster_count;
416
        /** Where the dependencies for this cluster group in m_deps_to_add start. */
417
        uint32_t m_deps_offset;
418
        /** How many dependencies to add. */
419
        uint32_t m_deps_count;
420
    };
421
422
    /** Information about all groups of Clusters to be merged. */
423
    struct GroupData
424
    {
425
        /** The groups of Clusters to be merged. */
426
        std::vector<GroupEntry> m_groups;
427
        /** Which clusters are to be merged. GroupEntry::m_cluster_offset indexes into this. */
428
        std::vector<Cluster*> m_group_clusters;
429
    };
430
431
    /** The collection of all Clusters in main or staged. */
432
    struct ClusterSet
433
    {
434
        /** The vectors of clusters, one vector per quality level. ClusterSetIndex indexes into each. */
435
        std::array<std::vector<std::unique_ptr<Cluster>>, int(QualityLevel::NONE)> m_clusters;
436
        /** Which removals have yet to be applied. */
437
        std::vector<GraphIndex> m_to_remove;
438
        /** Which dependencies are to be added ((parent,child) pairs). GroupData::m_deps_offset indexes
439
         *  into this. */
440
        std::vector<std::pair<GraphIndex, GraphIndex>> m_deps_to_add;
441
        /** Information about the merges to be performed, if known. */
442
        std::optional<GroupData> m_group_data = GroupData{};
443
        /** Which entries were removed in this ClusterSet (so they can be wiped on abort). This
444
         *  includes all entries which have an (R) removed locator at this level (staging only),
445
         *  plus optionally any transaction in m_unlinked. */
446
        std::vector<GraphIndex> m_removed;
447
        /** Total number of transactions in this graph (sum of all transaction counts in all
448
         *  Clusters, and for staging also those inherited from the main ClusterSet). */
449
        GraphIndex m_txcount{0};
450
        /** Total number of individually oversized transactions in the graph. */
451
        GraphIndex m_txcount_oversized{0};
452
        /** Whether this graph is oversized (if known). */
453
        std::optional<bool> m_oversized{false};
454
        /** The combined TotalMemoryUsage of all clusters in this level (only Clusters that
455
         *  are materialized; in staging, implicit Clusters from main are not counted), */
456
        size_t m_cluster_usage{0};
457
458
13.5k
        ClusterSet() noexcept = default;
459
    };
460
461
    /** The main ClusterSet. */
462
    ClusterSet m_main_clusterset;
463
    /** The staging ClusterSet, if any. */
464
    std::optional<ClusterSet> m_staging_clusterset;
465
    /** Next sequence number to assign to created Clusters. */
466
    uint64_t m_next_sequence_counter{0};
467
468
    /** Information about a chunk in the main graph. */
469
    struct ChunkData
470
    {
471
        /** The Entry which is the last transaction of the chunk. */
472
        mutable GraphIndex m_graph_index;
473
        /** How many transactions the chunk contains (-1 = singleton tail of cluster). */
474
        LinearizationIndex m_chunk_count;
475
476
        ChunkData(GraphIndex graph_index, LinearizationIndex chunk_count) noexcept :
477
30.9k
            m_graph_index{graph_index}, m_chunk_count{chunk_count} {}
478
    };
479
480
    /** Compare two Cluster* by their m_sequence value (while supporting nullptr). */
481
    static std::strong_ordering CompareClusters(Cluster* a, Cluster* b) noexcept
482
18.7k
    {
483
        // The nullptr pointer compares before everything else.
484
18.7k
        if (a == nullptr || b == nullptr) {
485
0
            return (a != nullptr) <=> (b != nullptr);
486
0
        }
487
        // If neither pointer is nullptr, compare the Clusters' sequence numbers.
488
18.7k
        Assume(a == b || a->m_sequence != b->m_sequence);
Line
Count
Source
128
30.7k
#define Assume(val) 
inline_assertion_check<false>(18.7k
val, std::source_location::current(),
#18.7k
val)
489
18.7k
        return a->m_sequence <=> b->m_sequence;
490
18.7k
    }
491
492
    /** Compare two entries (which must both exist within the main graph). */
493
    std::strong_ordering CompareMainTransactions(GraphIndex a, GraphIndex b) const noexcept
494
652k
    {
495
652k
        if (a == b) 
return std::strong_ordering::equal0
;
496
652k
        Assume(a < m_entries.size() && b < m_entries.size());
Line
Count
Source
128
1.30M
#define Assume(val) 
inline_assertion_check<false>(652k
val, std::source_location::current(),
#652k
val)
497
652k
        const auto& entry_a = m_entries[a];
498
652k
        const auto& entry_b = m_entries[b];
499
        // Compare chunk feerates, and return result if it differs.
500
652k
        auto feerate_cmp = ByRatio{entry_b.m_main_chunk_feerate} <=> ByRatio{entry_a.m_main_chunk_feerate};
501
652k
        if (feerate_cmp != 0) 
return feerate_cmp0
;
502
        // Compare equal-feerate chunk prefix size for comparing equal chunk feerates. This does two
503
        // things: it distinguishes equal-feerate chunks within the same cluster (because later
504
        // ones will always have a higher prefix size), and it may distinguish equal-feerate chunks
505
        // from distinct clusters.
506
652k
        if (entry_a.m_main_equal_feerate_chunk_prefix_size != entry_b.m_main_equal_feerate_chunk_prefix_size) {
507
487k
            return entry_a.m_main_equal_feerate_chunk_prefix_size <=> entry_b.m_main_equal_feerate_chunk_prefix_size;
508
487k
        }
509
        // Compare by maximum m_fallback_order element to order equal-feerate chunks in distinct
510
        // clusters, when the equal-feerate-prefix size is also the same.
511
164k
        const auto& locator_a = entry_a.m_locator[0];
512
164k
        const auto& locator_b = entry_b.m_locator[0];
513
164k
        Assume(locator_a.IsPresent() && locator_b.IsPresent());
Line
Count
Source
128
329k
#define Assume(val) 
inline_assertion_check<false>(164k
val, std::source_location::current(),
#164k
val)
514
164k
        if (locator_a.cluster != locator_b.cluster) {
515
164k
            auto fallback_cmp = m_fallback_order(*m_entries[entry_a.m_main_max_chunk_fallback].m_ref,
516
164k
                                                 *m_entries[entry_b.m_main_max_chunk_fallback].m_ref);
517
164k
            if (fallback_cmp != 0) return fallback_cmp;
518
            // This shouldn't be reachable as m_fallback_order defines a strong ordering.
519
0
            Assume(false);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
520
0
            return CompareClusters(locator_a.cluster, locator_b.cluster);
521
164k
        }
522
        // Within a single chunk, sort by position within cluster linearization.
523
0
        return entry_a.m_main_lin_index <=> entry_b.m_main_lin_index;
524
164k
    }
525
526
    /** Comparator for ChunkData objects in mining order. */
527
    class ChunkOrder
528
    {
529
        const TxGraphImpl* const m_graph;
530
    public:
531
925
        explicit ChunkOrder(const TxGraphImpl* graph) : m_graph(graph) {}
532
533
        bool operator()(const ChunkData& a, const ChunkData& b) const noexcept
534
130k
        {
535
130k
            return m_graph->CompareMainTransactions(a.m_graph_index, b.m_graph_index) < 0;
536
130k
        }
537
    };
538
539
    /** Definition for the mining index type. */
540
    using ChunkIndex = std::set<ChunkData, ChunkOrder>;
541
542
    /** Index of ChunkData objects, indexing the last transaction in each chunk in the main
543
     *  graph. */
544
    ChunkIndex m_main_chunkindex;
545
    /** Number of index-observing objects in existence (BlockBuilderImpls). */
546
    size_t m_main_chunkindex_observers{0};
547
    /** Cache of discarded ChunkIndex node handles to reuse, avoiding additional allocation. */
548
    std::vector<ChunkIndex::node_type> m_main_chunkindex_discarded;
549
550
    /** A Locator that describes whether, where, and in which Cluster an Entry appears.
551
     *  Every Entry has MAX_LEVELS locators, as it may appear in one Cluster per level.
552
     *
553
     *  Each level of a Locator is in one of three states:
554
     *
555
     *  - (P)resent: actually occurs in a Cluster at that level.
556
     *
557
     *  - (M)issing:
558
     *    - In the main graph:    the transaction does not exist in main.
559
     *    - In the staging graph: the transaction's existence is the same as in main. If it doesn't
560
     *                            exist in main, (M) in staging means it does not exist there
561
     *                            either. If it does exist in main, (M) in staging means the
562
     *                            cluster it is in has not been modified in staging, and thus the
563
     *                            transaction implicitly exists in staging too (without explicit
564
     *                            Cluster object; see PullIn() to create it in staging too).
565
     *
566
     *  - (R)emoved: only possible in staging; it means the transaction exists in main, but is
567
     *               removed in staging.
568
     *
569
     * The following combinations are possible:
570
     * - (M,M): the transaction doesn't exist in either graph.
571
     * - (P,M): the transaction exists in both, but only exists explicitly in a Cluster object in
572
     *          main. Its existence in staging is inherited from main.
573
     * - (P,P): the transaction exists in both, and is materialized in both. Thus, the clusters
574
     *          and/or their linearizations may be different in main and staging.
575
     * - (M,P): the transaction is added in staging, and does not exist in main.
576
     * - (P,R): the transaction exists in main, but is removed in staging.
577
     *
578
     * When staging does not exist, only (M,M) and (P,M) are possible.
579
     */
580
    struct Locator
581
    {
582
        /** Which Cluster the Entry appears in (nullptr = missing). */
583
        Cluster* cluster{nullptr};
584
        /** Where in the Cluster it appears (if cluster == nullptr: 0 = missing, -1 = removed). */
585
        DepGraphIndex index{0};
586
587
        /** Mark this Locator as missing (= same as lower level, or non-existing if level 0). */
588
18.1k
        void SetMissing() noexcept { cluster = nullptr; index = 0; }
589
        /** Mark this Locator as removed (not allowed in level 0). */
590
0
        void SetRemoved() noexcept { cluster = nullptr; index = DepGraphIndex(-1); }
591
        /** Mark this Locator as present, in the specified Cluster. */
592
74.5k
        void SetPresent(Cluster* c, DepGraphIndex i) noexcept { cluster = c; index = i; }
593
        /** Check if this Locator is missing. */
594
321k
        bool IsMissing() const noexcept { return cluster == nullptr && 
index == 0141k
; }
595
        /** Check if this Locator is removed. */
596
434k
        bool IsRemoved() const noexcept { return cluster == nullptr && 
index == DepGraphIndex(-1)254k
; }
597
        /** Check if this Locator is present (in some Cluster). */
598
2.06M
        bool IsPresent() const noexcept { return cluster != nullptr; }
599
    };
600
601
    /** Internal information about each transaction in a TxGraphImpl. */
602
    struct Entry
603
    {
604
        /** Pointer to the corresponding Ref object if any, or nullptr if unlinked. */
605
        Ref* m_ref{nullptr};
606
        /** Iterator to the corresponding ChunkData, if any, and m_main_chunkindex.end() otherwise.
607
         *  This is initialized on construction of the Entry, in AddTransaction. */
608
        ChunkIndex::iterator m_main_chunkindex_iterator;
609
        /** Which Cluster and position therein this Entry appears in. ([0] = main, [1] = staged). */
610
        Locator m_locator[MAX_LEVELS];
611
        /** The chunk feerate of this transaction in main (if present in m_locator[0]). */
612
        FeePerWeight m_main_chunk_feerate;
613
        /** The equal-feerate chunk prefix size of this transaction in main. If the transaction is
614
         *  part of chunk C in main, then this gives the sum of the sizes of all chunks in C's
615
         *  cluster, whose feerate is equal to that of C, which do not appear after C itself in
616
         *  the cluster's linearization.
617
         *  This provides a way to sort equal-feerate chunks across clusters, in a way that agrees
618
         *  with the within-cluster chunk ordering. */
619
        int32_t m_main_equal_feerate_chunk_prefix_size;
620
        /** The position this transaction has in the main linearization (if present). */
621
        LinearizationIndex m_main_lin_index;
622
        /** Of all transactions within this transaction's chunk in main (if present there), the
623
         *  maximal one according to m_fallback_order. */
624
        GraphIndex m_main_max_chunk_fallback = GraphIndex(-1);
625
    };
626
627
    /** The set of all transactions (in all levels combined). GraphIndex values index into this. */
628
    std::vector<Entry> m_entries;
629
630
    /** Set of Entries which have no linked Ref anymore. */
631
    std::vector<GraphIndex> m_unlinked;
632
633
public:
634
    /** Construct a new TxGraphImpl with the specified limits and fallback order. */
635
    explicit TxGraphImpl(
636
        DepGraphIndex max_cluster_count,
637
        uint64_t max_cluster_size,
638
        uint64_t acceptable_cost,
639
        const std::function<std::strong_ordering(const TxGraph::Ref&, const TxGraph::Ref&)>& fallback_order
640
    ) noexcept :
641
925
        m_max_cluster_count(max_cluster_count),
642
925
        m_max_cluster_size(max_cluster_size),
643
925
        m_acceptable_cost(acceptable_cost),
644
925
        m_fallback_order(fallback_order),
645
925
        m_main_chunkindex(ChunkOrder(this))
646
925
    {
647
925
        Assume(max_cluster_count >= 1);
Line
Count
Source
128
925
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
648
925
        Assume(max_cluster_count <= MAX_CLUSTER_COUNT_LIMIT);
Line
Count
Source
128
925
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
649
925
    }
650
651
    /** Destructor. */
652
    ~TxGraphImpl() noexcept;
653
654
    // Cannot move or copy (would invalidate TxGraphImpl* in Ref, MiningOrder, EvictionOrder).
655
    TxGraphImpl(const TxGraphImpl&) = delete;
656
    TxGraphImpl& operator=(const TxGraphImpl&) = delete;
657
    TxGraphImpl(TxGraphImpl&&) = delete;
658
    TxGraphImpl& operator=(TxGraphImpl&&) = delete;
659
660
    // Simple helper functions.
661
662
    /** Swap the Entry referred to by a and the one referred to by b. Gather main clusters to
663
     *  update afterwards. */
664
    void SwapIndexes(GraphIndex a, GraphIndex b, std::vector<Cluster*>& affected_main) noexcept;
665
    /** If idx exists in the specified level ClusterSet (explicitly, or in the level below and not
666
    *   removed), return the Cluster it is in. Otherwise, return nullptr. */
667
47.4k
    Cluster* FindCluster(GraphIndex idx, int level) const noexcept { return FindClusterAndLevel(idx, level).first; }
668
    /** Like FindCluster, but also return what level the match was found in (-1 if not found). */
669
    std::pair<Cluster*, int> FindClusterAndLevel(GraphIndex idx, int level) const noexcept;
670
    /** Extract a Cluster from its ClusterSet, and set its quality to QualityLevel::NONE. */
671
    std::unique_ptr<Cluster> ExtractCluster(int level, QualityLevel quality, ClusterSetIndex setindex) noexcept;
672
    /** Delete a Cluster. */
673
    void DeleteCluster(Cluster& cluster, int level) noexcept;
674
    /** Insert a Cluster into its ClusterSet. */
675
    ClusterSetIndex InsertCluster(int level, std::unique_ptr<Cluster>&& cluster, QualityLevel quality) noexcept;
676
    /** Change the QualityLevel of a Cluster (identified by old_quality and old_index). */
677
    void SetClusterQuality(int level, QualityLevel old_quality, ClusterSetIndex old_index, QualityLevel new_quality) noexcept;
678
    /** Get the index of the top level ClusterSet (staging if it exists, main otherwise). */
679
2.58M
    int GetTopLevel() const noexcept { return m_staging_clusterset.has_value(); }
680
    /** Get the specified level (staging if it exists and level is TOP, main otherwise). */
681
235k
    int GetSpecifiedLevel(Level level) const noexcept { return level == Level::TOP && 
m_staging_clusterset.has_value()8.04k
; }
682
    /** Get a reference to the ClusterSet at the specified level (which must exist). */
683
    ClusterSet& GetClusterSet(int level) noexcept;
684
    const ClusterSet& GetClusterSet(int level) const noexcept;
685
    /** Make a transaction not exist at a specified level. It must currently exist there.
686
     *  oversized_tx indicates whether the transaction is an individually-oversized one
687
     *  (OVERSIZED_SINGLETON). */
688
    void ClearLocator(int level, GraphIndex index, bool oversized_tx) noexcept;
689
    /** Find which Clusters in main conflict with ones in staging. */
690
    std::vector<Cluster*> GetConflicts() const noexcept;
691
    /** Clear an Entry's ChunkData. */
692
    void ClearChunkData(Entry& entry) noexcept;
693
    /** Give an Entry a ChunkData object. */
694
    void CreateChunkData(GraphIndex idx, LinearizationIndex chunk_count) noexcept;
695
    /** Create an empty GenericClusterImpl object. */
696
    std::unique_ptr<GenericClusterImpl> CreateEmptyGenericCluster() noexcept
697
2.04k
    {
698
2.04k
        return std::make_unique<GenericClusterImpl>(m_next_sequence_counter++);
699
2.04k
    }
700
    /** Create an empty SingletonClusterImpl object. */
701
    std::unique_ptr<SingletonClusterImpl> CreateEmptySingletonCluster() noexcept
702
12.7k
    {
703
12.7k
        return std::make_unique<SingletonClusterImpl>(m_next_sequence_counter++);
704
12.7k
    }
705
    /** Create an empty Cluster of the appropriate implementation for the specified (maximum) tx
706
     *  count. */
707
    std::unique_ptr<Cluster> CreateEmptyCluster(DepGraphIndex tx_count) noexcept
708
14.7k
    {
709
14.7k
        if (tx_count >= SingletonClusterImpl::MIN_INTENDED_TX_COUNT && tx_count <= SingletonClusterImpl::MAX_TX_COUNT) {
710
12.7k
            return CreateEmptySingletonCluster();
711
12.7k
        }
712
2.04k
        if (tx_count >= GenericClusterImpl::MIN_INTENDED_TX_COUNT && tx_count <= GenericClusterImpl::MAX_TX_COUNT) {
713
2.04k
            return CreateEmptyGenericCluster();
714
2.04k
        }
715
2.04k
        assert
(false)0
;
716
0
        return {};
717
0
    }
718
719
    // Functions for handling Refs.
720
721
    /** Only called by Ref's move constructor/assignment to update Ref locations. */
722
    void UpdateRef(GraphIndex idx, Ref& new_location) noexcept final
723
0
    {
724
0
        auto& entry = m_entries[idx];
725
0
        Assume(entry.m_ref != nullptr);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
726
0
        entry.m_ref = &new_location;
727
0
    }
728
729
    /** Only called by Ref::~Ref to unlink Refs, and Ref's move assignment. */
730
    void UnlinkRef(GraphIndex idx) noexcept final
731
12.6k
    {
732
12.6k
        auto& entry = m_entries[idx];
733
12.6k
        Assume(entry.m_ref != nullptr);
Line
Count
Source
128
12.6k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
734
12.6k
        Assume(m_main_chunkindex_observers == 0 || !entry.m_locator[0].IsPresent());
Line
Count
Source
128
12.6k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
735
        // Remove all chunk index entries for the affected cluster, to avoid any chunk indexes
736
        // referencing unlinked/destroyed Refs.
737
12.6k
        if (entry.m_locator[0].IsPresent()) {
738
8.04k
            entry.m_locator[0].cluster->RemoveChunkData(*this);
739
8.04k
        }
740
12.6k
        entry.m_ref = nullptr;
741
        // Mark the transaction as to be removed in all levels where it explicitly or implicitly
742
        // exists.
743
12.6k
        bool exists_anywhere{false};
744
12.6k
        bool exists{false};
745
25.3k
        for (int level = 0; level <= GetTopLevel(); 
++level12.6k
) {
746
12.6k
            if (entry.m_locator[level].IsPresent()) {
747
8.04k
                exists_anywhere = true;
748
8.04k
                exists = true;
749
8.04k
            } else 
if (4.61k
entry.m_locator[level].IsRemoved()4.61k
) {
750
0
                exists = false;
751
0
            }
752
12.6k
            if (exists) {
753
8.04k
                auto& clusterset = GetClusterSet(level);
754
8.04k
                clusterset.m_to_remove.push_back(idx);
755
                // Force recomputation of grouping data.
756
8.04k
                clusterset.m_group_data = std::nullopt;
757
                // Do not wipe the oversized state of main if staging exists. The reason for this
758
                // is that the alternative would mean that cluster merges may need to be applied to
759
                // a formerly-oversized main graph while staging exists (to satisfy chunk feerate
760
                // queries into main, for example), and such merges could conflict with pulls of
761
                // some of their constituents into staging.
762
8.04k
                if (level == GetTopLevel() && clusterset.m_oversized == true) {
763
0
                    clusterset.m_oversized = std::nullopt;
764
0
                }
765
8.04k
            }
766
12.6k
        }
767
12.6k
        m_unlinked.push_back(idx);
768
12.6k
        if (!exists_anywhere) 
Compact()4.61k
;
769
12.6k
    }
770
771
    // Functions related to various normalization/application steps.
772
    /** Get rid of unlinked Entry objects in m_entries, if possible (this changes the GraphIndex
773
     *  values for remaining Entry objects, so this only does something when no to-be-applied
774
     *  operations or staged removals referring to GraphIndexes remain). */
775
    void Compact() noexcept;
776
    /** If cluster is not in staging, copy it there, and return a pointer to it.
777
    *   Staging must exist, and this modifies the locators of its
778
    *   transactions from inherited (P,M) to explicit (P,P). */
779
    Cluster* PullIn(Cluster* cluster, int level) noexcept;
780
    /** Apply all removals queued up in m_to_remove to the relevant Clusters (which get a
781
     *  NEEDS_SPLIT* QualityLevel) up to the specified level. */
782
    void ApplyRemovals(int up_to_level) noexcept;
783
    /** Split an individual cluster. */
784
    void Split(Cluster& cluster, int level) noexcept;
785
    /** Split all clusters that need splitting up to the specified level. */
786
    void SplitAll(int up_to_level) noexcept;
787
    /** Populate m_group_data based on m_deps_to_add in the specified level. */
788
    void GroupClusters(int level) noexcept;
789
    /** Merge the specified clusters. */
790
    void Merge(std::span<Cluster*> to_merge, int level) noexcept;
791
    /** Apply all m_deps_to_add to the relevant Clusters in the specified level. */
792
    void ApplyDependencies(int level) noexcept;
793
    /** Make a specified Cluster have quality ACCEPTABLE or OPTIMAL. */
794
    void MakeAcceptable(Cluster& cluster, int level) noexcept;
795
    /** Make all Clusters at the specified level have quality ACCEPTABLE or OPTIMAL. */
796
    void MakeAllAcceptable(int level) noexcept;
797
798
    // Implementations for the public TxGraph interface.
799
800
    void AddTransaction(Ref& arg, const FeePerWeight& feerate) noexcept final;
801
    void RemoveTransaction(const Ref& arg) noexcept final;
802
    void AddDependency(const Ref& parent, const Ref& child) noexcept final;
803
    void SetTransactionFee(const Ref&, int64_t fee) noexcept final;
804
805
    bool DoWork(uint64_t max_cost) noexcept final;
806
807
    void StartStaging() noexcept final;
808
    void CommitStaging() noexcept final;
809
    void AbortStaging() noexcept final;
810
12.6k
    bool HaveStaging() const noexcept final { return m_staging_clusterset.has_value(); }
811
812
    bool Exists(const Ref& arg, Level level) noexcept final;
813
    FeePerWeight GetMainChunkFeerate(const Ref& arg) noexcept final;
814
    FeePerWeight GetIndividualFeerate(const Ref& arg) noexcept final;
815
    std::vector<Ref*> GetCluster(const Ref& arg, Level level) noexcept final;
816
    std::vector<Ref*> GetAncestors(const Ref& arg, Level level) noexcept final;
817
    std::vector<Ref*> GetDescendants(const Ref& arg, Level level) noexcept final;
818
    std::vector<Ref*> GetAncestorsUnion(std::span<const Ref* const> args, Level level) noexcept final;
819
    std::vector<Ref*> GetDescendantsUnion(std::span<const Ref* const> args, Level level) noexcept final;
820
    GraphIndex GetTransactionCount(Level level) noexcept final;
821
    bool IsOversized(Level level) noexcept final;
822
    std::strong_ordering CompareMainOrder(const Ref& a, const Ref& b) noexcept final;
823
    GraphIndex CountDistinctClusters(std::span<const Ref* const> refs, Level level) noexcept final;
824
    std::pair<std::vector<FeeFrac>, std::vector<FeeFrac>> GetMainStagingDiagrams() noexcept final;
825
    std::vector<Ref*> Trim() noexcept final;
826
827
    std::unique_ptr<BlockBuilder> GetBlockBuilder() noexcept final;
828
    std::pair<std::vector<Ref*>, FeePerWeight> GetWorstMainChunk() noexcept final;
829
830
    size_t GetMainMemoryUsage() noexcept final;
831
832
    void SanityCheck() const final;
833
};
834
835
TxGraphImpl::ClusterSet& TxGraphImpl::GetClusterSet(int level) noexcept
836
7.49M
{
837
7.49M
    if (level == 0) 
return m_main_clusterset7.40M
;
838
96.2k
    Assume(level == 1);
Line
Count
Source
128
96.2k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
839
96.2k
    Assume(m_staging_clusterset.has_value());
Line
Count
Source
128
96.2k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
840
96.2k
    return *m_staging_clusterset;
841
7.49M
}
842
843
const TxGraphImpl::ClusterSet& TxGraphImpl::GetClusterSet(int level) const noexcept
844
221k
{
845
221k
    if (level == 0) 
return m_main_clusterset213k
;
846
8.04k
    Assume(level == 1);
Line
Count
Source
128
8.04k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
847
8.04k
    Assume(m_staging_clusterset.has_value());
Line
Count
Source
128
8.04k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
848
8.04k
    return *m_staging_clusterset;
849
221k
}
850
851
/** Implementation of the TxGraph::BlockBuilder interface. */
852
class BlockBuilderImpl final : public TxGraph::BlockBuilder
853
{
854
    /** Which TxGraphImpl this object is doing block building for. It will have its
855
     *  m_main_chunkindex_observers incremented as long as this BlockBuilderImpl exists. */
856
    TxGraphImpl* const m_graph;
857
    /** Cluster sequence numbers which we're not including further transactions from. */
858
    std::unordered_set<uint64_t> m_excluded_clusters;
859
    /** Iterator to the current chunk in the chunk index. end() if nothing further remains. */
860
    TxGraphImpl::ChunkIndex::const_iterator m_cur_iter;
861
    /** Which cluster the current chunk belongs to, so we can exclude further transactions from it
862
     *  when that chunk is skipped. */
863
    Cluster* m_cur_cluster;
864
    /** Whether we know that m_cur_iter points to the last chunk of m_cur_cluster. */
865
    bool m_known_end_of_cluster;
866
867
    // Move m_cur_iter / m_cur_cluster to the next acceptable chunk.
868
    void Next() noexcept;
869
870
public:
871
    /** Construct a new BlockBuilderImpl to build blocks for the provided graph. */
872
    BlockBuilderImpl(TxGraphImpl& graph) noexcept;
873
874
    // Implement the public interface.
875
    ~BlockBuilderImpl() final;
876
    std::optional<std::pair<std::vector<TxGraph::Ref*>, FeePerWeight>> GetCurrentChunk() noexcept final;
877
    void Include() noexcept final;
878
    void Skip() noexcept final;
879
};
880
881
void TxGraphImpl::ClearChunkData(Entry& entry) noexcept
882
108k
{
883
108k
    if (entry.m_main_chunkindex_iterator != m_main_chunkindex.end()) {
884
30.9k
        Assume(m_main_chunkindex_observers == 0);
Line
Count
Source
128
30.9k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
885
        // If the Entry has a non-empty m_main_chunkindex_iterator, extract it, and move the handle
886
        // to the cache of discarded chunkindex entries.
887
30.9k
        m_main_chunkindex_discarded.emplace_back(m_main_chunkindex.extract(entry.m_main_chunkindex_iterator));
888
30.9k
        entry.m_main_chunkindex_iterator = m_main_chunkindex.end();
889
30.9k
    }
890
108k
}
891
892
void TxGraphImpl::CreateChunkData(GraphIndex idx, LinearizationIndex chunk_count) noexcept
893
30.9k
{
894
30.9k
    auto& entry = m_entries[idx];
895
    // Make sure to not create chunk data for unlinked entries, which would make invoking
896
    // m_fallback_order on them impossible.
897
30.9k
    Assume(entry.m_ref != nullptr);
Line
Count
Source
128
30.9k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
898
30.9k
    if (!m_main_chunkindex_discarded.empty()) {
899
        // Reuse an discarded node handle.
900
0
        auto& node = m_main_chunkindex_discarded.back().value();
901
0
        node.m_graph_index = idx;
902
0
        node.m_chunk_count = chunk_count;
903
0
        auto insert_result = m_main_chunkindex.insert(std::move(m_main_chunkindex_discarded.back()));
904
0
        Assume(insert_result.inserted);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
905
0
        entry.m_main_chunkindex_iterator = insert_result.position;
906
0
        m_main_chunkindex_discarded.pop_back();
907
30.9k
    } else {
908
        // Construct a new entry.
909
30.9k
        auto emplace_result = m_main_chunkindex.emplace(idx, chunk_count);
910
30.9k
        Assume(emplace_result.second);
Line
Count
Source
128
30.9k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
911
30.9k
        entry.m_main_chunkindex_iterator = emplace_result.first;
912
30.9k
    }
913
30.9k
}
914
915
size_t GenericClusterImpl::TotalMemoryUsage() const noexcept
916
36.9k
{
917
36.9k
    return // Dynamic memory allocated in this Cluster.
918
36.9k
           memusage::DynamicUsage(m_mapping) + memusage::DynamicUsage(m_linearization) +
919
           // Dynamic memory usage inside m_depgraph.
920
36.9k
           m_depgraph.DynamicMemoryUsage() +
921
           // Memory usage of the allocated Cluster itself.
922
36.9k
           memusage::MallocUsage(sizeof(GenericClusterImpl)) +
923
           // Memory usage of the ClusterSet::m_clusters entry.
924
36.9k
           sizeof(std::unique_ptr<Cluster>);
925
36.9k
}
926
927
size_t SingletonClusterImpl::TotalMemoryUsage() const noexcept
928
52.9k
{
929
52.9k
    return // Memory usage of the allocated SingletonClusterImpl itself.
930
52.9k
           memusage::MallocUsage(sizeof(SingletonClusterImpl)) +
931
           // Memory usage of the ClusterSet::m_clusters entry.
932
52.9k
           sizeof(std::unique_ptr<Cluster>);
933
52.9k
}
934
935
uint64_t GenericClusterImpl::GetTotalTxSize() const noexcept
936
26.8k
{
937
26.8k
    uint64_t ret{0};
938
123k
    for (auto i : m_linearization) {
939
123k
        ret += m_depgraph.FeeRate(i).size;
940
123k
    }
941
26.8k
    return ret;
942
26.8k
}
943
944
DepGraphIndex GenericClusterImpl::AppendTransaction(GraphIndex graph_idx, FeePerWeight feerate) noexcept
945
98
{
946
98
    Assume(graph_idx != GraphIndex(-1));
Line
Count
Source
128
98
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
947
98
    auto ret = m_depgraph.AddTransaction(feerate);
948
98
    m_mapping.push_back(graph_idx);
949
98
    m_linearization.push_back(ret);
950
98
    return ret;
951
98
}
952
953
DepGraphIndex SingletonClusterImpl::AppendTransaction(GraphIndex graph_idx, FeePerWeight feerate) noexcept
954
12.7k
{
955
12.7k
    Assume(!GetTxCount());
Line
Count
Source
128
12.7k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
956
12.7k
    m_graph_index = graph_idx;
957
12.7k
    m_feerate = feerate;
958
12.7k
    return 0;
959
12.7k
}
960
961
void GenericClusterImpl::AddDependencies(SetType parents, DepGraphIndex child) noexcept
962
98
{
963
98
    m_depgraph.AddDependencies(parents, child);
964
98
}
965
966
void SingletonClusterImpl::AddDependencies(SetType parents, DepGraphIndex child) noexcept
967
59
{
968
    // Singletons cannot have any dependencies.
969
59
    Assume(child == 0);
Line
Count
Source
128
59
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
970
59
    Assume(parents == SetType{} || parents == SetType::Fill(0));
Line
Count
Source
128
59
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
971
59
}
972
973
void GenericClusterImpl::ExtractTransactions(const std::function<void (DepGraphIndex, GraphIndex, FeePerWeight)>& visit1_fn, const std::function<void (DepGraphIndex, GraphIndex, SetType)>& visit2_fn) noexcept
974
0
{
975
0
    for (auto pos : m_linearization) {
976
0
        visit1_fn(pos, m_mapping[pos], FeePerWeight::FromFeeFrac(m_depgraph.FeeRate(pos)));
977
0
    }
978
0
    for (auto pos : m_linearization) {
979
0
        visit2_fn(pos, m_mapping[pos], m_depgraph.GetReducedParents(pos));
980
0
    }
981
    // Purge this Cluster, now that everything has been moved.
982
0
    m_depgraph = DepGraph<SetType>{};
983
0
    m_linearization.clear();
984
0
    m_mapping.clear();
985
0
}
986
987
void SingletonClusterImpl::ExtractTransactions(const std::function<void (DepGraphIndex, GraphIndex, FeePerWeight)>& visit1_fn, const std::function<void (DepGraphIndex, GraphIndex, SetType)>& visit2_fn) noexcept
988
7.55k
{
989
7.55k
    if (GetTxCount()) {
990
7.55k
        visit1_fn(0, m_graph_index, m_feerate);
991
7.55k
        visit2_fn(0, m_graph_index, SetType{});
992
7.55k
        m_graph_index = NO_GRAPH_INDEX;
993
7.55k
    }
994
7.55k
}
995
996
int GenericClusterImpl::GetLevel(const TxGraphImpl& graph) const noexcept
997
23.2k
{
998
    // GetLevel() does not work for empty Clusters.
999
23.2k
    if (!Assume(!m_linearization.empty())) 
return -10
;
Line
Count
Source
128
23.2k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1000
1001
    // Pick an arbitrary Entry that occurs in this Cluster.
1002
23.2k
    const auto& entry = graph.m_entries[m_mapping[m_linearization.front()]];
1003
    // See if there is a level whose Locator matches this Cluster, if so return that level.
1004
23.2k
    for (int level = 0; level < MAX_LEVELS; 
++level0
) {
1005
23.2k
        if (entry.m_locator[level].cluster == this) return level;
1006
23.2k
    }
1007
    // Given that we started with an Entry that occurs in this Cluster, one of its Locators must
1008
    // point back to it.
1009
23.2k
    assert
(false)0
;
1010
0
    return -1;
1011
0
}
1012
1013
int SingletonClusterImpl::GetLevel(const TxGraphImpl& graph) const noexcept
1014
16.3k
{
1015
    // GetLevel() does not work for empty Clusters.
1016
16.3k
    if (!Assume(GetTxCount())) 
return -10
;
Line
Count
Source
128
16.3k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1017
1018
    // Get the Entry in this Cluster.
1019
16.3k
    const auto& entry = graph.m_entries[m_graph_index];
1020
    // See if there is a level whose Locator matches this Cluster, if so return that level.
1021
16.3k
    for (int level = 0; level < MAX_LEVELS; 
++level0
) {
1022
16.3k
        if (entry.m_locator[level].cluster == this) return level;
1023
16.3k
    }
1024
    // Given that we started with an Entry that occurs in this Cluster, one of its Locators must
1025
    // point back to it.
1026
16.3k
    assert
(false)0
;
1027
0
    return -1;
1028
0
}
1029
1030
void TxGraphImpl::ClearLocator(int level, GraphIndex idx, bool oversized_tx) noexcept
1031
5.45k
{
1032
5.45k
    auto& entry = m_entries[idx];
1033
5.45k
    auto& clusterset = GetClusterSet(level);
1034
5.45k
    Assume(entry.m_locator[level].IsPresent());
Line
Count
Source
128
5.45k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1035
    // Change the locator from Present to Missing or Removed.
1036
5.45k
    if (level == 0 || 
!entry.m_locator[level - 1].IsPresent()0
) {
1037
5.45k
        entry.m_locator[level].SetMissing();
1038
5.45k
    } else {
1039
0
        entry.m_locator[level].SetRemoved();
1040
0
        clusterset.m_removed.push_back(idx);
1041
0
    }
1042
    // Update the transaction count.
1043
5.45k
    --clusterset.m_txcount;
1044
5.45k
    clusterset.m_txcount_oversized -= oversized_tx;
1045
    // If clearing main, adjust the status of Locators of this transaction in staging, if it exists.
1046
5.45k
    if (level == 0 && GetTopLevel() == 1) {
1047
0
        if (entry.m_locator[1].IsRemoved()) {
1048
0
            entry.m_locator[1].SetMissing();
1049
0
        } else if (!entry.m_locator[1].IsPresent()) {
1050
0
            --m_staging_clusterset->m_txcount;
1051
0
            m_staging_clusterset->m_txcount_oversized -= oversized_tx;
1052
0
        }
1053
0
    }
1054
5.45k
    if (level == 0) ClearChunkData(entry);
1055
5.45k
}
1056
1057
void GenericClusterImpl::RemoveChunkData(TxGraphImpl& graph) noexcept
1058
7.49k
{
1059
40.9k
    for (DepGraphIndex idx : m_linearization) {
1060
40.9k
        auto& entry = graph.m_entries[m_mapping[idx]];
1061
40.9k
        graph.ClearChunkData(entry);
1062
40.9k
    }
1063
7.49k
}
1064
1065
void SingletonClusterImpl::RemoveChunkData(TxGraphImpl& graph) noexcept
1066
549
{
1067
549
    if (GetTxCount() == 0) 
return0
;
1068
549
    auto& entry = graph.m_entries[m_graph_index];
1069
549
    graph.ClearChunkData(entry);
1070
549
}
1071
1072
void GenericClusterImpl::Updated(TxGraphImpl& graph, int level, bool rename) noexcept
1073
12.8k
{
1074
    // Update all the Locators for this Cluster's Entry objects.
1075
46.1k
    for (DepGraphIndex idx : m_linearization) {
1076
46.1k
        auto& entry = graph.m_entries[m_mapping[idx]];
1077
        // Discard any potential ChunkData prior to modifying the Cluster (as that could
1078
        // invalidate its ordering).
1079
46.1k
        if (level == 0 && !rename) 
graph.ClearChunkData(entry)45.9k
;
1080
46.1k
        entry.m_locator[level].SetPresent(this, idx);
1081
46.1k
    }
1082
    // If this is for the main graph (level = 0), and the Cluster's quality is ACCEPTABLE or
1083
    // OPTIMAL, compute its chunking and store its information in the Entry's m_main_lin_index
1084
    // and m_main_chunk_feerate. These fields are only accessed after making the entire graph
1085
    // ACCEPTABLE, so it is pointless to compute these if we haven't reached that quality level
1086
    // yet.
1087
    // When rename=true, this is always performed for level 0, to make sure the values inside the
1088
    // entries remain consistent with the chunk index (otherwise unrelated chunk index operations
1089
    // could cause the index to become corrupted, by inserting elements in the wrong place).
1090
12.8k
    if (level == 0 && (rename || 
IsAcceptable()12.7k
)) {
1091
5.68k
        auto chunking = ChunkLinearizationInfo(m_depgraph, m_linearization);
1092
5.68k
        LinearizationIndex lin_idx{0};
1093
        /** The sum of all chunk feerate FeeFracs with the same feerate as the current chunk,
1094
         *  up to and including the current chunk. */
1095
5.68k
        FeeFrac equal_feerate_chunk_feerate;
1096
        // Iterate over the chunks.
1097
28.6k
        for (unsigned chunk_idx = 0; chunk_idx < chunking.size(); 
++chunk_idx22.9k
) {
1098
22.9k
            auto& chunk = chunking[chunk_idx];
1099
22.9k
            auto chunk_count = chunk.transactions.Count();
1100
22.9k
            Assume(chunk_count > 0);
Line
Count
Source
128
22.9k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1101
            // Update equal_feerate_chunk_feerate to include this chunk, starting over when the
1102
            // feerate changed.
1103
22.9k
            if (ByRatio{chunk.feerate} < ByRatio{equal_feerate_chunk_feerate}) {
1104
0
                equal_feerate_chunk_feerate = chunk.feerate;
1105
22.9k
            } else {
1106
                // Note that this is adding fees to fees, and sizes to sizes, so the overall
1107
                // ratio remains the same; it's just accounting for the size of the added chunk.
1108
22.9k
                equal_feerate_chunk_feerate += chunk.feerate;
1109
22.9k
            }
1110
            // Determine the m_fallback_order maximum transaction in the chunk.
1111
22.9k
            auto it = chunk.transactions.begin();
1112
22.9k
            GraphIndex max_element = m_mapping[*it];
1113
22.9k
            ++it;
1114
22.9k
            while (it != chunk.transactions.end()) {
1115
0
                GraphIndex this_element = m_mapping[*it];
1116
0
                if (graph.m_fallback_order(*graph.m_entries[this_element].m_ref, *graph.m_entries[max_element].m_ref) > 0) {
1117
0
                    max_element = this_element;
1118
0
                }
1119
0
                ++it;
1120
0
            }
1121
            // Iterate over the transactions in the linearization, which must match those in chunk.
1122
22.9k
            while (true) {
1123
22.9k
                DepGraphIndex idx = m_linearization[lin_idx];
1124
22.9k
                GraphIndex graph_idx = m_mapping[idx];
1125
22.9k
                auto& entry = graph.m_entries[graph_idx];
1126
22.9k
                entry.m_main_lin_index = lin_idx++;
1127
22.9k
                entry.m_main_chunk_feerate = FeePerWeight::FromFeeFrac(chunk.feerate);
1128
22.9k
                entry.m_main_equal_feerate_chunk_prefix_size = equal_feerate_chunk_feerate.size;
1129
22.9k
                entry.m_main_max_chunk_fallback = max_element;
1130
22.9k
                Assume(chunk.transactions[idx]);
Line
Count
Source
128
22.9k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1131
22.9k
                chunk.transactions.Reset(idx);
1132
22.9k
                if (chunk.transactions.None()) {
1133
                    // Last transaction in the chunk.
1134
22.9k
                    if (chunk_count == 1 && chunk_idx + 1 == chunking.size()) {
1135
                        // If this is the final chunk of the cluster, and it contains just a single
1136
                        // transaction (which will always be true for the very common singleton
1137
                        // clusters), store the special value -1 as chunk count.
1138
5.68k
                        chunk_count = LinearizationIndex(-1);
1139
5.68k
                    }
1140
22.9k
                    if (!rename) 
graph.CreateChunkData(graph_idx, chunk_count)22.8k
;
1141
22.9k
                    break;
1142
22.9k
                }
1143
22.9k
            }
1144
22.9k
        }
1145
5.68k
    }
1146
12.8k
}
1147
1148
void SingletonClusterImpl::Updated(TxGraphImpl& graph, int level, bool rename) noexcept
1149
20.9k
{
1150
    // Don't do anything if this is empty.
1151
20.9k
    if (GetTxCount() == 0) 
return0
;
1152
1153
20.9k
    auto& entry = graph.m_entries[m_graph_index];
1154
    // Discard any potential ChunkData prior to modifying the Cluster (as that could
1155
    // invalidate its ordering).
1156
20.9k
    if (level == 0 && 
!rename8.24k
)
graph.ClearChunkData(entry)8.10k
;
1157
20.9k
    entry.m_locator[level].SetPresent(this, 0);
1158
    // If this is for the main graph (level = 0), compute its chunking and store its information in
1159
    // the Entry's m_main_lin_index and m_main_chunk_feerate.
1160
20.9k
    if (level == 0 && 
(8.24k
rename8.24k
||
IsAcceptable()8.10k
)) {
1161
8.24k
        entry.m_main_lin_index = 0;
1162
8.24k
        entry.m_main_chunk_feerate = m_feerate;
1163
8.24k
        entry.m_main_equal_feerate_chunk_prefix_size = m_feerate.size;
1164
8.24k
        entry.m_main_max_chunk_fallback = m_graph_index;
1165
        // Always use the special LinearizationIndex(-1), indicating singleton chunk at end of
1166
        // Cluster, here.
1167
8.24k
        if (!rename) 
graph.CreateChunkData(m_graph_index, LinearizationIndex(-1))8.10k
;
1168
8.24k
    }
1169
20.9k
}
1170
1171
void GenericClusterImpl::GetConflicts(const TxGraphImpl& graph, std::vector<Cluster*>& out) const noexcept
1172
0
{
1173
0
    for (auto i : m_linearization) {
1174
0
        auto& entry = graph.m_entries[m_mapping[i]];
1175
        // For every transaction Entry in this Cluster, if it also exists in a lower-level Cluster,
1176
        // then that Cluster conflicts.
1177
0
        if (entry.m_locator[0].IsPresent()) {
1178
0
            out.push_back(entry.m_locator[0].cluster);
1179
0
        }
1180
0
    }
1181
0
}
1182
1183
void SingletonClusterImpl::GetConflicts(const TxGraphImpl& graph, std::vector<Cluster*>& out) const noexcept
1184
8.04k
{
1185
    // Empty clusters have no conflicts.
1186
8.04k
    if (GetTxCount() == 0) 
return0
;
1187
1188
8.04k
    auto& entry = graph.m_entries[m_graph_index];
1189
    // If the transaction in this Cluster also exists in a lower-level Cluster, then that Cluster
1190
    // conflicts.
1191
8.04k
    if (entry.m_locator[0].IsPresent()) {
1192
0
        out.push_back(entry.m_locator[0].cluster);
1193
0
    }
1194
8.04k
}
1195
1196
std::vector<Cluster*> TxGraphImpl::GetConflicts() const noexcept
1197
8.04k
{
1198
8.04k
    Assume(GetTopLevel() == 1);
Line
Count
Source
128
8.04k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1199
8.04k
    auto& clusterset = GetClusterSet(1);
1200
8.04k
    std::vector<Cluster*> ret;
1201
    // All main Clusters containing transactions in m_removed (so (P,R) ones) are conflicts.
1202
8.04k
    for (auto i : clusterset.m_removed) {
1203
0
        auto& entry = m_entries[i];
1204
0
        if (entry.m_locator[0].IsPresent()) {
1205
0
            ret.push_back(entry.m_locator[0].cluster);
1206
0
        }
1207
0
    }
1208
    // Then go over all Clusters at this level, and find their conflicts (the (P,P) ones).
1209
64.3k
    for (int quality = 0; quality < int(QualityLevel::NONE); 
++quality56.3k
) {
1210
56.3k
        auto& clusters = clusterset.m_clusters[quality];
1211
56.3k
        for (const auto& cluster : clusters) {
1212
8.04k
            cluster->GetConflicts(*this, ret);
1213
8.04k
        }
1214
56.3k
    }
1215
    // Deduplicate the result (the same Cluster may appear multiple times).
1216
8.04k
    std::ranges::sort(ret, [](Cluster* a, Cluster* b) noexcept 
{ return CompareClusters(a, b) < 0; }0
);
1217
8.04k
    ret.erase(std::ranges::unique(ret).begin(), ret.end());
1218
8.04k
    return ret;
1219
8.04k
}
1220
1221
Cluster* GenericClusterImpl::CopyToStaging(TxGraphImpl& graph) const noexcept
1222
0
{
1223
    // Construct an empty Cluster.
1224
0
    auto ret = graph.CreateEmptyGenericCluster();
1225
0
    auto ptr = ret.get();
1226
    // Copy depgraph, mapping, and linearization.
1227
0
    ptr->m_depgraph = m_depgraph;
1228
0
    ptr->m_mapping = m_mapping;
1229
0
    ptr->m_linearization = m_linearization;
1230
    // Insert the new Cluster into the graph.
1231
0
    graph.InsertCluster(/*level=*/1, std::move(ret), m_quality);
1232
    // Update its Locators.
1233
0
    ptr->Updated(graph, /*level=*/1, /*rename=*/false);
1234
    // Update memory usage.
1235
0
    graph.GetClusterSet(/*level=*/1).m_cluster_usage += ptr->TotalMemoryUsage();
1236
0
    return ptr;
1237
0
}
1238
1239
Cluster* SingletonClusterImpl::CopyToStaging(TxGraphImpl& graph) const noexcept
1240
0
{
1241
    // Construct an empty Cluster.
1242
0
    auto ret = graph.CreateEmptySingletonCluster();
1243
0
    auto ptr = ret.get();
1244
    // Copy data.
1245
0
    ptr->m_graph_index = m_graph_index;
1246
0
    ptr->m_feerate = m_feerate;
1247
    // Insert the new Cluster into the graph.
1248
0
    graph.InsertCluster(/*level=*/1, std::move(ret), m_quality);
1249
    // Update its Locators.
1250
0
    ptr->Updated(graph, /*level=*/1, /*rename=*/false);
1251
    // Update memory usage.
1252
0
    graph.GetClusterSet(/*level=*/1).m_cluster_usage += ptr->TotalMemoryUsage();
1253
0
    return ptr;
1254
0
}
1255
1256
void GenericClusterImpl::ApplyRemovals(TxGraphImpl& graph, int level, std::span<GraphIndex>& to_remove) noexcept
1257
1.54k
{
1258
    // Iterate over the prefix of to_remove that applies to this cluster.
1259
1.54k
    Assume(!to_remove.empty());
Line
Count
Source
128
1.54k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1260
1.54k
    SetType todo;
1261
1.54k
    graph.GetClusterSet(level).m_cluster_usage -= TotalMemoryUsage();
1262
5.92k
    do {
1263
5.92k
        GraphIndex idx = to_remove.front();
1264
5.92k
        Assume(idx < graph.m_entries.size());
Line
Count
Source
128
5.92k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1265
5.92k
        auto& entry = graph.m_entries[idx];
1266
5.92k
        auto& locator = entry.m_locator[level];
1267
        // Stop once we hit an entry that applies to another Cluster.
1268
5.92k
        if (locator.cluster != this) 
break756
;
1269
        // - Remember it in a set of to-remove DepGraphIndexes.
1270
5.16k
        todo.Set(locator.index);
1271
        // - Remove from m_mapping. This isn't strictly necessary as unused positions in m_mapping
1272
        //   are just never accessed, but set it to -1 here to increase the ability to detect a bug
1273
        //   that causes it to be accessed regardless.
1274
5.16k
        m_mapping[locator.index] = GraphIndex(-1);
1275
        // - Remove its linearization index from the Entry (if in main).
1276
5.16k
        if (level == 0) {
1277
5.16k
            entry.m_main_lin_index = LinearizationIndex(-1);
1278
5.16k
        }
1279
        // - Mark it as missing/removed in the Entry's locator.
1280
5.16k
        graph.ClearLocator(level, idx, m_quality == QualityLevel::OVERSIZED_SINGLETON);
1281
5.16k
        to_remove = to_remove.subspan(1);
1282
5.16k
    } while(!to_remove.empty());
1283
1284
1.54k
    Assume(todo.Any());
Line
Count
Source
128
1.54k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1285
    // Wipe from the Cluster's DepGraph (this is O(n) regardless of the number of entries
1286
    // removed, so we benefit from batching all the removals).
1287
1.54k
    m_depgraph.RemoveTransactions(todo);
1288
1.54k
    m_mapping.resize(m_depgraph.PositionRange());
1289
1290
    // Filter removed transactions out of m_linearization.
1291
1.54k
    m_linearization.erase(std::remove_if(m_linearization.begin(), m_linearization.end(),
1292
5.48k
                                         [&](auto pos) { return todo[pos]; }),
1293
1.54k
                          m_linearization.end());
1294
1295
1.54k
    Compact();
1296
1.54k
    graph.GetClusterSet(level).m_cluster_usage += TotalMemoryUsage();
1297
1.54k
    auto new_quality = IsTopological() ? QualityLevel::NEEDS_SPLIT : 
QualityLevel::NEEDS_SPLIT_FIX0
;
1298
1.54k
    graph.SetClusterQuality(level, m_quality, m_setindex, new_quality);
1299
1.54k
    Updated(graph, /*level=*/level, /*rename=*/false);
1300
1.54k
}
1301
1302
void SingletonClusterImpl::ApplyRemovals(TxGraphImpl& graph, int level, std::span<GraphIndex>& to_remove) noexcept
1303
295
{
1304
    // We can only remove the one transaction this Cluster has.
1305
295
    Assume(!to_remove.empty());
Line
Count
Source
128
295
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1306
295
    Assume(GetTxCount());
Line
Count
Source
128
295
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1307
295
    Assume(to_remove.front() == m_graph_index);
Line
Count
Source
128
295
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1308
    // Pop all copies of m_graph_index from the front of to_remove (at least one, but there may be
1309
    // multiple).
1310
295
    do {
1311
295
        to_remove = to_remove.subspan(1);
1312
295
    } while (!to_remove.empty() && 
to_remove.front() == m_graph_index247
);
1313
    // Clear this cluster.
1314
295
    graph.ClearLocator(level, m_graph_index, m_quality == QualityLevel::OVERSIZED_SINGLETON);
1315
295
    m_graph_index = NO_GRAPH_INDEX;
1316
295
    graph.SetClusterQuality(level, m_quality, m_setindex, QualityLevel::NEEDS_SPLIT);
1317
    // No need to account for m_cluster_usage changes here, as SingletonClusterImpl has constant
1318
    // memory usage.
1319
295
}
1320
1321
void GenericClusterImpl::Clear(TxGraphImpl& graph, int level) noexcept
1322
0
{
1323
0
    Assume(GetTxCount());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1324
0
    graph.GetClusterSet(level).m_cluster_usage -= TotalMemoryUsage();
1325
0
    for (auto i : m_linearization) {
1326
0
        graph.ClearLocator(level, m_mapping[i], m_quality == QualityLevel::OVERSIZED_SINGLETON);
1327
0
    }
1328
0
    m_depgraph = {};
1329
0
    m_linearization.clear();
1330
0
    m_mapping.clear();
1331
0
}
1332
1333
void SingletonClusterImpl::Clear(TxGraphImpl& graph, int level) noexcept
1334
0
{
1335
0
    Assume(GetTxCount());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1336
0
    graph.GetClusterSet(level).m_cluster_usage -= TotalMemoryUsage();
1337
0
    graph.ClearLocator(level, m_graph_index, m_quality == QualityLevel::OVERSIZED_SINGLETON);
1338
0
    m_graph_index = NO_GRAPH_INDEX;
1339
0
}
1340
1341
void GenericClusterImpl::MoveToMain(TxGraphImpl& graph) noexcept
1342
0
{
1343
0
    for (auto i : m_linearization) {
1344
0
        GraphIndex idx = m_mapping[i];
1345
0
        auto& entry = graph.m_entries[idx];
1346
0
        entry.m_locator[1].SetMissing();
1347
0
    }
1348
0
    auto quality = m_quality;
1349
    // Subtract memory usage from staging and add it to main.
1350
0
    graph.GetClusterSet(/*level=*/1).m_cluster_usage -= TotalMemoryUsage();
1351
0
    graph.GetClusterSet(/*level=*/0).m_cluster_usage += TotalMemoryUsage();
1352
    // Remove cluster itself from staging and add it to main.
1353
0
    auto cluster = graph.ExtractCluster(1, quality, m_setindex);
1354
0
    graph.InsertCluster(/*level=*/0, std::move(cluster), quality);
1355
0
    Updated(graph, /*level=*/0, /*rename=*/false);
1356
0
}
1357
1358
void SingletonClusterImpl::MoveToMain(TxGraphImpl& graph) noexcept
1359
8.04k
{
1360
8.04k
    if (GetTxCount()) {
1361
8.04k
        auto& entry = graph.m_entries[m_graph_index];
1362
8.04k
        entry.m_locator[1].SetMissing();
1363
8.04k
    }
1364
8.04k
    auto quality = m_quality;
1365
8.04k
    graph.GetClusterSet(/*level=*/1).m_cluster_usage -= TotalMemoryUsage();
1366
8.04k
    auto cluster = graph.ExtractCluster(/*level=*/1, quality, m_setindex);
1367
8.04k
    graph.InsertCluster(/*level=*/0, std::move(cluster), quality);
1368
8.04k
    graph.GetClusterSet(/*level=*/0).m_cluster_usage += TotalMemoryUsage();
1369
8.04k
    Updated(graph, /*level=*/0, /*rename=*/false);
1370
8.04k
}
1371
1372
void GenericClusterImpl::Compact() noexcept
1373
7.10k
{
1374
7.10k
    m_linearization.shrink_to_fit();
1375
7.10k
    m_mapping.shrink_to_fit();
1376
7.10k
    m_depgraph.Compact();
1377
7.10k
}
1378
1379
void SingletonClusterImpl::Compact() noexcept
1380
59
{
1381
    // Nothing to compact; SingletonClusterImpl is constant size.
1382
59
}
1383
1384
void GenericClusterImpl::AppendChunkFeerates(std::vector<FeeFrac>& ret) const noexcept
1385
0
{
1386
0
    auto chunk_feerates = ChunkLinearization(m_depgraph, m_linearization);
1387
0
    ret.reserve(ret.size() + chunk_feerates.size());
1388
0
    ret.insert(ret.end(), chunk_feerates.begin(), chunk_feerates.end());
1389
0
}
1390
1391
void SingletonClusterImpl::AppendChunkFeerates(std::vector<FeeFrac>& ret) const noexcept
1392
0
{
1393
0
    if (GetTxCount()) {
1394
0
        ret.push_back(m_feerate);
1395
0
    }
1396
0
}
1397
1398
uint64_t GenericClusterImpl::AppendTrimData(std::vector<TrimTxData>& ret, std::vector<std::pair<GraphIndex, GraphIndex>>& deps) const noexcept
1399
0
{
1400
0
    Assume(IsAcceptable());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1401
0
    auto linchunking = ChunkLinearizationInfo(m_depgraph, m_linearization);
1402
0
    LinearizationIndex pos{0};
1403
0
    uint64_t size{0};
1404
0
    auto prev_index = GraphIndex(-1);
1405
    // Iterate over the chunks of this cluster's linearization.
1406
0
    for (const auto& [chunk, chunk_feerate] : linchunking) {
1407
        // Iterate over the transactions of that chunk, in linearization order.
1408
0
        auto chunk_tx_count = chunk.Count();
1409
0
        for (unsigned j = 0; j < chunk_tx_count; ++j) {
1410
0
            auto cluster_idx = m_linearization[pos];
1411
            // The transaction must appear in the chunk.
1412
0
            Assume(chunk[cluster_idx]);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1413
            // Construct a new element in ret.
1414
0
            auto& entry = ret.emplace_back();
1415
0
            entry.m_chunk_feerate = FeePerWeight::FromFeeFrac(chunk_feerate);
1416
0
            entry.m_index = m_mapping[cluster_idx];
1417
            // If this is not the first transaction of the cluster linearization, it has an
1418
            // implicit dependency on its predecessor.
1419
0
            if (pos != 0) deps.emplace_back(prev_index, entry.m_index);
1420
0
            prev_index = entry.m_index;
1421
0
            entry.m_tx_size = m_depgraph.FeeRate(cluster_idx).size;
1422
0
            size += entry.m_tx_size;
1423
0
            ++pos;
1424
0
        }
1425
0
    }
1426
0
    return size;
1427
0
}
1428
1429
uint64_t SingletonClusterImpl::AppendTrimData(std::vector<TrimTxData>& ret, std::vector<std::pair<GraphIndex, GraphIndex>>& deps) const noexcept
1430
0
{
1431
0
    if (!GetTxCount()) return 0;
1432
0
    auto& entry = ret.emplace_back();
1433
0
    entry.m_chunk_feerate = m_feerate;
1434
0
    entry.m_index = m_graph_index;
1435
0
    entry.m_tx_size = m_feerate.size;
1436
0
    return m_feerate.size;
1437
0
}
1438
1439
bool GenericClusterImpl::Split(TxGraphImpl& graph, int level) noexcept
1440
1.54k
{
1441
    // This function can only be called when the Cluster needs splitting.
1442
1.54k
    Assume(NeedsSplitting());
Line
Count
Source
128
1.54k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1443
    // Determine the new quality the split-off Clusters will have.
1444
1.54k
    QualityLevel new_quality = IsTopological() ? QualityLevel::NEEDS_RELINEARIZE : 
QualityLevel::NEEDS_FIX0
;
1445
    /** Which positions are still left in this Cluster. */
1446
1.54k
    auto todo = m_depgraph.Positions();
1447
    /** Mapping from transaction positions in this Cluster to the Cluster where it ends up, and
1448
     *  its position therein. */
1449
1.54k
    std::vector<std::pair<Cluster*, DepGraphIndex>> remap(m_depgraph.PositionRange());
1450
1.54k
    std::vector<Cluster*> new_clusters;
1451
1.54k
    bool first{true};
1452
    // Iterate over the connected components of this Cluster's m_depgraph.
1453
1.62k
    while (todo.Any()) {
1454
128
        auto component = m_depgraph.FindConnectedComponent(todo);
1455
128
        auto component_size = component.Count();
1456
128
        auto split_quality = component_size <= 1 ? 
QualityLevel::OPTIMAL59
:
new_quality69
;
1457
128
        if (first && component == todo && SetType::Fill(component_size) == component && 
component_size >= MIN_INTENDED_TX_COUNT66
) {
1458
            // The existing Cluster is an entire component, without holes. Leave it be, but update
1459
            // its quality. If there are holes, we continue, so that the Cluster is reconstructed
1460
            // without holes, reducing memory usage. If the component's size is below the intended
1461
            // transaction count for this Cluster implementation, continue so that it can get
1462
            // converted.
1463
43
            Assume(todo == m_depgraph.Positions());
Line
Count
Source
128
43
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1464
43
            graph.SetClusterQuality(level, m_quality, m_setindex, split_quality);
1465
            // If this made the quality ACCEPTABLE or OPTIMAL, we need to compute and cache its
1466
            // chunking.
1467
43
            Updated(graph, /*level=*/level, /*rename=*/false);
1468
43
            return false;
1469
43
        }
1470
85
        first = false;
1471
        // Construct a new Cluster to hold the found component.
1472
85
        auto new_cluster = graph.CreateEmptyCluster(component_size);
1473
85
        new_clusters.push_back(new_cluster.get());
1474
        // Remember that all the component's transactions go to this new Cluster. The positions
1475
        // will be determined below, so use -1 for now.
1476
157
        for (auto i : component) {
1477
157
            remap[i] = {new_cluster.get(), DepGraphIndex(-1)};
1478
157
        }
1479
85
        graph.InsertCluster(level, std::move(new_cluster), split_quality);
1480
85
        todo -= component;
1481
85
    }
1482
    // We have to split the Cluster up. Remove accounting for the existing one first.
1483
1.49k
    graph.GetClusterSet(level).m_cluster_usage -= TotalMemoryUsage();
1484
    // Redistribute the transactions.
1485
1.49k
    for (auto i : m_linearization) {
1486
        /** The cluster which transaction originally in position i is moved to. */
1487
157
        Cluster* new_cluster = remap[i].first;
1488
        // Copy the transaction to the new cluster's depgraph, and remember the position.
1489
157
        remap[i].second = new_cluster->AppendTransaction(m_mapping[i], FeePerWeight::FromFeeFrac(m_depgraph.FeeRate(i)));
1490
157
    }
1491
    // Redistribute the dependencies.
1492
1.49k
    for (auto i : m_linearization) {
1493
        /** The cluster transaction in position i is moved to. */
1494
157
        Cluster* new_cluster = remap[i].first;
1495
        // Copy its parents, translating positions.
1496
157
        SetType new_parents;
1497
157
        for (auto par : m_depgraph.GetReducedParents(i)) 
new_parents.Set(remap[par].second)72
;
1498
157
        new_cluster->AddDependencies(new_parents, remap[i].second);
1499
157
    }
1500
    // Update all the Locators of moved transactions, and memory usage.
1501
1.49k
    for (Cluster* new_cluster : new_clusters) {
1502
85
        new_cluster->Updated(graph, /*level=*/level, /*rename=*/false);
1503
85
        new_cluster->Compact();
1504
85
        graph.GetClusterSet(level).m_cluster_usage += new_cluster->TotalMemoryUsage();
1505
85
    }
1506
    // Wipe this Cluster, and return that it needs to be deleted.
1507
1.49k
    m_depgraph = DepGraph<SetType>{};
1508
1.49k
    m_mapping.clear();
1509
1.49k
    m_linearization.clear();
1510
1.49k
    return true;
1511
1.54k
}
1512
1513
bool SingletonClusterImpl::Split(TxGraphImpl& graph, int level) noexcept
1514
295
{
1515
295
    Assume(NeedsSplitting());
Line
Count
Source
128
295
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1516
295
    Assume(!GetTxCount());
Line
Count
Source
128
295
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1517
295
    graph.GetClusterSet(level).m_cluster_usage -= TotalMemoryUsage();
1518
295
    return true;
1519
295
}
1520
1521
void GenericClusterImpl::Merge(TxGraphImpl& graph, int level, Cluster& other) noexcept
1522
7.55k
{
1523
    /** Vector to store the positions in this Cluster for each position in other. */
1524
7.55k
    std::vector<DepGraphIndex> remap(other.GetDepGraphIndexRange());
1525
    // Iterate over all transactions in the other Cluster (the one being absorbed).
1526
7.55k
    other.ExtractTransactions([&](DepGraphIndex pos, GraphIndex idx, FeePerWeight feerate) noexcept {
1527
        // Copy the transaction into this Cluster, and remember its position.
1528
7.55k
        auto new_pos = m_depgraph.AddTransaction(feerate);
1529
        // Since this cluster must have been made hole-free before being merged into, all added
1530
        // transactions should appear at the end.
1531
7.55k
        Assume(new_pos == m_mapping.size());
Line
Count
Source
128
7.55k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1532
7.55k
        remap[pos] = new_pos;
1533
7.55k
        m_mapping.push_back(idx);
1534
7.55k
        m_linearization.push_back(new_pos);
1535
7.55k
    }, [&](DepGraphIndex pos, GraphIndex idx, SetType other_parents) noexcept {
1536
        // Copy the transaction's dependencies, translating them using remap.
1537
7.55k
        SetType parents;
1538
7.55k
        for (auto par : other_parents) {
1539
0
            parents.Set(remap[par]);
1540
0
        }
1541
7.55k
        m_depgraph.AddDependencies(parents, remap[pos]);
1542
        // Update the transaction's Locator. There is no need to call Updated() to update chunk
1543
        // feerates, as Updated() will be invoked by Cluster::ApplyDependencies on the resulting
1544
        // merged Cluster later anyway.
1545
7.55k
        auto& entry = graph.m_entries[idx];
1546
        // Discard any potential ChunkData prior to modifying the Cluster (as that could
1547
        // invalidate its ordering).
1548
7.55k
        if (level == 0) graph.ClearChunkData(entry);
1549
7.55k
        entry.m_locator[level].SetPresent(this, remap[pos]);
1550
7.55k
    });
1551
7.55k
}
1552
1553
void SingletonClusterImpl::Merge(TxGraphImpl&, int, Cluster&) noexcept
1554
0
{
1555
    // Nothing can be merged into a singleton; it should have been converted to GenericClusterImpl first.
1556
0
    Assume(false);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1557
0
}
1558
1559
void GenericClusterImpl::ApplyDependencies(TxGraphImpl& graph, int level, std::span<std::pair<GraphIndex, GraphIndex>> to_apply) noexcept
1560
5.53k
{
1561
    // Sort the list of dependencies to apply by child, so those can be applied in batch.
1562
5.53k
    std::ranges::sort(to_apply, [](auto& a, auto& b) 
{ return a.second < b.second; }0
);
1563
    // Iterate over groups of to-be-added dependencies with the same child.
1564
5.53k
    auto it = to_apply.begin();
1565
11.0k
    while (it != to_apply.end()) {
1566
5.53k
        auto& first_child = graph.m_entries[it->second].m_locator[level];
1567
5.53k
        const auto child_idx = first_child.index;
1568
        // Iterate over all to-be-added dependencies within that same child, gather the relevant
1569
        // parents.
1570
5.53k
        SetType parents;
1571
11.0k
        while (it != to_apply.end()) {
1572
5.53k
            auto& child = graph.m_entries[it->second].m_locator[level];
1573
5.53k
            auto& parent = graph.m_entries[it->first].m_locator[level];
1574
5.53k
            Assume(child.cluster == this && parent.cluster == this);
Line
Count
Source
128
11.0k
#define Assume(val) 
inline_assertion_check<false>(5.53k
val, std::source_location::current(),
#5.53k
val)
1575
5.53k
            if (child.index != child_idx) 
break0
;
1576
5.53k
            parents.Set(parent.index);
1577
5.53k
            ++it;
1578
5.53k
        }
1579
        // Push all dependencies to the underlying DepGraph. Note that this is O(N) in the size of
1580
        // the cluster, regardless of the number of parents being added, so batching them together
1581
        // has a performance benefit.
1582
5.53k
        m_depgraph.AddDependencies(parents, child_idx);
1583
5.53k
    }
1584
1585
    // Finally set the cluster to NEEDS_FIX, so its linearization is fixed the next time it is
1586
    // attempted to be made ACCEPTABLE.
1587
5.53k
    Assume(!NeedsSplitting());
Line
Count
Source
128
5.53k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1588
5.53k
    Assume(!IsOversized());
Line
Count
Source
128
5.53k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1589
5.53k
    graph.SetClusterQuality(level, m_quality, m_setindex, QualityLevel::NEEDS_FIX);
1590
1591
    // Finally push the changes to graph.m_entries.
1592
5.53k
    Updated(graph, /*level=*/level, /*rename=*/false);
1593
5.53k
}
1594
1595
void SingletonClusterImpl::ApplyDependencies(TxGraphImpl&, int, std::span<std::pair<GraphIndex, GraphIndex>>) noexcept
1596
0
{
1597
    // Nothing can actually be applied.
1598
0
    Assume(false);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1599
0
}
1600
1601
TxGraphImpl::~TxGraphImpl() noexcept
1602
926
{
1603
    // If Refs outlive the TxGraphImpl they refer to, unlink them, so that their destructor does not
1604
    // try to reach into a non-existing TxGraphImpl anymore.
1605
2.58k
    for (auto& entry : m_entries) {
1606
2.58k
        if (entry.m_ref != nullptr) {
1607
0
            GetRefGraph(*entry.m_ref) = nullptr;
1608
0
        }
1609
2.58k
    }
1610
926
}
1611
1612
std::unique_ptr<Cluster> TxGraphImpl::ExtractCluster(int level, QualityLevel quality, ClusterSetIndex setindex) noexcept
1613
30.4k
{
1614
30.4k
    Assume(quality != QualityLevel::NONE);
Line
Count
Source
128
30.4k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1615
1616
30.4k
    auto& clusterset = GetClusterSet(level);
1617
30.4k
    auto& quality_clusters = clusterset.m_clusters[int(quality)];
1618
30.4k
    Assume(setindex < quality_clusters.size());
Line
Count
Source
128
30.4k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1619
1620
    // Extract the Cluster-owning unique_ptr.
1621
30.4k
    std::unique_ptr<Cluster> ret = std::move(quality_clusters[setindex]);
1622
30.4k
    ret->m_quality = QualityLevel::NONE;
1623
30.4k
    ret->m_setindex = ClusterSetIndex(-1);
1624
1625
    // Clean up space in quality_cluster.
1626
30.4k
    auto max_setindex = quality_clusters.size() - 1;
1627
30.4k
    if (setindex != max_setindex) {
1628
        // If the cluster was not the last element of quality_clusters, move that to take its place.
1629
4.86k
        quality_clusters.back()->m_setindex = setindex;
1630
4.86k
        quality_clusters[setindex] = std::move(quality_clusters.back());
1631
4.86k
    }
1632
    // The last element of quality_clusters is now unused; drop it.
1633
30.4k
    quality_clusters.pop_back();
1634
1635
30.4k
    return ret;
1636
30.4k
}
1637
1638
ClusterSetIndex TxGraphImpl::InsertCluster(int level, std::unique_ptr<Cluster>&& cluster, QualityLevel quality) noexcept
1639
35.8k
{
1640
    // Cannot insert with quality level NONE (as that would mean not inserted).
1641
35.8k
    Assume(quality != QualityLevel::NONE);
Line
Count
Source
128
35.8k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1642
    // The passed-in Cluster must not currently be in the TxGraphImpl.
1643
35.8k
    Assume(cluster->m_quality == QualityLevel::NONE);
Line
Count
Source
128
35.8k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1644
1645
    // Append it at the end of the relevant TxGraphImpl::m_cluster.
1646
35.8k
    auto& clusterset = GetClusterSet(level);
1647
35.8k
    auto& quality_clusters = clusterset.m_clusters[int(quality)];
1648
35.8k
    ClusterSetIndex ret = quality_clusters.size();
1649
35.8k
    cluster->m_quality = quality;
1650
35.8k
    cluster->m_setindex = ret;
1651
35.8k
    quality_clusters.push_back(std::move(cluster));
1652
35.8k
    return ret;
1653
35.8k
}
1654
1655
void TxGraphImpl::SetClusterQuality(int level, QualityLevel old_quality, ClusterSetIndex old_index, QualityLevel new_quality) noexcept
1656
13.0k
{
1657
13.0k
    Assume(new_quality != QualityLevel::NONE);
Line
Count
Source
128
13.0k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1658
1659
    // Don't do anything if the quality did not change.
1660
13.0k
    if (old_quality == new_quality) 
return0
;
1661
    // Extract the cluster from where it currently resides.
1662
13.0k
    auto cluster_ptr = ExtractCluster(level, old_quality, old_index);
1663
    // And re-insert it where it belongs.
1664
13.0k
    InsertCluster(level, std::move(cluster_ptr), new_quality);
1665
13.0k
}
1666
1667
void TxGraphImpl::DeleteCluster(Cluster& cluster, int level) noexcept
1668
9.35k
{
1669
    // Extract the cluster from where it currently resides.
1670
9.35k
    auto cluster_ptr = ExtractCluster(level, cluster.m_quality, cluster.m_setindex);
1671
    // And throw it away.
1672
9.35k
    cluster_ptr.reset();
1673
9.35k
}
1674
1675
std::pair<Cluster*, int> TxGraphImpl::FindClusterAndLevel(GraphIndex idx, int level) const noexcept
1676
55.6k
{
1677
55.6k
    Assume(level >= 0 && level <= GetTopLevel());
Line
Count
Source
128
111k
#define Assume(val) 
inline_assertion_check<false>(55.6k
val, std::source_location::current(),
#55.6k
val)
1678
55.6k
    auto& entry = m_entries[idx];
1679
    // Search the entry's locators from top to bottom.
1680
72.2k
    for (int l = level; l >= 0; 
--l16.6k
) {
1681
        // If the locator is missing, dig deeper; it may exist at a lower level and therefore be
1682
        // implicitly existing at this level too.
1683
72.2k
        if (entry.m_locator[l].IsMissing()) 
continue16.6k
;
1684
        // If the locator has the entry marked as explicitly removed, stop.
1685
55.6k
        if (entry.m_locator[l].IsRemoved()) 
break0
;
1686
        // Otherwise, we have found the topmost ClusterSet that contains this entry.
1687
55.6k
        return {entry.m_locator[l].cluster, l};
1688
55.6k
    }
1689
    // If no non-empty locator was found, or an explicitly removed was hit, return nothing.
1690
0
    return {nullptr, -1};
1691
55.6k
}
1692
1693
Cluster* TxGraphImpl::PullIn(Cluster* cluster, int level) noexcept
1694
0
{
1695
0
    int to_level = GetTopLevel();
1696
0
    Assume(to_level == 1);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1697
0
    Assume(level <= to_level);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1698
    // Copy the Cluster from main to staging, if it's not already there.
1699
0
    if (level == 0) {
1700
        // Make the Cluster Acceptable before copying. This isn't strictly necessary, but doing it
1701
        // now avoids doing double work later.
1702
0
        MakeAcceptable(*cluster, level);
1703
0
        cluster = cluster->CopyToStaging(*this);
1704
0
    }
1705
0
    return cluster;
1706
0
}
1707
1708
void TxGraphImpl::ApplyRemovals(int up_to_level) noexcept
1709
628k
{
1710
628k
    Assume(up_to_level >= 0 && up_to_level <= GetTopLevel());
Line
Count
Source
128
1.25M
#define Assume(val) 
inline_assertion_check<false>(628k
val, std::source_location::current(),
#628k
val)
1711
1.27M
    for (int level = 0; level <= up_to_level; 
++level647k
) {
1712
647k
        auto& clusterset = GetClusterSet(level);
1713
647k
        auto& to_remove = clusterset.m_to_remove;
1714
        // Skip if there is nothing to remove in this level.
1715
647k
        if (to_remove.empty()) 
continue646k
;
1716
        // Pull in all Clusters that are not in staging.
1717
834
        if (level == 1) {
1718
0
            for (GraphIndex index : to_remove) {
1719
0
                auto [cluster, cluster_level] = FindClusterAndLevel(index, level);
1720
0
                if (cluster != nullptr) PullIn(cluster, cluster_level);
1721
0
            }
1722
0
        }
1723
        // Group the set of to-be-removed entries by Cluster::m_sequence.
1724
18.7k
        std::ranges::sort(to_remove, [&](GraphIndex a, GraphIndex b) noexcept {
1725
18.7k
            Cluster* cluster_a = m_entries[a].m_locator[level].cluster;
1726
18.7k
            Cluster* cluster_b = m_entries[b].m_locator[level].cluster;
1727
18.7k
            return CompareClusters(cluster_a, cluster_b) < 0;
1728
18.7k
        });
1729
        // Process per Cluster.
1730
834
        std::span to_remove_span{to_remove};
1731
2.67k
        while (!to_remove_span.empty()) {
1732
1.83k
            Cluster* cluster = m_entries[to_remove_span.front()].m_locator[level].cluster;
1733
1.83k
            if (cluster != nullptr) {
1734
                // If the first to_remove_span entry's Cluster exists, hand to_remove_span to it, so it
1735
                // can pop off whatever applies to it.
1736
1.83k
                cluster->ApplyRemovals(*this, level, to_remove_span);
1737
1.83k
            } else {
1738
                // Otherwise, skip this already-removed entry. This may happen when
1739
                // RemoveTransaction was called twice on the same Ref, for example.
1740
0
                to_remove_span = to_remove_span.subspan(1);
1741
0
            }
1742
1.83k
        }
1743
834
        to_remove.clear();
1744
834
    }
1745
628k
    Compact();
1746
628k
}
1747
1748
void TxGraphImpl::SwapIndexes(GraphIndex a, GraphIndex b, std::vector<Cluster*>& affected_main) noexcept
1749
807
{
1750
807
    Assume(a < m_entries.size());
Line
Count
Source
128
807
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1751
807
    Assume(b < m_entries.size());
Line
Count
Source
128
807
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1752
    // Swap the Entry objects.
1753
807
    std::swap(m_entries[a], m_entries[b]);
1754
    // Iterate over both objects.
1755
2.42k
    for (int i = 0; i < 2; 
++i1.61k
) {
1756
1.61k
        GraphIndex idx = i ? 
b807
:
a807
;
1757
1.61k
        Entry& entry = m_entries[idx];
1758
        // Update linked Ref, if any exists.
1759
1.61k
        if (entry.m_ref) 
GetRefIndex(*entry.m_ref) = idx807
;
1760
        // Update linked chunk index entries, if any exist.
1761
1.61k
        if (entry.m_main_chunkindex_iterator != m_main_chunkindex.end()) {
1762
651
            entry.m_main_chunkindex_iterator->m_graph_index = idx;
1763
651
        }
1764
        // Update the locators for both levels.
1765
4.84k
        for (int level = 0; level < MAX_LEVELS; 
++level3.22k
) {
1766
3.22k
            Locator& locator = entry.m_locator[level];
1767
3.22k
            if (locator.IsPresent()) {
1768
806
                locator.cluster->UpdateMapping(locator.index, idx);
1769
806
                if (level == 0) affected_main.push_back(locator.cluster);
1770
806
            }
1771
3.22k
        }
1772
1.61k
    }
1773
807
}
1774
1775
void TxGraphImpl::Compact() noexcept
1776
657k
{
1777
    // We cannot compact while any to-be-applied operations or staged removals remain as we'd need
1778
    // to rewrite them. It is easier to delay the compaction until they have been applied.
1779
657k
    if (!m_main_clusterset.m_deps_to_add.empty()) 
return5.53k
;
1780
651k
    if (!m_main_clusterset.m_to_remove.empty()) 
return0
;
1781
651k
    Assume(m_main_clusterset.m_removed.empty()); // non-staging m_removed is always empty
Line
Count
Source
128
651k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1782
651k
    if (m_staging_clusterset.has_value()) {
1783
24.6k
        if (!m_staging_clusterset->m_deps_to_add.empty()) 
return22.1k
;
1784
2.51k
        if (!m_staging_clusterset->m_to_remove.empty()) 
return0
;
1785
2.51k
        if (!m_staging_clusterset->m_removed.empty()) 
return0
;
1786
2.51k
    }
1787
1788
    // Release memory used by discarded ChunkData index entries.
1789
629k
    ClearShrink(m_main_chunkindex_discarded);
1790
1791
    // Sort the GraphIndexes that need to be cleaned up. They are sorted in reverse, so the last
1792
    // ones get processed first. This means earlier-processed GraphIndexes will not cause moving of
1793
    // later-processed ones during the "swap with end of m_entries" step below (which might
1794
    // invalidate them).
1795
629k
    std::ranges::sort(m_unlinked, std::greater{});
1796
1797
629k
    std::vector<Cluster*> affected_main;
1798
629k
    auto last = GraphIndex(-1);
1799
629k
    for (GraphIndex idx : m_unlinked) {
1800
        // m_unlinked should never contain the same GraphIndex twice (the code below would fail
1801
        // if so, because GraphIndexes get invalidated by removing them).
1802
10.0k
        Assume(idx != last);
Line
Count
Source
128
10.0k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1803
10.0k
        last = idx;
1804
1805
        // Make sure the entry is unlinked.
1806
10.0k
        Entry& entry = m_entries[idx];
1807
10.0k
        Assume(entry.m_ref == nullptr);
Line
Count
Source
128
10.0k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1808
        // Make sure the entry does not occur in the graph.
1809
30.2k
        for (int level = 0; level < MAX_LEVELS; 
++level20.1k
) {
1810
20.1k
            Assume(!entry.m_locator[level].IsPresent());
Line
Count
Source
128
20.1k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1811
20.1k
        }
1812
1813
        // Move the entry to the end.
1814
10.0k
        if (idx != m_entries.size() - 1) 
SwapIndexes(idx, m_entries.size() - 1, affected_main)807
;
1815
        // Drop the entry for idx, now that it is at the end.
1816
10.0k
        m_entries.pop_back();
1817
10.0k
    }
1818
1819
    // Update the affected clusters, to fixup Entry::m_main_max_chunk_fallback values which may
1820
    // have become outdated due to the compaction above.
1821
629k
    std::ranges::sort(affected_main);
1822
629k
    affected_main.erase(std::unique(affected_main.begin(), affected_main.end()), affected_main.end());
1823
629k
    for (Cluster* cluster : affected_main) {
1824
222
        cluster->Updated(*this, /*level=*/0, /*rename=*/true);
1825
222
    }
1826
629k
    m_unlinked.clear();
1827
629k
}
1828
1829
void TxGraphImpl::Split(Cluster& cluster, int level) noexcept
1830
1.83k
{
1831
    // To split a Cluster, first make sure all removals are applied (as we might need to split
1832
    // again afterwards otherwise).
1833
1.83k
    ApplyRemovals(level);
1834
1.83k
    bool del = cluster.Split(*this, level);
1835
1.83k
    if (del) {
1836
        // Cluster::Split reports whether the Cluster is to be deleted.
1837
1.79k
        DeleteCluster(cluster, level);
1838
1.79k
    }
1839
1.83k
}
1840
1841
void TxGraphImpl::SplitAll(int up_to_level) noexcept
1842
612k
{
1843
612k
    Assume(up_to_level >= 0 && up_to_level <= GetTopLevel());
Line
Count
Source
128
1.22M
#define Assume(val) 
inline_assertion_check<false>(612k
val, std::source_location::current(),
#612k
val)
1844
    // Before splitting all Cluster, first make sure all removals are applied.
1845
612k
    ApplyRemovals(up_to_level);
1846
1.23M
    for (int level = 0; level <= up_to_level; 
++level618k
) {
1847
1.23M
        for (auto quality : {QualityLevel::NEEDS_SPLIT_FIX, QualityLevel::NEEDS_SPLIT}) {
1848
1.23M
            auto& queue = GetClusterSet(level).m_clusters[int(quality)];
1849
1.23M
            while (!queue.empty()) {
1850
1.83k
                Split(*queue.back().get(), level);
1851
1.83k
            }
1852
1.23M
        }
1853
618k
    }
1854
612k
}
1855
1856
void TxGraphImpl::GroupClusters(int level) noexcept
1857
2.12M
{
1858
2.12M
    auto& clusterset = GetClusterSet(level);
1859
    // If the groupings have been computed already, nothing is left to be done.
1860
2.12M
    if (clusterset.m_group_data.has_value()) 
return2.12M
;
1861
1862
    // Before computing which Clusters need to be merged together, first apply all removals and
1863
    // split the Clusters into connected components. If we would group first, we might end up
1864
    // with inefficient and/or oversized Clusters which just end up being split again anyway.
1865
6.37k
    SplitAll(level);
1866
1867
    /** Annotated clusters: an entry for each Cluster, together with the sequence number for the
1868
     *  representative for the partition it is in (initially its own, later that of the
1869
     *  to-be-merged group). */
1870
6.37k
    std::vector<std::pair<Cluster*, uint64_t>> an_clusters;
1871
    /** Annotated dependencies: an entry for each m_deps_to_add entry (excluding ones that apply
1872
     *  to removed transactions), together with the sequence number of the representative root of
1873
     *  Clusters it applies to (initially that of the child Cluster, later that of the
1874
     *  to-be-merged group). */
1875
6.37k
    std::vector<std::pair<std::pair<GraphIndex, GraphIndex>, uint64_t>> an_deps;
1876
1877
    // Construct an an_clusters entry for every oversized cluster, including ones from levels below,
1878
    // as they may be inherited in this one.
1879
18.2k
    for (int level_iter = 0; level_iter <= level; 
++level_iter11.9k
) {
1880
11.9k
        for (auto& cluster : GetClusterSet(level_iter).m_clusters[int(QualityLevel::OVERSIZED_SINGLETON)]) {
1881
0
            auto graph_idx = cluster->GetClusterEntry(0);
1882
0
            auto cur_cluster = FindCluster(graph_idx, level);
1883
0
            if (cur_cluster == nullptr) continue;
1884
0
            an_clusters.emplace_back(cur_cluster, cur_cluster->m_sequence);
1885
0
        }
1886
11.9k
    }
1887
1888
    // Construct a an_clusters entry for every parent and child in the to-be-applied dependencies,
1889
    // and an an_deps entry for each dependency to be applied.
1890
6.37k
    an_deps.reserve(clusterset.m_deps_to_add.size());
1891
6.37k
    for (const auto& [par, chl] : clusterset.m_deps_to_add) {
1892
5.53k
        auto par_cluster = FindCluster(par, level);
1893
5.53k
        auto chl_cluster = FindCluster(chl, level);
1894
        // Skip dependencies for which the parent or child transaction is removed.
1895
5.53k
        if (par_cluster == nullptr || chl_cluster == nullptr) 
continue0
;
1896
5.53k
        an_clusters.emplace_back(par_cluster, par_cluster->m_sequence);
1897
        // Do not include a duplicate when parent and child are identical, as it'll be removed
1898
        // below anyway.
1899
5.53k
        if (chl_cluster != par_cluster) an_clusters.emplace_back(chl_cluster, chl_cluster->m_sequence);
1900
        // Add entry to an_deps, using the child sequence number.
1901
5.53k
        an_deps.emplace_back(std::pair{par, chl}, chl_cluster->m_sequence);
1902
5.53k
    }
1903
    // Sort and deduplicate an_clusters, so we end up with a sorted list of all involved Clusters
1904
    // to which dependencies apply, or which are oversized.
1905
11.0k
    std::ranges::sort(an_clusters, [](auto& a, auto& b) noexcept { return a.second < b.second; });
1906
6.37k
    an_clusters.erase(std::ranges::unique(an_clusters).begin(), an_clusters.end());
1907
    // Sort an_deps by applying the same order to the involved child cluster.
1908
6.37k
    std::ranges::sort(an_deps, [&](auto& a, auto& b) noexcept 
{ return a.second < b.second; }0
);
1909
1910
    // Run the union-find algorithm to find partitions of the input Clusters which need to be
1911
    // grouped together. See https://en.wikipedia.org/wiki/Disjoint-set_data_structure.
1912
6.37k
    {
1913
        /** Each PartitionData entry contains information about a single input Cluster. */
1914
6.37k
        struct PartitionData
1915
6.37k
        {
1916
            /** The sequence number of the cluster this holds information for. */
1917
6.37k
            uint64_t sequence;
1918
            /** All PartitionData entries belonging to the same partition are organized in a tree.
1919
             *  Each element points to its parent, or to itself if it is the root. The root is then
1920
             *  a representative for the entire tree, and can be found by walking upwards from any
1921
             *  element. */
1922
6.37k
            PartitionData* parent;
1923
            /** (only if this is a root, so when parent == this) An upper bound on the height of
1924
             *  tree for this partition. */
1925
6.37k
            unsigned rank;
1926
6.37k
        };
1927
        /** Information about each input Cluster. Sorted by Cluster::m_sequence. */
1928
6.37k
        std::vector<PartitionData> partition_data;
1929
1930
        /** Given a Cluster, find its corresponding PartitionData. */
1931
11.0k
        auto locate_fn = [&](uint64_t sequence) noexcept -> PartitionData* {
1932
11.0k
            auto it = std::lower_bound(partition_data.begin(), partition_data.end(), sequence,
1933
22.1k
                                       [](auto& a, uint64_t seq) noexcept { return a.sequence < seq; });
1934
11.0k
            Assume(it != partition_data.end());
Line
Count
Source
128
11.0k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1935
11.0k
            Assume(it->sequence == sequence);
Line
Count
Source
128
11.0k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1936
11.0k
            return &*it;
1937
11.0k
        };
1938
1939
        /** Given a PartitionData, find the root of the tree it is in (its representative). */
1940
22.1k
        static constexpr auto find_root_fn = [](PartitionData* data) noexcept -> PartitionData* {
1941
27.6k
            while (data->parent != data) {
1942
                // Replace pointers to parents with pointers to grandparents.
1943
                // See https://en.wikipedia.org/wiki/Disjoint-set_data_structure#Finding_set_representatives.
1944
5.53k
                auto par = data->parent;
1945
5.53k
                data->parent = par->parent;
1946
5.53k
                data = par;
1947
5.53k
            }
1948
22.1k
            return data;
1949
22.1k
        };
1950
1951
        /** Given two PartitionDatas, union the partitions they are in, and return their
1952
         *  representative. */
1953
6.37k
        static constexpr auto union_fn = [](PartitionData* arg1, PartitionData* arg2) noexcept {
1954
            // Find the roots of the trees, and bail out if they are already equal (which would
1955
            // mean they are in the same partition already).
1956
5.53k
            auto rep1 = find_root_fn(arg1);
1957
5.53k
            auto rep2 = find_root_fn(arg2);
1958
5.53k
            if (rep1 == rep2) 
return rep10
;
1959
            // Pick the lower-rank root to become a child of the higher-rank one.
1960
            // See https://en.wikipedia.org/wiki/Disjoint-set_data_structure#Union_by_rank.
1961
5.53k
            if (rep1->rank < rep2->rank) 
std::swap(rep1, rep2)0
;
1962
5.53k
            rep2->parent = rep1;
1963
5.53k
            rep1->rank += (rep1->rank == rep2->rank);
1964
5.53k
            return rep1;
1965
5.53k
        };
1966
1967
        // Start by initializing every Cluster as its own singleton partition.
1968
6.37k
        partition_data.resize(an_clusters.size());
1969
17.4k
        for (size_t i = 0; i < an_clusters.size(); 
++i11.0k
) {
1970
11.0k
            partition_data[i].sequence = an_clusters[i].first->m_sequence;
1971
11.0k
            partition_data[i].parent = &partition_data[i];
1972
11.0k
            partition_data[i].rank = 0;
1973
11.0k
        }
1974
1975
        // Run through all parent/child pairs in an_deps, and union the partitions their Clusters
1976
        // are in.
1977
6.37k
        Cluster* last_chl_cluster{nullptr};
1978
6.37k
        PartitionData* last_partition{nullptr};
1979
6.37k
        for (const auto& [dep, _] : an_deps) {
1980
5.53k
            auto [par, chl] = dep;
1981
5.53k
            auto par_cluster = FindCluster(par, level);
1982
5.53k
            auto chl_cluster = FindCluster(chl, level);
1983
5.53k
            Assume(chl_cluster != nullptr && par_cluster != nullptr);
Line
Count
Source
128
11.0k
#define Assume(val) 
inline_assertion_check<false>(5.53k
val, std::source_location::current(),
#5.53k
val)
1984
            // Nothing to do if parent and child are in the same Cluster.
1985
5.53k
            if (par_cluster == chl_cluster) 
continue0
;
1986
5.53k
            Assume(par != chl);
Line
Count
Source
128
5.53k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
1987
5.53k
            if (chl_cluster == last_chl_cluster) {
1988
                // If the child Clusters is the same as the previous iteration, union with the
1989
                // tree they were in, avoiding the need for another lookup. Note that an_deps
1990
                // is sorted by child Cluster, so batches with the same child are expected.
1991
0
                last_partition = union_fn(locate_fn(par_cluster->m_sequence), last_partition);
1992
5.53k
            } else {
1993
5.53k
                last_chl_cluster = chl_cluster;
1994
5.53k
                last_partition = union_fn(locate_fn(par_cluster->m_sequence), locate_fn(chl_cluster->m_sequence));
1995
5.53k
            }
1996
5.53k
        }
1997
1998
        // Update the sequence numbers in an_clusters and an_deps to be those of the partition
1999
        // representative.
2000
6.37k
        auto deps_it = an_deps.begin();
2001
17.4k
        for (size_t i = 0; i < partition_data.size(); 
++i11.0k
) {
2002
11.0k
            auto& data = partition_data[i];
2003
            // Find the sequence of the representative of the partition Cluster i is in, and store
2004
            // it with the Cluster.
2005
11.0k
            auto rep_seq = find_root_fn(&data)->sequence;
2006
11.0k
            an_clusters[i].second = rep_seq;
2007
            // Find all dependencies whose child Cluster is Cluster i, and annotate them with rep.
2008
16.6k
            while (deps_it != an_deps.end()) {
2009
11.0k
                auto [par, chl] = deps_it->first;
2010
11.0k
                auto chl_cluster = FindCluster(chl, level);
2011
11.0k
                Assume(chl_cluster != nullptr);
Line
Count
Source
128
11.0k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2012
11.0k
                if (chl_cluster->m_sequence > data.sequence) 
break5.53k
;
2013
5.53k
                deps_it->second = rep_seq;
2014
5.53k
                ++deps_it;
2015
5.53k
            }
2016
11.0k
        }
2017
6.37k
    }
2018
2019
    // Sort both an_clusters and an_deps by sequence number of the representative of the
2020
    // partition they are in, grouping all those applying to the same partition together.
2021
6.37k
    std::ranges::sort(an_deps, [](auto& a, auto& b) noexcept 
{ return a.second < b.second; }0
);
2022
11.0k
    std::ranges::sort(an_clusters, [](auto& a, auto& b) noexcept { return a.second < b.second; });
2023
2024
    // Translate the resulting cluster groups to the m_group_data structure, and the dependencies
2025
    // back to m_deps_to_add.
2026
6.37k
    clusterset.m_group_data = GroupData{};
2027
6.37k
    clusterset.m_group_data->m_group_clusters.reserve(an_clusters.size());
2028
6.37k
    clusterset.m_deps_to_add.clear();
2029
6.37k
    clusterset.m_deps_to_add.reserve(an_deps.size());
2030
6.37k
    clusterset.m_oversized = false;
2031
6.37k
    auto an_deps_it = an_deps.begin();
2032
6.37k
    auto an_clusters_it = an_clusters.begin();
2033
11.9k
    while (an_clusters_it != an_clusters.end()) {
2034
        // Process all clusters/dependencies belonging to the partition with representative rep.
2035
5.53k
        auto rep = an_clusters_it->second;
2036
        // Create and initialize a new GroupData entry for the partition.
2037
5.53k
        auto& new_entry = clusterset.m_group_data->m_groups.emplace_back();
2038
5.53k
        new_entry.m_cluster_offset = clusterset.m_group_data->m_group_clusters.size();
2039
5.53k
        new_entry.m_cluster_count = 0;
2040
5.53k
        new_entry.m_deps_offset = clusterset.m_deps_to_add.size();
2041
5.53k
        new_entry.m_deps_count = 0;
2042
5.53k
        uint32_t total_count{0};
2043
5.53k
        uint64_t total_size{0};
2044
        // Add all its clusters to it (copying those from an_clusters to m_group_clusters).
2045
16.6k
        while (an_clusters_it != an_clusters.end() && 
an_clusters_it->second == rep11.0k
) {
2046
11.0k
            clusterset.m_group_data->m_group_clusters.push_back(an_clusters_it->first);
2047
11.0k
            total_count += an_clusters_it->first->GetTxCount();
2048
11.0k
            total_size += an_clusters_it->first->GetTotalTxSize();
2049
11.0k
            ++an_clusters_it;
2050
11.0k
            ++new_entry.m_cluster_count;
2051
11.0k
        }
2052
        // Add all its dependencies to it (copying those back from an_deps to m_deps_to_add).
2053
11.0k
        while (an_deps_it != an_deps.end() && 
an_deps_it->second == rep5.53k
) {
2054
5.53k
            clusterset.m_deps_to_add.push_back(an_deps_it->first);
2055
5.53k
            ++an_deps_it;
2056
5.53k
            ++new_entry.m_deps_count;
2057
5.53k
        }
2058
        // Detect oversizedness.
2059
5.53k
        if (total_count > m_max_cluster_count || total_size > m_max_cluster_size) {
2060
0
            clusterset.m_oversized = true;
2061
0
        }
2062
5.53k
    }
2063
6.37k
    Assume(an_deps_it == an_deps.end());
Line
Count
Source
128
6.37k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2064
6.37k
    Assume(an_clusters_it == an_clusters.end());
Line
Count
Source
128
6.37k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2065
6.37k
    Compact();
2066
6.37k
}
2067
2068
void TxGraphImpl::Merge(std::span<Cluster*> to_merge, int level) noexcept
2069
5.53k
{
2070
5.53k
    Assume(!to_merge.empty());
Line
Count
Source
128
5.53k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2071
    // Nothing to do if a group consists of just a single Cluster.
2072
5.53k
    if (to_merge.size() == 1) 
return0
;
2073
2074
    // Move the largest Cluster to the front of to_merge. As all transactions in other to-be-merged
2075
    // Clusters will be moved to that one, putting the largest one first minimizes the number of
2076
    // moves.
2077
5.53k
    size_t max_size_pos{0};
2078
5.53k
    DepGraphIndex max_size = to_merge[max_size_pos]->GetTxCount();
2079
5.53k
    GetClusterSet(level).m_cluster_usage -= to_merge[max_size_pos]->TotalMemoryUsage();
2080
5.53k
    DepGraphIndex total_size = max_size;
2081
11.0k
    for (size_t i = 1; i < to_merge.size(); 
++i5.53k
) {
2082
5.53k
        GetClusterSet(level).m_cluster_usage -= to_merge[i]->TotalMemoryUsage();
2083
5.53k
        DepGraphIndex size = to_merge[i]->GetTxCount();
2084
5.53k
        total_size += size;
2085
5.53k
        if (size > max_size) {
2086
0
            max_size_pos = i;
2087
0
            max_size = size;
2088
0
        }
2089
5.53k
    }
2090
5.53k
    if (max_size_pos != 0) 
std::swap(to_merge[0], to_merge[max_size_pos])0
;
2091
2092
5.53k
    size_t start_idx = 1;
2093
5.53k
    Cluster* into_cluster = to_merge[0];
2094
5.53k
    if (total_size > into_cluster->GetMaxTxCount()) {
2095
        // The into_merge cluster is too small to fit all transactions being merged. Construct a
2096
        // a new Cluster using an implementation that matches the total size, and merge everything
2097
        // in there.
2098
2.02k
        auto new_cluster = CreateEmptyCluster(total_size);
2099
2.02k
        into_cluster = new_cluster.get();
2100
2.02k
        InsertCluster(level, std::move(new_cluster), QualityLevel::OPTIMAL);
2101
2.02k
        start_idx = 0;
2102
2.02k
    }
2103
2104
    // Merge all further Clusters in the group into the result (first one, or new one), and delete
2105
    // them.
2106
13.0k
    for (size_t i = start_idx; i < to_merge.size(); 
++i7.55k
) {
2107
7.55k
        into_cluster->Merge(*this, level, *to_merge[i]);
2108
7.55k
        DeleteCluster(*to_merge[i], level);
2109
7.55k
    }
2110
5.53k
    into_cluster->Compact();
2111
5.53k
    GetClusterSet(level).m_cluster_usage += into_cluster->TotalMemoryUsage();
2112
5.53k
}
2113
2114
void TxGraphImpl::ApplyDependencies(int level) noexcept
2115
2.12M
{
2116
2.12M
    auto& clusterset = GetClusterSet(level);
2117
    // Do not bother computing groups if we already know the result will be oversized.
2118
2.12M
    if (clusterset.m_oversized == true) 
return0
;
2119
    // Compute the groups of to-be-merged Clusters (which also applies all removals, and splits).
2120
2.12M
    GroupClusters(level);
2121
2.12M
    Assume(clusterset.m_group_data.has_value());
Line
Count
Source
128
2.12M
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2122
    // Nothing to do if there are no dependencies to be added.
2123
2.12M
    if (clusterset.m_deps_to_add.empty()) 
return2.11M
;
2124
    // Dependencies cannot be applied if it would result in oversized clusters.
2125
5.53k
    if (clusterset.m_oversized == true) 
return0
;
2126
2127
    // For each group of to-be-merged Clusters.
2128
5.53k
    for (const auto& group_entry : clusterset.m_group_data->m_groups) {
2129
5.53k
        auto cluster_span = std::span{clusterset.m_group_data->m_group_clusters}
2130
5.53k
                                .subspan(group_entry.m_cluster_offset, group_entry.m_cluster_count);
2131
        // Pull in all the Clusters that contain dependencies.
2132
5.53k
        if (level == 1) {
2133
0
            for (Cluster*& cluster : cluster_span) {
2134
0
                cluster = PullIn(cluster, cluster->GetLevel(*this));
2135
0
            }
2136
0
        }
2137
        // Invoke Merge() to merge them into a single Cluster.
2138
5.53k
        Merge(cluster_span, level);
2139
        // Actually apply all to-be-added dependencies (all parents and children from this grouping
2140
        // belong to the same Cluster at this point because of the merging above).
2141
5.53k
        auto deps_span = std::span{clusterset.m_deps_to_add}
2142
5.53k
                             .subspan(group_entry.m_deps_offset, group_entry.m_deps_count);
2143
5.53k
        Assume(!deps_span.empty());
Line
Count
Source
128
5.53k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2144
5.53k
        const auto& loc = m_entries[deps_span[0].second].m_locator[level];
2145
5.53k
        Assume(loc.IsPresent());
Line
Count
Source
128
5.53k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2146
5.53k
        loc.cluster->ApplyDependencies(*this, level, deps_span);
2147
5.53k
    }
2148
2149
    // Wipe the list of to-be-added dependencies now that they are applied.
2150
5.53k
    clusterset.m_deps_to_add.clear();
2151
5.53k
    Compact();
2152
    // Also no further Cluster mergings are needed (note that we clear, but don't set to
2153
    // std::nullopt, as that would imply the groupings are unknown).
2154
5.53k
    clusterset.m_group_data = GroupData{};
2155
5.53k
}
2156
2157
std::pair<uint64_t, bool> GenericClusterImpl::Relinearize(TxGraphImpl& graph, int level, uint64_t max_cost) noexcept
2158
5.60k
{
2159
    // We can only relinearize Clusters that do not need splitting.
2160
5.60k
    Assume(!NeedsSplitting());
Line
Count
Source
128
5.60k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2161
    // No work is required for Clusters which are already optimally linearized.
2162
5.60k
    if (IsOptimal()) 
return {0, false}0
;
2163
    // Invoke the actual linearization algorithm (passing in the existing one).
2164
5.60k
    uint64_t rng_seed = graph.m_rng.rand64();
2165
5.60k
    const auto fallback_order = [&](DepGraphIndex a, DepGraphIndex b) noexcept {
2166
0
        const auto ref_a = graph.m_entries[m_mapping[a]].m_ref;
2167
0
        const auto ref_b = graph.m_entries[m_mapping[b]].m_ref;
2168
0
        return graph.m_fallback_order(*ref_a, *ref_b);
2169
0
    };
2170
5.60k
    auto [linearization, optimal, cost] = Linearize(
2171
5.60k
        /*depgraph=*/m_depgraph,
2172
5.60k
        /*max_cost=*/max_cost,
2173
5.60k
        /*rng_seed=*/rng_seed,
2174
5.60k
        /*fallback_order=*/fallback_order,
2175
5.60k
        /*old_linearization=*/m_linearization,
2176
5.60k
        /*is_topological=*/IsTopological());
2177
    // Postlinearize to improve the linearization (if optimal, only the sub-chunk order).
2178
    // This also guarantees that all chunks are connected (even when non-optimal).
2179
5.60k
    PostLinearize(m_depgraph, linearization);
2180
    // Update the linearization.
2181
5.60k
    m_linearization = std::move(linearization);
2182
    // Update the Cluster's quality.
2183
5.60k
    bool improved = false;
2184
5.60k
    if (optimal) {
2185
5.60k
        graph.SetClusterQuality(level, m_quality, m_setindex, QualityLevel::OPTIMAL);
2186
5.60k
        improved = true;
2187
5.60k
    } else 
if (0
max_cost >= graph.m_acceptable_cost0
&&
!IsAcceptable()0
) {
2188
0
        graph.SetClusterQuality(level, m_quality, m_setindex, QualityLevel::ACCEPTABLE);
2189
0
        improved = true;
2190
0
    } else if (!IsTopological()) {
2191
0
        graph.SetClusterQuality(level, m_quality, m_setindex, QualityLevel::NEEDS_RELINEARIZE);
2192
0
        improved = true;
2193
0
    }
2194
    // Update the Entry objects.
2195
5.60k
    Updated(graph, /*level=*/level, /*rename=*/false);
2196
5.60k
    return {cost, improved};
2197
5.60k
}
2198
2199
std::pair<uint64_t, bool> SingletonClusterImpl::Relinearize(TxGraphImpl& graph, int level, uint64_t max_cost) noexcept
2200
0
{
2201
    // All singletons are optimal, oversized, or need splitting. Each of these precludes
2202
    // Relinearize from being called.
2203
0
    assert(false);
2204
0
    return {0, false};
2205
0
}
2206
2207
void TxGraphImpl::MakeAcceptable(Cluster& cluster, int level) noexcept
2208
1.04M
{
2209
    // Relinearize the Cluster if needed.
2210
1.04M
    if (!cluster.NeedsSplitting() && !cluster.IsAcceptable() && 
!cluster.IsOversized()43
) {
2211
43
        cluster.Relinearize(*this, level, m_acceptable_cost);
2212
43
    }
2213
1.04M
}
2214
2215
void TxGraphImpl::MakeAllAcceptable(int level) noexcept
2216
398k
{
2217
398k
    ApplyDependencies(level);
2218
398k
    auto& clusterset = GetClusterSet(level);
2219
398k
    if (clusterset.m_oversized == true) 
return0
;
2220
796k
    
for (auto quality : {QualityLevel::NEEDS_FIX, QualityLevel::NEEDS_RELINEARIZE})398k
{
2221
796k
        auto& queue = clusterset.m_clusters[int(quality)];
2222
796k
        while (!queue.empty()) {
2223
0
            MakeAcceptable(*queue.back().get(), level);
2224
0
        }
2225
796k
    }
2226
398k
}
2227
2228
2.04k
GenericClusterImpl::GenericClusterImpl(uint64_t sequence) noexcept : Cluster{sequence} {}
2229
2230
void TxGraphImpl::AddTransaction(Ref& arg, const FeePerWeight& feerate) noexcept
2231
12.6k
{
2232
12.6k
    Assume(m_main_chunkindex_observers == 0 || GetTopLevel() != 0);
Line
Count
Source
128
12.6k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2233
12.6k
    Assume(feerate.size > 0);
Line
Count
Source
128
12.6k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2234
12.6k
    Assume(GetRefGraph(arg) == nullptr);
Line
Count
Source
128
12.6k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2235
    // Construct a new Entry, and link it with the Ref.
2236
12.6k
    auto idx = m_entries.size();
2237
12.6k
    m_entries.emplace_back();
2238
12.6k
    auto& entry = m_entries.back();
2239
12.6k
    entry.m_main_chunkindex_iterator = m_main_chunkindex.end();
2240
12.6k
    entry.m_ref = &arg;
2241
12.6k
    GetRefGraph(arg) = this;
2242
12.6k
    GetRefIndex(arg) = idx;
2243
    // Construct a new singleton Cluster (which is necessarily optimally linearized).
2244
12.6k
    bool oversized = uint64_t(feerate.size) > m_max_cluster_size;
2245
12.6k
    auto cluster = CreateEmptyCluster(1);
2246
12.6k
    cluster->AppendTransaction(idx, feerate);
2247
12.6k
    auto cluster_ptr = cluster.get();
2248
12.6k
    int level = GetTopLevel();
2249
12.6k
    auto& clusterset = GetClusterSet(level);
2250
12.6k
    InsertCluster(level, std::move(cluster), oversized ? 
QualityLevel::OVERSIZED_SINGLETON0
: QualityLevel::OPTIMAL);
2251
12.6k
    cluster_ptr->Updated(*this, /*level=*/level, /*rename=*/false);
2252
12.6k
    clusterset.m_cluster_usage += cluster_ptr->TotalMemoryUsage();
2253
12.6k
    ++clusterset.m_txcount;
2254
    // Deal with individually oversized transactions.
2255
12.6k
    if (oversized) {
2256
0
        ++clusterset.m_txcount_oversized;
2257
0
        clusterset.m_oversized = true;
2258
0
        clusterset.m_group_data = std::nullopt;
2259
0
    }
2260
12.6k
}
2261
2262
void TxGraphImpl::RemoveTransaction(const Ref& arg) noexcept
2263
0
{
2264
    // Don't do anything if the Ref is empty (which may be indicative of the transaction already
2265
    // having been removed).
2266
0
    if (GetRefGraph(arg) == nullptr) return;
2267
0
    Assume(GetRefGraph(arg) == this);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2268
0
    Assume(m_main_chunkindex_observers == 0 || GetTopLevel() != 0);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2269
    // Find the Cluster the transaction is in, and stop if it isn't in any.
2270
0
    int level = GetTopLevel();
2271
0
    auto cluster = FindCluster(GetRefIndex(arg), level);
2272
0
    if (cluster == nullptr) return;
2273
    // Remember that the transaction is to be removed.
2274
0
    auto& clusterset = GetClusterSet(level);
2275
0
    clusterset.m_to_remove.push_back(GetRefIndex(arg));
2276
    // Wipe m_group_data (as it will need to be recomputed).
2277
0
    clusterset.m_group_data.reset();
2278
0
    if (clusterset.m_oversized == true) clusterset.m_oversized = std::nullopt;
2279
0
}
2280
2281
void TxGraphImpl::AddDependency(const Ref& parent, const Ref& child) noexcept
2282
5.53k
{
2283
    // Don't do anything if either Ref is empty (which may be indicative of it having already been
2284
    // removed).
2285
5.53k
    if (GetRefGraph(parent) == nullptr || GetRefGraph(child) == nullptr) 
return0
;
2286
5.53k
    Assume(GetRefGraph(parent) == this && GetRefGraph(child) == this);
Line
Count
Source
128
11.0k
#define Assume(val) 
inline_assertion_check<false>(5.53k
val, std::source_location::current(),
#5.53k
val)
2287
5.53k
    Assume(m_main_chunkindex_observers == 0 || GetTopLevel() != 0);
Line
Count
Source
128
5.53k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2288
    // Don't do anything if this is a dependency on self.
2289
5.53k
    if (GetRefIndex(parent) == GetRefIndex(child)) 
return0
;
2290
    // Find the Cluster the parent and child transaction are in, and stop if either appears to be
2291
    // already removed.
2292
5.53k
    int level = GetTopLevel();
2293
5.53k
    auto par_cluster = FindCluster(GetRefIndex(parent), level);
2294
5.53k
    if (par_cluster == nullptr) 
return0
;
2295
5.53k
    auto chl_cluster = FindCluster(GetRefIndex(child), level);
2296
5.53k
    if (chl_cluster == nullptr) 
return0
;
2297
    // Remember that this dependency is to be applied.
2298
5.53k
    auto& clusterset = GetClusterSet(level);
2299
5.53k
    clusterset.m_deps_to_add.emplace_back(GetRefIndex(parent), GetRefIndex(child));
2300
    // Wipe m_group_data (as it will need to be recomputed).
2301
5.53k
    clusterset.m_group_data.reset();
2302
5.53k
    if (clusterset.m_oversized == false) clusterset.m_oversized = std::nullopt;
2303
5.53k
}
2304
2305
bool TxGraphImpl::Exists(const Ref& arg, Level level_select) noexcept
2306
0
{
2307
0
    if (GetRefGraph(arg) == nullptr) return false;
2308
0
    Assume(GetRefGraph(arg) == this);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2309
0
    size_t level = GetSpecifiedLevel(level_select);
2310
    // Make sure the transaction isn't scheduled for removal.
2311
0
    ApplyRemovals(level);
2312
0
    auto cluster = FindCluster(GetRefIndex(arg), level);
2313
0
    return cluster != nullptr;
2314
0
}
2315
2316
void GenericClusterImpl::GetAncestorRefs(const TxGraphImpl& graph, std::span<std::pair<Cluster*, DepGraphIndex>>& args, std::vector<TxGraph::Ref*>& output) noexcept
2317
30
{
2318
    /** The union of all ancestors to be returned. */
2319
30
    SetType ancestors_union;
2320
    // Process elements from the front of args, as long as they apply.
2321
60
    while (!args.empty()) {
2322
30
        if (args.front().first != this) 
break0
;
2323
30
        ancestors_union |= m_depgraph.Ancestors(args.front().second);
2324
30
        args = args.subspan(1);
2325
30
    }
2326
30
    Assume(ancestors_union.Any());
Line
Count
Source
128
30
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2327
    // Translate all ancestors (in arbitrary order) to Refs (if they have any), and return them.
2328
48
    for (auto idx : ancestors_union) {
2329
48
        const auto& entry = graph.m_entries[m_mapping[idx]];
2330
48
        Assume(entry.m_ref != nullptr);
Line
Count
Source
128
48
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2331
48
        output.push_back(entry.m_ref);
2332
48
    }
2333
30
}
2334
2335
void SingletonClusterImpl::GetAncestorRefs(const TxGraphImpl& graph, std::span<std::pair<Cluster*, DepGraphIndex>>& args, std::vector<TxGraph::Ref*>& output) noexcept
2336
693
{
2337
693
    Assume(GetTxCount());
Line
Count
Source
128
693
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2338
1.38k
    while (!args.empty()) {
2339
693
        if (args.front().first != this) 
break0
;
2340
693
        args = args.subspan(1);
2341
693
    }
2342
693
    const auto& entry = graph.m_entries[m_graph_index];
2343
693
    Assume(entry.m_ref != nullptr);
Line
Count
Source
128
693
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2344
693
    output.push_back(entry.m_ref);
2345
693
}
2346
2347
void GenericClusterImpl::GetDescendantRefs(const TxGraphImpl& graph, std::span<std::pair<Cluster*, DepGraphIndex>>& args, std::vector<TxGraph::Ref*>& output) noexcept
2348
7.47k
{
2349
    /** The union of all descendants to be returned. */
2350
7.47k
    SetType descendants_union;
2351
    // Process elements from the front of args, as long as they apply.
2352
14.9k
    while (!args.empty()) {
2353
7.47k
        if (args.front().first != this) 
break0
;
2354
7.47k
        descendants_union |= m_depgraph.Descendants(args.front().second);
2355
7.47k
        args = args.subspan(1);
2356
7.47k
    }
2357
7.47k
    Assume(descendants_union.Any());
Line
Count
Source
128
7.47k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2358
    // Translate all descendants (in arbitrary order) to Refs (if they have any), and return them.
2359
23.4k
    for (auto idx : descendants_union) {
2360
23.4k
        const auto& entry = graph.m_entries[m_mapping[idx]];
2361
23.4k
        Assume(entry.m_ref != nullptr);
Line
Count
Source
128
23.4k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2362
23.4k
        output.push_back(entry.m_ref);
2363
23.4k
    }
2364
7.47k
}
2365
2366
void SingletonClusterImpl::GetDescendantRefs(const TxGraphImpl& graph, std::span<std::pair<Cluster*, DepGraphIndex>>& args, std::vector<TxGraph::Ref*>& output) noexcept
2367
550
{
2368
    // In a singleton cluster, the ancestors or descendants are always just the entire cluster.
2369
550
    GetAncestorRefs(graph, args, output);
2370
550
}
2371
2372
bool GenericClusterImpl::GetClusterRefs(TxGraphImpl& graph, std::span<TxGraph::Ref*> range, LinearizationIndex start_pos) noexcept
2373
85.1k
{
2374
    // Translate the transactions in the Cluster (in linearization order, starting at start_pos in
2375
    // the linearization) to Refs, and fill them in range.
2376
85.1k
    for (auto& ref : range) {
2377
85.1k
        Assume(start_pos < m_linearization.size());
Line
Count
Source
128
85.1k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2378
85.1k
        const auto& entry = graph.m_entries[m_mapping[m_linearization[start_pos++]]];
2379
85.1k
        Assume(entry.m_ref != nullptr);
Line
Count
Source
128
85.1k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2380
85.1k
        ref = entry.m_ref;
2381
85.1k
    }
2382
    // Return whether start_pos has advanced to the end of the Cluster.
2383
85.1k
    return start_pos == m_linearization.size();
2384
85.1k
}
2385
2386
bool SingletonClusterImpl::GetClusterRefs(TxGraphImpl& graph, std::span<TxGraph::Ref*> range, LinearizationIndex start_pos) noexcept
2387
0
{
2388
0
    Assume(!range.empty());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2389
0
    Assume(GetTxCount());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2390
0
    Assume(start_pos == 0);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2391
0
    const auto& entry = graph.m_entries[m_graph_index];
2392
0
    Assume(entry.m_ref != nullptr);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2393
0
    range[0] = entry.m_ref;
2394
0
    return true;
2395
0
}
2396
2397
FeePerWeight GenericClusterImpl::GetIndividualFeerate(DepGraphIndex idx) noexcept
2398
0
{
2399
0
    return FeePerWeight::FromFeeFrac(m_depgraph.FeeRate(idx));
2400
0
}
2401
2402
FeePerWeight SingletonClusterImpl::GetIndividualFeerate(DepGraphIndex idx) noexcept
2403
0
{
2404
0
    Assume(GetTxCount());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2405
0
    Assume(idx == 0);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2406
0
    return m_feerate;
2407
0
}
2408
2409
void GenericClusterImpl::MakeStagingTransactionsMissing(TxGraphImpl& graph) noexcept
2410
0
{
2411
    // Mark all transactions of a Cluster missing, needed when aborting staging, so that the
2412
    // corresponding Locators don't retain references into aborted Clusters.
2413
0
    for (auto ci : m_linearization) {
2414
0
        GraphIndex idx = m_mapping[ci];
2415
0
        auto& entry = graph.m_entries[idx];
2416
0
        entry.m_locator[1].SetMissing();
2417
0
    }
2418
0
}
2419
2420
void SingletonClusterImpl::MakeStagingTransactionsMissing(TxGraphImpl& graph) noexcept
2421
4.61k
{
2422
4.61k
    if (GetTxCount()) {
2423
4.61k
        auto& entry = graph.m_entries[m_graph_index];
2424
4.61k
        entry.m_locator[1].SetMissing();
2425
4.61k
    }
2426
4.61k
}
2427
2428
std::vector<TxGraph::Ref*> TxGraphImpl::GetAncestors(const Ref& arg, Level level_select) noexcept
2429
173
{
2430
    // Return the empty vector if the Ref is empty.
2431
173
    if (GetRefGraph(arg) == nullptr) 
return {}0
;
2432
173
    Assume(GetRefGraph(arg) == this);
Line
Count
Source
128
173
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2433
    // Apply all removals and dependencies, as the result might be incorrect otherwise.
2434
173
    size_t level = GetSpecifiedLevel(level_select);
2435
173
    ApplyDependencies(level);
2436
    // Ancestry cannot be known if unapplied dependencies remain.
2437
173
    Assume(GetClusterSet(level).m_deps_to_add.empty());
Line
Count
Source
128
173
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2438
    // Find the Cluster the argument is in, and return the empty vector if it isn't in any.
2439
173
    auto [cluster, cluster_level] = FindClusterAndLevel(GetRefIndex(arg), level);
2440
173
    if (cluster == nullptr) 
return {}0
;
2441
    // Dispatch to the Cluster.
2442
173
    std::pair<Cluster*, DepGraphIndex> match = {cluster, m_entries[GetRefIndex(arg)].m_locator[cluster_level].index};
2443
173
    auto matches = std::span(&match, 1);
2444
173
    std::vector<TxGraph::Ref*> ret;
2445
173
    cluster->GetAncestorRefs(*this, matches, ret);
2446
173
    return ret;
2447
173
}
2448
2449
std::vector<TxGraph::Ref*> TxGraphImpl::GetDescendants(const Ref& arg, Level level_select) noexcept
2450
8.02k
{
2451
    // Return the empty vector if the Ref is empty.
2452
8.02k
    if (GetRefGraph(arg) == nullptr) 
return {}0
;
2453
8.02k
    Assume(GetRefGraph(arg) == this);
Line
Count
Source
128
8.02k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2454
    // Apply all removals and dependencies, as the result might be incorrect otherwise.
2455
8.02k
    size_t level = GetSpecifiedLevel(level_select);
2456
8.02k
    ApplyDependencies(level);
2457
    // Ancestry cannot be known if unapplied dependencies remain.
2458
8.02k
    Assume(GetClusterSet(level).m_deps_to_add.empty());
Line
Count
Source
128
8.02k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2459
    // Find the Cluster the argument is in, and return the empty vector if it isn't in any.
2460
8.02k
    auto [cluster, cluster_level] = FindClusterAndLevel(GetRefIndex(arg), level);
2461
8.02k
    if (cluster == nullptr) 
return {}0
;
2462
    // Dispatch to the Cluster.
2463
8.02k
    std::pair<Cluster*, DepGraphIndex> match = {cluster, m_entries[GetRefIndex(arg)].m_locator[cluster_level].index};
2464
8.02k
    auto matches = std::span(&match, 1);
2465
8.02k
    std::vector<TxGraph::Ref*> ret;
2466
8.02k
    cluster->GetDescendantRefs(*this, matches, ret);
2467
8.02k
    return ret;
2468
8.02k
}
2469
2470
std::vector<TxGraph::Ref*> TxGraphImpl::GetAncestorsUnion(std::span<const Ref* const> args, Level level_select) noexcept
2471
0
{
2472
    // Apply all dependencies, as the result might be incorrect otherwise.
2473
0
    size_t level = GetSpecifiedLevel(level_select);
2474
0
    ApplyDependencies(level);
2475
    // Ancestry cannot be known if unapplied dependencies remain.
2476
0
    Assume(GetClusterSet(level).m_deps_to_add.empty());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2477
2478
    // Translate args to matches.
2479
0
    std::vector<std::pair<Cluster*, DepGraphIndex>> matches;
2480
0
    matches.reserve(args.size());
2481
0
    for (auto arg : args) {
2482
0
        Assume(arg);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2483
        // Skip empty Refs.
2484
0
        if (GetRefGraph(*arg) == nullptr) continue;
2485
0
        Assume(GetRefGraph(*arg) == this);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2486
        // Find the Cluster the argument is in, and skip if none is found.
2487
0
        auto [cluster, cluster_level] = FindClusterAndLevel(GetRefIndex(*arg), level);
2488
0
        if (cluster == nullptr) continue;
2489
        // Append to matches.
2490
0
        matches.emplace_back(cluster, m_entries[GetRefIndex(*arg)].m_locator[cluster_level].index);
2491
0
    }
2492
    // Group by Cluster.
2493
0
    std::ranges::sort(matches, [](auto& a, auto& b) noexcept { return CompareClusters(a.first, b.first) < 0; });
2494
    // Dispatch to the Clusters.
2495
0
    std::span match_span(matches);
2496
0
    std::vector<TxGraph::Ref*> ret;
2497
0
    while (!match_span.empty()) {
2498
0
        match_span.front().first->GetAncestorRefs(*this, match_span, ret);
2499
0
    }
2500
0
    return ret;
2501
0
}
2502
2503
std::vector<TxGraph::Ref*> TxGraphImpl::GetDescendantsUnion(std::span<const Ref* const> args, Level level_select) noexcept
2504
0
{
2505
    // Apply all dependencies, as the result might be incorrect otherwise.
2506
0
    size_t level = GetSpecifiedLevel(level_select);
2507
0
    ApplyDependencies(level);
2508
    // Ancestry cannot be known if unapplied dependencies remain.
2509
0
    Assume(GetClusterSet(level).m_deps_to_add.empty());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2510
2511
    // Translate args to matches.
2512
0
    std::vector<std::pair<Cluster*, DepGraphIndex>> matches;
2513
0
    matches.reserve(args.size());
2514
0
    for (auto arg : args) {
2515
0
        Assume(arg);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2516
        // Skip empty Refs.
2517
0
        if (GetRefGraph(*arg) == nullptr) continue;
2518
0
        Assume(GetRefGraph(*arg) == this);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2519
        // Find the Cluster the argument is in, and skip if none is found.
2520
0
        auto [cluster, cluster_level] = FindClusterAndLevel(GetRefIndex(*arg), level);
2521
0
        if (cluster == nullptr) continue;
2522
        // Append to matches.
2523
0
        matches.emplace_back(cluster, m_entries[GetRefIndex(*arg)].m_locator[cluster_level].index);
2524
0
    }
2525
    // Group by Cluster.
2526
0
    std::ranges::sort(matches, [](auto& a, auto& b) noexcept { return CompareClusters(a.first, b.first) < 0; });
2527
    // Dispatch to the Clusters.
2528
0
    std::span match_span(matches);
2529
0
    std::vector<TxGraph::Ref*> ret;
2530
0
    while (!match_span.empty()) {
2531
0
        match_span.front().first->GetDescendantRefs(*this, match_span, ret);
2532
0
    }
2533
0
    return ret;
2534
0
}
2535
2536
std::vector<TxGraph::Ref*> TxGraphImpl::GetCluster(const Ref& arg, Level level_select) noexcept
2537
0
{
2538
    // Return the empty vector if the Ref is empty (which may be indicative of the transaction
2539
    // having been removed already.
2540
0
    if (GetRefGraph(arg) == nullptr) return {};
2541
0
    Assume(GetRefGraph(arg) == this);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2542
    // Apply all removals and dependencies, as the result might be incorrect otherwise.
2543
0
    size_t level = GetSpecifiedLevel(level_select);
2544
0
    ApplyDependencies(level);
2545
    // Cluster linearization cannot be known if unapplied dependencies remain.
2546
0
    Assume(GetClusterSet(level).m_deps_to_add.empty());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2547
    // Find the Cluster the argument is in, and return the empty vector if it isn't in any.
2548
0
    auto [cluster, cluster_level] = FindClusterAndLevel(GetRefIndex(arg), level);
2549
0
    if (cluster == nullptr) return {};
2550
    // Make sure the Cluster has an acceptable quality level, and then dispatch to it.
2551
0
    MakeAcceptable(*cluster, cluster_level);
2552
0
    std::vector<TxGraph::Ref*> ret(cluster->GetTxCount());
2553
0
    cluster->GetClusterRefs(*this, ret, 0);
2554
0
    return ret;
2555
0
}
2556
2557
TxGraph::GraphIndex TxGraphImpl::GetTransactionCount(Level level_select) noexcept
2558
0
{
2559
0
    size_t level = GetSpecifiedLevel(level_select);
2560
0
    ApplyRemovals(level);
2561
0
    return GetClusterSet(level).m_txcount;
2562
0
}
2563
2564
FeePerWeight TxGraphImpl::GetIndividualFeerate(const Ref& arg) noexcept
2565
0
{
2566
    // Return the empty FeePerWeight if the passed Ref is empty.
2567
0
    if (GetRefGraph(arg) == nullptr) return {};
2568
0
    Assume(GetRefGraph(arg) == this);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2569
    // Find the cluster the argument is in (the level does not matter as individual feerates will
2570
    // be identical if it occurs in both), and return the empty FeePerWeight if it isn't in any.
2571
0
    Cluster* cluster{nullptr};
2572
0
    int level;
2573
0
    for (level = 0; level <= GetTopLevel(); ++level) {
2574
        // Apply removals, so that we can correctly report FeePerWeight{} for non-existing
2575
        // transactions.
2576
0
        ApplyRemovals(level);
2577
0
        if (m_entries[GetRefIndex(arg)].m_locator[level].IsPresent()) {
2578
0
            cluster = m_entries[GetRefIndex(arg)].m_locator[level].cluster;
2579
0
            break;
2580
0
        }
2581
0
    }
2582
0
    if (cluster == nullptr) return {};
2583
    // Dispatch to the Cluster.
2584
0
    return cluster->GetIndividualFeerate(m_entries[GetRefIndex(arg)].m_locator[level].index);
2585
0
}
2586
2587
FeePerWeight TxGraphImpl::GetMainChunkFeerate(const Ref& arg) noexcept
2588
0
{
2589
    // Return the empty FeePerWeight if the passed Ref is empty.
2590
0
    if (GetRefGraph(arg) == nullptr) return {};
2591
0
    Assume(GetRefGraph(arg) == this);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2592
    // Apply all removals and dependencies, as the result might be inaccurate otherwise.
2593
0
    ApplyDependencies(/*level=*/0);
2594
    // Chunk feerates cannot be accurately known if unapplied dependencies remain.
2595
0
    Assume(m_main_clusterset.m_deps_to_add.empty());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2596
    // Find the cluster the argument is in, and return the empty FeePerWeight if it isn't in any.
2597
0
    auto [cluster, cluster_level] = FindClusterAndLevel(GetRefIndex(arg), /*level=*/0);
2598
0
    if (cluster == nullptr) return {};
2599
    // Make sure the Cluster has an acceptable quality level, and then return the transaction's
2600
    // chunk feerate.
2601
0
    MakeAcceptable(*cluster, cluster_level);
2602
0
    const auto& entry = m_entries[GetRefIndex(arg)];
2603
0
    return entry.m_main_chunk_feerate;
2604
0
}
2605
2606
bool TxGraphImpl::IsOversized(Level level_select) noexcept
2607
224k
{
2608
224k
    size_t level = GetSpecifiedLevel(level_select);
2609
224k
    auto& clusterset = GetClusterSet(level);
2610
224k
    if (clusterset.m_oversized.has_value()) {
2611
        // Return cached value if known.
2612
218k
        return *clusterset.m_oversized;
2613
218k
    }
2614
5.53k
    ApplyRemovals(level);
2615
5.53k
    if (clusterset.m_txcount_oversized > 0) {
2616
0
        clusterset.m_oversized = true;
2617
5.53k
    } else {
2618
        // Find which Clusters will need to be merged together, as that is where the oversize
2619
        // property is assessed.
2620
5.53k
        GroupClusters(level);
2621
5.53k
    }
2622
5.53k
    Assume(clusterset.m_oversized.has_value());
Line
Count
Source
128
5.53k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2623
5.53k
    return *clusterset.m_oversized;
2624
224k
}
2625
2626
void TxGraphImpl::StartStaging() noexcept
2627
12.6k
{
2628
    // Staging cannot already exist.
2629
12.6k
    Assume(!m_staging_clusterset.has_value());
Line
Count
Source
128
12.6k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2630
    // Apply all remaining dependencies in main before creating a staging graph. Once staging
2631
    // exists, we cannot merge Clusters anymore (because of interference with Clusters being
2632
    // pulled into staging), so to make sure all inspectors are available (if not oversized), do
2633
    // all merging work now. Call SplitAll() first, so that even if ApplyDependencies does not do
2634
    // any thing due to knowing the result is oversized, splitting is still performed.
2635
12.6k
    SplitAll(0);
2636
12.6k
    ApplyDependencies(0);
2637
    // Construct the staging ClusterSet.
2638
12.6k
    m_staging_clusterset.emplace();
2639
    // Copy statistics, precomputed data, and to-be-applied dependencies (only if oversized) to
2640
    // the new graph. To-be-applied removals will always be empty at this point.
2641
12.6k
    m_staging_clusterset->m_txcount = m_main_clusterset.m_txcount;
2642
12.6k
    m_staging_clusterset->m_txcount_oversized = m_main_clusterset.m_txcount_oversized;
2643
12.6k
    m_staging_clusterset->m_deps_to_add = m_main_clusterset.m_deps_to_add;
2644
12.6k
    m_staging_clusterset->m_group_data = m_main_clusterset.m_group_data;
2645
12.6k
    m_staging_clusterset->m_oversized = m_main_clusterset.m_oversized;
2646
12.6k
    Assume(m_staging_clusterset->m_oversized.has_value());
Line
Count
Source
128
12.6k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2647
12.6k
    m_staging_clusterset->m_cluster_usage = 0;
2648
12.6k
}
2649
2650
void TxGraphImpl::AbortStaging() noexcept
2651
4.57k
{
2652
    // Staging must exist.
2653
4.57k
    Assume(m_staging_clusterset.has_value());
Line
Count
Source
128
4.57k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2654
    // Mark all removed transactions as Missing (so the staging locator for these transactions
2655
    // can be reused if another staging is created).
2656
4.57k
    for (auto idx : m_staging_clusterset->m_removed) {
2657
0
        m_entries[idx].m_locator[1].SetMissing();
2658
0
    }
2659
    // Do the same with the non-removed transactions in staging Clusters.
2660
36.6k
    for (int quality = 0; quality < int(QualityLevel::NONE); 
++quality32.0k
) {
2661
32.0k
        for (auto& cluster : m_staging_clusterset->m_clusters[quality]) {
2662
4.61k
            cluster->MakeStagingTransactionsMissing(*this);
2663
4.61k
        }
2664
32.0k
    }
2665
    // Destroy the staging ClusterSet.
2666
4.57k
    m_staging_clusterset.reset();
2667
4.57k
    Compact();
2668
4.57k
    if (!m_main_clusterset.m_group_data.has_value()) {
2669
        // In case m_oversized in main was kept after a Ref destruction while staging exists, we
2670
        // need to re-evaluate m_oversized now.
2671
0
        if (m_main_clusterset.m_to_remove.empty() && m_main_clusterset.m_txcount_oversized > 0) {
2672
            // It is possible that a Ref destruction caused a removal in main while staging existed.
2673
            // In this case, m_txcount_oversized may be inaccurate.
2674
0
            m_main_clusterset.m_oversized = true;
2675
0
        } else {
2676
0
            m_main_clusterset.m_oversized = std::nullopt;
2677
0
        }
2678
0
    }
2679
4.57k
}
2680
2681
void TxGraphImpl::CommitStaging() noexcept
2682
8.04k
{
2683
    // Staging must exist.
2684
8.04k
    Assume(m_staging_clusterset.has_value());
Line
Count
Source
128
8.04k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2685
8.04k
    Assume(m_main_chunkindex_observers == 0);
Line
Count
Source
128
8.04k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2686
    // Get rid of removed transactions in staging before moving to main, so they do not need to be
2687
    // added to the chunk index there. Doing so is impossible if they were unlinked, and thus have
2688
    // no Ref anymore to pass to the fallback comparator.
2689
8.04k
    ApplyRemovals(/*up_to_level=*/1);
2690
    // Delete all conflicting Clusters in main, to make place for moving the staging ones
2691
    // there. All of these have been copied to staging in PullIn().
2692
8.04k
    auto conflicts = GetConflicts();
2693
8.04k
    for (Cluster* conflict : conflicts) {
2694
0
        conflict->Clear(*this, /*level=*/0);
2695
0
        DeleteCluster(*conflict, /*level=*/0);
2696
0
    }
2697
    // Mark the removed transactions as Missing (so the staging locator for these transactions
2698
    // can be reused if another staging is created).
2699
8.04k
    for (auto idx : m_staging_clusterset->m_removed) {
2700
0
        m_entries[idx].m_locator[1].SetMissing();
2701
0
    }
2702
    // Then move all Clusters in staging to main.
2703
64.3k
    for (int quality = 0; quality < int(QualityLevel::NONE); 
++quality56.3k
) {
2704
56.3k
        auto& stage_sets = m_staging_clusterset->m_clusters[quality];
2705
64.3k
        while (!stage_sets.empty()) {
2706
8.04k
            stage_sets.back()->MoveToMain(*this);
2707
8.04k
        }
2708
56.3k
    }
2709
    // Move all statistics, precomputed data, and to-be-applied removals and dependencies.
2710
8.04k
    m_main_clusterset.m_deps_to_add = std::move(m_staging_clusterset->m_deps_to_add);
2711
8.04k
    m_main_clusterset.m_to_remove = std::move(m_staging_clusterset->m_to_remove);
2712
8.04k
    m_main_clusterset.m_group_data = std::move(m_staging_clusterset->m_group_data);
2713
8.04k
    m_main_clusterset.m_oversized = std::move(m_staging_clusterset->m_oversized);
2714
8.04k
    m_main_clusterset.m_txcount = std::move(m_staging_clusterset->m_txcount);
2715
8.04k
    m_main_clusterset.m_txcount_oversized = std::move(m_staging_clusterset->m_txcount_oversized);
2716
    // Delete the old staging graph, after all its information was moved to main.
2717
8.04k
    m_staging_clusterset.reset();
2718
8.04k
    Compact();
2719
8.04k
}
2720
2721
void GenericClusterImpl::SetFee(TxGraphImpl& graph, int level, DepGraphIndex idx, int64_t fee) noexcept
2722
0
{
2723
    // Make sure the specified DepGraphIndex exists in this Cluster.
2724
0
    Assume(m_depgraph.Positions()[idx]);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2725
    // Bail out if the fee isn't actually being changed.
2726
0
    if (m_depgraph.FeeRate(idx).fee == fee) return;
2727
    // Update the fee, remember that relinearization will be necessary, and update the Entries
2728
    // in the same Cluster.
2729
0
    m_depgraph.FeeRate(idx).fee = fee;
2730
0
    if (m_quality == QualityLevel::OVERSIZED_SINGLETON) {
2731
        // Nothing to do.
2732
0
    } else if (IsAcceptable()) {
2733
0
        graph.SetClusterQuality(level, m_quality, m_setindex, QualityLevel::NEEDS_RELINEARIZE);
2734
0
    }
2735
0
    Updated(graph, /*level=*/level, /*rename=*/false);
2736
0
}
2737
2738
void SingletonClusterImpl::SetFee(TxGraphImpl& graph, int level, DepGraphIndex idx, int64_t fee) noexcept
2739
0
{
2740
0
    Assume(GetTxCount());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2741
0
    Assume(idx == 0);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2742
0
    m_feerate.fee = fee;
2743
0
    Updated(graph, /*level=*/level, /*rename=*/false);
2744
0
}
2745
2746
void TxGraphImpl::SetTransactionFee(const Ref& ref, int64_t fee) noexcept
2747
0
{
2748
    // Don't do anything if the passed Ref is empty.
2749
0
    if (GetRefGraph(ref) == nullptr) return;
2750
0
    Assume(GetRefGraph(ref) == this);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2751
0
    Assume(m_main_chunkindex_observers == 0);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2752
    // Find the entry, its locator, and inform its Cluster about the new feerate, if any.
2753
0
    auto& entry = m_entries[GetRefIndex(ref)];
2754
0
    for (int level = 0; level < MAX_LEVELS; ++level) {
2755
0
        auto& locator = entry.m_locator[level];
2756
0
        if (locator.IsPresent()) {
2757
0
            locator.cluster->SetFee(*this, level, locator.index, fee);
2758
0
        }
2759
0
    }
2760
0
}
2761
2762
std::strong_ordering TxGraphImpl::CompareMainOrder(const Ref& a, const Ref& b) noexcept
2763
521k
{
2764
    // The references must not be empty.
2765
521k
    Assume(GetRefGraph(a) == this);
Line
Count
Source
128
521k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2766
521k
    Assume(GetRefGraph(b) == this);
Line
Count
Source
128
521k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2767
    // Apply dependencies in main.
2768
521k
    ApplyDependencies(0);
2769
521k
    Assume(m_main_clusterset.m_deps_to_add.empty());
Line
Count
Source
128
521k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2770
    // Make both involved Clusters acceptable, so chunk feerates are relevant.
2771
521k
    const auto& entry_a = m_entries[GetRefIndex(a)];
2772
521k
    const auto& entry_b = m_entries[GetRefIndex(b)];
2773
521k
    const auto& locator_a = entry_a.m_locator[0];
2774
521k
    const auto& locator_b = entry_b.m_locator[0];
2775
521k
    Assume(locator_a.IsPresent());
Line
Count
Source
128
521k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2776
521k
    Assume(locator_b.IsPresent());
Line
Count
Source
128
521k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2777
521k
    MakeAcceptable(*locator_a.cluster, /*level=*/0);
2778
521k
    MakeAcceptable(*locator_b.cluster, /*level=*/0);
2779
    // Invoke comparison logic.
2780
521k
    return CompareMainTransactions(GetRefIndex(a), GetRefIndex(b));
2781
521k
}
2782
2783
TxGraph::GraphIndex TxGraphImpl::CountDistinctClusters(std::span<const Ref* const> refs, Level level_select) noexcept
2784
3.12k
{
2785
3.12k
    size_t level = GetSpecifiedLevel(level_select);
2786
3.12k
    ApplyDependencies(level);
2787
3.12k
    auto& clusterset = GetClusterSet(level);
2788
3.12k
    Assume(clusterset.m_deps_to_add.empty());
Line
Count
Source
128
3.12k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2789
    // Build a vector of Clusters that the specified Refs occur in.
2790
3.12k
    std::vector<Cluster*> clusters;
2791
3.12k
    clusters.reserve(refs.size());
2792
3.12k
    for (const Ref* ref : refs) {
2793
3.12k
        Assume(ref);
Line
Count
Source
128
3.12k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2794
3.12k
        if (GetRefGraph(*ref) == nullptr) 
continue0
;
2795
3.12k
        Assume(GetRefGraph(*ref) == this);
Line
Count
Source
128
3.12k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2796
3.12k
        auto cluster = FindCluster(GetRefIndex(*ref), level);
2797
3.12k
        if (cluster != nullptr) clusters.push_back(cluster);
2798
3.12k
    }
2799
    // Count the number of distinct elements in clusters.
2800
3.12k
    std::ranges::sort(clusters, [](Cluster* a, Cluster* b) noexcept 
{ return CompareClusters(a, b) < 0; }0
);
2801
3.12k
    Cluster* last{nullptr};
2802
3.12k
    GraphIndex ret{0};
2803
3.12k
    for (Cluster* cluster : clusters) {
2804
3.12k
        ret += (cluster != last);
2805
3.12k
        last = cluster;
2806
3.12k
    }
2807
3.12k
    return ret;
2808
3.12k
}
2809
2810
std::pair<std::vector<FeeFrac>, std::vector<FeeFrac>> TxGraphImpl::GetMainStagingDiagrams() noexcept
2811
0
{
2812
0
    Assume(m_staging_clusterset.has_value());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2813
0
    MakeAllAcceptable(0);
2814
0
    Assume(m_main_clusterset.m_deps_to_add.empty()); // can only fail if main is oversized
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2815
0
    MakeAllAcceptable(1);
2816
0
    Assume(m_staging_clusterset->m_deps_to_add.empty()); // can only fail if staging is oversized
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
2817
    // For all Clusters in main which conflict with Clusters in staging (i.e., all that are removed
2818
    // by, or replaced in, staging), gather their chunk feerates.
2819
0
    auto main_clusters = GetConflicts();
2820
0
    std::vector<FeeFrac> main_feerates, staging_feerates;
2821
0
    for (Cluster* cluster : main_clusters) {
2822
0
        cluster->AppendChunkFeerates(main_feerates);
2823
0
    }
2824
    // Do the same for the Clusters in staging themselves.
2825
0
    for (int quality = 0; quality < int(QualityLevel::NONE); ++quality) {
2826
0
        for (const auto& cluster : m_staging_clusterset->m_clusters[quality]) {
2827
0
            cluster->AppendChunkFeerates(staging_feerates);
2828
0
        }
2829
0
    }
2830
    // Sort both by decreasing feerate to obtain diagrams, and return them.
2831
0
    std::ranges::sort(main_feerates, std::greater<ByRatioNegSize<FeeFrac>>{});
2832
0
    std::ranges::sort(staging_feerates, std::greater<ByRatioNegSize<FeeFrac>>{});
2833
0
    return std::make_pair(std::move(main_feerates), std::move(staging_feerates));
2834
0
}
2835
2836
void GenericClusterImpl::SanityCheck(const TxGraphImpl& graph, int level) const
2837
23.2k
{
2838
    // There must be an m_mapping for each m_depgraph position (including holes).
2839
23.2k
    assert(m_depgraph.PositionRange() == m_mapping.size());
2840
    // The linearization for this Cluster must contain every transaction once.
2841
23.2k
    assert(m_depgraph.TxCount() == m_linearization.size());
2842
    // Unless a split is to be applied, the cluster cannot have any holes.
2843
23.2k
    if (!NeedsSplitting()) {
2844
23.2k
        assert(m_depgraph.Positions() == SetType::Fill(m_depgraph.TxCount()));
2845
23.2k
    }
2846
2847
    // Compute the chunking of m_linearization.
2848
23.2k
    auto linchunking = ChunkLinearizationInfo(m_depgraph, m_linearization);
2849
23.2k
    unsigned chunk_num{0};
2850
2851
    // Verify m_linearization.
2852
23.2k
    SetType m_done;
2853
23.2k
    LinearizationIndex linindex{0};
2854
23.2k
    DepGraphIndex chunk_pos{0}; //!< position within the current chunk
2855
23.2k
    assert(m_depgraph.IsAcyclic());
2856
23.2k
    if (m_linearization.empty()) 
return0
;
2857
23.2k
    FeeFrac equal_feerate_prefix = linchunking[chunk_num].feerate;
2858
108k
    for (auto lin_pos : m_linearization) {
2859
108k
        assert(lin_pos < m_mapping.size());
2860
108k
        const auto& entry = graph.m_entries[m_mapping[lin_pos]];
2861
        // Check that the linearization is topological.
2862
108k
        m_done.Set(lin_pos);
2863
108k
        if (IsTopological()) {
2864
108k
            assert(m_done.IsSupersetOf(m_depgraph.Ancestors(lin_pos)));
2865
108k
        }
2866
        // Check that the Entry has a locator pointing back to this Cluster & position within it.
2867
108k
        assert(entry.m_locator[level].cluster == this);
2868
108k
        assert(entry.m_locator[level].index == lin_pos);
2869
        // For main-level entries, check linearization position and chunk feerate.
2870
108k
        if (level == 0 && IsAcceptable()) {
2871
108k
            assert(entry.m_main_lin_index == linindex);
2872
108k
            ++linindex;
2873
108k
            if (!linchunking[chunk_num].transactions[lin_pos]) {
2874
                // First transaction of a new chunk.
2875
85.0k
                ++chunk_num;
2876
85.0k
                assert(chunk_num < linchunking.size());
2877
85.0k
                chunk_pos = 0;
2878
85.0k
                if (ByRatio{linchunking[chunk_num].feerate} < ByRatio{equal_feerate_prefix}) {
2879
0
                    equal_feerate_prefix = linchunking[chunk_num].feerate;
2880
85.0k
                } else {
2881
85.0k
                    assert(ByRatio{linchunking[chunk_num].feerate} == ByRatio{equal_feerate_prefix});
2882
85.0k
                    equal_feerate_prefix += linchunking[chunk_num].feerate;
2883
85.0k
                }
2884
85.0k
            }
2885
108k
            assert(entry.m_main_chunk_feerate == linchunking[chunk_num].feerate);
2886
108k
            assert(entry.m_main_equal_feerate_chunk_prefix_size == equal_feerate_prefix.size);
2887
            // Verify that an entry in the chunk index exists for every chunk-ending transaction.
2888
108k
            ++chunk_pos;
2889
108k
            if (graph.m_main_clusterset.m_to_remove.empty()) {
2890
108k
                bool is_chunk_end = (chunk_pos == linchunking[chunk_num].transactions.Count());
2891
108k
                assert((entry.m_main_chunkindex_iterator != graph.m_main_chunkindex.end()) == is_chunk_end);
2892
108k
                if (is_chunk_end) {
2893
108k
                    auto& chunk_data = *entry.m_main_chunkindex_iterator;
2894
108k
                    if (m_done == m_depgraph.Positions() && 
chunk_pos == 123.2k
) {
2895
23.2k
                        assert(chunk_data.m_chunk_count == LinearizationIndex(-1));
2896
85.0k
                    } else {
2897
85.0k
                        assert(chunk_data.m_chunk_count == chunk_pos);
2898
85.0k
                    }
2899
108k
                }
2900
108k
            }
2901
            // If this Cluster has an acceptable quality level, its chunks must be connected.
2902
108k
            assert(m_depgraph.IsConnected(linchunking[chunk_num].transactions));
2903
108k
        }
2904
108k
    }
2905
    // Verify that each element of m_depgraph occurred in m_linearization.
2906
23.2k
    assert(m_done == m_depgraph.Positions());
2907
23.2k
}
2908
2909
void SingletonClusterImpl::SanityCheck(const TxGraphImpl& graph, int level) const
2910
16.3k
{
2911
    // All singletons are optimal, oversized, or need splitting.
2912
16.3k
    Assume(IsOptimal() || IsOversized() || NeedsSplitting());
Line
Count
Source
128
32.6k
#define Assume(val) 
inline_assertion_check<false>(16.3k
val, std::source_location::current(),
#16.3k
val)
2913
16.3k
    if (GetTxCount()) {
2914
16.3k
        const auto& entry = graph.m_entries[m_graph_index];
2915
        // Check that the Entry has a locator pointing back to this Cluster & position within it.
2916
16.3k
        assert(entry.m_locator[level].cluster == this);
2917
16.3k
        assert(entry.m_locator[level].index == 0);
2918
        // For main-level entries, check linearization position and chunk feerate.
2919
16.3k
        if (level == 0 && IsAcceptable()) {
2920
16.3k
            assert(entry.m_main_lin_index == 0);
2921
16.3k
            assert(entry.m_main_chunk_feerate == m_feerate);
2922
16.3k
            assert(entry.m_main_equal_feerate_chunk_prefix_size == m_feerate.size);
2923
16.3k
            if (graph.m_main_clusterset.m_to_remove.empty()) {
2924
16.3k
                assert(entry.m_main_chunkindex_iterator != graph.m_main_chunkindex.end());
2925
16.3k
                auto& chunk_data = *entry.m_main_chunkindex_iterator;
2926
16.3k
                assert(chunk_data.m_chunk_count == LinearizationIndex(-1));
2927
16.3k
            }
2928
16.3k
        }
2929
16.3k
    }
2930
16.3k
}
2931
2932
void TxGraphImpl::SanityCheck() const
2933
213k
{
2934
    /** Which GraphIndexes ought to occur in m_unlinked, based on m_entries. */
2935
213k
    std::set<GraphIndex> expected_unlinked;
2936
    /** Which Clusters ought to occur in ClusterSet::m_clusters, based on m_entries. */
2937
213k
    std::set<const Cluster*> expected_clusters[MAX_LEVELS];
2938
    /** Which GraphIndexes ought to occur in ClusterSet::m_removed, based on m_entries. */
2939
213k
    std::set<GraphIndex> expected_removed[MAX_LEVELS];
2940
    /** Which Cluster::m_sequence values have been encountered. */
2941
213k
    std::set<uint64_t> sequences;
2942
    /** Which GraphIndexes ought to occur in m_main_chunkindex, based on m_entries. */
2943
213k
    std::set<GraphIndex> expected_chunkindex;
2944
    /** Whether compaction is possible in the current state. */
2945
213k
    bool compact_possible{true};
2946
2947
    // Go over all Entry objects in m_entries.
2948
337k
    for (GraphIndex idx = 0; idx < m_entries.size(); 
++idx124k
) {
2949
124k
        const auto& entry = m_entries[idx];
2950
124k
        if (entry.m_ref == nullptr) {
2951
            // Unlinked Entry must have indexes appear in m_unlinked.
2952
0
            expected_unlinked.insert(idx);
2953
124k
        } else {
2954
            // Every non-unlinked Entry must have a Ref that points back to it.
2955
124k
            assert(GetRefGraph(*entry.m_ref) == this);
2956
124k
            assert(GetRefIndex(*entry.m_ref) == idx);
2957
124k
        }
2958
124k
        if (entry.m_main_chunkindex_iterator != m_main_chunkindex.end()) {
2959
            // Remember which entries we see a chunkindex entry for.
2960
124k
            assert(entry.m_locator[0].IsPresent());
2961
124k
            expected_chunkindex.insert(idx);
2962
124k
        }
2963
        // Verify the Entry m_locators.
2964
124k
        bool was_present{false}, was_removed{false};
2965
374k
        for (int level = 0; level < MAX_LEVELS; 
++level249k
) {
2966
249k
            const auto& locator = entry.m_locator[level];
2967
            // Every Locator must be in exactly one of these 3 states.
2968
249k
            assert(locator.IsMissing() + locator.IsRemoved() + locator.IsPresent() == 1);
2969
249k
            if (locator.IsPresent()) {
2970
                // Once removed, a transaction cannot be revived.
2971
124k
                assert(!was_removed);
2972
                // Verify that the Cluster agrees with where the Locator claims the transaction is.
2973
124k
                assert(locator.cluster->GetClusterEntry(locator.index) == idx);
2974
                // Remember that we expect said Cluster to appear in the ClusterSet::m_clusters.
2975
124k
                expected_clusters[level].insert(locator.cluster);
2976
124k
                was_present = true;
2977
124k
            } else if (locator.IsRemoved()) {
2978
                // Level 0 (main) cannot have IsRemoved locators (IsMissing there means non-existing).
2979
0
                assert(level > 0);
2980
                // A Locator can only be IsRemoved if it was IsPresent before, and only once.
2981
0
                assert(was_present && !was_removed);
2982
                // Remember that we expect this GraphIndex to occur in the ClusterSet::m_removed.
2983
0
                expected_removed[level].insert(idx);
2984
0
                was_removed = true;
2985
0
            }
2986
249k
        }
2987
124k
    }
2988
2989
    // For all levels (0 = main, 1 = staged)...
2990
426k
    
for (int level = 0; 213k
level <= GetTopLevel();
++level213k
) {
2991
213k
        assert(level < MAX_LEVELS);
2992
213k
        auto& clusterset = GetClusterSet(level);
2993
213k
        std::set<const Cluster*> actual_clusters;
2994
213k
        size_t recomputed_cluster_usage{0};
2995
2996
        // For all quality levels...
2997
1.70M
        for (int qual = 0; qual < int(QualityLevel::NONE); 
++qual1.49M
) {
2998
1.49M
            QualityLevel quality{qual};
2999
1.49M
            const auto& quality_clusters = clusterset.m_clusters[qual];
3000
            // ... for all clusters in them ...
3001
1.53M
            for (ClusterSetIndex setindex = 0; setindex < quality_clusters.size(); 
++setindex39.6k
) {
3002
39.6k
                const auto& cluster = *quality_clusters[setindex];
3003
                // The number of transactions in a Cluster cannot exceed m_max_cluster_count.
3004
39.6k
                assert(cluster.GetTxCount() <= m_max_cluster_count);
3005
                // The level must match the Cluster's own idea of what level it is in (but GetLevel
3006
                // can only be called for non-empty Clusters).
3007
39.6k
                assert(cluster.GetTxCount() == 0 || level == cluster.GetLevel(*this));
3008
                // The sum of their sizes cannot exceed m_max_cluster_size, unless it is an
3009
                // individually oversized transaction singleton. Note that groups of to-be-merged
3010
                // clusters which would exceed this limit are marked oversized, which means they
3011
                // are never applied.
3012
39.6k
                assert(cluster.IsOversized() || cluster.GetTotalTxSize() <= m_max_cluster_size);
3013
                // OVERSIZED clusters are singletons.
3014
39.6k
                assert(!cluster.IsOversized() || cluster.GetTxCount() == 1);
3015
                // Transaction counts cannot exceed the Cluster implementation's maximum
3016
                // supported transactions count.
3017
39.6k
                assert(cluster.GetTxCount() <= cluster.GetMaxTxCount());
3018
                // Unless a Split is yet to be applied, the number of transactions must not be
3019
                // below the Cluster implementation's intended transaction count.
3020
39.6k
                if (!cluster.NeedsSplitting()) {
3021
39.6k
                    assert(cluster.GetTxCount() >= cluster.GetMinIntendedTxCount());
3022
39.6k
                }
3023
3024
                // Check the sequence number.
3025
39.6k
                assert(cluster.m_sequence < m_next_sequence_counter);
3026
39.6k
                assert(!sequences.contains(cluster.m_sequence));
3027
39.6k
                sequences.insert(cluster.m_sequence);
3028
                // Remember we saw this Cluster (only if it is non-empty; empty Clusters aren't
3029
                // expected to be referenced by the Entry vector).
3030
39.6k
                if (cluster.GetTxCount() != 0) {
3031
39.6k
                    actual_clusters.insert(&cluster);
3032
39.6k
                }
3033
                // Sanity check the cluster, according to the Cluster's internal rules.
3034
39.6k
                cluster.SanityCheck(*this, level);
3035
                // Check that the cluster's quality and setindex matches its position in the quality list.
3036
39.6k
                assert(cluster.m_quality == quality);
3037
39.6k
                assert(cluster.m_setindex == setindex);
3038
                // Count memory usage.
3039
39.6k
                recomputed_cluster_usage += cluster.TotalMemoryUsage();
3040
39.6k
            }
3041
1.49M
        }
3042
3043
        // Verify memory usage.
3044
213k
        assert(clusterset.m_cluster_usage == recomputed_cluster_usage);
3045
3046
        // Verify that all to-be-removed transactions have valid identifiers.
3047
213k
        for (GraphIndex idx : clusterset.m_to_remove) {
3048
0
            assert(idx < m_entries.size());
3049
            // We cannot assert that all m_to_remove transactions are still present: ~Ref on a
3050
            // (P,M) transaction (present in main, inherited in staging) will cause an m_to_remove
3051
            // addition in both main and staging, but a subsequence ApplyRemovals in main will
3052
            // cause it to disappear from staging too, leaving the m_to_remove in place.
3053
0
        }
3054
3055
        // Verify that all to-be-added dependencies have valid identifiers.
3056
213k
        for (auto [par_idx, chl_idx] : clusterset.m_deps_to_add) {
3057
0
            assert(par_idx != chl_idx);
3058
0
            assert(par_idx < m_entries.size());
3059
0
            assert(chl_idx < m_entries.size());
3060
0
        }
3061
3062
        // Verify that the actually encountered clusters match the ones occurring in Entry vector.
3063
213k
        assert(actual_clusters == expected_clusters[level]);
3064
3065
        // Verify that the contents of m_removed matches what was expected based on the Entry vector.
3066
213k
        std::set<GraphIndex> actual_removed(clusterset.m_removed.begin(), clusterset.m_removed.end());
3067
213k
        for (auto i : expected_unlinked) {
3068
            // If a transaction exists in both main and staging, and is removed from staging (adding
3069
            // it to m_removed there), and consequently destroyed (wiping the locator completely),
3070
            // it can remain in m_removed despite not having an IsRemoved() locator. Exclude those
3071
            // transactions from the comparison here.
3072
0
            actual_removed.erase(i);
3073
0
            expected_removed[level].erase(i);
3074
0
        }
3075
3076
213k
        assert(actual_removed == expected_removed[level]);
3077
3078
        // If any GraphIndex entries remain in this ClusterSet, compact is not possible.
3079
213k
        if (!clusterset.m_deps_to_add.empty()) 
compact_possible = false0
;
3080
213k
        if (!clusterset.m_to_remove.empty()) 
compact_possible = false0
;
3081
213k
        if (!clusterset.m_removed.empty()) 
compact_possible = false0
;
3082
3083
        // For non-top levels, m_oversized must be known (as it cannot change until the level
3084
        // on top is gone).
3085
213k
        if (level < GetTopLevel()) assert(clusterset.m_oversized.has_value());
3086
213k
    }
3087
3088
    // Verify that the contents of m_unlinked matches what was expected based on the Entry vector.
3089
213k
    std::set<GraphIndex> actual_unlinked(m_unlinked.begin(), m_unlinked.end());
3090
213k
    assert(actual_unlinked == expected_unlinked);
3091
3092
    // If compaction was possible, it should have been performed already, and m_unlinked must be
3093
    // empty (to prevent memory leaks due to an ever-growing m_entries vector).
3094
213k
    if (compact_possible) {
3095
213k
        assert(actual_unlinked.empty());
3096
213k
    }
3097
3098
    // Finally, check the chunk index.
3099
213k
    std::set<GraphIndex> actual_chunkindex;
3100
213k
    FeeFrac last_chunk_feerate;
3101
213k
    for (const auto& chunk : m_main_chunkindex) {
3102
124k
        GraphIndex idx = chunk.m_graph_index;
3103
124k
        actual_chunkindex.insert(idx);
3104
124k
        auto chunk_feerate = m_entries[idx].m_main_chunk_feerate;
3105
124k
        if (!last_chunk_feerate.IsEmpty()) {
3106
110k
            assert(ByRatio{last_chunk_feerate} >= ByRatio{FeeFrac{chunk_feerate}});
3107
110k
        }
3108
124k
        last_chunk_feerate = chunk_feerate;
3109
124k
    }
3110
213k
    assert(actual_chunkindex == expected_chunkindex);
3111
213k
}
3112
3113
bool TxGraphImpl::DoWork(uint64_t max_cost) noexcept
3114
194k
{
3115
194k
    uint64_t cost_done{0};
3116
    // First linearize everything in NEEDS_RELINEARIZE to an acceptable level. If more budget
3117
    // remains after that, try to make everything optimal.
3118
584k
    for (QualityLevel quality : {QualityLevel::NEEDS_FIX, QualityLevel::NEEDS_RELINEARIZE, QualityLevel::ACCEPTABLE}) {
3119
        // First linearize staging, if it exists, then main.
3120
1.16M
        for (int level = GetTopLevel(); level >= 0; 
--level584k
) {
3121
            // Do not modify main if it has any observers.
3122
584k
            if (level == 0 && m_main_chunkindex_observers != 0) 
continue0
;
3123
584k
            ApplyDependencies(level);
3124
584k
            auto& clusterset = GetClusterSet(level);
3125
            // Do not modify oversized levels.
3126
584k
            if (clusterset.m_oversized == true) 
continue0
;
3127
584k
            auto& queue = clusterset.m_clusters[int(quality)];
3128
590k
            while (!queue.empty()) {
3129
5.56k
                if (cost_done >= max_cost) 
return false0
;
3130
                // Randomize the order in which we process, so that if the first cluster somehow
3131
                // needs more work than what max_cost allows, we don't keep spending it on the same
3132
                // one.
3133
5.56k
                auto pos = m_rng.randrange<size_t>(queue.size());
3134
5.56k
                auto cost_now = max_cost - cost_done;
3135
5.56k
                if (quality == QualityLevel::NEEDS_FIX || 
quality == QualityLevel::NEEDS_RELINEARIZE26
) {
3136
                    // If we're working with clusters that need relinearization still, only perform
3137
                    // up to m_acceptable_cost work. If they become ACCEPTABLE, and we still
3138
                    // have budget after all other clusters are ACCEPTABLE too, we'll spend the
3139
                    // remaining budget on trying to make them OPTIMAL.
3140
5.56k
                    cost_now = std::min(cost_now, m_acceptable_cost);
3141
5.56k
                }
3142
5.56k
                auto [cost, improved] = queue[pos].get()->Relinearize(*this, level, cost_now);
3143
5.56k
                cost_done += cost;
3144
                // If no improvement was made to the Cluster, it means we've essentially run out of
3145
                // budget. Even though it may be the case that cost_done < max_cost still, the
3146
                // linearizer decided there wasn't enough budget left to attempt anything with.
3147
                // To avoid an infinite loop that keeps trying clusters with minuscule budgets,
3148
                // stop here too.
3149
5.56k
                if (!improved) 
return false0
;
3150
5.56k
            }
3151
584k
        }
3152
584k
    }
3153
    // All possible work has been performed, so we can return true. Note that this does *not* mean
3154
    // that all clusters are optimally linearized now. It may be that there is nothing to do left
3155
    // because all non-optimal clusters are in oversized and/or observer-bearing levels.
3156
194k
    return true;
3157
194k
}
3158
3159
void BlockBuilderImpl::Next() noexcept
3160
124k
{
3161
    // Don't do anything if we're already done.
3162
124k
    if (m_cur_iter == m_graph->m_main_chunkindex.end()) 
return0
;
3163
124k
    while (true) {
3164
        // Advance the pointer, and stop if we reach the end.
3165
124k
        ++m_cur_iter;
3166
124k
        m_cur_cluster = nullptr;
3167
124k
        if (m_cur_iter == m_graph->m_main_chunkindex.end()) 
break14.3k
;
3168
        // Find the cluster pointed to by m_cur_iter.
3169
110k
        const auto& chunk_data = *m_cur_iter;
3170
110k
        const auto& chunk_end_entry = m_graph->m_entries[chunk_data.m_graph_index];
3171
110k
        m_cur_cluster = chunk_end_entry.m_locator[0].cluster;
3172
110k
        m_known_end_of_cluster = false;
3173
        // If we previously skipped a chunk from this cluster we cannot include more from it.
3174
110k
        if (!m_excluded_clusters.contains(m_cur_cluster->m_sequence)) break;
3175
110k
    }
3176
124k
}
3177
3178
std::optional<std::pair<std::vector<TxGraph::Ref*>, FeePerWeight>> BlockBuilderImpl::GetCurrentChunk() noexcept
3179
522k
{
3180
522k
    std::optional<std::pair<std::vector<TxGraph::Ref*>, FeePerWeight>> ret;
3181
    // Populate the return value if we are not done.
3182
522k
    if (m_cur_iter != m_graph->m_main_chunkindex.end()) {
3183
124k
        ret.emplace();
3184
124k
        const auto& chunk_data = *m_cur_iter;
3185
124k
        const auto& chunk_end_entry = m_graph->m_entries[chunk_data.m_graph_index];
3186
124k
        if (chunk_data.m_chunk_count == LinearizationIndex(-1)) {
3187
            // Special case in case just a single transaction remains, avoiding the need to
3188
            // dispatch to and dereference Cluster.
3189
39.6k
            ret->first.resize(1);
3190
39.6k
            Assume(chunk_end_entry.m_ref != nullptr);
Line
Count
Source
128
39.6k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3191
39.6k
            ret->first[0] = chunk_end_entry.m_ref;
3192
39.6k
            m_known_end_of_cluster = true;
3193
85.1k
        } else {
3194
85.1k
            Assume(m_cur_cluster);
Line
Count
Source
128
85.1k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3195
85.1k
            ret->first.resize(chunk_data.m_chunk_count);
3196
85.1k
            auto start_pos = chunk_end_entry.m_main_lin_index + 1 - chunk_data.m_chunk_count;
3197
85.1k
            m_known_end_of_cluster = m_cur_cluster->GetClusterRefs(*m_graph, ret->first, start_pos);
3198
            // If the chunk size was 1 and at end of cluster, then the special case above should
3199
            // have been used.
3200
85.1k
            Assume(!m_known_end_of_cluster || chunk_data.m_chunk_count > 1);
Line
Count
Source
128
85.1k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3201
85.1k
        }
3202
124k
        ret->second = chunk_end_entry.m_main_chunk_feerate;
3203
124k
    }
3204
522k
    return ret;
3205
522k
}
3206
3207
398k
BlockBuilderImpl::BlockBuilderImpl(TxGraphImpl& graph) noexcept : m_graph(&graph)
3208
398k
{
3209
    // Make sure all clusters in main are up to date, and acceptable.
3210
398k
    m_graph->MakeAllAcceptable(0);
3211
    // There cannot remain any inapplicable dependencies (only possible if main is oversized).
3212
398k
    Assume(m_graph->m_main_clusterset.m_deps_to_add.empty());
Line
Count
Source
128
398k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3213
    // Remember that this object is observing the graph's index, so that we can detect concurrent
3214
    // modifications.
3215
398k
    ++m_graph->m_main_chunkindex_observers;
3216
    // Find the first chunk.
3217
398k
    m_cur_iter = m_graph->m_main_chunkindex.begin();
3218
398k
    m_cur_cluster = nullptr;
3219
398k
    if (m_cur_iter != m_graph->m_main_chunkindex.end()) {
3220
        // Find the cluster pointed to by m_cur_iter.
3221
14.3k
        const auto& chunk_data = *m_cur_iter;
3222
14.3k
        const auto& chunk_end_entry = m_graph->m_entries[chunk_data.m_graph_index];
3223
14.3k
        m_cur_cluster = chunk_end_entry.m_locator[0].cluster;
3224
14.3k
    }
3225
398k
}
3226
3227
BlockBuilderImpl::~BlockBuilderImpl()
3228
398k
{
3229
398k
    Assume(m_graph->m_main_chunkindex_observers > 0);
Line
Count
Source
128
398k
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3230
    // Permit modifications to the main graph again after destroying the BlockBuilderImpl.
3231
398k
    --m_graph->m_main_chunkindex_observers;
3232
398k
}
3233
3234
void BlockBuilderImpl::Include() noexcept
3235
124k
{
3236
    // The actual inclusion of the chunk is done by the calling code. All we have to do is switch
3237
    // to the next chunk.
3238
124k
    Next();
3239
124k
}
3240
3241
void BlockBuilderImpl::Skip() noexcept
3242
0
{
3243
    // When skipping a chunk we need to not include anything more of the cluster, as that could make
3244
    // the result topologically invalid. However, don't do this if the chunk is known to be the last
3245
    // chunk of the cluster. This may significantly reduce the size of m_excluded_clusters,
3246
    // especially when many singleton clusters are ignored.
3247
0
    if (m_cur_cluster != nullptr && !m_known_end_of_cluster) {
3248
0
        m_excluded_clusters.insert(m_cur_cluster->m_sequence);
3249
0
    }
3250
0
    Next();
3251
0
}
3252
3253
std::unique_ptr<TxGraph::BlockBuilder> TxGraphImpl::GetBlockBuilder() noexcept
3254
398k
{
3255
398k
    return std::make_unique<BlockBuilderImpl>(*this);
3256
398k
}
3257
3258
std::pair<std::vector<TxGraph::Ref*>, FeePerWeight> TxGraphImpl::GetWorstMainChunk() noexcept
3259
0
{
3260
0
    std::pair<std::vector<Ref*>, FeePerWeight> ret;
3261
    // Make sure all clusters in main are up to date, and acceptable.
3262
0
    MakeAllAcceptable(0);
3263
0
    Assume(m_main_clusterset.m_deps_to_add.empty());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3264
    // If the graph is not empty, populate ret.
3265
0
    if (!m_main_chunkindex.empty()) {
3266
0
        const auto& chunk_data = *m_main_chunkindex.rbegin();
3267
0
        const auto& chunk_end_entry = m_entries[chunk_data.m_graph_index];
3268
0
        Cluster* cluster = chunk_end_entry.m_locator[0].cluster;
3269
0
        if (chunk_data.m_chunk_count == LinearizationIndex(-1) || chunk_data.m_chunk_count == 1)  {
3270
            // Special case for singletons.
3271
0
            ret.first.resize(1);
3272
0
            Assume(chunk_end_entry.m_ref != nullptr);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3273
0
            ret.first[0] = chunk_end_entry.m_ref;
3274
0
        } else {
3275
0
            ret.first.resize(chunk_data.m_chunk_count);
3276
0
            auto start_pos = chunk_end_entry.m_main_lin_index + 1 - chunk_data.m_chunk_count;
3277
0
            cluster->GetClusterRefs(*this, ret.first, start_pos);
3278
0
            std::reverse(ret.first.begin(), ret.first.end());
3279
0
        }
3280
0
        ret.second = chunk_end_entry.m_main_chunk_feerate;
3281
0
    }
3282
0
    return ret;
3283
0
}
3284
3285
std::vector<TxGraph::Ref*> TxGraphImpl::Trim() noexcept
3286
0
{
3287
0
    int level = GetTopLevel();
3288
0
    Assume(m_main_chunkindex_observers == 0 || level != 0);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3289
0
    std::vector<TxGraph::Ref*> ret;
3290
3291
    // Compute the groups of to-be-merged Clusters (which also applies all removals, and splits).
3292
0
    auto& clusterset = GetClusterSet(level);
3293
0
    if (clusterset.m_oversized == false) return ret;
3294
0
    GroupClusters(level);
3295
0
    Assume(clusterset.m_group_data.has_value());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3296
    // Nothing to do if not oversized.
3297
0
    Assume(clusterset.m_oversized.has_value());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3298
0
    if (clusterset.m_oversized == false) return ret;
3299
3300
    // In this function, would-be clusters (as precomputed in m_group_data by GroupClusters) are
3301
    // trimmed by removing transactions in them such that the resulting clusters satisfy the size
3302
    // and count limits.
3303
    //
3304
    // It works by defining for each would-be cluster a rudimentary linearization: at every point
3305
    // the highest-chunk-feerate remaining transaction is picked among those with no unmet
3306
    // dependencies. "Dependency" here means either a to-be-added dependency (m_deps_to_add), or
3307
    // an implicit dependency added between any two consecutive transaction in their current
3308
    // cluster linearization. So it can be seen as a "merge sort" of the chunks of the clusters,
3309
    // but respecting the dependencies being added.
3310
    //
3311
    // This rudimentary linearization is computed lazily, by putting all eligible (no unmet
3312
    // dependencies) transactions in a heap, and popping the highest-feerate one from it. Along the
3313
    // way, the counts and sizes of the would-be clusters up to that point are tracked (by
3314
    // partitioning the involved transactions using a union-find structure). Any transaction whose
3315
    // addition would cause a violation is removed, along with all their descendants.
3316
    //
3317
    // A next invocation of GroupClusters (after applying the removals) will compute the new
3318
    // resulting clusters, and none of them will violate the limits.
3319
3320
    /** All dependencies (both to be added ones, and implicit ones between consecutive transactions
3321
     *  in existing cluster linearizations), sorted by parent. */
3322
0
    std::vector<std::pair<GraphIndex, GraphIndex>> deps_by_parent;
3323
    /** Same, but sorted by child. */
3324
0
    std::vector<std::pair<GraphIndex, GraphIndex>> deps_by_child;
3325
    /** Information about all transactions involved in a Cluster group to be trimmed, sorted by
3326
     *  GraphIndex. It contains entries both for transactions that have already been included,
3327
     *  and ones that have not yet been. */
3328
0
    std::vector<TrimTxData> trim_data;
3329
    /** Iterators into trim_data, treated as a max heap according to cmp_fn below. Each entry is
3330
     *  a transaction that has not yet been included yet, but all its ancestors have. */
3331
0
    std::vector<std::vector<TrimTxData>::iterator> trim_heap;
3332
    /** The list of representatives of the partitions a given transaction depends on. */
3333
0
    std::vector<TrimTxData*> current_deps;
3334
3335
    /** Function to define the ordering of trim_heap. */
3336
0
    static constexpr auto cmp_fn = [](auto a, auto b) noexcept {
3337
        // Sort by increasing chunk feerate, and then by decreasing size.
3338
        // We do not need to sort by cluster or within clusters, because due to the implicit
3339
        // dependency between consecutive linearization elements, no two transactions from the
3340
        // same Cluster will ever simultaneously be in the heap.
3341
0
        return ByRatioNegSize{a->m_chunk_feerate} < ByRatioNegSize{b->m_chunk_feerate};
3342
0
    };
3343
3344
    /** Given a TrimTxData entry, find the representative of the partition it is in. */
3345
0
    static constexpr auto find_fn = [](TrimTxData* arg) noexcept {
3346
0
        while (arg != arg->m_uf_parent) {
3347
            // Replace pointer to parent with pointer to grandparent (path splitting).
3348
            // See https://en.wikipedia.org/wiki/Disjoint-set_data_structure#Finding_set_representatives.
3349
0
            auto par = arg->m_uf_parent;
3350
0
            arg->m_uf_parent = par->m_uf_parent;
3351
0
            arg = par;
3352
0
        }
3353
0
        return arg;
3354
0
    };
3355
3356
    /** Given two TrimTxData entries, union the partitions they are in, and return the
3357
     *  representative. */
3358
0
    static constexpr auto union_fn = [](TrimTxData* arg1, TrimTxData* arg2) noexcept {
3359
        // Replace arg1 and arg2 by their representatives.
3360
0
        auto rep1 = find_fn(arg1);
3361
0
        auto rep2 = find_fn(arg2);
3362
        // Bail out if both representatives are the same, because that means arg1 and arg2 are in
3363
        // the same partition already.
3364
0
        if (rep1 == rep2) return rep1;
3365
        // Pick the lower-count root to become a child of the higher-count one.
3366
        // See https://en.wikipedia.org/wiki/Disjoint-set_data_structure#Union_by_size.
3367
0
        if (rep1->m_uf_count < rep2->m_uf_count) std::swap(rep1, rep2);
3368
0
        rep2->m_uf_parent = rep1;
3369
        // Add the statistics of arg2 (which is no longer a representative) to those of arg1 (which
3370
        // is now the representative for both).
3371
0
        rep1->m_uf_size += rep2->m_uf_size;
3372
0
        rep1->m_uf_count += rep2->m_uf_count;
3373
0
        return rep1;
3374
0
    };
3375
3376
    /** Get iterator to TrimTxData entry for a given index. */
3377
0
    auto locate_fn = [&](GraphIndex index) noexcept {
3378
0
        auto it = std::lower_bound(trim_data.begin(), trim_data.end(), index, [](TrimTxData& elem, GraphIndex idx) noexcept {
3379
0
            return elem.m_index < idx;
3380
0
        });
3381
0
        Assume(it != trim_data.end() && it->m_index == index);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3382
0
        return it;
3383
0
    };
3384
3385
    // For each group of to-be-merged Clusters.
3386
0
    for (const auto& group_data : clusterset.m_group_data->m_groups) {
3387
0
        trim_data.clear();
3388
0
        trim_heap.clear();
3389
0
        deps_by_child.clear();
3390
0
        deps_by_parent.clear();
3391
3392
        // Gather trim data and implicit dependency data from all involved Clusters.
3393
0
        auto cluster_span = std::span{clusterset.m_group_data->m_group_clusters}
3394
0
                                .subspan(group_data.m_cluster_offset, group_data.m_cluster_count);
3395
0
        uint64_t size{0};
3396
0
        for (Cluster* cluster : cluster_span) {
3397
0
            MakeAcceptable(*cluster, cluster->GetLevel(*this));
3398
0
            size += cluster->AppendTrimData(trim_data, deps_by_child);
3399
0
        }
3400
        // If this group of Clusters does not violate any limits, continue to the next group.
3401
0
        if (trim_data.size() <= m_max_cluster_count && size <= m_max_cluster_size) continue;
3402
        // Sort the trim data by GraphIndex. In what follows, we will treat this sorted vector as
3403
        // a map from GraphIndex to TrimTxData via locate_fn, and its ordering will not change
3404
        // anymore.
3405
0
        std::ranges::sort(trim_data, [](auto& a, auto& b) noexcept { return a.m_index < b.m_index; });
3406
3407
        // Add the explicitly added dependencies to deps_by_child.
3408
0
        deps_by_child.insert(deps_by_child.end(),
3409
0
                             clusterset.m_deps_to_add.begin() + group_data.m_deps_offset,
3410
0
                             clusterset.m_deps_to_add.begin() + group_data.m_deps_offset + group_data.m_deps_count);
3411
3412
        // Sort deps_by_child by child transaction GraphIndex. The order will not be changed
3413
        // anymore after this.
3414
0
        std::ranges::sort(deps_by_child, [](auto& a, auto& b) noexcept { return a.second < b.second; });
3415
        // Fill m_parents_count and m_parents_offset in trim_data, as well as m_deps_left, and
3416
        // initially populate trim_heap. Because of the sort above, all dependencies involving the
3417
        // same child are grouped together, so a single linear scan suffices.
3418
0
        auto deps_it = deps_by_child.begin();
3419
0
        for (auto trim_it = trim_data.begin(); trim_it != trim_data.end(); ++trim_it) {
3420
0
            trim_it->m_parent_offset = deps_it - deps_by_child.begin();
3421
0
            trim_it->m_deps_left = 0;
3422
0
            while (deps_it != deps_by_child.end() && deps_it->second == trim_it->m_index) {
3423
0
                ++trim_it->m_deps_left;
3424
0
                ++deps_it;
3425
0
            }
3426
0
            trim_it->m_parent_count = trim_it->m_deps_left;
3427
            // If this transaction has no unmet dependencies, and is not oversized, add it to the
3428
            // heap (just append for now, the heapification happens below).
3429
0
            if (trim_it->m_deps_left == 0 && trim_it->m_tx_size <= m_max_cluster_size) {
3430
0
                trim_heap.push_back(trim_it);
3431
0
            }
3432
0
        }
3433
0
        Assume(deps_it == deps_by_child.end());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3434
3435
        // Construct deps_by_parent, sorted by parent transaction GraphIndex. The order will not be
3436
        // changed anymore after this.
3437
0
        deps_by_parent = deps_by_child;
3438
0
        std::ranges::sort(deps_by_parent, [](auto& a, auto& b) noexcept { return a.first < b.first; });
3439
        // Fill m_children_offset and m_children_count in trim_data. Because of the sort above, all
3440
        // dependencies involving the same parent are grouped together, so a single linear scan
3441
        // suffices.
3442
0
        deps_it = deps_by_parent.begin();
3443
0
        for (auto& trim_entry : trim_data) {
3444
0
            trim_entry.m_children_count = 0;
3445
0
            trim_entry.m_children_offset = deps_it - deps_by_parent.begin();
3446
0
            while (deps_it != deps_by_parent.end() && deps_it->first == trim_entry.m_index) {
3447
0
                ++trim_entry.m_children_count;
3448
0
                ++deps_it;
3449
0
            }
3450
0
        }
3451
0
        Assume(deps_it == deps_by_parent.end());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3452
3453
        // Build a heap of all transactions with 0 unmet dependencies.
3454
0
        std::make_heap(trim_heap.begin(), trim_heap.end(), cmp_fn);
3455
3456
        // Iterate over to-be-included transactions, and convert them to included transactions, or
3457
        // skip them if adding them would violate resource limits of the would-be cluster.
3458
        //
3459
        // It is possible that the heap empties without ever hitting either cluster limit, in case
3460
        // the implied graph (to be added dependencies plus implicit dependency between each
3461
        // original transaction and its predecessor in the linearization it came from) contains
3462
        // cycles. Such cycles will be removed entirely, because each of the transactions in the
3463
        // cycle permanently have unmet dependencies. However, this cannot occur in real scenarios
3464
        // where Trim() is called to deal with reorganizations that would violate cluster limits,
3465
        // as all added dependencies are in the same direction (from old mempool transactions to
3466
        // new from-block transactions); cycles require dependencies in both directions to be
3467
        // added.
3468
0
        while (!trim_heap.empty()) {
3469
            // Move the best remaining transaction to the end of trim_heap.
3470
0
            std::pop_heap(trim_heap.begin(), trim_heap.end(), cmp_fn);
3471
            // Pop it, and find its TrimTxData.
3472
0
            auto& entry = *trim_heap.back();
3473
0
            trim_heap.pop_back();
3474
3475
            // Initialize it as a singleton partition.
3476
0
            entry.m_uf_parent = &entry;
3477
0
            entry.m_uf_count = 1;
3478
0
            entry.m_uf_size = entry.m_tx_size;
3479
3480
            // Find the distinct transaction partitions this entry depends on.
3481
0
            current_deps.clear();
3482
0
            for (auto& [par, chl] : std::span{deps_by_child}.subspan(entry.m_parent_offset, entry.m_parent_count)) {
3483
0
                Assume(chl == entry.m_index);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3484
0
                current_deps.push_back(find_fn(&*locate_fn(par)));
3485
0
            }
3486
0
            std::ranges::sort(current_deps);
3487
0
            current_deps.erase(std::ranges::unique(current_deps).begin(), current_deps.end());
3488
3489
            // Compute resource counts.
3490
0
            uint32_t new_count = 1;
3491
0
            uint64_t new_size = entry.m_tx_size;
3492
0
            for (TrimTxData* ptr : current_deps) {
3493
0
                new_count += ptr->m_uf_count;
3494
0
                new_size += ptr->m_uf_size;
3495
0
            }
3496
            // Skip the entry if this would violate any limit.
3497
0
            if (new_count > m_max_cluster_count || new_size > m_max_cluster_size) continue;
3498
3499
            // Union the partitions this transaction and all its dependencies are in together.
3500
0
            auto rep = &entry;
3501
0
            for (TrimTxData* ptr : current_deps) rep = union_fn(ptr, rep);
3502
            // Mark the entry as included (so the loop below will not remove the transaction).
3503
0
            entry.m_deps_left = uint32_t(-1);
3504
            // Mark each to-be-added dependency involving this transaction as parent satisfied.
3505
0
            for (auto& [par, chl] : std::span{deps_by_parent}.subspan(entry.m_children_offset, entry.m_children_count)) {
3506
0
                Assume(par == entry.m_index);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3507
0
                auto chl_it = locate_fn(chl);
3508
                // Reduce the number of unmet dependencies of chl_it, and if that brings the number
3509
                // to zero, add it to the heap of includable transactions.
3510
0
                Assume(chl_it->m_deps_left > 0);
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3511
0
                if (--chl_it->m_deps_left == 0) {
3512
0
                    trim_heap.push_back(chl_it);
3513
0
                    std::push_heap(trim_heap.begin(), trim_heap.end(), cmp_fn);
3514
0
                }
3515
0
            }
3516
0
        }
3517
3518
        // Remove all the transactions that were not processed above. Because nothing gets
3519
        // processed until/unless all its dependencies are met, this automatically guarantees
3520
        // that if a transaction is removed, all its descendants, or would-be descendants, are
3521
        // removed as well.
3522
0
        for (const auto& trim_entry : trim_data) {
3523
0
            if (trim_entry.m_deps_left != uint32_t(-1)) {
3524
0
                ret.push_back(m_entries[trim_entry.m_index].m_ref);
3525
0
                clusterset.m_to_remove.push_back(trim_entry.m_index);
3526
0
            }
3527
0
        }
3528
0
    }
3529
0
    clusterset.m_group_data.reset();
3530
0
    clusterset.m_oversized = false;
3531
0
    Assume(!ret.empty());
Line
Count
Source
128
0
#define Assume(val) inline_assertion_check<false>(val, std::source_location::current(), #val)
3532
0
    return ret;
3533
0
}
3534
3535
size_t TxGraphImpl::GetMainMemoryUsage() noexcept
3536
593k
{
3537
    // Make sure splits/merges are applied, as memory usage may not be representative otherwise.
3538
593k
    SplitAll(/*up_to_level=*/0);
3539
593k
    ApplyDependencies(/*level=*/0);
3540
    // Compute memory usage
3541
593k
    size_t usage = /* From clusters */
3542
593k
                   m_main_clusterset.m_cluster_usage +
3543
                   /* From Entry objects. */
3544
593k
                   sizeof(Entry) * m_main_clusterset.m_txcount +
3545
                   /* From the chunk index. */
3546
593k
                   memusage::DynamicUsage(m_main_chunkindex);
3547
593k
    return usage;
3548
593k
}
3549
3550
} // namespace
3551
3552
TxGraph::Ref::~Ref()
3553
12.6k
{
3554
12.6k
    if (m_graph) {
3555
        // Inform the TxGraph about the Ref being destroyed.
3556
12.6k
        m_graph->UnlinkRef(m_index);
3557
12.6k
        m_graph = nullptr;
3558
12.6k
    }
3559
12.6k
}
3560
3561
TxGraph::Ref::Ref(Ref&& other) noexcept
3562
0
{
3563
    // Inform the TxGraph of other that its Ref is being moved.
3564
0
    if (other.m_graph) other.m_graph->UpdateRef(other.m_index, *this);
3565
    // Actually move the contents.
3566
0
    std::swap(m_graph, other.m_graph);
3567
0
    std::swap(m_index, other.m_index);
3568
0
}
3569
3570
std::unique_ptr<TxGraph> MakeTxGraph(
3571
    unsigned max_cluster_count,
3572
    uint64_t max_cluster_size,
3573
    uint64_t acceptable_cost,
3574
    const std::function<std::strong_ordering(const TxGraph::Ref&, const TxGraph::Ref&)>& fallback_order) noexcept
3575
925
{
3576
925
    return std::make_unique<TxGraphImpl>(
3577
925
        /*max_cluster_count=*/max_cluster_count,
3578
925
        /*max_cluster_size=*/max_cluster_size,
3579
925
        /*acceptable_cost=*/acceptable_cost,
3580
925
        /*fallback_order=*/fallback_order);
3581
925
}