Coverage Report

Created: 2025-02-21 14:36

/root/bitcoin/src/index/base.cpp
Line
Count
Source (jump to first uncovered line)
1
// Copyright (c) 2017-present The Bitcoin Core developers
2
// Distributed under the MIT software license, see the accompanying
3
// file COPYING or http://www.opensource.org/licenses/mit-license.php.
4
5
#include <chainparams.h>
6
#include <common/args.h>
7
#include <index/base.h>
8
#include <interfaces/chain.h>
9
#include <kernel/chain.h>
10
#include <logging.h>
11
#include <node/abort.h>
12
#include <node/blockstorage.h>
13
#include <node/context.h>
14
#include <node/database_args.h>
15
#include <node/interface_ui.h>
16
#include <tinyformat.h>
17
#include <util/string.h>
18
#include <util/thread.h>
19
#include <util/translation.h>
20
#include <validation.h> // For g_chainman
21
22
#include <string>
23
#include <utility>
24
25
constexpr uint8_t DB_BEST_BLOCK{'B'};
26
27
constexpr auto SYNC_LOG_INTERVAL{30s};
28
constexpr auto SYNC_LOCATOR_WRITE_INTERVAL{30s};
29
30
template <typename... Args>
31
void BaseIndex::FatalErrorf(util::ConstevalFormatString<sizeof...(Args)> fmt, const Args&... args)
32
0
{
33
0
    auto message = tfm::format(fmt, args...);
34
0
    node::AbortNode(m_chain->context()->shutdown_request, m_chain->context()->exit_status, Untranslated(message), m_chain->context()->warnings.get());
35
0
}
Unexecuted instantiation: _ZN9BaseIndex11FatalErrorfIJA5_cNSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEEEEEvN4util21ConstevalFormatStringIXsZT_EEEDpRKT_
Unexecuted instantiation: _ZN9BaseIndex11FatalErrorfIJA15_ciEEEvN4util21ConstevalFormatStringIXsZT_EEEDpRKT_
Unexecuted instantiation: _ZN9BaseIndex11FatalErrorfIJA15_cNSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEEEEEvN4util21ConstevalFormatStringIXsZT_EEEDpRKT_
Unexecuted instantiation: _ZN9BaseIndex11FatalErrorfIJA18_cNSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEEEEEvN4util21ConstevalFormatStringIXsZT_EEEDpRKT_
36
37
CBlockLocator GetLocator(interfaces::Chain& chain, const uint256& block_hash)
38
0
{
39
0
    CBlockLocator locator;
40
0
    bool found = chain.findBlock(block_hash, interfaces::FoundBlock().locator(locator));
41
0
    assert(found);
42
0
    assert(!locator.IsNull());
43
0
    return locator;
44
0
}
45
46
BaseIndex::DB::DB(const fs::path& path, size_t n_cache_size, bool f_memory, bool f_wipe, bool f_obfuscate) :
47
0
    CDBWrapper{DBParams{
48
0
        .path = path,
49
0
        .cache_bytes = n_cache_size,
50
0
        .memory_only = f_memory,
51
0
        .wipe_data = f_wipe,
52
0
        .obfuscate = f_obfuscate,
53
0
        .options = [] { DBOptions options; node::ReadDatabaseArgs(gArgs, options); return options; }()}}
54
0
{}
55
56
bool BaseIndex::DB::ReadBestBlock(CBlockLocator& locator) const
57
0
{
58
0
    bool success = Read(DB_BEST_BLOCK, locator);
59
0
    if (!success) {
60
0
        locator.SetNull();
61
0
    }
62
0
    return success;
63
0
}
64
65
void BaseIndex::DB::WriteBestBlock(CDBBatch& batch, const CBlockLocator& locator)
66
0
{
67
0
    batch.Write(DB_BEST_BLOCK, locator);
68
0
}
69
70
BaseIndex::BaseIndex(std::unique_ptr<interfaces::Chain> chain, std::string name)
71
0
    : m_chain{std::move(chain)}, m_name{std::move(name)} {}
72
73
BaseIndex::~BaseIndex()
74
0
{
75
0
    Interrupt();
76
0
    Stop();
77
0
}
78
79
bool BaseIndex::Init()
80
0
{
81
0
    AssertLockNotHeld(cs_main);
82
83
    // May need reset if index is being restarted.
84
0
    m_interrupt.reset();
85
86
    // m_chainstate member gives indexing code access to node internals. It is
87
    // removed in followup https://github.com/bitcoin/bitcoin/pull/24230
88
0
    m_chainstate = WITH_LOCK(::cs_main,
89
0
        return &m_chain->context()->chainman->GetChainstateForIndexing());
90
    // Register to validation interface before setting the 'm_synced' flag, so that
91
    // callbacks are not missed once m_synced is true.
92
0
    m_chain->context()->validation_signals->RegisterValidationInterface(this);
93
94
0
    CBlockLocator locator;
95
0
    if (!GetDB().ReadBestBlock(locator)) {
96
0
        locator.SetNull();
97
0
    }
98
99
0
    LOCK(cs_main);
100
0
    CChain& index_chain = m_chainstate->m_chain;
101
102
0
    if (locator.IsNull()) {
103
0
        SetBestBlockIndex(nullptr);
104
0
    } else {
105
        // Setting the best block to the locator's top block. If it is not part of the
106
        // best chain, we will rewind to the fork point during index sync
107
0
        const CBlockIndex* locator_index{m_chainstate->m_blockman.LookupBlockIndex(locator.vHave.at(0))};
108
0
        if (!locator_index) {
109
0
            return InitError(Untranslated(strprintf("%s: best block of the index not found. Please rebuild the index.", GetName())));
110
0
        }
111
0
        SetBestBlockIndex(locator_index);
112
0
    }
113
114
    // Child init
115
0
    const CBlockIndex* start_block = m_best_block_index.load();
116
0
    if (!CustomInit(start_block ? std::make_optional(interfaces::BlockRef{start_block->GetBlockHash(), start_block->nHeight}) : std::nullopt)) {
117
0
        return false;
118
0
    }
119
120
    // Note: this will latch to true immediately if the user starts up with an empty
121
    // datadir and an index enabled. If this is the case, indexation will happen solely
122
    // via `BlockConnected` signals until, possibly, the next restart.
123
0
    m_synced = start_block == index_chain.Tip();
124
0
    m_init = true;
125
0
    return true;
126
0
}
127
128
static const CBlockIndex* NextSyncBlock(const CBlockIndex* pindex_prev, CChain& chain) EXCLUSIVE_LOCKS_REQUIRED(cs_main)
129
0
{
130
0
    AssertLockHeld(cs_main);
131
132
0
    if (!pindex_prev) {
133
0
        return chain.Genesis();
134
0
    }
135
136
0
    const CBlockIndex* pindex = chain.Next(pindex_prev);
137
0
    if (pindex) {
138
0
        return pindex;
139
0
    }
140
141
0
    return chain.Next(chain.FindFork(pindex_prev));
142
0
}
143
144
void BaseIndex::Sync()
145
0
{
146
0
    const CBlockIndex* pindex = m_best_block_index.load();
147
0
    if (!m_synced) {
148
0
        std::chrono::steady_clock::time_point last_log_time{0s};
149
0
        std::chrono::steady_clock::time_point last_locator_write_time{0s};
150
0
        while (true) {
151
0
            if (m_interrupt) {
152
0
                LogPrintf("%s: m_interrupt set; exiting ThreadSync\n", GetName());
153
154
0
                SetBestBlockIndex(pindex);
155
                // No need to handle errors in Commit. If it fails, the error will be already be
156
                // logged. The best way to recover is to continue, as index cannot be corrupted by
157
                // a missed commit to disk for an advanced index state.
158
0
                Commit();
159
0
                return;
160
0
            }
161
162
0
            const CBlockIndex* pindex_next = WITH_LOCK(cs_main, return NextSyncBlock(pindex, m_chainstate->m_chain));
163
            // If pindex_next is null, it means pindex is the chain tip, so
164
            // commit data indexed so far.
165
0
            if (!pindex_next) {
166
0
                SetBestBlockIndex(pindex);
167
                // No need to handle errors in Commit. See rationale above.
168
0
                Commit();
169
170
                // If pindex is still the chain tip after committing, exit the
171
                // sync loop. It is important for cs_main to be locked while
172
                // setting m_synced = true, otherwise a new block could be
173
                // attached while m_synced is still false, and it would not be
174
                // indexed.
175
0
                LOCK(::cs_main);
176
0
                pindex_next = NextSyncBlock(pindex, m_chainstate->m_chain);
177
0
                if (!pindex_next) {
178
0
                    m_synced = true;
179
0
                    break;
180
0
                }
181
0
            }
182
0
            if (pindex_next->pprev != pindex && !Rewind(pindex, pindex_next->pprev)) {
183
0
                FatalErrorf("%s: Failed to rewind index %s to a previous chain tip", __func__, GetName());
184
0
                return;
185
0
            }
186
0
            pindex = pindex_next;
187
188
189
0
            CBlock block;
190
0
            interfaces::BlockInfo block_info = kernel::MakeBlockInfo(pindex);
191
0
            if (!m_chainstate->m_blockman.ReadBlock(block, *pindex)) {
192
0
                FatalErrorf("%s: Failed to read block %s from disk",
193
0
                           __func__, pindex->GetBlockHash().ToString());
194
0
                return;
195
0
            } else {
196
0
                block_info.data = &block;
197
0
            }
198
0
            if (!CustomAppend(block_info)) {
199
0
                FatalErrorf("%s: Failed to write block %s to index database",
200
0
                           __func__, pindex->GetBlockHash().ToString());
201
0
                return;
202
0
            }
203
204
0
            auto current_time{std::chrono::steady_clock::now()};
205
0
            if (last_log_time + SYNC_LOG_INTERVAL < current_time) {
206
0
                LogPrintf("Syncing %s with block chain from height %d\n",
207
0
                          GetName(), pindex->nHeight);
208
0
                last_log_time = current_time;
209
0
            }
210
211
0
            if (last_locator_write_time + SYNC_LOCATOR_WRITE_INTERVAL < current_time) {
212
0
                SetBestBlockIndex(pindex);
213
0
                last_locator_write_time = current_time;
214
                // No need to handle errors in Commit. See rationale above.
215
0
                Commit();
216
0
            }
217
0
        }
218
0
    }
219
220
0
    if (pindex) {
221
0
        LogPrintf("%s is enabled at height %d\n", GetName(), pindex->nHeight);
222
0
    } else {
223
0
        LogPrintf("%s is enabled\n", GetName());
224
0
    }
225
0
}
226
227
bool BaseIndex::Commit()
228
0
{
229
    // Don't commit anything if we haven't indexed any block yet
230
    // (this could happen if init is interrupted).
231
0
    bool ok = m_best_block_index != nullptr;
232
0
    if (ok) {
233
0
        CDBBatch batch(GetDB());
234
0
        ok = CustomCommit(batch);
235
0
        if (ok) {
236
0
            GetDB().WriteBestBlock(batch, GetLocator(*m_chain, m_best_block_index.load()->GetBlockHash()));
237
0
            ok = GetDB().WriteBatch(batch);
238
0
        }
239
0
    }
240
0
    if (!ok) {
241
0
        LogError("%s: Failed to commit latest %s state\n", __func__, GetName());
242
0
        return false;
243
0
    }
244
0
    return true;
245
0
}
246
247
bool BaseIndex::Rewind(const CBlockIndex* current_tip, const CBlockIndex* new_tip)
248
0
{
249
0
    assert(current_tip == m_best_block_index);
250
0
    assert(current_tip->GetAncestor(new_tip->nHeight) == new_tip);
251
252
0
    if (!CustomRewind({current_tip->GetBlockHash(), current_tip->nHeight}, {new_tip->GetBlockHash(), new_tip->nHeight})) {
253
0
        return false;
254
0
    }
255
256
    // In the case of a reorg, ensure persisted block locator is not stale.
257
    // Pruning has a minimum of 288 blocks-to-keep and getting the index
258
    // out of sync may be possible but a users fault.
259
    // In case we reorg beyond the pruned depth, ReadBlock would
260
    // throw and lead to a graceful shutdown
261
0
    SetBestBlockIndex(new_tip);
262
0
    if (!Commit()) {
263
        // If commit fails, revert the best block index to avoid corruption.
264
0
        SetBestBlockIndex(current_tip);
265
0
        return false;
266
0
    }
267
268
0
    return true;
269
0
}
270
271
void BaseIndex::BlockConnected(ChainstateRole role, const std::shared_ptr<const CBlock>& block, const CBlockIndex* pindex)
272
0
{
273
    // Ignore events from the assumed-valid chain; we will process its blocks
274
    // (sequentially) after it is fully verified by the background chainstate. This
275
    // is to avoid any out-of-order indexing.
276
    //
277
    // TODO at some point we could parameterize whether a particular index can be
278
    // built out of order, but for now just do the conservative simple thing.
279
0
    if (role == ChainstateRole::ASSUMEDVALID) {
280
0
        return;
281
0
    }
282
283
    // Ignore BlockConnected signals until we have fully indexed the chain.
284
0
    if (!m_synced) {
285
0
        return;
286
0
    }
287
288
0
    const CBlockIndex* best_block_index = m_best_block_index.load();
289
0
    if (!best_block_index) {
290
0
        if (pindex->nHeight != 0) {
291
0
            FatalErrorf("%s: First block connected is not the genesis block (height=%d)",
292
0
                       __func__, pindex->nHeight);
293
0
            return;
294
0
        }
295
0
    } else {
296
        // Ensure block connects to an ancestor of the current best block. This should be the case
297
        // most of the time, but may not be immediately after the sync thread catches up and sets
298
        // m_synced. Consider the case where there is a reorg and the blocks on the stale branch are
299
        // in the ValidationInterface queue backlog even after the sync thread has caught up to the
300
        // new chain tip. In this unlikely event, log a warning and let the queue clear.
301
0
        if (best_block_index->GetAncestor(pindex->nHeight - 1) != pindex->pprev) {
302
0
            LogPrintf("%s: WARNING: Block %s does not connect to an ancestor of "
303
0
                      "known best chain (tip=%s); not updating index\n",
304
0
                      __func__, pindex->GetBlockHash().ToString(),
305
0
                      best_block_index->GetBlockHash().ToString());
306
0
            return;
307
0
        }
308
0
        if (best_block_index != pindex->pprev && !Rewind(best_block_index, pindex->pprev)) {
309
0
            FatalErrorf("%s: Failed to rewind index %s to a previous chain tip",
310
0
                       __func__, GetName());
311
0
            return;
312
0
        }
313
0
    }
314
0
    interfaces::BlockInfo block_info = kernel::MakeBlockInfo(pindex, block.get());
315
0
    if (CustomAppend(block_info)) {
316
        // Setting the best block index is intentionally the last step of this
317
        // function, so BlockUntilSyncedToCurrentChain callers waiting for the
318
        // best block index to be updated can rely on the block being fully
319
        // processed, and the index object being safe to delete.
320
0
        SetBestBlockIndex(pindex);
321
0
    } else {
322
0
        FatalErrorf("%s: Failed to write block %s to index",
323
0
                   __func__, pindex->GetBlockHash().ToString());
324
0
        return;
325
0
    }
326
0
}
327
328
void BaseIndex::ChainStateFlushed(ChainstateRole role, const CBlockLocator& locator)
329
0
{
330
    // Ignore events from the assumed-valid chain; we will process its blocks
331
    // (sequentially) after it is fully verified by the background chainstate.
332
0
    if (role == ChainstateRole::ASSUMEDVALID) {
333
0
        return;
334
0
    }
335
336
0
    if (!m_synced) {
337
0
        return;
338
0
    }
339
340
0
    const uint256& locator_tip_hash = locator.vHave.front();
341
0
    const CBlockIndex* locator_tip_index;
342
0
    {
343
0
        LOCK(cs_main);
344
0
        locator_tip_index = m_chainstate->m_blockman.LookupBlockIndex(locator_tip_hash);
345
0
    }
346
347
0
    if (!locator_tip_index) {
348
0
        FatalErrorf("%s: First block (hash=%s) in locator was not found",
349
0
                   __func__, locator_tip_hash.ToString());
350
0
        return;
351
0
    }
352
353
    // This checks that ChainStateFlushed callbacks are received after BlockConnected. The check may fail
354
    // immediately after the sync thread catches up and sets m_synced. Consider the case where
355
    // there is a reorg and the blocks on the stale branch are in the ValidationInterface queue
356
    // backlog even after the sync thread has caught up to the new chain tip. In this unlikely
357
    // event, log a warning and let the queue clear.
358
0
    const CBlockIndex* best_block_index = m_best_block_index.load();
359
0
    if (best_block_index->GetAncestor(locator_tip_index->nHeight) != locator_tip_index) {
360
0
        LogPrintf("%s: WARNING: Locator contains block (hash=%s) not on known best "
361
0
                  "chain (tip=%s); not writing index locator\n",
362
0
                  __func__, locator_tip_hash.ToString(),
363
0
                  best_block_index->GetBlockHash().ToString());
364
0
        return;
365
0
    }
366
367
    // No need to handle errors in Commit. If it fails, the error will be already be logged. The
368
    // best way to recover is to continue, as index cannot be corrupted by a missed commit to disk
369
    // for an advanced index state.
370
0
    Commit();
371
0
}
372
373
bool BaseIndex::BlockUntilSyncedToCurrentChain() const
374
0
{
375
0
    AssertLockNotHeld(cs_main);
376
377
0
    if (!m_synced) {
378
0
        return false;
379
0
    }
380
381
0
    {
382
        // Skip the queue-draining stuff if we know we're caught up with
383
        // m_chain.Tip().
384
0
        LOCK(cs_main);
385
0
        const CBlockIndex* chain_tip = m_chainstate->m_chain.Tip();
386
0
        const CBlockIndex* best_block_index = m_best_block_index.load();
387
0
        if (best_block_index->GetAncestor(chain_tip->nHeight) == chain_tip) {
388
0
            return true;
389
0
        }
390
0
    }
391
392
0
    LogPrintf("%s: %s is catching up on block notifications\n", __func__, GetName());
393
0
    m_chain->context()->validation_signals->SyncWithValidationInterfaceQueue();
394
0
    return true;
395
0
}
396
397
void BaseIndex::Interrupt()
398
0
{
399
0
    m_interrupt();
400
0
}
401
402
bool BaseIndex::StartBackgroundSync()
403
0
{
404
0
    if (!m_init) throw std::logic_error("Error: Cannot start a non-initialized index");
405
406
0
    m_thread_sync = std::thread(&util::TraceThread, GetName(), [this] { Sync(); });
407
0
    return true;
408
0
}
409
410
void BaseIndex::Stop()
411
0
{
412
0
    if (m_chain->context()->validation_signals) {
413
0
        m_chain->context()->validation_signals->UnregisterValidationInterface(this);
414
0
    }
415
416
0
    if (m_thread_sync.joinable()) {
417
0
        m_thread_sync.join();
418
0
    }
419
0
}
420
421
IndexSummary BaseIndex::GetSummary() const
422
0
{
423
0
    IndexSummary summary{};
424
0
    summary.name = GetName();
425
0
    summary.synced = m_synced;
426
0
    if (const auto& pindex = m_best_block_index.load()) {
427
0
        summary.best_block_height = pindex->nHeight;
428
0
        summary.best_block_hash = pindex->GetBlockHash();
429
0
    } else {
430
0
        summary.best_block_height = 0;
431
0
        summary.best_block_hash = m_chain->getBlockHash(0);
432
0
    }
433
0
    return summary;
434
0
}
435
436
void BaseIndex::SetBestBlockIndex(const CBlockIndex* block)
437
0
{
438
0
    assert(!m_chainstate->m_blockman.IsPruneMode() || AllowPrune());
439
440
0
    if (AllowPrune() && block) {
441
0
        node::PruneLockInfo prune_lock;
442
0
        prune_lock.height_first = block->nHeight;
443
0
        WITH_LOCK(::cs_main, m_chainstate->m_blockman.UpdatePruneLock(GetName(), prune_lock));
444
0
    }
445
446
    // Intentionally set m_best_block_index as the last step in this function,
447
    // after updating prune locks above, and after making any other references
448
    // to *this, so the BlockUntilSyncedToCurrentChain function (which checks
449
    // m_best_block_index as an optimization) can be used to wait for the last
450
    // BlockConnected notification and safely assume that prune locks are
451
    // updated and that the index object is safe to delete.
452
0
    m_best_block_index = block;
453
0
}