// Copyright (c) 2014-2019, The Monero Project // Copyright (c) 2018, The Loki Project // // All rights reserved. // // Redistribution and use in source and binary forms, with or without modification, are // permitted provided that the following conditions are met: // // 1. Redistributions of source code must retain the above copyright notice, this list of // conditions and the following disclaimer. // // 2. Redistributions in binary form must reproduce the above copyright notice, this list // of conditions and the following disclaimer in the documentation and/or other // materials provided with the distribution. // // 3. Neither the name of the copyright holder nor the names of its contributors may be // used to endorse or promote products derived from this software without specific // prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY // EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF // MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL // THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, // PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS // INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, // STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF // THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. #include #include #include #include #include #include #include #include "blocks/blocks.h" #include "bootstrap_file.h" #include "bootstrap_serialization.h" #include "common/fs-format.h" #include "common/hex.h" #include "cryptonote_basic/cryptonote_format_utils.h" #include "cryptonote_core/cryptonote_core.h" #include "cryptonote_core/uptime_proof.h" #include "cryptonote_protocol/quorumnet.h" #include "epee/misc_log_ex.h" #include "logging/oxen_logger.h" #include "serialization/binary_utils.h" using namespace blockchain_utils; namespace { // CONFIG bool opt_batch = true; bool opt_verify = true; // use add_new_block, which does verification before calling add_block bool opt_resume = true; bool opt_testnet = true; bool opt_devnet = true; // number of blocks per batch transaction // adjustable through command-line argument according to available RAM #if ARCH_WIDTH != 32 uint64_t db_batch_size = 20000; #else // set a lower default batch size, pending possible LMDB issue with large transaction size uint64_t db_batch_size = 100; #endif // when verifying, use a smaller default batch size so progress is more // frequently saved uint64_t db_batch_size_verify = 5000; std::string refresh_string = "\r \r"; const command_line::arg_descriptor arg_recalculate_difficulty = { "recalculate-difficulty", "Recalculate per-block difficulty starting from the height specified", // This is now enabled by default because the network broke at 526483 because of divergent // difficulty values (and the chain that kept going violated the correct difficulty, and got // checkpointed multiple times because enough of the network followed it). false}; } // namespace namespace po = boost::program_options; using namespace cryptonote; static auto logcat = log::Cat("bcutil"); // db_mode: safe, fast, fastest int get_db_flags_from_mode(const std::string& db_mode) { int db_flags = 0; if (db_mode == "safe") db_flags = DBF_SAFE; else if (db_mode == "fast") db_flags = DBF_FAST; else if (db_mode == "fastest") db_flags = DBF_FASTEST; return db_flags; } int pop_blocks(cryptonote::core& core, int num_blocks) { bool use_batch = opt_batch; if (use_batch) core.get_blockchain_storage().get_db().batch_start(); try { core.get_blockchain_storage().pop_blocks(num_blocks); if (use_batch) { core.get_blockchain_storage().get_db().batch_stop(); core.get_blockchain_storage().get_db().show_stats(); } } catch (const std::exception& e) { // There was an error, so don't commit pending data. // Destructor will abort write txn. } return num_blocks; } int check_flush(cryptonote::core& core, std::vector& blocks, bool force) { if (blocks.empty()) return 0; if (!force && blocks.size() < db_batch_size) return 0; // wait till we can verify a full HOH without extra, for speed uint64_t new_height = core.get_blockchain_storage().get_db().height() + blocks.size(); if (!force && new_height % HASH_OF_HASHES_STEP) return 0; std::vector hashes; for (const auto& b : blocks) { cryptonote::block block; if (!parse_and_validate_block_from_blob(b.block, block)) { log::error( logcat, "Failed to parse block: {}", tools::type_to_hex(get_blob_hash(b.block))); core.cleanup_handle_incoming_blocks(); return 1; } hashes.push_back(cryptonote::get_block_hash(block)); } core.prevalidate_block_hashes(core.get_blockchain_storage().get_db().height(), hashes); // TODO(doyle): Checkpointing std::vector pblocks; if (!core.prepare_handle_incoming_blocks(blocks, pblocks)) { log::error(logcat, "Failed to prepare to add blocks"); return 1; } if (!pblocks.empty() && pblocks.size() != blocks.size()) { log::error(logcat, "Unexpected parsed blocks size"); core.cleanup_handle_incoming_blocks(); return 1; } size_t blockidx = 0; for (const block_complete_entry& block_entry : blocks) { // process transactions for (auto& tx_blob : block_entry.txs) { tx_verification_context tvc{}; core.handle_incoming_tx(tx_blob, tvc, tx_pool_options::from_block()); if (tvc.m_verifivation_failed) { log::error( logcat, "transaction verification failed, tx_id = {}", tools::type_to_hex(get_blob_hash(tx_blob))); core.cleanup_handle_incoming_blocks(); return 1; } } // process block block_verification_context bvc{}; core.handle_incoming_block( block_entry.block, pblocks.empty() ? NULL : &pblocks[blockidx++], bvc, nullptr /*checkpoint*/, false); // <--- process block if (bvc.m_verifivation_failed) { log::error( logcat, "Block verification failed, id = {}", tools::type_to_hex(get_blob_hash(block_entry.block))); core.cleanup_handle_incoming_blocks(); return 1; } if (bvc.m_marked_as_orphaned) { log::error(logcat, "Block received at sync phase was marked as orphaned"); core.cleanup_handle_incoming_blocks(); return 1; } } // each download block if (!core.cleanup_handle_incoming_blocks()) return 1; blocks.clear(); return 0; } int import_from_file( cryptonote::core& core, const fs::path& import_file_path, uint64_t block_stop = 0) { // Reset stats, in case we're using newly created db, accumulating stats // from addition of genesis block. // This aligns internal db counts with importer counts. core.get_blockchain_storage().get_db().reset_stats(); if (std::error_code ec; !fs::exists(import_file_path, ec)) { log::error(logcat, "bootstrap file not found: {}", import_file_path); return false; } uint64_t start_height = 1, seek_height; if (opt_resume) start_height = core.get_blockchain_storage().get_current_blockchain_height(); seek_height = start_height; BootstrapFile bootstrap; std::streampos pos; // BootstrapFile bootstrap(import_file_path); uint64_t total_source_blocks = bootstrap.count_blocks(import_file_path, pos, seek_height); log::info( logcat, "bootstrap file last block number: {} (zero-based height) total blocks: {}", total_source_blocks - 1, total_source_blocks); if (total_source_blocks - 1 <= start_height) { return false; } std::cout << "\nPreparing to read blocks...\n\n"; fs::ifstream import_file{import_file_path, std::ios::binary}; if (import_file.fail()) { log::error(logcat, "import_file.open() fail"); return false; } // 4 byte magic + (currently) 1024 byte header structures bootstrap.seek_to_first_chunk(import_file); char buffer1[1024]; char buffer_block[BUFFER_SIZE]; block b; transaction tx; int quit = 0; uint64_t bytes_read; // Note that a new blockchain will start with block number 0 (total blocks: 1) // due to genesis block being added at initialization. if (!block_stop) { block_stop = total_source_blocks - 1; } // These are what we'll try to use, and they don't have to be a determination // from source and destination blockchains, but those are the defaults. log::info(logcat, "start block: {} stop block: {}", start_height, block_stop); bool use_batch = opt_batch && !opt_verify; log::info(logcat, "Reading blockchain from bootstrap file..."); std::cout << "\n"; std::vector blocks; uint64_t h = 0; uint64_t num_imported = 0; // Skip to start_height before we start adding. { bool q2 = false; import_file.seekg(pos); bytes_read = bootstrap.count_bytes(import_file, start_height - seek_height, h, q2); if (q2) { quit = 2; goto quitting; } h = start_height; } if (use_batch) { uint64_t bytes, h2; bool q2; pos = import_file.tellg(); bytes = bootstrap.count_bytes(import_file, db_batch_size, h2, q2); if (import_file.eof()) import_file.clear(); import_file.seekg(pos); core.get_blockchain_storage().get_db().batch_start(db_batch_size, bytes); } while (!quit) { uint32_t chunk_size; import_file.read(buffer1, sizeof(chunk_size)); // TODO: bootstrap.read_chunk(); if (!import_file) { std::cout << refresh_string; log::info(logcat, "End of file reached"); quit = 1; break; } bytes_read += sizeof(chunk_size); try { serialization::parse_binary(std::string_view{buffer1, sizeof(chunk_size)}, chunk_size); } catch (const std::exception& e) { throw std::runtime_error("Error in deserialization of chunk size: "s + e.what()); } log::debug(logcat, "chunk_size: {}", chunk_size); if (chunk_size > BUFFER_SIZE) { log::warning( logcat, "WARNING: chunk_size {} > BUFFER_SIZE {}", chunk_size, BUFFER_SIZE); throw std::runtime_error("Aborting: chunk size exceeds buffer size"); } if (chunk_size > CHUNK_SIZE_WARNING_THRESHOLD) { log::info(logcat, "NOTE: chunk_size {} > {}", chunk_size, CHUNK_SIZE_WARNING_THRESHOLD); } else if (chunk_size == 0) { log::error(logcat, "ERROR: chunk_size == 0"); return 2; } import_file.read(buffer_block, chunk_size); if (!import_file) { if (import_file.eof()) { std::cout << refresh_string; log::info(logcat, "End of file reached - file was truncated"); quit = 1; break; } else { log::error( logcat, "ERROR: unexpected end of file: bytes read before error: {} of chunk_size " "{}", import_file.gcount(), chunk_size); return 2; } } bytes_read += chunk_size; log::debug(logcat, "Total bytes read: {}", bytes_read); if (h > block_stop) { std::cout << refresh_string << "block " << h - 1 << " / " << block_stop << "\n" << std::endl; log::info( logcat, "Specified block number reached - stopping. block: {} total blocks: {}", h - 1, h); quit = 1; break; } try { bootstrap::block_package bp; try { serialization::parse_binary(std::string_view{buffer_block, chunk_size}, bp); } catch (const std::exception& e) { throw std::runtime_error("Error in deserialization of chunk"s + e.what()); } int display_interval = 1000; int progress_interval = 10; // NOTE: use of NUM_BLOCKS_PER_CHUNK is a placeholder in case multi-block chunks are // later supported. for (size_t chunk_ind = 0; chunk_ind < NUM_BLOCKS_PER_CHUNK; ++chunk_ind) { ++h; if ((h - 1) % display_interval == 0) { std::cout << refresh_string; log::debug(logcat, "loading block number {}", h - 1); } else { log::debug(logcat, "loading block number {}", h - 1); } b = bp.block; log::debug(logcat, "block prev_id: {}\n", b.prev_id); if ((h - 1) % progress_interval == 0) { std::cout << refresh_string << "block " << h - 1 << " / " << block_stop << "\r" << std::flush; } if (opt_verify) { std::string block; cryptonote::block_to_blob(bp.block, block); std::vector txs; for (const auto& tx : bp.txs) { txs.push_back(std::string()); cryptonote::tx_to_blob(tx, txs.back()); } blocks.push_back({block, txs}); int ret = check_flush(core, blocks, false); if (ret) { quit = 2; // make sure we don't commit partial block data break; } } else { std::vector> txs; std::vector archived_txs; archived_txs = bp.txs; // tx number 1: coinbase tx // tx number 2 onwards: archived_txs for (const transaction& tx : archived_txs) { // add blocks with verification. // for Blockchain and blockchain_storage add_new_block(). // for add_block() method, without (much) processing. // don't add coinbase transaction to txs. // // because add_block() calls // add_transaction(blk_hash, blk.miner_tx) first, and // then a for loop for the transactions in txs. txs.push_back(std::make_pair(tx, tx_to_blob(tx))); } size_t block_weight; difficulty_type cumulative_difficulty; uint64_t coins_generated; block_weight = bp.block_weight; cumulative_difficulty = bp.cumulative_difficulty; coins_generated = bp.coins_generated; try { uint64_t long_term_block_weight = core.get_blockchain_storage().get_next_long_term_block_weight( block_weight); core.get_blockchain_storage().get_db().add_block( std::make_pair(b, block_to_blob(b)), block_weight, long_term_block_weight, cumulative_difficulty, coins_generated, txs); } catch (const std::exception& e) { std::cout << refresh_string; log::error(logcat, "Error adding block to blockchain: {}", e.what()); quit = 2; // make sure we don't commit partial block data break; } if (use_batch) { if ((h - 1) % db_batch_size == 0) { uint64_t bytes, h2; bool q2; std::cout << refresh_string; // zero-based height std::cout << "\n[- batch commit at height " << h - 1 << " -]\n"; core.get_blockchain_storage().get_db().batch_stop(); pos = import_file.tellg(); bytes = bootstrap.count_bytes(import_file, db_batch_size, h2, q2); import_file.seekg(pos); core.get_blockchain_storage().get_db().batch_start( db_batch_size, bytes); std::cout << "\n"; core.get_blockchain_storage().get_db().show_stats(); } } } ++num_imported; } } catch (const std::exception& e) { std::cout << refresh_string; log::error(logcat, "exception while reading from file, height={}: {}", h, e.what()); return 2; } } // while quitting: import_file.close(); if (opt_verify) { int ret = check_flush(core, blocks, true); if (ret) return ret; } if (use_batch) { if (quit > 1) { // There was an error, so don't commit pending data. // Destructor will abort write txn. } else { core.get_blockchain_storage().get_db().batch_stop(); } } core.get_blockchain_storage().get_db().show_stats(); log::info(logcat, "Number of blocks imported: {}", num_imported); if (h > 0) // TODO: if there was an error, the last added block is probably at zero-based height h-2 log::info(logcat, "Finished at block: {} total blocks: {}", h - 1, h); std::cout << "\n"; return 0; } int main(int argc, char* argv[]) { TRY_ENTRY(); epee::string_tools::set_module_name_and_folder(argv[0]); uint64_t num_blocks = 0; uint64_t block_stop = 0; std::string m_config_folder; std::string db_arg_str; tools::on_startup(); auto opt_size = command_line::boost_option_sizes(); po::options_description desc_cmd_only("Command line options", opt_size.first, opt_size.second); po::options_description desc_cmd_sett( "Command line options and settings options", opt_size.first, opt_size.second); const command_line::arg_descriptor arg_input_file = { "input-file", "Specify input file", "", true}; const command_line::arg_descriptor arg_log_level = { "log-level", "0-4 or categories", ""}; const command_line::arg_descriptor arg_block_stop = { "block-stop", "Stop at block number", block_stop}; const command_line::arg_descriptor arg_batch_size = {"batch-size", "", db_batch_size}; const command_line::arg_descriptor arg_pop_blocks = { "pop-blocks", "Remove blocks from end of blockchain", num_blocks}; const command_line::arg_descriptor arg_count_blocks = { "count-blocks", "Count blocks in bootstrap file and exit", false}; const command_line::arg_descriptor arg_noverify = { "dangerous-unverified-import", "Blindly trust the import file and use potentially malicious blocks and transactions " "during import (only enable if you exported the file yourself)", false}; const command_line::arg_descriptor arg_batch = { "batch", "Batch transactions for faster import", true}; const command_line::arg_descriptor arg_resume = { "resume", "Resume from current height if output database already exists", true}; command_line::add_arg(desc_cmd_sett, arg_input_file); command_line::add_arg(desc_cmd_sett, arg_log_level); command_line::add_arg(desc_cmd_sett, arg_batch_size); command_line::add_arg(desc_cmd_sett, arg_block_stop); command_line::add_arg(desc_cmd_only, arg_count_blocks); command_line::add_arg(desc_cmd_only, arg_pop_blocks); command_line::add_arg(desc_cmd_only, command_line::arg_help); command_line::add_arg(desc_cmd_only, arg_recalculate_difficulty); // call add_options() directly for these arguments since // command_line helpers support only boolean switch, not boolean argument desc_cmd_sett.add_options()( arg_noverify.name, make_semantic(arg_noverify), arg_noverify.description)( arg_batch.name, make_semantic(arg_batch), arg_batch.description)( arg_resume.name, make_semantic(arg_resume), arg_resume.description); po::options_description desc_options("Allowed options"); desc_options.add(desc_cmd_only).add(desc_cmd_sett); cryptonote::core::init_options(desc_options); po::variables_map vm; bool r = command_line::handle_error_helper(desc_options, [&]() { po::store(po::parse_command_line(argc, argv, desc_options), vm); po::notify(vm); return true; }); if (!r) return 1; opt_verify = !command_line::get_arg(vm, arg_noverify); opt_batch = command_line::get_arg(vm, arg_batch); opt_resume = command_line::get_arg(vm, arg_resume); block_stop = command_line::get_arg(vm, arg_block_stop); db_batch_size = command_line::get_arg(vm, arg_batch_size); if (command_line::get_arg(vm, command_line::arg_help)) { std::cout << "Oxen '" << OXEN_RELEASE_NAME << "' (v" << OXEN_VERSION_FULL << ")\n\n"; std::cout << desc_options << std::endl; return 1; } if (!opt_batch && !command_line::is_arg_defaulted(vm, arg_batch_size)) { std::cerr << "Error: batch-size set, but batch option not enabled\n"; return 1; } if (!db_batch_size) { std::cerr << "Error: batch-size must be > 0\n"; return 1; } if (opt_verify && command_line::is_arg_defaulted(vm, arg_batch_size)) { // usually want batch size default lower if verify on, so progress can be // frequently saved. // // currently, with Windows, default batch size is low, so ignore // default db_batch_size_verify unless it's even lower if (db_batch_size > db_batch_size_verify) { db_batch_size = db_batch_size_verify; } } opt_testnet = command_line::get_arg(vm, cryptonote::arg_testnet_on); opt_devnet = command_line::get_arg(vm, cryptonote::arg_devnet_on); if (opt_testnet && opt_devnet) { std::cerr << "Error: Can't specify more than one of --testnet and --devnet\n"; return 1; } m_config_folder = command_line::get_arg(vm, cryptonote::arg_data_dir); auto log_file_path = m_config_folder + "oxen-blockchain-import.log"; log::Level log_level; if (auto level = oxen::logging::parse_level(command_line::get_arg(vm, arg_log_level).c_str())) { log_level = *level; } else { std::cerr << "Incorrect log level: " << command_line::get_arg(vm, arg_log_level).c_str() << std::endl; throw std::runtime_error{"Incorrect log level"}; } oxen::logging::init(log_file_path, log_level); log::info(logcat, "Starting..."); fs::path import_file_path; if (command_line::has_arg(vm, arg_input_file)) import_file_path = fs::u8path(command_line::get_arg(vm, arg_input_file)); else import_file_path = fs::u8path(m_config_folder) / "export" / BLOCKCHAIN_RAW; if (command_line::has_arg(vm, arg_count_blocks)) { BootstrapFile bootstrap; bootstrap.count_blocks(import_file_path); return 0; } log::info(logcat, "database: LMDB"); log::info(logcat, "verify: {}", opt_verify); if (opt_batch) { log::info(logcat, "batch: {} batch size: {}", opt_batch, db_batch_size); } else { log::info(logcat, "batch: {}", opt_batch); } log::info(logcat, "resume: {}", opt_resume); log::info(logcat, "nettype: {}", (opt_testnet ? "testnet" : opt_devnet ? "devnet" : "mainnet")); log::info(logcat, "bootstrap file path: {}", import_file_path); log::info(logcat, "database path: {}", m_config_folder); if (!opt_verify) { log::warning( logcat, fg(fmt::terminal_color::red), "\n\ Import is set to proceed WITHOUT VERIFICATION.\n\ This is a DANGEROUS operation: if the file was tampered with in transit, or obtained from a malicious source,\n\ you could end up with a compromised database. It is recommended to NOT use {}.\n\ *****************************************************************************************\n\ You have 90 seconds to press ^C or terminate this program before unverified import starts\n\ *****************************************************************************************", arg_noverify.name); sleep(90); } // TODO: currently using cryptonote_protocol stub for this kind of test, use real validation of // relayed objects cryptonote::core core; try { #if defined(PER_BLOCK_CHECKPOINT) const GetCheckpointsCallback& get_checkpoints = blocks::GetCheckpointsData; #else const GetCheckpointsCallback& get_checkpoints = nullptr; #endif quorumnet::init_core_callbacks(); if (!core.init(vm, nullptr, get_checkpoints)) { std::cerr << "Failed to initialize core\n"; return 1; } core.get_blockchain_storage().get_db().set_batch_transactions(true); if (!command_line::is_arg_defaulted(vm, arg_pop_blocks)) { num_blocks = command_line::get_arg(vm, arg_pop_blocks); log::info( logcat, "height: {}", core.get_blockchain_storage().get_current_blockchain_height()); pop_blocks(core, num_blocks); log::info( logcat, "height: {}", core.get_blockchain_storage().get_current_blockchain_height()); return 0; } if (command_line::get_arg(vm, arg_recalculate_difficulty)) core.get_blockchain_storage().get_db().fixup(core.get_nettype()); import_from_file(core, import_file_path, block_stop); // ensure db closed // - transactions properly checked and handled // - disk sync if needed // core.deinit(); } catch (const DB_ERROR& e) { std::cout << std::string("Error loading blockchain db: ") + e.what() + " -- shutting down now\n"; core.deinit(); return 1; } return 0; CATCH_ENTRY("Import error", 1); }