int main(int argc, char* argv[]) { uint32_t log_level = 0; uint64_t block_stop = 0; bool blocks_dat = false; tools::sanitize_locale(); boost::filesystem::path default_data_path {tools::get_default_data_dir()}; boost::filesystem::path default_testnet_data_path {default_data_path / "testnet"}; boost::filesystem::path output_file_path; po::options_description desc_cmd_only("Command line options"); po::options_description desc_cmd_sett("Command line options and settings options"); const command_line::arg_descriptor<std::string> arg_output_file = {"output-file", "Specify output file", "", true}; const command_line::arg_descriptor<uint32_t> arg_log_level = {"log-level", "", log_level}; const command_line::arg_descriptor<uint64_t> arg_block_stop = {"block-stop", "Stop at block number", block_stop}; const command_line::arg_descriptor<bool> arg_testnet_on = { "testnet" , "Run on testnet." , false }; const command_line::arg_descriptor<bool> arg_blocks_dat = {"blocksdat", "Output in blocks.dat format", blocks_dat}; command_line::add_arg(desc_cmd_sett, command_line::arg_data_dir, default_data_path.string()); command_line::add_arg(desc_cmd_sett, command_line::arg_testnet_data_dir, default_testnet_data_path.string()); command_line::add_arg(desc_cmd_sett, arg_output_file); command_line::add_arg(desc_cmd_sett, arg_testnet_on); command_line::add_arg(desc_cmd_sett, arg_log_level); command_line::add_arg(desc_cmd_sett, arg_block_stop); command_line::add_arg(desc_cmd_sett, arg_blocks_dat); command_line::add_arg(desc_cmd_only, command_line::arg_help); po::options_description desc_options("Allowed options"); desc_options.add(desc_cmd_only).add(desc_cmd_sett); po::variables_map vm; bool r = command_line::handle_error_helper(desc_options, [&]() { po::store(po::parse_command_line(argc, argv, desc_options), vm); po::notify(vm); return true; }); if (! r) return 1; if (command_line::get_arg(vm, command_line::arg_help)) { std::cout << CRYPTONOTE_NAME << " v" << MONERO_VERSION_FULL << ENDL << ENDL; std::cout << desc_options << std::endl; return 1; } log_level = command_line::get_arg(vm, arg_log_level); block_stop = command_line::get_arg(vm, arg_block_stop); log_space::get_set_log_detalisation_level(true, log_level); log_space::log_singletone::add_logger(LOGGER_CONSOLE, NULL, NULL); LOG_PRINT_L0("Starting..."); LOG_PRINT_L0("Setting log level = " << log_level); bool opt_testnet = command_line::get_arg(vm, arg_testnet_on); bool opt_blocks_dat = command_line::get_arg(vm, arg_blocks_dat); std::string m_config_folder; auto data_dir_arg = opt_testnet ? command_line::arg_testnet_data_dir : command_line::arg_data_dir; m_config_folder = command_line::get_arg(vm, data_dir_arg); if (command_line::has_arg(vm, arg_output_file)) output_file_path = boost::filesystem::path(command_line::get_arg(vm, arg_output_file)); else output_file_path = boost::filesystem::path(m_config_folder) / "export" / BLOCKCHAIN_RAW; LOG_PRINT_L0("Export output file: " << output_file_path.string()); // If we wanted to use the memory pool, we would set up a fake_core. #if SOURCE_DB == DB_MEMORY // blockchain_storage* core_storage = NULL; // tx_memory_pool m_mempool(*core_storage); // is this fake anyway? just passing in NULL! so m_mempool can't be used anyway, right? // core_storage = new blockchain_storage(&m_mempool); blockchain_storage* core_storage = new blockchain_storage(NULL); LOG_PRINT_L0("Initializing source blockchain (in-memory database)"); r = core_storage->init(m_config_folder, opt_testnet); #else // Use Blockchain instead of lower-level BlockchainDB for two reasons: // 1. Blockchain has the init() method for easy setup // 2. exporter needs to use get_current_blockchain_height(), get_block_id_by_height(), get_block_by_hash() // // cannot match blockchain_storage setup above with just one line, // e.g. // Blockchain* core_storage = new Blockchain(NULL); // because unlike blockchain_storage constructor, which takes a pointer to // tx_memory_pool, Blockchain's constructor takes tx_memory_pool object. LOG_PRINT_L0("Initializing source blockchain (BlockchainDB)"); Blockchain* core_storage = NULL; tx_memory_pool m_mempool(*core_storage); core_storage = new Blockchain(m_mempool); BlockchainDB* db = new BlockchainLMDB(); boost::filesystem::path folder(m_config_folder); folder /= db->get_db_name(); int lmdb_flags = 0; lmdb_flags |= MDB_RDONLY; const std::string filename = folder.string(); LOG_PRINT_L0("Loading blockchain from folder " << filename << " ..."); try { db->open(filename, lmdb_flags); } catch (const std::exception& e) { LOG_PRINT_L0("Error opening database: " << e.what()); return 1; } r = core_storage->init(db, opt_testnet); #endif CHECK_AND_ASSERT_MES(r, false, "Failed to initialize source blockchain storage"); LOG_PRINT_L0("Source blockchain storage initialized OK"); LOG_PRINT_L0("Exporting blockchain raw data..."); if (opt_blocks_dat) { BlocksdatFile blocksdat; r = blocksdat.store_blockchain_raw(core_storage, NULL, output_file_path, block_stop); } else { BootstrapFile bootstrap; r = bootstrap.store_blockchain_raw(core_storage, NULL, output_file_path, block_stop); } CHECK_AND_ASSERT_MES(r, false, "Failed to export blockchain raw data"); LOG_PRINT_L0("Blockchain raw data exported OK"); }
int main(int argc, char* argv[]) { #if defined(BLOCKCHAIN_DB) && (BLOCKCHAIN_DB == DB_MEMORY) std::string default_db_engine = "memory"; #else std::string default_db_engine = "lmdb"; #endif uint32_t log_level = LOG_LEVEL_0; uint64_t num_blocks = 0; uint64_t block_stop = 0; std::string m_config_folder; std::string db_arg_str; tools::sanitize_locale(); boost::filesystem::path default_data_path {tools::get_default_data_dir()}; boost::filesystem::path default_testnet_data_path {default_data_path / "testnet"}; std::string import_file_path; po::options_description desc_cmd_only("Command line options"); po::options_description desc_cmd_sett("Command line options and settings options"); const command_line::arg_descriptor<std::string> arg_input_file = {"input-file", "Specify input file", "", true}; const command_line::arg_descriptor<uint32_t> arg_log_level = {"log-level", "", log_level}; const command_line::arg_descriptor<uint64_t> arg_block_stop = {"block-stop", "Stop at block number", block_stop}; const command_line::arg_descriptor<uint64_t> arg_batch_size = {"batch-size", "", db_batch_size}; const command_line::arg_descriptor<uint64_t> arg_pop_blocks = {"pop-blocks", "Remove blocks from end of blockchain", num_blocks}; const command_line::arg_descriptor<bool> arg_testnet_on = { "testnet" , "Run on testnet." , false }; const command_line::arg_descriptor<bool> arg_count_blocks = { "count-blocks" , "Count blocks in bootstrap file and exit" , false }; const command_line::arg_descriptor<std::string> arg_database = { "database", "available: memory, lmdb" , default_db_engine }; const command_line::arg_descriptor<bool> arg_verify = {"verify", "Verify blocks and transactions during import", true}; const command_line::arg_descriptor<bool> arg_batch = {"batch", "Batch transactions for faster import", true}; const command_line::arg_descriptor<bool> arg_resume = {"resume", "Resume from current height if output database already exists", true}; command_line::add_arg(desc_cmd_sett, command_line::arg_data_dir, default_data_path.string()); command_line::add_arg(desc_cmd_sett, command_line::arg_testnet_data_dir, default_testnet_data_path.string()); command_line::add_arg(desc_cmd_sett, arg_input_file); command_line::add_arg(desc_cmd_sett, arg_testnet_on); command_line::add_arg(desc_cmd_sett, arg_log_level); command_line::add_arg(desc_cmd_sett, arg_database); command_line::add_arg(desc_cmd_sett, arg_batch_size); command_line::add_arg(desc_cmd_sett, arg_block_stop); command_line::add_arg(desc_cmd_only, arg_count_blocks); command_line::add_arg(desc_cmd_only, arg_pop_blocks); command_line::add_arg(desc_cmd_only, command_line::arg_help); // call add_options() directly for these arguments since // command_line helpers support only boolean switch, not boolean argument desc_cmd_sett.add_options() (arg_verify.name, make_semantic(arg_verify), arg_verify.description) (arg_batch.name, make_semantic(arg_batch), arg_batch.description) (arg_resume.name, make_semantic(arg_resume), arg_resume.description) ; po::options_description desc_options("Allowed options"); desc_options.add(desc_cmd_only).add(desc_cmd_sett); po::variables_map vm; bool r = command_line::handle_error_helper(desc_options, [&]() { po::store(po::parse_command_line(argc, argv, desc_options), vm); po::notify(vm); return true; }); if (! r) return 1; log_level = command_line::get_arg(vm, arg_log_level); opt_verify = command_line::get_arg(vm, arg_verify); opt_batch = command_line::get_arg(vm, arg_batch); opt_resume = command_line::get_arg(vm, arg_resume); block_stop = command_line::get_arg(vm, arg_block_stop); db_batch_size = command_line::get_arg(vm, arg_batch_size); if (command_line::get_arg(vm, command_line::arg_help)) { std::cout << "Monero '" << MONERO_RELEASE_NAME << "' (v" << MONERO_VERSION_FULL << ")" << ENDL << ENDL; std::cout << desc_options << std::endl; return 1; } if (! opt_batch && ! vm["batch-size"].defaulted()) { std::cerr << "Error: batch-size set, but batch option not enabled" << ENDL; return 1; } if (! db_batch_size) { std::cerr << "Error: batch-size must be > 0" << ENDL; return 1; } if (opt_verify && vm["batch-size"].defaulted()) { // usually want batch size default lower if verify on, so progress can be // frequently saved. // // currently, with Windows, default batch size is low, so ignore // default db_batch_size_verify unless it's even lower if (db_batch_size > db_batch_size_verify) { db_batch_size = db_batch_size_verify; } } std::vector<std::string> db_engines {"memory", "lmdb"}; opt_testnet = command_line::get_arg(vm, arg_testnet_on); auto data_dir_arg = opt_testnet ? command_line::arg_testnet_data_dir : command_line::arg_data_dir; m_config_folder = command_line::get_arg(vm, data_dir_arg); db_arg_str = command_line::get_arg(vm, arg_database); log_space::get_set_log_detalisation_level(true, log_level); log_space::log_singletone::add_logger(LOGGER_CONSOLE, NULL, NULL); LOG_PRINT_L0("Starting..."); LOG_PRINT_L0("Setting log level = " << log_level); boost::filesystem::path fs_import_file_path; if (command_line::has_arg(vm, arg_input_file)) fs_import_file_path = boost::filesystem::path(command_line::get_arg(vm, arg_input_file)); else fs_import_file_path = boost::filesystem::path(m_config_folder) / "export" / BLOCKCHAIN_RAW; import_file_path = fs_import_file_path.string(); if (command_line::has_arg(vm, arg_count_blocks)) { BootstrapFile bootstrap; bootstrap.count_blocks(import_file_path); return 0; } std::string db_engine; int mdb_flags = 0; int res = 0; res = parse_db_arguments(db_arg_str, db_engine, mdb_flags); if (res) { std::cerr << "Error parsing database argument(s)" << ENDL; return 1; } if (std::find(db_engines.begin(), db_engines.end(), db_engine) == db_engines.end()) { std::cerr << "Invalid database engine: " << db_engine << std::endl; return 1; } LOG_PRINT_L0("database: " << db_engine); LOG_PRINT_L0("verify: " << std::boolalpha << opt_verify << std::noboolalpha); if (opt_batch) { LOG_PRINT_L0("batch: " << std::boolalpha << opt_batch << std::noboolalpha << " batch size: " << db_batch_size); } else { LOG_PRINT_L0("batch: " << std::boolalpha << opt_batch << std::noboolalpha); } LOG_PRINT_L0("resume: " << std::boolalpha << opt_resume << std::noboolalpha); LOG_PRINT_L0("testnet: " << std::boolalpha << opt_testnet << std::noboolalpha); LOG_PRINT_L0("bootstrap file path: " << import_file_path); LOG_PRINT_L0("database path: " << m_config_folder); try { // fake_core needed for verification to work when enabled. // // NOTE: don't need fake_core method of doing things when we're going to call // BlockchainDB add_block() directly and have available the 3 block // properties to do so. Both ways work, but fake core isn't necessary in that // circumstance. // for multi_db_runtime: #if !defined(BLOCKCHAIN_DB) if (db_engine == "lmdb") { fake_core_lmdb simple_core(m_config_folder, opt_testnet, opt_batch, mdb_flags); import_from_file(simple_core, import_file_path, block_stop); } else if (db_engine == "memory") { fake_core_memory simple_core(m_config_folder, opt_testnet); import_from_file(simple_core, import_file_path, block_stop); } else { std::cerr << "database engine unrecognized" << ENDL; return 1; } // for multi_db_compile: #else if (db_engine != default_db_engine) { std::cerr << "Invalid database engine for compiled version: " << db_engine << std::endl; return 1; } #if BLOCKCHAIN_DB == DB_LMDB fake_core_lmdb simple_core(m_config_folder, opt_testnet, opt_batch, mdb_flags); #else fake_core_memory simple_core(m_config_folder, opt_testnet); #endif if (! vm["pop-blocks"].defaulted()) { num_blocks = command_line::get_arg(vm, arg_pop_blocks); LOG_PRINT_L0("height: " << simple_core.m_storage.get_current_blockchain_height()); pop_blocks(simple_core, num_blocks); LOG_PRINT_L0("height: " << simple_core.m_storage.get_current_blockchain_height()); return 0; } import_from_file(simple_core, import_file_path, block_stop); #endif } catch (const DB_ERROR& e) { std::cout << std::string("Error loading blockchain db: ") + e.what() + " -- shutting down now" << ENDL; return 1; } // destructors called at exit: // // ensure db closed // - transactions properly checked and handled // - disk sync if needed // // fake_core object's destructor is called when it goes out of scope. For an // LMDB fake_core, it calls Blockchain::deinit() on its object, which in turn // calls delete on its BlockchainDB derived class' object, which closes its // files. return 0; }
int main(int argc, char* argv[]) { TRY_ENTRY(); epee::string_tools::set_module_name_and_folder(argv[0]); std::string default_db_type = "lmdb"; std::string available_dbs = cryptonote::blockchain_db_types(", "); available_dbs = "available: " + available_dbs; uint32_t log_level = 0; uint64_t num_blocks = 0; uint64_t block_stop = 0; std::string m_config_folder; std::string db_arg_str; tools::on_startup(); std::string import_file_path; po::options_description desc_cmd_only("Command line options"); po::options_description desc_cmd_sett("Command line options and settings options"); const command_line::arg_descriptor<std::string> arg_input_file = {"input-file", "Specify input file", "", true}; const command_line::arg_descriptor<std::string> arg_log_level = {"log-level", "0-4 or categories", ""}; const command_line::arg_descriptor<uint64_t> arg_block_stop = {"block-stop", "Stop at block number", block_stop}; const command_line::arg_descriptor<uint64_t> arg_batch_size = {"batch-size", "", db_batch_size}; const command_line::arg_descriptor<uint64_t> arg_pop_blocks = {"pop-blocks", "Remove blocks from end of blockchain", num_blocks}; const command_line::arg_descriptor<bool> arg_drop_hf = {"drop-hard-fork", "Drop hard fork subdbs", false}; const command_line::arg_descriptor<bool> arg_count_blocks = { "count-blocks" , "Count blocks in bootstrap file and exit" , false }; const command_line::arg_descriptor<std::string> arg_database = { "database", available_dbs.c_str(), default_db_type }; const command_line::arg_descriptor<bool> arg_noverify = {"dangerous-unverified-import", "Blindly trust the import file and use potentially malicious blocks and transactions during import (only enable if you exported the file yourself)", false}; const command_line::arg_descriptor<bool> arg_batch = {"batch", "Batch transactions for faster import", true}; const command_line::arg_descriptor<bool> arg_resume = {"resume", "Resume from current height if output database already exists", true}; command_line::add_arg(desc_cmd_sett, arg_input_file); command_line::add_arg(desc_cmd_sett, arg_log_level); command_line::add_arg(desc_cmd_sett, arg_database); command_line::add_arg(desc_cmd_sett, arg_batch_size); command_line::add_arg(desc_cmd_sett, arg_block_stop); command_line::add_arg(desc_cmd_only, arg_count_blocks); command_line::add_arg(desc_cmd_only, arg_pop_blocks); command_line::add_arg(desc_cmd_only, arg_drop_hf); command_line::add_arg(desc_cmd_only, command_line::arg_help); // call add_options() directly for these arguments since // command_line helpers support only boolean switch, not boolean argument desc_cmd_sett.add_options() (arg_noverify.name, make_semantic(arg_noverify), arg_noverify.description) (arg_batch.name, make_semantic(arg_batch), arg_batch.description) (arg_resume.name, make_semantic(arg_resume), arg_resume.description) ; po::options_description desc_options("Allowed options"); desc_options.add(desc_cmd_only).add(desc_cmd_sett); cryptonote::core::init_options(desc_options); po::variables_map vm; bool r = command_line::handle_error_helper(desc_options, [&]() { po::store(po::parse_command_line(argc, argv, desc_options), vm); po::notify(vm); return true; }); if (! r) return 1; opt_verify = !command_line::get_arg(vm, arg_noverify); opt_batch = command_line::get_arg(vm, arg_batch); opt_resume = command_line::get_arg(vm, arg_resume); block_stop = command_line::get_arg(vm, arg_block_stop); db_batch_size = command_line::get_arg(vm, arg_batch_size); if (command_line::get_arg(vm, command_line::arg_help)) { std::cout << "Monero '" << MONERO_RELEASE_NAME << "' (v" << MONERO_VERSION_FULL << ")" << ENDL << ENDL; std::cout << desc_options << std::endl; return 1; } if (! opt_batch && !command_line::is_arg_defaulted(vm, arg_batch_size)) { std::cerr << "Error: batch-size set, but batch option not enabled" << ENDL; return 1; } if (! db_batch_size) { std::cerr << "Error: batch-size must be > 0" << ENDL; return 1; } if (opt_verify && command_line::is_arg_defaulted(vm, arg_batch_size)) { // usually want batch size default lower if verify on, so progress can be // frequently saved. // // currently, with Windows, default batch size is low, so ignore // default db_batch_size_verify unless it's even lower if (db_batch_size > db_batch_size_verify) { db_batch_size = db_batch_size_verify; } } opt_testnet = command_line::get_arg(vm, cryptonote::arg_testnet_on); opt_stagenet = command_line::get_arg(vm, cryptonote::arg_stagenet_on); if (opt_testnet && opt_stagenet) { std::cerr << "Error: Can't specify more than one of --testnet and --stagenet" << ENDL; return 1; } m_config_folder = command_line::get_arg(vm, cryptonote::arg_data_dir); db_arg_str = command_line::get_arg(vm, arg_database); mlog_configure(mlog_get_default_log_path("monero-blockchain-import.log"), true); if (!command_line::is_arg_defaulted(vm, arg_log_level)) mlog_set_log(command_line::get_arg(vm, arg_log_level).c_str()); else mlog_set_log(std::string(std::to_string(log_level) + ",bcutil:INFO").c_str()); MINFO("Starting..."); boost::filesystem::path fs_import_file_path; if (command_line::has_arg(vm, arg_input_file)) fs_import_file_path = boost::filesystem::path(command_line::get_arg(vm, arg_input_file)); else fs_import_file_path = boost::filesystem::path(m_config_folder) / "export" / BLOCKCHAIN_RAW; import_file_path = fs_import_file_path.string(); if (command_line::has_arg(vm, arg_count_blocks)) { BootstrapFile bootstrap; bootstrap.count_blocks(import_file_path); return 0; } std::string db_type; int db_flags = 0; int res = 0; res = parse_db_arguments(db_arg_str, db_type, db_flags); if (res) { std::cerr << "Error parsing database argument(s)" << ENDL; return 1; } if (!cryptonote::blockchain_valid_db_type(db_type)) { std::cerr << "Invalid database type: " << db_type << std::endl; return 1; } MINFO("database: " << db_type); MINFO("database flags: " << db_flags); MINFO("verify: " << std::boolalpha << opt_verify << std::noboolalpha); if (opt_batch) { MINFO("batch: " << std::boolalpha << opt_batch << std::noboolalpha << " batch size: " << db_batch_size); } else { MINFO("batch: " << std::boolalpha << opt_batch << std::noboolalpha); } MINFO("resume: " << std::boolalpha << opt_resume << std::noboolalpha); MINFO("nettype: " << (opt_testnet ? "testnet" : opt_stagenet ? "stagenet" : "mainnet")); MINFO("bootstrap file path: " << import_file_path); MINFO("database path: " << m_config_folder); if (!opt_verify) { MCLOG_RED(el::Level::Warning, "global", "\n" "Import is set to proceed WITHOUT VERIFICATION.\n" "This is a DANGEROUS operation: if the file was tampered with in transit, or obtained from a malicious source,\n" "you could end up with a compromised database. It is recommended to NOT use " << arg_noverify.name << ".\n" "*****************************************************************************************\n" "You have 90 seconds to press ^C or terminate this program before unverified import starts\n" "*****************************************************************************************"); sleep(90); } cryptonote::cryptonote_protocol_stub pr; //TODO: stub only for this kind of test, make real validation of relayed objects cryptonote::core core(&pr); try { core.disable_dns_checkpoints(true); #if defined(PER_BLOCK_CHECKPOINT) const GetCheckpointsCallback& get_checkpoints = blocks::GetCheckpointsData; #else const GetCheckpointsCallback& get_checkpoints = nullptr; #endif if (!core.init(vm, nullptr, get_checkpoints)) { std::cerr << "Failed to initialize core" << ENDL; return 1; } core.get_blockchain_storage().get_db().set_batch_transactions(true); if (!command_line::is_arg_defaulted(vm, arg_pop_blocks)) { num_blocks = command_line::get_arg(vm, arg_pop_blocks); MINFO("height: " << core.get_blockchain_storage().get_current_blockchain_height()); pop_blocks(core, num_blocks); MINFO("height: " << core.get_blockchain_storage().get_current_blockchain_height()); return 0; } if (!command_line::is_arg_defaulted(vm, arg_drop_hf)) { MINFO("Dropping hard fork tables..."); core.get_blockchain_storage().get_db().drop_hard_fork_info(); core.deinit(); return 0; } import_from_file(core, import_file_path, block_stop); // ensure db closed // - transactions properly checked and handled // - disk sync if needed // core.deinit(); } catch (const DB_ERROR& e) { std::cout << std::string("Error loading blockchain db: ") + e.what() + " -- shutting down now" << ENDL; core.deinit(); return 1; } return 0; CATCH_ENTRY("Import error", 1); }
int import_from_file(FakeCore& simple_core, const std::string& import_file_path, uint64_t block_stop=0) { #if !defined(BLOCKCHAIN_DB) static_assert(std::is_same<fake_core_memory, FakeCore>::value || std::is_same<fake_core_lmdb, FakeCore>::value, "FakeCore constraint error"); #endif #if !defined(BLOCKCHAIN_DB) || (BLOCKCHAIN_DB == DB_LMDB) if (std::is_same<fake_core_lmdb, FakeCore>::value) { // Reset stats, in case we're using newly created db, accumulating stats // from addition of genesis block. // This aligns internal db counts with importer counts. simple_core.m_storage.get_db().reset_stats(); } #endif boost::filesystem::path fs_import_file_path(import_file_path); boost::system::error_code ec; if (!boost::filesystem::exists(fs_import_file_path, ec)) { LOG_PRINT_L0("bootstrap file not found: " << fs_import_file_path); return false; } BootstrapFile bootstrap; // BootstrapFile bootstrap(import_file_path); uint64_t total_source_blocks = bootstrap.count_blocks(import_file_path); LOG_PRINT_L0("bootstrap file last block number: " << total_source_blocks-1 << " (zero-based height) total blocks: " << total_source_blocks); std::cout << ENDL; std::cout << "Preparing to read blocks..." << ENDL; std::cout << ENDL; std::ifstream import_file; import_file.open(import_file_path, std::ios_base::binary | std::ifstream::in); uint64_t h = 0; uint64_t num_imported = 0; if (import_file.fail()) { LOG_PRINT_L0("import_file.open() fail"); return false; } // 4 byte magic + (currently) 1024 byte header structures bootstrap.seek_to_first_chunk(import_file); std::string str1; char buffer1[1024]; char buffer_block[BUFFER_SIZE]; block b; transaction tx; int quit = 0; uint64_t bytes_read = 0; uint64_t start_height = 1; if (opt_resume) start_height = simple_core.m_storage.get_current_blockchain_height(); // Note that a new blockchain will start with block number 0 (total blocks: 1) // due to genesis block being added at initialization. if (! block_stop) { block_stop = total_source_blocks - 1; } // These are what we'll try to use, and they don't have to be a determination // from source and destination blockchains, but those are the defaults. LOG_PRINT_L0("start block: " << start_height << " stop block: " << block_stop); bool use_batch = false; if (opt_batch) { if (simple_core.support_batch) use_batch = true; else LOG_PRINT_L0("WARNING: batch transactions enabled but unsupported or unnecessary for this database engine - ignoring"); } if (use_batch) simple_core.batch_start(db_batch_size); LOG_PRINT_L0("Reading blockchain from bootstrap file..."); std::cout << ENDL; // Within the loop, we skip to start_height before we start adding. // TODO: Not a bottleneck, but we can use what's done in count_blocks() and // only do the chunk size reads, skipping the chunk content reads until we're // at start_height. while (! quit) { uint32_t chunk_size; import_file.read(buffer1, sizeof(chunk_size)); // TODO: bootstrap.read_chunk(); if (! import_file) { std::cout << refresh_string; LOG_PRINT_L0("End of file reached"); quit = 1; break; } bytes_read += sizeof(chunk_size); str1.assign(buffer1, sizeof(chunk_size)); if (! ::serialization::parse_binary(str1, chunk_size)) { throw std::runtime_error("Error in deserialization of chunk size"); } LOG_PRINT_L3("chunk_size: " << chunk_size); if (chunk_size > BUFFER_SIZE) { LOG_PRINT_L0("WARNING: chunk_size " << chunk_size << " > BUFFER_SIZE " << BUFFER_SIZE); throw std::runtime_error("Aborting: chunk size exceeds buffer size"); } if (chunk_size > 100000) { LOG_PRINT_L0("NOTE: chunk_size " << chunk_size << " > 100000"); } else if (chunk_size == 0) { LOG_PRINT_L0("ERROR: chunk_size == 0"); return 2; } import_file.read(buffer_block, chunk_size); if (! import_file) { LOG_PRINT_L0("ERROR: unexpected end of file: bytes read before error: " << import_file.gcount() << " of chunk_size " << chunk_size); return 2; } bytes_read += chunk_size; LOG_PRINT_L3("Total bytes read: " << bytes_read); if (h + NUM_BLOCKS_PER_CHUNK < start_height + 1) { h += NUM_BLOCKS_PER_CHUNK; continue; } if (h > block_stop) { std::cout << refresh_string << "block " << h-1 << " / " << block_stop << std::flush; std::cout << ENDL << ENDL; LOG_PRINT_L0("Specified block number reached - stopping. block: " << h-1 << " total blocks: " << h); quit = 1; break; } try { str1.assign(buffer_block, chunk_size); bootstrap::block_package bp; if (! ::serialization::parse_binary(str1, bp)) throw std::runtime_error("Error in deserialization of chunk"); int display_interval = 1000; int progress_interval = 10; // NOTE: use of NUM_BLOCKS_PER_CHUNK is a placeholder in case multi-block chunks are later supported. for (int chunk_ind = 0; chunk_ind < NUM_BLOCKS_PER_CHUNK; ++chunk_ind) { ++h; if ((h-1) % display_interval == 0) { std::cout << refresh_string; LOG_PRINT_L0("loading block number " << h-1); } else { LOG_PRINT_L3("loading block number " << h-1); } b = bp.block; LOG_PRINT_L2("block prev_id: " << b.prev_id << ENDL); if ((h-1) % progress_interval == 0) { std::cout << refresh_string << "block " << h-1 << " / " << block_stop << std::flush; } std::vector<transaction> txs; std::vector<transaction> archived_txs; archived_txs = bp.txs; // std::cout << refresh_string; // LOG_PRINT_L1("txs: " << archived_txs.size()); // if archived_txs is invalid // { // std::cout << refresh_string; // LOG_PRINT_RED_L0("exception while de-archiving txs, height=" << h); // quit = 1; // break; // } // tx number 1: coinbase tx // tx number 2 onwards: archived_txs unsigned int tx_num = 1; for (const transaction& tx : archived_txs) { ++tx_num; // if tx is invalid // { // LOG_PRINT_RED_L0("exception while indexing tx from txs, height=" << h <<", tx_num=" << tx_num); // quit = 1; // break; // } // std::cout << refresh_string; // LOG_PRINT_L1("tx hash: " << get_transaction_hash(tx)); // crypto::hash hsh = null_hash; // size_t blob_size = 0; // NOTE: all tx hashes except for coinbase tx are available in the block data // get_transaction_hash(tx, hsh, blob_size); // LOG_PRINT_L0("tx " << tx_num << " " << hsh << " : " << ENDL); // LOG_PRINT_L0(obj_to_json_str(tx) << ENDL); // add blocks with verification. // for Blockchain and blockchain_storage add_new_block(). if (opt_verify) { // crypto::hash hsh = null_hash; // size_t blob_size = 0; // get_transaction_hash(tx, hsh, blob_size); // we'd need to get the starting heights from the daemon // to be correct once voting kicks in uint64_t v2height = opt_testnet ? 624634 : 1009827; uint8_t version = h < v2height ? 1 : 2; tx_verification_context tvc = AUTO_VAL_INIT(tvc); bool r = true; r = simple_core.m_pool.add_tx(tx, tvc, true, true, version); if (!r) { LOG_PRINT_RED_L0("failed to add transaction to transaction pool, height=" << h <<", tx_num=" << tx_num); quit = 1; break; } } else { // for add_block() method, without (much) processing. // don't add coinbase transaction to txs. // // because add_block() calls // add_transaction(blk_hash, blk.miner_tx) first, and // then a for loop for the transactions in txs. txs.push_back(tx); } } if (opt_verify) { block_verification_context bvc = boost::value_initialized<block_verification_context>(); simple_core.m_storage.add_new_block(b, bvc); if (bvc.m_verifivation_failed) { LOG_PRINT_L0("Failed to add block to blockchain, verification failed, height = " << h); LOG_PRINT_L0("skipping rest of file"); // ok to commit previously batched data because it failed only in // verification of potential new block with nothing added to batch // yet quit = 1; break; } if (! bvc.m_added_to_main_chain) { LOG_PRINT_L0("Failed to add block to blockchain, height = " << h); LOG_PRINT_L0("skipping rest of file"); // make sure we don't commit partial block data quit = 2; break; } } else { size_t block_size; difficulty_type cumulative_difficulty; uint64_t coins_generated; block_size = bp.block_size; cumulative_difficulty = bp.cumulative_difficulty; coins_generated = bp.coins_generated; // std::cout << refresh_string; // LOG_PRINT_L2("block_size: " << block_size); // LOG_PRINT_L2("cumulative_difficulty: " << cumulative_difficulty); // LOG_PRINT_L2("coins_generated: " << coins_generated); try { simple_core.add_block(b, block_size, cumulative_difficulty, coins_generated, txs); } catch (const std::exception& e) { std::cout << refresh_string; LOG_PRINT_RED_L0("Error adding block to blockchain: " << e.what()); quit = 2; // make sure we don't commit partial block data break; } } ++num_imported; if (use_batch) { if ((h-1) % db_batch_size == 0) { std::cout << refresh_string; // zero-based height std::cout << ENDL << "[- batch commit at height " << h-1 << " -]" << ENDL; simple_core.batch_stop(); simple_core.batch_start(db_batch_size); std::cout << ENDL; #if !defined(BLOCKCHAIN_DB) || (BLOCKCHAIN_DB == DB_LMDB) simple_core.m_storage.get_db().show_stats(); #endif } } } } catch (const std::exception& e) { std::cout << refresh_string; LOG_PRINT_RED_L0("exception while reading from file, height=" << h); return 2; } } // while import_file.close(); if (use_batch) { if (quit > 1) { // There was an error, so don't commit pending data. // Destructor will abort write txn. } else { simple_core.batch_stop(); } #if !defined(BLOCKCHAIN_DB) || (BLOCKCHAIN_DB == DB_LMDB) simple_core.m_storage.get_db().show_stats(); #endif LOG_PRINT_L0("Number of blocks imported: " << num_imported); if (h > 0) // TODO: if there was an error, the last added block is probably at zero-based height h-2 LOG_PRINT_L0("Finished at block: " << h-1 << " total blocks: " << h); } std::cout << ENDL; return 0; }
int import_from_file(cryptonote::core& core, const std::string& import_file_path, uint64_t block_stop=0) { // Reset stats, in case we're using newly created db, accumulating stats // from addition of genesis block. // This aligns internal db counts with importer counts. core.get_blockchain_storage().get_db().reset_stats(); boost::filesystem::path fs_import_file_path(import_file_path); boost::system::error_code ec; if (!boost::filesystem::exists(fs_import_file_path, ec)) { MFATAL("bootstrap file not found: " << fs_import_file_path); return false; } uint64_t start_height = 1, seek_height; if (opt_resume) start_height = core.get_blockchain_storage().get_current_blockchain_height(); seek_height = start_height; BootstrapFile bootstrap; std::streampos pos; // BootstrapFile bootstrap(import_file_path); uint64_t total_source_blocks = bootstrap.count_blocks(import_file_path, pos, seek_height); MINFO("bootstrap file last block number: " << total_source_blocks-1 << " (zero-based height) total blocks: " << total_source_blocks); if (total_source_blocks-1 <= start_height) { return false; } std::cout << ENDL; std::cout << "Preparing to read blocks..." << ENDL; std::cout << ENDL; std::ifstream import_file; import_file.open(import_file_path, std::ios_base::binary | std::ifstream::in); uint64_t h = 0; uint64_t num_imported = 0; if (import_file.fail()) { MFATAL("import_file.open() fail"); return false; } // 4 byte magic + (currently) 1024 byte header structures uint8_t major_version, minor_version; bootstrap.seek_to_first_chunk(import_file, major_version, minor_version); std::string str1; char buffer1[1024]; char buffer_block[BUFFER_SIZE]; block b; transaction tx; int quit = 0; uint64_t bytes_read; // Note that a new blockchain will start with block number 0 (total blocks: 1) // due to genesis block being added at initialization. if (! block_stop) { block_stop = total_source_blocks - 1; } // These are what we'll try to use, and they don't have to be a determination // from source and destination blockchains, but those are the defaults. MINFO("start block: " << start_height << " stop block: " << block_stop); bool use_batch = opt_batch && !opt_verify; MINFO("Reading blockchain from bootstrap file..."); std::cout << ENDL; std::vector<block_complete_entry> blocks; // Skip to start_height before we start adding. { bool q2 = false; import_file.seekg(pos); bytes_read = bootstrap.count_bytes(import_file, start_height-seek_height, h, q2); if (q2) { quit = 2; goto quitting; } h = start_height; } if (use_batch) { uint64_t bytes, h2; bool q2; pos = import_file.tellg(); bytes = bootstrap.count_bytes(import_file, db_batch_size, h2, q2); if (import_file.eof()) import_file.clear(); import_file.seekg(pos); core.get_blockchain_storage().get_db().batch_start(db_batch_size, bytes); } while (! quit) { uint32_t chunk_size; import_file.read(buffer1, sizeof(chunk_size)); // TODO: bootstrap.read_chunk(); if (! import_file) { std::cout << refresh_string; MINFO("End of file reached"); quit = 1; break; } bytes_read += sizeof(chunk_size); str1.assign(buffer1, sizeof(chunk_size)); if (! ::serialization::parse_binary(str1, chunk_size)) { throw std::runtime_error("Error in deserialization of chunk size"); } MDEBUG("chunk_size: " << chunk_size); if (chunk_size > BUFFER_SIZE) { MWARNING("WARNING: chunk_size " << chunk_size << " > BUFFER_SIZE " << BUFFER_SIZE); throw std::runtime_error("Aborting: chunk size exceeds buffer size"); } if (chunk_size > CHUNK_SIZE_WARNING_THRESHOLD) { MINFO("NOTE: chunk_size " << chunk_size << " > " << CHUNK_SIZE_WARNING_THRESHOLD); } else if (chunk_size == 0) { MFATAL("ERROR: chunk_size == 0"); return 2; } import_file.read(buffer_block, chunk_size); if (! import_file) { if (import_file.eof()) { std::cout << refresh_string; MINFO("End of file reached - file was truncated"); quit = 1; break; } else { MFATAL("ERROR: unexpected end of file: bytes read before error: " << import_file.gcount() << " of chunk_size " << chunk_size); return 2; } } bytes_read += chunk_size; MDEBUG("Total bytes read: " << bytes_read); if (h > block_stop) { std::cout << refresh_string << "block " << h-1 << " / " << block_stop << "\r" << std::flush; std::cout << ENDL << ENDL; MINFO("Specified block number reached - stopping. block: " << h-1 << " total blocks: " << h); quit = 1; break; } try { str1.assign(buffer_block, chunk_size); bootstrap::block_package bp; bool res; if (major_version == 0) { bootstrap::block_package_1 bp1; res = ::serialization::parse_binary(str1, bp1); if (res) { bp.block = std::move(bp1.block); bp.txs = std::move(bp1.txs); bp.block_weight = bp1.block_weight; bp.cumulative_difficulty = bp1.cumulative_difficulty; bp.coins_generated = bp1.coins_generated; } } else res = ::serialization::parse_binary(str1, bp); if (!res) throw std::runtime_error("Error in deserialization of chunk"); int display_interval = 1000; int progress_interval = 10; // NOTE: use of NUM_BLOCKS_PER_CHUNK is a placeholder in case multi-block chunks are later supported. for (int chunk_ind = 0; chunk_ind < NUM_BLOCKS_PER_CHUNK; ++chunk_ind) { ++h; if ((h-1) % display_interval == 0) { std::cout << refresh_string; MDEBUG("loading block number " << h-1); } else { MDEBUG("loading block number " << h-1); } b = bp.block; MDEBUG("block prev_id: " << b.prev_id << ENDL); if ((h-1) % progress_interval == 0) { std::cout << refresh_string << "block " << h-1 << " / " << block_stop << "\r" << std::flush; } if (opt_verify) { cryptonote::blobdata block; cryptonote::block_to_blob(bp.block, block); std::vector<cryptonote::blobdata> txs; for (const auto &tx: bp.txs) { txs.push_back(cryptonote::blobdata()); cryptonote::tx_to_blob(tx, txs.back()); } blocks.push_back({block, txs}); int ret = check_flush(core, blocks, false); if (ret) { quit = 2; // make sure we don't commit partial block data break; } } else { std::vector<std::pair<transaction, blobdata>> txs; std::vector<transaction> archived_txs; archived_txs = bp.txs; // tx number 1: coinbase tx // tx number 2 onwards: archived_txs for (const transaction &tx : archived_txs) { // add blocks with verification. // for Blockchain and blockchain_storage add_new_block(). // for add_block() method, without (much) processing. // don't add coinbase transaction to txs. // // because add_block() calls // add_transaction(blk_hash, blk.miner_tx) first, and // then a for loop for the transactions in txs. txs.push_back(std::make_pair(tx, tx_to_blob(tx))); } size_t block_weight; difficulty_type cumulative_difficulty; uint64_t coins_generated; block_weight = bp.block_weight; cumulative_difficulty = bp.cumulative_difficulty; coins_generated = bp.coins_generated; try { uint64_t long_term_block_weight = core.get_blockchain_storage().get_next_long_term_block_weight(block_weight); core.get_blockchain_storage().get_db().add_block(std::make_pair(b, block_to_blob(b)), block_weight, long_term_block_weight, cumulative_difficulty, coins_generated, txs); } catch (const std::exception& e) { std::cout << refresh_string; MFATAL("Error adding block to blockchain: " << e.what()); quit = 2; // make sure we don't commit partial block data break; } if (use_batch) { if ((h-1) % db_batch_size == 0) { uint64_t bytes, h2; bool q2; std::cout << refresh_string; // zero-based height std::cout << ENDL << "[- batch commit at height " << h-1 << " -]" << ENDL; core.get_blockchain_storage().get_db().batch_stop(); pos = import_file.tellg(); bytes = bootstrap.count_bytes(import_file, db_batch_size, h2, q2); import_file.seekg(pos); core.get_blockchain_storage().get_db().batch_start(db_batch_size, bytes); std::cout << ENDL; core.get_blockchain_storage().get_db().show_stats(); } } } ++num_imported; } } catch (const std::exception& e) { std::cout << refresh_string; MFATAL("exception while reading from file, height=" << h << ": " << e.what()); return 2; } } // while quitting: import_file.close(); if (opt_verify) { int ret = check_flush(core, blocks, true); if (ret) return ret; } if (use_batch) { if (quit > 1) { // There was an error, so don't commit pending data. // Destructor will abort write txn. } else { core.get_blockchain_storage().get_db().batch_stop(); } } core.get_blockchain_storage().get_db().show_stats(); MINFO("Number of blocks imported: " << num_imported); if (h > 0) // TODO: if there was an error, the last added block is probably at zero-based height h-2 MINFO("Finished at block: " << h-1 << " total blocks: " << h); std::cout << ENDL; return 0; }