{
printf("EnvShutdown exception: %s (%d)\n", e.what(), e.get_errno());
}
- DbEnv(0).remove(GetDataDir().c_str(), 0);
+ DbEnv(0).remove(GetDataDir().string().c_str(), 0);
}
class CDBInit
instance_of_cdbinit;
-CDB::CDB(const char* pszFile, const char* pszMode) : pdb(NULL)
+CDB::CDB(const char *pszFile, const char* pszMode) : pdb(NULL)
{
int ret;
if (pszFile == NULL)
if (fCreate)
nFlags |= DB_CREATE;
- CRITICAL_BLOCK(cs_db)
{
+ LOCK(cs_db);
if (!fDbEnvInit)
{
if (fShutdown)
return;
- string strDataDir = GetDataDir();
- string strLogDir = strDataDir + "/database";
- filesystem::create_directory(strLogDir.c_str());
- string strErrorFile = strDataDir + "/db.log";
- printf("dbenv.open strLogDir=%s strErrorFile=%s\n", strLogDir.c_str(), strErrorFile.c_str());
+ filesystem::path pathDataDir = GetDataDir();
+ filesystem::path pathLogDir = pathDataDir / "database";
+ filesystem::create_directory(pathLogDir);
+ filesystem::path pathErrorFile = pathDataDir / "db.log";
+ printf("dbenv.open LogDir=%s ErrorFile=%s\n", pathLogDir.string().c_str(), pathErrorFile.string().c_str());
int nDbCache = GetArg("-dbcache", 25);
- dbenv.set_lg_dir(strLogDir.c_str());
+ dbenv.set_lg_dir(pathLogDir.string().c_str());
dbenv.set_cachesize(nDbCache / 1024, (nDbCache % 1024)*1048576, 1);
- dbenv.set_lg_bsize(10485760);
- dbenv.set_lg_max(104857600);
+ dbenv.set_lg_bsize(1048576);
+ dbenv.set_lg_max(10485760);
dbenv.set_lk_max_locks(10000);
dbenv.set_lk_max_objects(10000);
- dbenv.set_errfile(fopen(strErrorFile.c_str(), "a")); /// debug
+ dbenv.set_errfile(fopen(pathErrorFile.string().c_str(), "a")); /// debug
dbenv.set_flags(DB_AUTO_COMMIT, 1);
- ret = dbenv.open(strDataDir.c_str(),
+ dbenv.log_set_config(DB_LOG_AUTO_REMOVE, 1);
+ ret = dbenv.open(pathDataDir.string().c_str(),
DB_CREATE |
DB_INIT_LOCK |
DB_INIT_LOG |
{
delete pdb;
pdb = NULL;
- CRITICAL_BLOCK(cs_db)
+ {
+ LOCK(cs_db);
--mapFileUseCount[strFile];
+ }
strFile = "";
throw runtime_error(strprintf("CDB() : can't open database file %s, error %d", pszFile, ret));
}
nMinutes = 1;
if (strFile == "addr.dat")
nMinutes = 2;
- if (strFile == "blkindex.dat" && IsInitialBlockDownload() && nBestHeight % 5000 != 0)
- nMinutes = 1;
- dbenv.txn_checkpoint(0, nMinutes, 0);
+ if (strFile == "blkindex.dat")
+ nMinutes = 2;
+ if (strFile == "blkindex.dat" && IsInitialBlockDownload())
+ nMinutes = 5;
+
+ dbenv.txn_checkpoint(nMinutes ? GetArg("-dblogsize", 100)*1024 : 0, nMinutes, 0);
- CRITICAL_BLOCK(cs_db)
+ {
+ LOCK(cs_db);
--mapFileUseCount[strFile];
+ }
}
void static CloseDb(const string& strFile)
{
- CRITICAL_BLOCK(cs_db)
{
+ LOCK(cs_db);
if (mapDb[strFile] != NULL)
{
// Close the database handle
{
while (!fShutdown)
{
- CRITICAL_BLOCK(cs_db)
{
+ LOCK(cs_db);
if (!mapFileUseCount.count(strFile) || mapFileUseCount[strFile] == 0)
{
// Flush log data to the dat file
printf("DBFlush(%s)%s\n", fShutdown ? "true" : "false", fDbEnvInit ? "" : " db not started");
if (!fDbEnvInit)
return;
- CRITICAL_BLOCK(cs_db)
{
+ LOCK(cs_db);
map<string, int>::iterator mi = mapFileUseCount.begin();
while (mi != mapFileUseCount.end())
{
ReadBestInvalidWork(bnBestInvalidWork);
// Verify blocks in the best chain
+ int nCheckLevel = GetArg("-checklevel", 1);
+ int nCheckDepth = GetArg( "-checkblocks", 2500);
+ if (nCheckDepth == 0)
+ nCheckDepth = 1000000000; // suffices until the year 19000
+ if (nCheckDepth > nBestHeight)
+ nCheckDepth = nBestHeight;
+ printf("Verifying last %i blocks at level %i\n", nCheckDepth, nCheckLevel);
CBlockIndex* pindexFork = NULL;
+ map<pair<unsigned int, unsigned int>, CBlockIndex*> mapBlockPos;
for (CBlockIndex* pindex = pindexBest; pindex && pindex->pprev; pindex = pindex->pprev)
{
- if (pindex->nHeight < nBestHeight-2500 && !mapArgs.count("-checkblocks"))
+ if (pindex->nHeight < nBestHeight-nCheckDepth)
break;
CBlock block;
if (!block.ReadFromDisk(pindex))
return error("LoadBlockIndex() : block.ReadFromDisk failed");
- if (!block.CheckBlock())
+ // check level 1: verify block validity
+ if (nCheckLevel>0 && !block.CheckBlock())
{
printf("LoadBlockIndex() : *** found bad block at %d, hash=%s\n", pindex->nHeight, pindex->GetBlockHash().ToString().c_str());
pindexFork = pindex->pprev;
}
+ // check level 2: verify transaction index validity
+ if (nCheckLevel>1)
+ {
+ pair<unsigned int, unsigned int> pos = make_pair(pindex->nFile, pindex->nBlockPos);
+ mapBlockPos[pos] = pindex;
+ BOOST_FOREACH(const CTransaction &tx, block.vtx)
+ {
+ uint256 hashTx = tx.GetHash();
+ CTxIndex txindex;
+ if (ReadTxIndex(hashTx, txindex))
+ {
+ // check level 3: checker transaction hashes
+ if (nCheckLevel>2 || pindex->nFile != txindex.pos.nFile || pindex->nBlockPos != txindex.pos.nBlockPos)
+ {
+ // either an error or a duplicate transaction
+ CTransaction txFound;
+ if (!txFound.ReadFromDisk(txindex.pos))
+ {
+ printf("LoadBlockIndex() : *** cannot read mislocated transaction %s\n", hashTx.ToString().c_str());
+ pindexFork = pindex->pprev;
+ }
+ else
+ if (txFound.GetHash() != hashTx) // not a duplicate tx
+ {
+ printf("LoadBlockIndex(): *** invalid tx position for %s\n", hashTx.ToString().c_str());
+ pindexFork = pindex->pprev;
+ }
+ }
+ // check level 4: check whether spent txouts were spent within the main chain
+ int nOutput = 0;
+ if (nCheckLevel>3)
+ {
+ BOOST_FOREACH(const CDiskTxPos &txpos, txindex.vSpent)
+ {
+ if (!txpos.IsNull())
+ {
+ pair<unsigned int, unsigned int> posFind = make_pair(txpos.nFile, txpos.nBlockPos);
+ if (!mapBlockPos.count(posFind))
+ {
+ printf("LoadBlockIndex(): *** found bad spend at %d, hashBlock=%s, hashTx=%s\n", pindex->nHeight, pindex->GetBlockHash().ToString().c_str(), hashTx.ToString().c_str());
+ pindexFork = pindex->pprev;
+ }
+ // check level 6: check whether spent txouts were spent by a valid transaction that consume them
+ if (nCheckLevel>5)
+ {
+ CTransaction txSpend;
+ if (!txSpend.ReadFromDisk(txpos))
+ {
+ printf("LoadBlockIndex(): *** cannot read spending transaction of %s:%i from disk\n", hashTx.ToString().c_str(), nOutput);
+ pindexFork = pindex->pprev;
+ }
+ else if (!txSpend.CheckTransaction())
+ {
+ printf("LoadBlockIndex(): *** spending transaction of %s:%i is invalid\n", hashTx.ToString().c_str(), nOutput);
+ pindexFork = pindex->pprev;
+ }
+ else
+ {
+ bool fFound = false;
+ BOOST_FOREACH(const CTxIn &txin, txSpend.vin)
+ if (txin.prevout.hash == hashTx && txin.prevout.n == nOutput)
+ fFound = true;
+ if (!fFound)
+ {
+ printf("LoadBlockIndex(): *** spending transaction of %s:%i does not spend it\n", hashTx.ToString().c_str(), nOutput);
+ pindexFork = pindex->pprev;
+ }
+ }
+ }
+ }
+ nOutput++;
+ }
+ }
+ }
+ // check level 5: check whether all prevouts are marked spent
+ if (nCheckLevel>4)
+ {
+ BOOST_FOREACH(const CTxIn &txin, tx.vin)
+ {
+ CTxIndex txindex;
+ if (ReadTxIndex(txin.prevout.hash, txindex))
+ if (txindex.vSpent.size()-1 < txin.prevout.n || txindex.vSpent[txin.prevout.n].IsNull())
+ {
+ printf("LoadBlockIndex(): *** found unspent prevout %s:%i in %s\n", txin.prevout.hash.ToString().c_str(), txin.prevout.n, hashTx.ToString().c_str());
+ pindexFork = pindex->pprev;
+ }
+ }
+ }
+ }
+ }
}
if (pindexFork)
{
// CAddrDB
//
-bool CAddrDB::WriteAddress(const CAddress& addr)
-{
- return Write(make_pair(string("addr"), addr.GetKey()), addr);
-}
-
bool CAddrDB::WriteAddrman(const CAddrMan& addrman)
{
return Write(string("addrman"), addrman);
}
-bool CAddrDB::EraseAddress(const CAddress& addr)
+bool CAddrDB::LoadAddresses()
{
- return Erase(make_pair(string("addr"), addr.GetKey()));
-}
-
-bool CAddrDB::LoadAddresses(bool &fUpdate)
-{
- bool fAddrMan = false;
if (Read(string("addrman"), addrman))
{
printf("Loaded %i addresses\n", addrman.size());
- fAddrMan = true;
+ return true;
}
+
+ // Read pre-0.6 addr records
vector<CAddress> vAddr;
+ vector<vector<unsigned char> > vDelete;
// Get cursor
Dbc* pcursor = GetCursor();
ssKey >> strType;
if (strType == "addr")
{
- if (fAddrMan)
- fUpdate = true;
- else
- {
- CAddress addr;
- ssValue >> addr;
- vAddr.push_back(addr);
- }
-
+ CAddress addr;
+ ssValue >> addr;
+ vAddr.push_back(addr);
}
}
pcursor->close();
- if (!fAddrMan)
- {
- addrman.Add(vAddr, CNetAddr("0.0.0.0"));
- printf("Loaded %i addresses\n", addrman.size());
- }
+ addrman.Add(vAddr, CNetAddr("0.0.0.0"));
+ printf("Loaded %i addresses\n", addrman.size());
+
+ // Note: old records left; we ran into hangs-on-startup
+ // bugs for some users who (we think) were running after
+ // an unclean shutdown.
return true;
}
bool LoadAddresses()
{
- bool fUpdate = false;
- bool fRet = CAddrDB("cr+").LoadAddresses(fUpdate);
- if (fUpdate)
- CDB::Rewrite("addr.dat", "\004addr");
- return fRet;
+ return CAddrDB("cr+").LoadAddresses();
}
bool fIsEncrypted = false;
//// todo: shouldn't we catch exceptions and try to recover and continue?
- CRITICAL_BLOCK(pwallet->cs_wallet)
{
+ LOCK(pwallet->cs_wallet);
int nMinVersion = 0;
if (Read((string)"minversion", nMinVersion))
{
return DB_NEED_REWRITE;
if (nFileVersion < CLIENT_VERSION) // Update
- {
- // Get rid of old debug.log file in current directory
- if (nFileVersion <= 105 && !pszSetDataDir[0])
- unlink("debug.log");
-
WriteVersion(CLIENT_VERSION);
- }
return DB_LOAD_OK;
}
if (nLastFlushed != nWalletDBUpdated && GetTime() - nLastWalletUpdate >= 2)
{
- TRY_CRITICAL_BLOCK(cs_db)
+ TRY_LOCK(cs_db,lockDb);
+ if (lockDb)
{
// Don't do this if any databases are in use
int nRefCount = 0;
return false;
while (!fShutdown)
{
- CRITICAL_BLOCK(cs_db)
{
+ LOCK(cs_db);
if (!mapFileUseCount.count(wallet.strWalletFile) || mapFileUseCount[wallet.strWalletFile] == 0)
{
// Flush log data to the dat file
mapFileUseCount.erase(wallet.strWalletFile);
// Copy wallet.dat
- filesystem::path pathSrc(GetDataDir() + "/" + wallet.strWalletFile);
+ filesystem::path pathSrc = GetDataDir() / wallet.strWalletFile;
filesystem::path pathDest(strDest);
if (filesystem::is_directory(pathDest))
- pathDest = pathDest / wallet.strWalletFile;
+ pathDest /= wallet.strWalletFile;
try {
#if BOOST_VERSION >= 104000