Bitcoin Forum
May 12, 2024, 11:43:58 AM *
News: Latest Bitcoin Core release: 27.0 [Torrent]
 
   Home   Help Search Login Register More  
Pages: [1]
  Print  
Author Topic: Data directory over NFS  (Read 1496 times)
myfirst (OP)
Full Member
***
Offline Offline

Activity: 156
Merit: 100


Crypto Currency Developer


View Profile
May 19, 2015, 04:39:56 PM
 #1

Is it safe to keep the blockchain data and wallet.dat file on a NFS mounted share?
What happens if the NFS server temporarily becomes unavailable?  Will the databases get damaged?
1715514238
Hero Member
*
Offline Offline

Posts: 1715514238

View Profile Personal Message (Offline)

Ignore
1715514238
Reply with quote  #2

1715514238
Report to moderator
1715514238
Hero Member
*
Offline Offline

Posts: 1715514238

View Profile Personal Message (Offline)

Ignore
1715514238
Reply with quote  #2

1715514238
Report to moderator
The grue lurks in the darkest places of the earth. Its favorite diet is adventurers, but its insatiable appetite is tempered by its fear of light. No grue has ever been seen by the light of day, and few have survived its fearsome jaws to tell the tale.
Advertised sites are not endorsed by the Bitcoin Forum. They may be unsafe, untrustworthy, or illegal in your jurisdiction.
1715514238
Hero Member
*
Offline Offline

Posts: 1715514238

View Profile Personal Message (Offline)

Ignore
1715514238
Reply with quote  #2

1715514238
Report to moderator
1715514238
Hero Member
*
Offline Offline

Posts: 1715514238

View Profile Personal Message (Offline)

Ignore
1715514238
Reply with quote  #2

1715514238
Report to moderator
1715514238
Hero Member
*
Offline Offline

Posts: 1715514238

View Profile Personal Message (Offline)

Ignore
1715514238
Reply with quote  #2

1715514238
Report to moderator
2112
Legendary
*
Offline Offline

Activity: 2128
Merit: 1068



View Profile
May 19, 2015, 10:42:48 PM
 #2

Most likely not safe. The LevelDB isn't really 100% ACID, especially the chainstate directory is prone to corruptions if the node crashes.

Edit: on the other hand if you have a tight backup schedule or the underlying file system supports snapshots then it is OK to run them over NFS because neither the blocks nor chainstate LevelDB databases are really critical and can be always rebuilt after the corruption. It all depends on what exactly is your tolerance profile for outages.

Please comment, critique, criticize or ridicule BIP 2112: https://bitcointalk.org/index.php?topic=54382.0
Long-term mining prognosis: https://bitcointalk.org/index.php?topic=91101.0
myfirst (OP)
Full Member
***
Offline Offline

Activity: 156
Merit: 100


Crypto Currency Developer


View Profile
May 21, 2015, 01:01:34 AM
Last edit: May 21, 2015, 01:28:18 AM by myfirst
 #3

Most likely not safe. The LevelDB isn't really 100% ACID, especially the chainstate directory is prone to corruptions if the node crashes.

Edit: on the other hand if you have a tight backup schedule or the underlying file system supports snapshots then it is OK to run them over NFS because neither the blocks nor chainstate LevelDB databases are really critical and can be always rebuilt after the corruption. It all depends on what exactly is your tolerance profile for outages.

I have 3 servers that host nodes for several different block chains (altcoins) and 1 NFS server used to store the blockchain and wallet data files.

I've certainly starting to run into issues with the altcoins based on 0.8.x and 0.9.x codebase.  So far Bitcoin 0.10.1 seems to do okay.  The NFS server kernel panicked while the nodes where still running and it seems that the nodes continue to 'write' to the databases even though the mount is offline.  The operating system has 'paused' all operations to that mount yet the running mode does not seem to honor that.

The end result is devastating at times and fits in to these 3 categories:

  • wallet.dat will become damaged, so will backups
  • "Large valid fork found" error
  • "Error loading blkindex.dat"

In all cases you don't know something is wrong until after you restart the node, which make matters worse.

wallet.dat

I run wallet.dat backups every hour thinking I would be safe.  However when wallet.dat gets damaged, the backup is also damaged.  This is strange and the worst case.  Thankfully I've only seen it once and happen to have had an older backup that was not damaged.

With both the original wallet.dat and the backup version you will see something like this in debug.log indicating it's hosed.

Code:
2015-05-21 00:39:12  block index            5175ms
2015-05-21 00:39:12 init message: Loading wallet...
2015-05-21 00:39:12 ERROR: CTransaction::CheckTransaction() : vout empty
2015-05-21 00:39:22 Error reading wallet database: CPrivKey pubkey inconsistency
2015-05-21 00:39:24 Error loading wallet.dat: Wallet corrupted
 wallet                12226ms
2015-05-21 00:39:24 init message: Rescanning.
2015-05-21 00:42:24 AddToWallet ....<redacted txid>  
2015-05-21 00:42:24 AddToWallet ....<redacted txid>
2015-05-21 00:42:25 AddToWallet ....<redacted txid>  
...
2015-05-21 00:43:27  rescan               242591ms
2015-05-21 00:43:27 init message: Loading addresses...
2015-05-21 00:43:27 Loaded 19079 addresses from peers.dat  62ms
2015-05-21 00:43:32 Shutdown : In progress...
2015-05-21 00:43:32 Flush(false)
2015-05-21 00:43:32 wallet.dat refcount=0
2015-05-21 00:43:32 wallet.dat checkpoint
2015-05-21 00:43:32 wallet.dat detach
2015-05-21 00:43:33 wallet.dat closed
2015-05-21 00:43:33 DBFlush(false) ended             301ms
2015-05-21 00:43:33 StopNode()
2015-05-21 00:43:33 Flushed 19079 addresses to peers.dat  47ms
2015-05-21 00:43:33 Committing 252 changed transactions to coin database...
2015-05-21 00:43:33 Flush(true)
2015-05-21 00:43:33 wallet.dat refcount=0
2015-05-21 00:43:33 wallet.dat checkpoint
2015-05-21 00:43:33 wallet.dat detach
2015-05-21 00:43:33 wallet.dat closed
2015-05-21 00:43:33 DBFlush(true) ended             240ms
2015-05-21 00:43:33 Shutdown : done

and in db.log

Code:
Page 1229: unreferenced page
Page 1229: totally zeroed page
Page 1230: unreferenced page
Page 1230: totally zeroed page
Page 1231: unreferenced page
Page 1231: totally zeroed page
wallet.dat: DB_VERIFY_BAD: Database verification failed

The question is, why is the backup corrupted?

Large valid fork found

This error I've only seen with 0.9.x version, it's not the worst thing in the world as long as you catch it.  As with all cases the error does not appear until after the node is restarted.

Code:
2015-05-21 00:18:20 CheckForkWarningConditions: Warning: Large valid fork found
  forking the chain at height 2887908 (0000000000017225f9c08e76062d60aa616a14781e6e89a82febe6bd90167052)
  lasting to height 2888067 (000000000000f6f16950d0af208bf8087a4c726cb7d935329bd74985f02a9b75).
Chain state database corruption likely.

Code:
 Warning: The network does not appear to fully agree! Some miners appear to be experiencing issues.

The node continues to run in this state until you do a reindex.  Very scary.

Error loading blkindex.dat

This has been a very common issue, especially with older versions (0.8.x and whatever Peercoin is based off of).  In some versions I can delete  blkindex.dat and it will reindex on it's own, in other version I have to run with the 'reindex' command.  Worst case here is I need to re-download the chain.


debug.log will look like this, the node will attempt to start then shutdown:

Code:
05/21/15 00:10:28 Verifying last 2500 blocks at level 1
05/21/15 00:10:28 ERROR: CheckProofOfWork() : nBits below minimum work
05/21/15 00:10:28 ERROR: CBlock::ReadFromDisk() : errors in block header
05/21/15 00:10:28 ERROR: LoadBlockIndex() : block.ReadFromDisk failed
05/21/15 00:10:28  block index           88932ms
05/21/15 00:10:28 Loading wallet...
05/21/15 00:10:28 nFileVersion = 80501
05/21/15 00:10:28 Error loading blkindex.dat
 wallet                  282ms
05/21/15 00:10:28 Loading addresses...
05/21/15 00:10:28 Loaded 11294 addresses from peers.dat  72ms
05/21/15 00:10:28 mapBlockIndex.size() = 1509863
05/21/15 00:10:28 nBestHeight = 1508834
05/21/15 00:10:28 setKeyPool.size() = 101
05/21/15 00:10:28 mapWallet.size() = 0
05/21/15 00:10:28 mapAddressBook.size() = 1
05/21/15 00:10:28 Done loading
05/21/15 00:10:28 Emerald: Error loading blkindex.dat

Here is another example where I had to run reindex manually

Code:
2015-05-20 23:56:10 Opened LevelDB successfully
2015-05-20 23:56:49 LoadBlockIndexDB(): last block file = 4
2015-05-20 23:56:49 LoadBlockIndexDB(): last block file info: CBlockFileInfo(blocks=72806, size=32576240, heights=1128729...1199945, time=2015-04-17...2015-05-20)
2015-05-20 23:56:49 LoadBlockIndexDB(): transaction index disabled
2015-05-20 23:56:49 LoadBlockIndexDB(): hashBestChain=000000000000521f11b8f41aa943f28e6ff5f2bd304d0dc7cef217c442812790  height=1199945 date=2015-05-20 23:38:48
2015-05-20 23:56:49 init message: Verifying blocks...
2015-05-20 23:56:49 Verifying last 288 blocks at level 3
2015-05-20 23:56:52 ERROR: CheckProofOfWork() : nBits below minimum work
2015-05-20 23:56:52 ERROR: ReadBlockFromDisk(CBlock&, CDiskBlockPos&) : errors in block header
2015-05-20 23:56:52 ERROR: VerifyDB() : *** ReadBlockFromDisk failed at 1199799, hash=000000000000befd4121befe5272aff3767133e6205d073a87086876273a55c7
2015-05-20 23:56:52 : Corrupted block database detected.

Do you want to rebuild the block database now?
2015-05-20 23:56:52 Aborted block database rebuild. Exiting.
2015-05-20 23:56:52 Flush(false)
2015-05-20 23:56:52 DBFlush(false) ended               0ms
2015-05-20 23:56:52 StopNode()
2015-05-20 23:56:52 Flushed 0 addresses to peers.dat  23ms
2015-05-20 23:56:52 Committing 161 changed transactions to coin database...
2015-05-20 23:56:52 Flush(true)
2015-05-20 23:56:52 DBFlush(true) ended               0ms

With one coin there was no blkindex.dat file present (even with a valid data directory) yet it complains it can't load it.  This one I had to re-download the chain.

Amazingly (and thankfully) the Bitcoin blockchain has not suffered corruption as of yet.

The worst part about all this is that not even your backups are safe.  

Besides NOT using NFS (not an option for the time being), is there a way to minimize the damage?  Would taking hourly volume snapshots on the NFS server be of any value?  It runs ZFS.
achow101_alt
Sr. Member
****
Offline Offline

Activity: 268
Merit: 256


View Profile
May 21, 2015, 01:10:27 AM
 #4

Since you are running multiple nodes, the nodes will write and overwrite things in the data directory. This causes corruption because the nodes don't know what the other nodes wrote and the sudden change of data along with simultaneous reading and writing done by the nodes to the data directory causes all of these issues. The data is then constantly changed and corrupted, causing you to need to constantly reindex or redownload the blockchain. Since all of the nodes are also accessing the same wallet.dat file, they cause it to be corrupted like the blockchain database files. The only weird thing is that the backups were also corrupted. You should keep the backups on a separate machine just in case.

Tip Me!: 1AQx99s7q1wVinbgXbA48BaZQVWpHe5gYM | My PGP Key: Fingerprint 0x17565732E08E5E41
myfirst (OP)
Full Member
***
Offline Offline

Activity: 156
Merit: 100


Crypto Currency Developer


View Profile
May 21, 2015, 01:16:54 AM
 #5

Since you are running multiple nodes, the nodes will write and overwrite things in the data directory.

The nodes each write to their own data directory on the same NFS server but independent of each other.  The nodes are for various different altcoins.

Since all of the nodes are also accessing the same wallet.dat file, they cause it to be corrupted like the blockchain database files. The only weird thing is that the backups were also corrupted. You should keep the backups on a separate machine just in case.

It's not the same wallet.dat file, each node has it's own wallet.dat and blockchain.  The wallet.dat is backed up to the local filesystem, not the NFS mount.
achow101_alt
Sr. Member
****
Offline Offline

Activity: 268
Merit: 256


View Profile
May 21, 2015, 01:39:50 AM
 #6

Since you are running multiple nodes, the nodes will write and overwrite things in the data directory.

The nodes each write to their own data directory on the same NFS server but independent of each other.  The nodes are for various different altcoins.

Since all of the nodes are also accessing the same wallet.dat file, they cause it to be corrupted like the blockchain database files. The only weird thing is that the backups were also corrupted. You should keep the backups on a separate machine just in case.

It's not the same wallet.dat file, each node has it's own wallet.dat and blockchain.  The wallet.dat is backed up to the local filesystem, not the NFS mount.
Sorry, I misunderstood your problem.

I think that there may be open issues with some of the altcoins. You should check the git repos and see if there are any issues that are similar to yours. Also, it would be helpful if we knew what the coins you have on this NFS.

Regarding your corrupted wallet.dat, perhaps your backup had accidentally backed up a corrupted version of the wallet before you noticed?

Tip Me!: 1AQx99s7q1wVinbgXbA48BaZQVWpHe5gYM | My PGP Key: Fingerprint 0x17565732E08E5E41
myfirst (OP)
Full Member
***
Offline Offline

Activity: 156
Merit: 100


Crypto Currency Developer


View Profile
May 21, 2015, 02:12:51 AM
 #7

I think that there may be open issues with some of the altcoins. You should check the git repos and see if there are any issues that are similar to yours. Also, it would be helpful if we knew what the coins you have on this NFS.

The problematic ones have been:

  • Peercoin,Joulecoin,Emerald <= damaged blockindex.dat
  • Zetacoin <= the forking issue
  • Mooncoin <= hosed wallet.dat (only happened once)

Emerald and Peercoin have been the most troublesome.  I have several others on NFS (probably more than 20 total at the moment) like unobtanium, litecoin, dash, earthcoin, etc... None of those have had problems.

Regarding your corrupted wallet.dat, perhaps your backup had accidentally backed up a corrupted version of the wallet before you noticed?

That is most likely the problem. The wallet.dat backups run automatically every hour via RPC 'backupwallet()'.  When the NFS server reboots and the node has not been restarted, the 'backupwallet' command will result in a damaged wallet.

I think what I will end up doing it setting up a script that will trigger the nodes to restart if NFS goes offline and online.  One thing I have not tried is to keep NFS offline indefinitely.  I am wondering how long before the node crashes.  Might try that on a test server.
2112
Legendary
*
Offline Offline

Activity: 2128
Merit: 1068



View Profile
May 21, 2015, 04:24:08 AM
 #8

I think you are doing things wrong. I understand that your problems are actually with BerkeleyDB.

1) always hard mount the NFS directories with the initial mount done in background if the server is not available (NFS mount options "hard,bg")

2) do the backups from the client side, not from the server. This way the backup program sees the most recent data. The backups done on the server are only consistent when the

3) change the initialization options of BerkeleyDB to make the memory mapped files visible in the file system (I think the files have names like "_db.000"). I don't remember if that can be done with simply adding flags to the "DB_CONFIG" or if it requires recompilation. With those files visible you can run the db_utils on the client while coins clients are running and make fully consistent backups using db_backup and do live consistency checks with db_verify. Just remember to add "set_lg_dir database" to the DB_CONFIG so the BerkeleyDB utilities can find the live  logs.

With those caveats I've successfully run the BerkeleyDB-based programs over NFS for very many years with the central storage & build server, longer than the Bitcoin was in existence.

Please comment, critique, criticize or ridicule BIP 2112: https://bitcointalk.org/index.php?topic=54382.0
Long-term mining prognosis: https://bitcointalk.org/index.php?topic=91101.0
myfirst (OP)
Full Member
***
Offline Offline

Activity: 156
Merit: 100


Crypto Currency Developer


View Profile
May 22, 2015, 12:06:55 AM
 #9

I think you are doing things wrong. I understand that your problems are actually with BerkeleyDB.

1) always hard mount the NFS directories with the initial mount done in background if the server is not available (NFS mount options "hard,bg")

I'll experiment with those options.  I think on my OS 'hard' is default.

2) do the backups from the client side, not from the server. This way the backup program sees the most recent data. The backups done on the server are only consistent when the

I think that is what I am doing when I use the RPC function?:

Code:
2015-05-21 23:55:02 Using data directory /mnt/storage01/blockchains/bitcoin
2015-05-21 23:55:02 Using config file /usr/local/etc/bitcoin.conf
2015-05-21 23:55:02 Using at most 20 connections (176283 file descriptors available)
2015-05-21 23:55:02 Using 4 threads for script verification
2015-05-21 23:55:02 Binding RPC on address :: port 8332 (IPv4+IPv6 bind any: 1)
2015-05-21 23:55:02 Using wallet wallet.dat
2015-05-21 23:55:02 init message: Verifying wallet...
2015-05-21 23:55:02 CDBEnv::Open : LogDir=/mnt/storage01/blockchains/bitcoin/database ErrorFile=/mnt/storage01/blockchains/bitcoin/db.log
2015-05-21 23:55:02 Bound to [::]:8333
2015-05-21 23:55:02 Bound to 0.0.0.0:8333
2015-05-21 23:55:02 init message: Loading block index...
2015-05-21 23:55:02 Opening LevelDB in /mnt/storage01/blockchains/bitcoin/blocks/index
2015-05-21 23:55:02 Opened LevelDB successfully
2015-05-21 23:55:02 Opening LevelDB in /mnt/storage01/blockchains/bitcoin/chainstate
2015-05-21 23:55:03 Opened LevelDB successfully
...
2015-05-21 24:25:04 copied wallet.dat to /var/backups/wallets/Bitcoin.dat

3) change the initialization options of BerkeleyDB to make the memory mapped files visible in the file system (I think the files have names like "_db.000"). I don't remember if that can be done with simply adding flags to the "DB_CONFIG" or if it requires recompilation. With those files visible you can run the db_utils on the client while coins clients are running and make fully consistent backups using db_backup and do live consistency checks with db_verify. Just remember to add "set_lg_dir database" to the DB_CONFIG so the BerkeleyDB utilities can find the live  logs.

This is interesting, I think peercoin does that already since I've seen those files.  Anyway, where can I get more information about trying this suggestion.
myfirst (OP)
Full Member
***
Offline Offline

Activity: 156
Merit: 100


Crypto Currency Developer


View Profile
May 22, 2015, 12:19:28 AM
 #10

Bummer, just had another wallet go bad:

Code:
05/21/15 22:20:02 KolschCoin version v1.0.0.0 ()
05/21/15 22:20:02 Using OpenSSL version OpenSSL 0.9.8zd-freebsd 8 Jan 2015
05/21/15 22:20:02 Default data directory /.KolschCoin
05/21/15 22:20:02 Used data directory /mnt/storage02/blockchains/KolschCoin
05/21/15 22:20:02 dbenv.open LogDir=/mnt/storage02/blockchains/KolschCoin/database ErrorFile=/mnt/storage02/blockchains/KolschCoin/db.log
05/21/15 22:20:02 Renamed wallet.dat to wallet.1432246802.bak
05/21/15 22:20:02 Error: Salvage found errors, all data may not be recoverable.
05/21/15 22:20:02 Salvage(aggressive) found no records in wallet.1432246802.bak.
05/21/15 22:20:02 KolschCoin: wallet.dat corrupt, salvage failed
05/21/15 22:20:02 Flush(false)
05/21/15 22:20:02 DBFlush(false) ended               5ms
05/21/15 22:20:02 StopNode()
05/21/15 22:20:02 Flushed 0 addresses to peers.dat  14ms
05/21/15 22:20:02 Flush(true)
05/21/15 22:20:02 DBFlush(true) ended               4ms
05/21/15 22:20:03 KolschCoin exited
2112
Legendary
*
Offline Offline

Activity: 2128
Merit: 1068



View Profile
May 25, 2015, 08:30:32 PM
 #11

This is interesting, I think peercoin does that already since I've seen those files.  Anyway, where can I get more information about trying this suggestion.
Actually, after checking, there is a command line/*coin.conf for this in the recent code: -privdb=0 .

But no matter what you do you will have to read the Berkeyley DB documentation anyway.

While I didn't run the most recent Bitcoin clients over NFS, I did quite a bit of load testing in times of version 0.6.3 (and thereabouts) and did not have anything like your problems. My professional opinion is that you are suffering from some sort of hardware problem or misconfiguration. NFS is quite fragile and tends to stress the infrastructure, especially if run over UDP. Check your port mapper RPC error statistics, verify the network switches for correct configuration of flow control, verify the correct operation of checksum offload in the drivers, etc.

I didn't do anything serious then, the coin daemons were simply the second thing that popped into my mind when I had to think for a load test for our revamped/reinstalled NAS & SAN hardware running under Solaris 9 and 10. The first tests were actually "cp -rpv" of a terabytes-sized collection of Linux ISOs followed by "md5sum".  There were some errors, but in particular all my wallet.dat-s survived without any damage.

Please comment, critique, criticize or ridicule BIP 2112: https://bitcointalk.org/index.php?topic=54382.0
Long-term mining prognosis: https://bitcointalk.org/index.php?topic=91101.0
Pages: [1]
  Print  
 
Jump to:  

Powered by MySQL Powered by PHP Powered by SMF 1.1.19 | SMF © 2006-2009, Simple Machines Valid XHTML 1.0! Valid CSS!