Hi All,
I was replacing some network equipment and decided to update the software on the linux machine running my node. I restarted the machine after the updates and noticed my used disk space had been reset from near 1.6 TB to about 500 MB (I did not pull up the dashboard right away, it probably started at 0). Now every time I stop and start the node, or remove and pull and start the node, the used disk space goes back to 0. The only change I made with direct impact to the storage node is that I changed the version tag in my run command from “arm” to “beta”. Any time I take the node down, I check that my run command is up to date, so I made this change to align with the setup instructions. When I updated system software, Docker CE was updated as well.
Background:
-Node was v0.26.2 both before and after the restart/update.
-Running on rock64 4GB. Debian Stretch. Docker CE version 5:19.03.5~3-0~debian-stretch
-Node has been active since alpha launch.
-Node was approximately 50% full.
-Node appears online, and other than the used space resetting appears to be working fine.
Run Command:
docker run -d --restart unless-stopped -p 28967:28967 -p 14002:14002 \
--log-opt max-size=50m --log-opt max-file=10 \
-e WALLET="{removed}" \
-e EMAIL="{removed}" \
-e ADDRESS="{removed}:28967" \
-e BANDWIDTH="36.8TB" \
-e STORAGE="3.28TB" \
--mount type=bind,source="/etc/storj-cert",destination=/app/identity \
--mount type=bind,source="/mnt/storj1/v3alpha",destination=/app/config \
--name storagenode storjlabs/storagenode:beta;
I checked that the mount location has r/w access, so no problem there. As far as I know the mount type=bind
option should prevent the node from running if the mount location isn’t available. I am fearful that something has happened to the DB, although bandwidth usage is consistent with before the restart. I always shut down the node with the stop -t 300
option.
What I did to get here:
- Stopped the node:
docker stop -t 300 storagenode
- Upgraded software with
sudo apt update
thensudo apt dist-upgrade
- Restarted the machine
- Ran stop command again in case the node had auto-started on restart (it had not)
docker stop -t 300 storagenode
- Removed container
docker rm storagenode
- Pulled image (but using the beta tag instead of the arm tag I was previously using)
docker pull storjlabs/storagenode:beta
- Started the node with the run command detailed above
- Noticed disk space was reduced to around 500 MB from 1.6 TB
- Dumped the logs (see below)
Here are the first 50 lines from the log after I noticed the used disk discrepancy:
2019-12-11T02:36:14.668Z e[34mINFOe[0m Configuration loaded from: /app/config/config.yaml
2019-12-11T02:36:14.677Z e[34mINFOe[0m Operator email: {removed}
2019-12-11T02:36:14.677Z e[34mINFOe[0m operator wallet: {removed}
2019-12-11T02:36:15.734Z e[34mINFOe[0m version running on version v0.26.2
2019-12-11T02:36:16.659Z e[34mINFOe[0m db.migration Database Version {"version": 25}
2019-12-11T02:36:16.705Z e[34mINFOe[0m bandwidth Performing bandwidth usage rollups
2019-12-11T02:36:16.708Z e[34mINFOe[0m Node 12dqTFg2cPRT4fv4nhxjVvwqbwxumMwknLBBn1wVkqCXSG4qC1m started
2019-12-11T02:36:16.709Z e[34mINFOe[0m Public server started on [::]:28967
2019-12-11T02:36:16.709Z e[34mINFOe[0m Private server started on 127.0.0.1:7778
2019-12-11T02:36:16.709Z e[34mINFOe[0m contact:chore Storagenode contact chore starting up
2019-12-11T02:36:16.741Z e[34mINFOe[0m version running on version v0.26.2
2019-12-11T02:36:25.117Z e[34mINFOe[0m orders.118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW sending {"count": 418}
2019-12-11T02:36:25.120Z e[34mINFOe[0m orders.121RTSDpyNZVcEU84Ticf2L1ntiuUimbWgfATz21tuvgk3vzoA6 sending {"count": 294}
2019-12-11T02:36:25.120Z e[34mINFOe[0m orders.12L9ZFwhzVpuEKMUNUqkaTLGzwY9G24tbiigLiXpmZWKwmcNDDs sending {"count": 172}
2019-12-11T02:36:25.122Z e[34mINFOe[0m orders.12EayRS2V1kEsWESU9QMRseFhdxYxKicsiFmxrsLZHeLUtdps3S sending {"count": 205}
2019-12-11T02:36:25.773Z e[34mINFOe[0m orders.12EayRS2V1kEsWESU9QMRseFhdxYxKicsiFmxrsLZHeLUtdps3S finished
2019-12-11T02:36:26.340Z e[34mINFOe[0m orders.12L9ZFwhzVpuEKMUNUqkaTLGzwY9G24tbiigLiXpmZWKwmcNDDs finished
2019-12-11T02:36:27.284Z e[34mINFOe[0m orders.121RTSDpyNZVcEU84Ticf2L1ntiuUimbWgfATz21tuvgk3vzoA6 finished
2019-12-11T02:36:28.814Z e[34mINFOe[0m orders.118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW finished
2019-12-11T02:36:29.083Z e[34mINFOe[0m piecestore:monitor Remaining Bandwidth {"bytes": 36295416042240}
2019-12-11T02:36:38.822Z e[31mERRORe[0m orders archiving orders {"error": "ordersdb error: database is locked", "errorVerbose": "ordersdb error: database is locked\n\tstorj.io/storj/storagenode/storagenodedb.(*ordersDB).archiveOne:238\n\tstorj.io/storj/storagenode/storagenodedb.(*ordersDB).Archive:202\n\tstorj.io/storj/storagenode/orders.(*Service).handleBatches.func2:213\n\tstorj.io/storj/storagenode/orders.(*Service).handleBatches:230\n\tstorj.io/storj/storagenode/orders.(*Service).sendOrders.func1:164\n\tgolang.org/x/sync/errgroup.(*Group).Go.func1:57"}
2019-12-11T02:41:03.726Z e[34mINFOe[0m piecestore upload started {"Piece ID": "D4SNT57PIYRMYASVTQ3NTDTFWAEVVNQZAQXOYH2CWL5WYUXWNDGA", "Satellite ID": "118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW", "Action": "PUT"}
2019-12-11T02:41:04.295Z e[34mINFOe[0m piecestore upload started {"Piece ID": "X24LLYGBIROPEUM7RDCUFBBI56AUI5LIRWEDMLBEQKMTPUB5GLEA", "Satellite ID": "12L9ZFwhzVpuEKMUNUqkaTLGzwY9G24tbiigLiXpmZWKwmcNDDs", "Action": "PUT"}
2019-12-11T02:41:04.569Z e[34mINFOe[0m piecestore upload started {"Piece ID": "AX6OEGJB23V3R7Z3PJ6QSP7POSNGXWTYPBUPEBGN7SSR5LFUGYUA", "Satellite ID": "118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW", "Action": "PUT"}
2019-12-11T02:41:06.979Z e[34mINFOe[0m piecestore upload started {"Piece ID": "BAAR4XKJEZQERC2JBPLTZQPHSRPNCDPK3XEEHJTPDWHCCKURZ3AA", "Satellite ID": "121RTSDpyNZVcEU84Ticf2L1ntiuUimbWgfATz21tuvgk3vzoA6", "Action": "PUT"}
2019-12-11T02:41:07.290Z e[34mINFOe[0m piecestore upload started {"Piece ID": "E3H4YCFSE5ATF6HG4QMHAYQ6OTR3RBAOSGZD2UQHMZTKUWXQNLZQ", "Satellite ID": "12L9ZFwhzVpuEKMUNUqkaTLGzwY9G24tbiigLiXpmZWKwmcNDDs", "Action": "PUT"}
2019-12-11T02:41:13.907Z e[34mINFOe[0m piecestore upload started {"Piece ID": "DPPIY7HCKI43LKP5TXVRCMAHI3YXPIFRKPY33JJLYLDW7USHC7AA", "Satellite ID": "12EayRS2V1kEsWESU9QMRseFhdxYxKicsiFmxrsLZHeLUtdps3S", "Action": "PUT"}
2019-12-11T02:41:14.236Z e[34mINFOe[0m piecestore upload started {"Piece ID": "LGK7I6ZXXVELNSU537ER6FDV4ZHJZQSKQT444TMIUC7F7JR4EYGA", "Satellite ID": "12L9ZFwhzVpuEKMUNUqkaTLGzwY9G24tbiigLiXpmZWKwmcNDDs", "Action": "PUT"}
2019-12-11T02:41:15.775Z e[31mERRORe[0m piecestore failed to add order {"error": "ordersdb error: database is locked", "errorVerbose": "ordersdb error: database is locked\n\tstorj.io/storj/storagenode/storagenodedb.(*ordersDB).Enqueue:52\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).saveOrder:625\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).doUpload:379\n\tstorj.io/storj/storagenode/piecestore.(*drpcEndpoint).Upload:176\n\tstorj.io/storj/pkg/pb.DRPCPiecestoreDescription.Method.func1:830\n\tstorj.io/drpc/drpcserver.(*Server).doHandle:175\n\tstorj.io/drpc/drpcserver.(*Server).HandleRPC:153\n\tstorj.io/drpc/drpcserver.(*Server).ServeOne:114\n\tstorj.io/drpc/drpcserver.(*Server).Serve.func2:147\n\tstorj.io/drpc/drpcctx.(*Tracker).track:51"}
2019-12-11T02:41:15.776Z e[34mINFOe[0m piecestore uploaded {"Piece ID": "X24LLYGBIROPEUM7RDCUFBBI56AUI5LIRWEDMLBEQKMTPUB5GLEA", "Satellite ID": "12L9ZFwhzVpuEKMUNUqkaTLGzwY9G24tbiigLiXpmZWKwmcNDDs", "Action": "PUT"}
2019-12-11T02:41:15.978Z e[31mERRORe[0m piecestore failed to add order {"error": "ordersdb error: database is locked", "errorVerbose": "ordersdb error: database is locked\n\tstorj.io/storj/storagenode/storagenodedb.(*ordersDB).Enqueue:52\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).saveOrder:625\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).doUpload:379\n\tstorj.io/storj/storagenode/piecestore.(*drpcEndpoint).Upload:176\n\tstorj.io/storj/pkg/pb.DRPCPiecestoreDescription.Method.func1:830\n\tstorj.io/drpc/drpcserver.(*Server).doHandle:175\n\tstorj.io/drpc/drpcserver.(*Server).HandleRPC:153\n\tstorj.io/drpc/drpcserver.(*Server).ServeOne:114\n\tstorj.io/drpc/drpcserver.(*Server).Serve.func2:147\n\tstorj.io/drpc/drpcctx.(*Tracker).track:51"}
2019-12-11T02:41:15.979Z e[34mINFOe[0m piecestore uploaded {"Piece ID": "D4SNT57PIYRMYASVTQ3NTDTFWAEVVNQZAQXOYH2CWL5WYUXWNDGA", "Satellite ID": "118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW", "Action": "PUT"}
2019-12-11T02:41:17.968Z e[31mERRORe[0m piecestore failed to add order {"error": "ordersdb error: database is locked", "errorVerbose": "ordersdb error: database is locked\n\tstorj.io/storj/storagenode/storagenodedb.(*ordersDB).Enqueue:52\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).saveOrder:625\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).doUpload:379\n\tstorj.io/storj/storagenode/piecestore.(*drpcEndpoint).Upload:176\n\tstorj.io/storj/pkg/pb.DRPCPiecestoreDescription.Method.func1:830\n\tstorj.io/drpc/drpcserver.(*Server).doHandle:175\n\tstorj.io/drpc/drpcserver.(*Server).HandleRPC:153\n\tstorj.io/drpc/drpcserver.(*Server).ServeOne:114\n\tstorj.io/drpc/drpcserver.(*Server).Serve.func2:147\n\tstorj.io/drpc/drpcctx.(*Tracker).track:51"}
2019-12-11T02:41:17.969Z e[34mINFOe[0m piecestore uploaded {"Piece ID": "E3H4YCFSE5ATF6HG4QMHAYQ6OTR3RBAOSGZD2UQHMZTKUWXQNLZQ", "Satellite ID": "12L9ZFwhzVpuEKMUNUqkaTLGzwY9G24tbiigLiXpmZWKwmcNDDs", "Action": "PUT"}
2019-12-11T02:41:18.068Z e[31mERRORe[0m piecestore failed to add order {"error": "ordersdb error: database is locked", "errorVerbose": "ordersdb error: database is locked\n\tstorj.io/storj/storagenode/storagenodedb.(*ordersDB).Enqueue:52\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).saveOrder:625\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).doUpload:379\n\tstorj.io/storj/storagenode/piecestore.(*drpcEndpoint).Upload:176\n\tstorj.io/storj/pkg/pb.DRPCPiecestoreDescription.Method.func1:830\n\tstorj.io/drpc/drpcserver.(*Server).doHandle:175\n\tstorj.io/drpc/drpcserver.(*Server).HandleRPC:153\n\tstorj.io/drpc/drpcserver.(*Server).ServeOne:114\n\tstorj.io/drpc/drpcserver.(*Server).Serve.func2:147\n\tstorj.io/drpc/drpcctx.(*Tracker).track:51"}
2019-12-11T02:41:18.069Z e[34mINFOe[0m piecestore uploaded {"Piece ID": "BAAR4XKJEZQERC2JBPLTZQPHSRPNCDPK3XEEHJTPDWHCCKURZ3AA", "Satellite ID": "121RTSDpyNZVcEU84Ticf2L1ntiuUimbWgfATz21tuvgk3vzoA6", "Action": "PUT"}
2019-12-11T02:41:20.313Z e[34mINFOe[0m piecestore upload started {"Piece ID": "R6PUHX5QAQVBG7S3OJ7GS622MCIBRLZ22SETDPWYQQBNZEASE77A", "Satellite ID": "118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW", "Action": "PUT"}
2019-12-11T02:41:20.882Z e[34mINFOe[0m piecestore upload started {"Piece ID": "EDA2UVKEFRZEVLG7MMU5RSF7BDA66A3BV6QH4JAT3GZMHZALVNTQ", "Satellite ID": "118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW", "Action": "PUT"}
2019-12-11T02:41:22.701Z e[34mINFOe[0m piecestore upload started {"Piece ID": "NR2BLBST3WFF6JSZDNIUTCHRXBP3XNPVNXWMU3LWBXBG3FCWKBLQ", "Satellite ID": "121RTSDpyNZVcEU84Ticf2L1ntiuUimbWgfATz21tuvgk3vzoA6", "Action": "PUT"}
2019-12-11T02:41:23.706Z e[34mINFOe[0m piecestore upload started {"Piece ID": "SY5LEW5BK5EUCR6F2Q7DOOYTA2LITAG623K54CNHKIUNUXEVDP4Q", "Satellite ID": "12EayRS2V1kEsWESU9QMRseFhdxYxKicsiFmxrsLZHeLUtdps3S", "Action": "PUT"}
2019-12-11T02:41:24.164Z e[34mINFOe[0m piecestore download started {"Piece ID": "LXLPJAOZE7ZLCCTOBD2S7D5RZ47I2EBRNYUI4QC5BQ4ALNMLQOHA", "Satellite ID": "121RTSDpyNZVcEU84Ticf2L1ntiuUimbWgfATz21tuvgk3vzoA6", "Action": "GET"}
2019-12-11T02:41:24.864Z e[31mERRORe[0m piecestore failed to add order {"error": "ordersdb error: database is locked", "errorVerbose": "ordersdb error: database is locked\n\tstorj.io/storj/storagenode/storagenodedb.(*ordersDB).Enqueue:52\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).saveOrder:625\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).doUpload:379\n\tstorj.io/storj/storagenode/piecestore.(*drpcEndpoint).Upload:176\n\tstorj.io/storj/pkg/pb.DRPCPiecestoreDescription.Method.func1:830\n\tstorj.io/drpc/drpcserver.(*Server).doHandle:175\n\tstorj.io/drpc/drpcserver.(*Server).HandleRPC:153\n\tstorj.io/drpc/drpcserver.(*Server).ServeOne:114\n\tstorj.io/drpc/drpcserver.(*Server).Serve.func2:147\n\tstorj.io/drpc/drpcctx.(*Tracker).track:51"}
2019-12-11T02:41:24.865Z e[34mINFOe[0m piecestore uploaded {"Piece ID": "LGK7I6ZXXVELNSU537ER6FDV4ZHJZQSKQT444TMIUC7F7JR4EYGA", "Satellite ID": "12L9ZFwhzVpuEKMUNUqkaTLGzwY9G24tbiigLiXpmZWKwmcNDDs", "Action": "PUT"}
2019-12-11T02:41:25.272Z e[31mERRORe[0m piecestore failed to add order {"error": "ordersdb error: database is locked", "errorVerbose": "ordersdb error: database is locked\n\tstorj.io/storj/storagenode/storagenodedb.(*ordersDB).Enqueue:52\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).saveOrder:625\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).doUpload:379\n\tstorj.io/storj/storagenode/piecestore.(*drpcEndpoint).Upload:176\n\tstorj.io/storj/pkg/pb.DRPCPiecestoreDescription.Method.func1:830\n\tstorj.io/drpc/drpcserver.(*Server).doHandle:175\n\tstorj.io/drpc/drpcserver.(*Server).HandleRPC:153\n\tstorj.io/drpc/drpcserver.(*Server).ServeOne:114\n\tstorj.io/drpc/drpcserver.(*Server).Serve.func2:147\n\tstorj.io/drpc/drpcctx.(*Tracker).track:51"}
2019-12-11T02:41:25.272Z e[34mINFOe[0m piecestore uploaded {"Piece ID": "AX6OEGJB23V3R7Z3PJ6QSP7POSNGXWTYPBUPEBGN7SSR5LFUGYUA", "Satellite ID": "118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW", "Action": "PUT"}
2019-12-11T02:41:26.586Z e[34mINFOe[0m piecestore upload started {"Piece ID": "ZQPLL4AUZLLZ3ITYDFKYJJVG3BAXFEX3EDPSA3NKXKWP3Q2ZKOOQ", "Satellite ID": "118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW", "Action": "PUT"}
2019-12-11T02:41:26.748Z e[31mERRORe[0m piecestore failed to add order {"error": "ordersdb error: database is locked", "errorVerbose": "ordersdb error: database is locked\n\tstorj.io/storj/storagenode/storagenodedb.(*ordersDB).Enqueue:52\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).saveOrder:625\n\tstorj.io/storj/storagenode/piecestore.(*Endpoint).doUpload:379\n\tstorj.io/storj/storagenode/piecestore.(*drpcEndpoint).Upload:176\n\tstorj.io/storj/pkg/pb.DRPCPiecestoreDescription.Method.func1:830\n\tstorj.io/drpc/drpcserver.(*Server).doHandle:175\n\tstorj.io/drpc/drpcserver.(*Server).HandleRPC:153\n\tstorj.io/drpc/drpcserver.(*Server).ServeOne:114\n\tstorj.io/drpc/drpcserver.(*Server).Serve.func2:147\n\tstorj.io/drpc/drpcctx.(*Tracker).track:51"}
2019-12-11T02:41:26.749Z e[34mINFOe[0m piecestore uploaded {"Piece ID": "DPPIY7HCKI43LKP5TXVRCMAHI3YXPIFRKPY33JJLYLDW7USHC7AA", "Satellite ID": "12EayRS2V1kEsWESU9QMRseFhdxYxKicsiFmxrsLZHeLUtdps3S", "Action": "PUT"}
2019-12-11T02:41:27.923Z e[34mINFOe[0m piecestore download started {"Piece ID": "O2YRJLBWK4SZGIFE6QMR34W7IM32MIX3CCKSBBT2RNQECVM4BWZQ", "Satellite ID": "121RTSDpyNZVcEU84Ticf2L1ntiuUimbWgfATz21tuvgk3vzoA6", "Action": "GET"}
2019-12-11T02:41:31.130Z e[34mINFOe[0m piecestore upload started {"Piece ID": "RKZPRUEABWNBT5W2PX4ROI33FAKRQI6V7NKTUJD2LWTQNOFCY6FA", "Satellite ID": "118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW", "Action": "PUT"}
I tried restarting the container with no effect, the used disk space reset. I tried removing and recreating the container, the used disk space reset. When I did this the only thing that changed is now I get this at the beginning of the logs (which was not present after the initial restart):
2019-12-11T04:17:38.642Z INFO piecestore:monitor Remaining Bandwidth {"bytes": 36293612318976}
2019-12-11T04:17:38.642Z WARN piecestore:monitor Disk space is less than requested. Allocating space {"bytes": 1868984774656}
I checked, and the space is still being used on the disk. I have currently stopped the node in the hopes that I might be able to fix it without collecting too much other data in the mean time. I’m not a total newbie at this, and I think I ran down all of the usual problems/troubleshooting steps. If you made it this far, thanks for your time with this lengthy post.
(edit: fixed typos)