Until today I monitored my logs manually trough docker logs --tail 20 storagenode.
What I’ve seen is, that my satellite uploads work just fine (upload starts and finishes properly).
Only my satellite downloads seem to have a problem, because all Downloads I’ve seen have failed.
So I found the successrate.sh script , let it run and got a bit shocked:
To my system:
Host is a Unraid System
Ubuntu 19.10 runs in a VM with a 4TB HDD dedicated only for Storj (nothing else runs or is stored on that hdd).
UPS for the entire System (Server, Switch, Router, Modem)
Internet:
Fibreoptic with 400 Mbit / 200 Mbit and a Ping of 12 ms (what I actually receive directly in my Ubuntu VM)
Static IP
I then thought that my hdd might be stressed, so I tested it:
Those numbers look just fine. The downloads fail because others beat your node in the race to upload to the client. My download success rate hovers around 60%. There are many factors that affect this, but it is not something to worry about. It is part of the normal operation of the system.
You can check out this thread for some comparisons, although older posts may not be a good reflection of the current behaviour of the network.
thanks for you feedback. You’re right. That seems like I loose the race very often. But that is not acceptable for me and the hardware / ISP I use as well that I get paid by successfull uploads to the clients.
So I tried to optimized my server settings and network connections, so that my Ubuntu VM is privileged in my arrangement and that seems to help:
I have been seeing all of those same error messages for download fails from the same satellite. Although, my download success rate has been hovering between 20-30% since the big uptick in download traffic this month.
Just within the last day, I’ve noticed my download success rate has jumped back up to ~50%, but still not great. Upload success rate hovering around 90-95%.
Moscow
--- A L L S A T E L L I T E S---
118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW : 46,64% bandwidth, 88,43% effeciency
12L9ZFwhzVpuEKMUNUqkaTLGzwY9G24tbiigLiXpmZWKwmcNDDs : 16,53% bandwidth, 99,59% effeciency
12EayRS2V1kEsWESU9QMRseFhdxYxKicsiFmxrsLZHeLUtdps3S : 18,83% bandwidth, 99,6% effeciency
121RTSDpyNZVcEU84Ticf2L1ntiuUimbWgfATz21tuvgk3vzoA6 : 18,00% bandwidth, 99,64% effeciency
Name Value
---- -----
1. DOWNLOAD [-------------------------- ] 86,672 617565 of 712532
2. UPLOAD [----------------------------- ] 98,2502 1400738 of 1425687
3. GET_AUDIT [------------------------------] 99,9936 31406 of 31408
4. GET [-------------------------- ] 86,0256 561099 of 652248
5. GET_REPAIR [-------------------------- ] 86,7849 25060 of 28876
6. PUT [----------------------------- ] 98,2497 1393415 of 1418241
7. PUT_REPAIR [------------------------------] 98,3481 7323 of 7446
9. EFFECIENCY [---------------------------- ] 94,392 2018303 of 2138219
France
--- A L L S A T E L L I T E S---
118UWpMCHzs6CvSgWd9BfFVjw5K9pZbJjkfZJexMtSkmKxvvAW : 38,79% bandwidth, 99,43% effeciency
12L9ZFwhzVpuEKMUNUqkaTLGzwY9G24tbiigLiXpmZWKwmcNDDs : 20,26% bandwidth, 99,94% effeciency
12EayRS2V1kEsWESU9QMRseFhdxYxKicsiFmxrsLZHeLUtdps3S : 24,55% bandwidth, 99,77% effeciency
121RTSDpyNZVcEU84Ticf2L1ntiuUimbWgfATz21tuvgk3vzoA6 : 16,40% bandwidth, 99,96% effeciency
Name Value
---- -----
1. DOWNLOAD [------------------------------] 99,6591 411881 of 413290
2. UPLOAD [------------------------------] 99,7256 855775 of 858130
3. GET_AUDIT [------------------------------] 100 8338 of 8338
4. GET [------------------------------] 99,6521 403541 of 404950
5. GET_REPAIR [------------------------------] 100 2 of 2
6. PUT [------------------------------] 99,7408 850262 of 852472
7. PUT_REPAIR [----------------------------- ] 97,4373 5513 of 5658
9. EFFECIENCY [------------------------------] 99,704 1267656 of 1271420
So I copied a 800 MB Log file to my PC. Running the success rate script and realized the amount of RAM it’s using to just process the file. 7 GB. I had about 1 GB left of RAM and I have 16 GB of RAM in total.
I have now setup a task scheduler job on my NAS Server to execute a batch file that will delete the logs on the 1st of ever month so I can avoid this next time.
I am now am just waiting for the script to finish. Think it’s 2-3 months of data.
Here are my results, looking pretty good.
I would recommend keeping the old log files and executing the script on them using another PC instead. You never know when your node might get in trouble & log files can help diagnose any issue.