the recommended max size is 24TB of each node… however this is more on a per ip subnet basis.
at 24TB we suspect, not enough live data on how the tardigrade users will work with their data… however if we say 5% is deleted each month on avg…
then when you get to 20TB that would mean 1TB of data is deleted each month, over the last month i got about 500gb ingress, sure atm the node is growing… but since the network is new a good deal of the data will be test data, so nobody knows how it will work out eventually.
you can set it however you like, however the max capacity ends up being a ratio of the % of data deleted per month vs new ingress to the network… minimum recommended is 500GB to 24TB + 10% for trash and basic node data operations… tho maybe at the higher end of the spectrum one can go a lot lower in the over allocation % but recommended is 10%
like say if you got a 40TB node… then 4TB free space for whatever it has to do seems … excessive to say the least…
sometimes data will flow in fast, sometimes slow … it’s not easy to predict, and i bet storj labs likes it that way… atleast with their test data… the network is slowly getting some media attention and has a few interesting partnerships, and performance seems to be great… so should just be a matter of time before some serious data starts to flood in…
even having two nodes you run into the same ratio issue, the only way to get around that is using multiple ip’s to ingress the ingress and thus reduce the net drain from deletions over time.
but the network has be very quite lately… so presently it’s barely an advantage…
and in regard to unraid it sounds very much like a replication type deal… maybe with some layered storage involved… which is without a doubt a very cool solution… very redundant on a file to file basis because one would select the level of backup / replication one wants for each file or folder.
it has some advantages ofc the redundant data stored would then to my understanding take up more space than when using raid solutions… but the iops and such would be much greater, without a doubt a worthwhile setup to be running especially if one has very specific and variable data redundancy demands and high utilization maybe…
kinda guessing a bit here… but should be its strong suits if it indeed is a layered storage with replication type redundancy.
sounds cool, i did strongly consider such a solution when i was deciding on my storage solution, looked at windows storage spaces at the time because i wasn’t really into the whole linux aspect of things at the time…
ended up learning about zfs and just had to have it because of all the amazing things about that… checksum, copy on write, l2arc, slog and such goodness.
can be a bit heavy to work with sometimes tho…