Hi, I would like to clarify how the garbage collector works.
I have this in my config file:
What I would expect:
When the reported size is above 600 GB, the garbage collector starts to remove data until the reported size is 600 * 0.9 = 540 GB.
What I actually see:
Garbage collector starts to remove data but stops on ~20 GB of reportedly used data. I have no idea why so little. It removed 580 GB of data, ~97 % of all data.
Why it happens so? It does a similar thing when I set 30 GB of repo size (it stops on ~2 GB or so). How I could define it does not need to remove data too much? My version is the newest 0.9.0 but it happened the same thing on 0.7.0 as well.
Another issue is the garbage collector is very disk-intensive and IPFS starts to timeout – so for several hours, the IPFS daemon is almost unusable. So this useless deleting is pain even more. What works is to rename the data folder, start to (slowly) delete, and let IPFS start again. But I have plenty of disk space, I can participate instead of deleting all and downloading some stuff again.
What should I do? Thanks.