Replies: 2 comments 6 replies
-
In borg 1.x there is a (very high) design limitation on the max. archive size, Other than that, there are some timing and memory usage influences:
I maybe would try to avoid creating "monster repos" and prefer multiple smaller repos, using some natural partitioning of your data. Sometimes there are also other reasons for partitioning, e.g. for pruning one wants to have system file archives separate from user file archives. In borg 1.x there are some reasons for not using 1 repo for multiple borg clients (see FAQ). |
Beta Was this translation helpful? Give feedback.
-
@awgcooper I am having similar thoughts about large archives. Having a 10TiB archive (for maximum deduplication) I conclude, that eg. the time for Therefor splitting in logical sub units is my goal, accepting loss of some deduplication efficiency. On the other hand, as with a future Last thing is only "safe(r)" if errors in one repository cannot be "populated" to another repository with the above mentioned commands. (as Just there is a thought @ThomasWaldmann, concerning |
Beta Was this translation helpful? Give feedback.
-
I want to try and understand the specific risks associated with having an archive size that is 'large' (unless the answers are the same for both, let's assume two sizes for the purpose of this question: 2-5TB and ≥5TB). I get the basic point that the larger the archive, the larger amount of irrecoverable data there is should the archive become severely corrupted or unreadable. However, is there a causal connection between archive size and the probability of it becoming corrupted, i.e. an archive of size X is A% more likely to become corrupted than one of 0.5X, once X exceeds a certain size. I'm using the term 'corrupted' here to mean 'archive-wide corruption' which is, I appreciate, a pretty bad scenario.
In terms of other potential downsides:
Anything else specific?
Beta Was this translation helpful? Give feedback.
All reactions