[pbs-devel] [pve-devel] [PATCH v2 proxmox-backup-qemu 05/11] access: use bigger cache and LRU chunk reader
Stefan Reiter
s.reiter at proxmox.com
Wed Mar 17 14:37:44 CET 2021
On 16/03/2021 21:17, Thomas Lamprecht wrote:
> On 03.03.21 10:56, Stefan Reiter wrote:
>> Values chosen by fair dice roll, seems to be a good sweet spot on my
>> machine where any less causes performance degradation but any more
>> doesn't really make it go any faster.
>>
>> Keep in mind that those values are per drive in an actual restore.
>>
>> Signed-off-by: Stefan Reiter <s.reiter at proxmox.com>
>> ---
>>
>> Depends on new proxmox-backup.
>>
>> v2:
>> * unchanged
>>
>> src/restore.rs | 5 +++--
>> 1 file changed, 3 insertions(+), 2 deletions(-)
>>
>> diff --git a/src/restore.rs b/src/restore.rs
>> index 0790d7f..a1acce4 100644
>> --- a/src/restore.rs
>> +++ b/src/restore.rs
>> @@ -218,15 +218,16 @@ impl RestoreTask {
>>
>> let index = client.download_fixed_index(&manifest, &archive_name).await?;
>> let archive_size = index.index_bytes();
>> - let most_used = index.find_most_used_chunks(8);
>> + let most_used = index.find_most_used_chunks(16); // 64 MB most used cache
>
>
>
>>
>> let file_info = manifest.lookup_file_info(&archive_name)?;
>>
>> - let chunk_reader = RemoteChunkReader::new(
>> + let chunk_reader = RemoteChunkReader::new_lru_cached(
>> Arc::clone(&client),
>> self.crypt_config.clone(),
>> file_info.chunk_crypt_mode(),
>> most_used,
>> + 64, // 256 MB LRU cache
>
> how does this work with low(er) memory situations? Lots of people do not over
> dimension their memory that much, and especially the need for mass-recovery could
> seem to correlate with reduced resource availability (a node failed, now I need
> to restore X backups on my <test/old/other-already-in-use> node, so multiple
> restore jobs may run in parallel, and they all may have even multiple disks,
> so tens of GiB of memory just for the cache are not that unlikely.
This is a seperate function from the regular restore, so it currently
only affects live-restore. This is not an operation you would usually do
under memory constraints anyway, and regular restore is unaffected if
you just want the data.
Upcoming single-file restore too though, I suppose, where it might make
more sense...
>
> How is the behavior, hard failure if memory is not available? Also, some archives
> may be smaller than 256 MiB (EFI disk??) so there it'd be weird to have 256 cache
> and get 64 of most used chunks if that's all/more than it would actually need to
> be..
Yes, if memory is unavailable it is a hard error. Memory should not be
pre-allocated however, so restoring this way will only ever use as much
memory as the disk size (not accounting for overhead).
>
> There may be the reversed situation too, beefy fast node with lots of memory
> and restore is used as recovery or migration but network bw/latency to PBS is not
> that good - so bigger cache could be wanted.
The reason I chose the numbers I did was that I couldn't see any real
performance benefits by going higher, though I didn't specifically test
with slow networking.
I don't believe more cache would improve the situation there though,
this is mostly to avoid random access from the guest and the linear
access from the block-stream operation to interfere with each other, and
allow multiple smaller guest reads within the same chunk to be served
quickly.
>
> Maybe we could get the available memory and use that as hint, I mean as memory
> usage can be highly dynamic it will never be perfect, but better than just ignoring
> it..
If anything, I'd make it user-configurable - I don't think a heuristic
would be a good choice here.
This way we could also set it smaller for single-file restore for
example - on the other hand, that adds another parameter to the already
somewhat cluttered QEMU<->Rust interface.
>
>> );
>>
>> let reader = AsyncIndexReader::new(index, chunk_reader);
>>
>
More information about the pbs-devel
mailing list