Closed amisevsk closed 2 months ago
Performance wise, it's not really much faster assuming the disk is fast -- 77s vs 81s for ghcr.io/jozu-ai/llama-2:7b-chat-q4_0
on my machine, but it should avoid some storage issues we run into with larger models.
Description
Avoid requiring a data copy during the packing process to cut down on peak storage usage and slightly increase speed. Previously, the process was
.tar.gz
in$TMPDIR
and read digest from file.tar.gz
into local OCI storage (requires copying).tar.gz
from step 1.Since OCI storage is content-addressible, instead of pushing into OCI storage during step 2, we instead move the temporary
.tar.gz
directly into its expected location and verify that the OCI store has picked it up correctly.Linked issues
N/A