Hello everyone! I am wondering whether it's better to share one big dump (e.g. latest english Wikipedia dump as one file) or may small files which constitute the dump file.
Arguments for big dumps: performance: one hash in the network.
Arguments for small files: gnome-4.5.tar.gz is already in the network, wouldn't be deduplicated if a Linux-ISO containing the file would be introduced, too. Only new additions need new storage space.
This is similar to the dichotomy of Debian's way to link packages if possible (both reducing disk space/RAM requirements) opposed to Docker images.
Hello everyone! I am wondering whether it's better to share one big dump (e.g. latest english Wikipedia dump as one file) or may small files which constitute the dump file.
Arguments for big dumps: performance: one hash in the network.
Arguments for small files: gnome-4.5.tar.gz is already in the network, wouldn't be deduplicated if a Linux-ISO containing the file would be introduced, too. Only new additions need new storage space.
This is similar to the dichotomy of Debian's way to link packages if possible (both reducing disk space/RAM requirements) opposed to Docker images.
What do you think?