Open abkfenris opened 3 years ago
Removing tensorflow only slims things down another half gig
➜ docker images
REPOSITORY TAG IMAGE ID CREATED SIZE
ohw-no-py-tensorflow latest 200d425d469f 52 seconds ago 5.3GB
ohw-cache-apt latest c646a0031f14 9 hours ago 5.81GB
ohw-cache latest 274e85773a32 9 hours ago 5.81GB
ohw latest d2014651c42b 10 hours ago 8.27GB
causing the build to explode.
What do you mean by exploding? We are not able to upload that?
PS: let's remove tensorflow!
I meant size in this case, but also didn't have permissions for uploading to Docker Hub.
I'll remove tensorflow in #23
no_link
is no longer in the dataframe, so it should now be
import json
from pathlib import Path
import pandas as pd
pkg_files = Path("/opt/conda/conda-meta/").glob("*.json")
paths = []
for pkg_file in pkg_files:
with pkg_file.open() as f:
pkg = json.load(f)
paths += pkg["paths_data"]["paths"]
df = pd.DataFrame(paths)
df = df.drop(
[
"path_type",
"sha256",
"sha256_in_prefix",
# "no_link",
"file_mode",
"prefix_placeholder",
],
axis=1,
)
df = df.dropna()
df = df.sort_values("size_in_bytes", ascending=False)
df
It's also useful to include mamba in the environment for mamba repoquery
to get dependencies.
Here's a way to start analyzing what packages are causing the build to explode.
Caching some of the build with #23
@ocefpaf