Closed aditya2803 closed 2 years ago
hi @aditya2803, indeed vHive didn't have the firecracker-containerd folks' fix. We are about to merge a new version vHive/Firecracker snapshots with this #465. The code in that PR does have more recent firecracker-containerd binary, which is already tested but the docs are not updated. You can use that branch before the PR is merged
Hi @ustiugov thanks for your suggestion. I cloned the PR branch and rebuilt the stack. However, the functions are still not getting deployed properly. Below is the error I get:
WARN[0600] Failed to deploy function pyaes-1, configs/knative_workloads/pyaes.yaml: exit status 1
Creating service 'pyaes-1' in namespace 'default':
0.019s The Route is still working to reflect the latest desired specification.
0.097s Configuration "pyaes-1" is waiting for a Revision to become ready.
Error: timeout: service 'pyaes-1' not ready after 600 seconds
Run 'kn --help' for usage
INFO[0600] Deployed function pyaes-1
WARN[0600] Failed to deploy function pyaes-0, configs/knative_workloads/pyaes.yaml: exit status 1
Creating service 'pyaes-0' in namespace 'default':
0.029s The Route is still working to reflect the latest desired specification.
0.123s Configuration "pyaes-0" is waiting for a Revision to become ready.
Error: timeout: service 'pyaes-0' not ready after 600 seconds
Run 'kn --help' for usage
INFO[0600] Deployed function pyaes-0
WARN[0600] Failed to deploy function rnn-serving-1, configs/knative_workloads/rnn_serving.yaml: exit status 1
Creating service 'rnn-serving-1' in namespace 'default':
0.088s The Route is still working to reflect the latest desired specification.
0.163s ...
0.179s Configuration "rnn-serving-1" is waiting for a Revision to become ready.
Error: timeout: service 'rnn-serving-1' not ready after 600 seconds
Run 'kn --help' for usage
WARN[0600] Failed to deploy function rnn-serving-0, configs/knative_workloads/rnn_serving.yaml: exit status 1
Creating service 'rnn-serving-0' in namespace 'default':
0.126s The Route is still working to reflect the latest desired specification.
0.169s ...
0.185s Configuration "rnn-serving-0" is waiting for a Revision to become ready.
Error: timeout: service 'rnn-serving-0' not ready after 600 seconds
Run 'kn --help' for usage
WARN[0600] Failed to deploy function helloworld-0, configs/knative_workloads/helloworld.yaml: exit status 1
Creating service 'helloworld-0' in namespace 'default':
0.086s The Route is still working to reflect the latest desired specification.
0.161s ...
0.183s Configuration "helloworld-0" is waiting for a Revision to become ready.
Error: timeout: service 'helloworld-0' not ready after 600 seconds
Run 'kn --help' for usage
INFO[0600] Deployed function helloworld-0
INFO[0600] Deployed function rnn-serving-0
INFO[0600] Deployed function rnn-serving-1
WARN[1200] Failed to deploy function rnn-serving-2, configs/knative_workloads/rnn_serving.yaml: exit status 1
Creating service 'rnn-serving-2' in namespace 'default':
0.025s The Route is still working to reflect the latest desired specification.
0.059s ...
0.087s Configuration "rnn-serving-2" is waiting for a Revision to become ready.
Error: timeout: service 'rnn-serving-2' not ready after 600 seconds
Run 'kn --help' for usage
INFO[1200] Deployed function rnn-serving-2
INFO[1200] Deployment finished
Output of kubectl describe revision/helloworld-0-00001
Name: helloworld-0-00001
Namespace: default
Labels: serving.knative.dev/configuration=helloworld-0
serving.knative.dev/configurationGeneration=1
serving.knative.dev/configurationUID=700a477f-63f1-445f-bb64-291d3b62016b
serving.knative.dev/routingState=active
serving.knative.dev/service=helloworld-0
serving.knative.dev/serviceUID=aa954d1d-975f-416f-901c-1e68572c26e4
Annotations: autoscaling.knative.dev/target: 1
serving.knative.dev/creator: kubernetes-admin
serving.knative.dev/routes: helloworld-0
serving.knative.dev/routingStateModified: 2022-05-20T17:45:45Z
API Version: serving.knative.dev/v1
Kind: Revision
Metadata:
Creation Timestamp: 2022-05-20T17:45:45Z
Generation: 1
Managed Fields:
API Version: serving.knative.dev/v1
Fields Type: FieldsV1
fieldsV1:
f:metadata:
f:annotations:
.:
f:autoscaling.knative.dev/target:
f:serving.knative.dev/creator:
f:serving.knative.dev/routes:
f:serving.knative.dev/routingStateModified:
f:labels:
.:
f:serving.knative.dev/configuration:
f:serving.knative.dev/configurationGeneration:
f:serving.knative.dev/configurationUID:
f:serving.knative.dev/routingState:
f:serving.knative.dev/service:
f:serving.knative.dev/serviceUID:
f:ownerReferences:
.:
k:{"uid":"700a477f-63f1-445f-bb64-291d3b62016b"}:
f:spec:
.:
f:containerConcurrency:
f:containers:
f:enableServiceLinks:
f:timeoutSeconds:
Manager: Go-http-client
Operation: Update
Time: 2022-05-20T17:45:45Z
API Version: serving.knative.dev/v1
Fields Type: FieldsV1
fieldsV1:
f:status:
.:
f:actualReplicas:
f:conditions:
f:containerStatuses:
f:observedGeneration:
Manager: Go-http-client
Operation: Update
Subresource: status
Time: 2022-05-20T17:45:45Z
Owner References:
API Version: serving.knative.dev/v1
Block Owner Deletion: true
Controller: true
Kind: Configuration
Name: helloworld-0
UID: 700a477f-63f1-445f-bb64-291d3b62016b
Resource Version: 6933
UID: d547b289-d825-4c9c-9e09-dd1398b6cc12
Spec:
Container Concurrency: 0
Containers:
Env:
Name: GUEST_PORT
Value: 50051
Name: GUEST_IMAGE
Value: ghcr.io/ease-lab/helloworld:var_workload
Image: crccheck/hello-world:latest
Name: user-container
Ports:
Container Port: 50051
Name: h2c
Protocol: TCP
Readiness Probe:
Success Threshold: 1
Tcp Socket:
Port: 0
Resources:
Enable Service Links: false
Timeout Seconds: 300
Status:
Actual Replicas: 0
Conditions:
Last Transition Time: 2022-05-20T17:56:15Z
Message: The target is not receiving traffic.
Reason: NoTraffic
Severity: Info
Status: False
Type: Active
Last Transition Time: 2022-05-20T17:45:45Z
Reason: Deploying
Status: Unknown
Type: ContainerHealthy
Last Transition Time: 2022-05-20T17:56:15Z
Message: Initial scale was never achieved
Reason: ProgressDeadlineExceeded
Status: False
Type: Ready
Last Transition Time: 2022-05-20T17:56:15Z
Message: Initial scale was never achieved
Reason: ProgressDeadlineExceeded
Status: False
Type: ResourcesAvailable
Container Statuses:
Name: user-container
Observed Generation: 1
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning InternalError 34m revision-controller failed to update deployment "helloworld-0-00001-deployment": Operation cannot be fulfilled on deployments.apps "helloworld-0-00001-deployment": the object has been modified; please apply your changes to the latest version and try again
Note that I am using the #481 fix in my local code, as suggested by you.
Another observation is that running the script with the 'stock-only' option results in proper deployment of the functions. It is with firecracker-containerd (default) that the issue comes up.
Services in case of using stock-only:
NAME URL LATEST AGE CONDITIONS READY REASON
helloworld-0 http://helloworld-0.default.192.168.1.240.sslip.io helloworld-0-00001 7m40s 3 OK / 3 True
pyaes-0 http://pyaes-0.default.192.168.1.240.sslip.io pyaes-0-00001 7m40s 3 OK / 3 True
pyaes-1 http://pyaes-1.default.192.168.1.240.sslip.io pyaes-1-00001 7m40s 3 OK / 3 True
rnn-serving-0 http://rnn-serving-0.default.192.168.1.240.sslip.io rnn-serving-0-00001 7m40s 3 OK / 3 True
rnn-serving-1 http://rnn-serving-1.default.192.168.1.240.sslip.io rnn-serving-1-00001 7m40s 3 OK / 3 True
rnn-serving-2 http://rnn-serving-2.default.192.168.1.240.sslip.io rnn-serving-2-00001 7m28s 3 OK / 3 True
@aditya2803 I cannot say much without lower-level logs in the firecracker setup (vHive, containerd, firecracker-containerd). The vhive CRI test worked in that branch. Try deploying a new cluster on a fresh node
also, the YAML of the workloads do not suit the stock-only setup. You need to use YAML files in a conventional Knative format (you can take them from their website).
@ustiugov Here are the logs:
vhive.stdout logs
time="2022-05-21T12:21:48.816889649Z" level=info msg="Creating containerd client"
time="2022-05-21T12:21:48.817591298Z" level=info msg="Created containerd client"
time="2022-05-21T12:21:48.817639598Z" level=info msg="Creating firecracker client"
time="2022-05-21T12:21:48.817793247Z" level=info msg="Created firecracker client"
time="2022-05-21T12:21:48.825213285Z" level=info msg="Creating image manager"
time="2022-05-21T12:21:48.829549076Z" level=info msg="Registering bridges for tap manager"
time="2022-05-21T12:21:48.831582308Z" level=info msg="Listening on port:3334"
time="2022-05-21T12:21:48.831614107Z" level=info msg="Listening on port:3333"
time="2022-05-21T12:22:48.816984934Z" level=info msg="HEARTBEAT: number of active VMs: 0"
time="2022-05-21T12:22:48.832103671Z" level=info msg="FuncPool heartbeat: ==== Stats by cold functions ====\nfID, #started, #served\n==================================="
time="2022-05-21T12:23:48.816425650Z" level=info msg="HEARTBEAT: number of active VMs: 0"
time="2022-05-21T12:23:48.831493369Z" level=info msg="FuncPool heartbeat: ==== Stats by cold functions ====\nfID, #started, #served\n==================================="
time="2022-05-21T12:24:47.880301696Z" level=warning msg="Failed to Fetch k8s dns clusterIP exit status 1\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n\n"
time="2022-05-21T12:24:47.880342182Z" level=warning msg="Using google dns 8.8.8.8\n"
time="2022-05-21T12:24:47.881345044Z" level=warning msg="Failed to Fetch k8s dns clusterIP exit status 1\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n\n"
time="2022-05-21T12:24:47.881372977Z" level=warning msg="Using google dns 8.8.8.8\n"
time="2022-05-21T12:24:48.816480060Z" level=info msg="HEARTBEAT: number of active VMs: 5"
time="2022-05-21T12:24:48.831606453Z" level=info msg="FuncPool heartbeat: ==== Stats by cold functions ====\nfID, #started, #served\n==================================="
time="2022-05-21T12:24:51.338213803Z" level=warning msg="Failed to Fetch k8s dns clusterIP exit status 1\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n\n"
time="2022-05-21T12:24:51.338253999Z" level=warning msg="Using google dns 8.8.8.8\n"
time="2022-05-21T12:25:11.326207014Z" level=warning msg="Failed to Fetch k8s dns clusterIP exit status 1\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n\n"
time="2022-05-21T12:25:11.326246148Z" level=warning msg="Using google dns 8.8.8.8\n"
time="2022-05-21T12:25:11.327834634Z" level=warning msg="Failed to Fetch k8s dns clusterIP exit status 1\nThe connection to the server localhost:8080 was refused - did you specify the right host or port?\n\n"
time="2022-05-21T12:25:11.327854311Z" level=warning msg="Using google dns 8.8.8.8\n"
time="2022-05-21T12:25:48.816850231Z" level=info msg="HEARTBEAT: number of active VMs: 5"
time="2022-05-21T12:25:48.831974752Z" level=info msg="FuncPool heartbeat: ==== Stats by cold functions ====\nfID, #started, #served\n==================================="
time="2022-05-21T12:26:48.817059802Z" level=info msg="HEARTBEAT: number of active VMs: 5"
time="2022-05-21T12:26:48.832181982Z" level=info msg="FuncPool heartbeat: ==== Stats by cold functions ====\nfID, #started, #served\n==================================="
time="2022-05-21T12:27:48.817262702Z" level=info msg="HEARTBEAT: number of active VMs: 5"
time="2022-05-21T12:27:48.832380083Z" level=info msg="FuncPool heartbeat: ==== Stats by cold functions ====\nfID, #started, #served\n==================================="
time="20
containerd.stderr logs
time="2022-05-21T12:21:18.695175301Z" level=info msg="starting containerd" revision=de8046a5501db9e0e478e1c10cbcfb21af4c6b2d version=v1.6.2
time="2022-05-21T12:21:18.708869114Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1
time="2022-05-21T12:21:18.709312206Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1
time="2022-05-21T12:21:18.711055793Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.16.0-rc1inuma+\\n\"): skip plugin" type=io.containerd.snapshotter.v1
time="2022-05-21T12:21:18.711121015Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1
time="2022-05-21T12:21:18.711370815Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
time="2022-05-21T12:21:18.711395631Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1
time="2022-05-21T12:21:18.711415078Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured"
time="2022-05-21T12:21:18.711429785Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1
time="2022-05-21T12:21:18.711461485Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1
time="2022-05-21T12:21:18.711697609Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1
time="2022-05-21T12:21:18.711862709Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
time="2022-05-21T12:21:18.711885732Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1
time="2022-05-21T12:21:18.711908104Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured"
time="2022-05-21T12:21:18.711926228Z" level=info msg="metadata content store policy set" policy=shared
time="2022-05-21T12:21:28.712415211Z" level=warning msg="waiting for response from boltdb open" plugin=bolt
firecracker.stderr logs
time="2022-05-21T12:21:29Z" level=warning msg="deprecated version : `1`, please switch to version `2`"
time="2022-05-21T12:21:29.745823459Z" level=info msg="starting containerd" revision=19c96c059d7a95e8eb7f27b4e2847c4a84898698 version=1.5.5+unknown
time="2022-05-21T12:21:29.763711658Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1
time="2022-05-21T12:21:29.763799332Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1
time="2022-05-21T12:21:29.763855197Z" level=info msg="initializing pool device \"fc-dev-thinpool\""
time="2022-05-21T12:21:29.765041918Z" level=info msg="using dmsetup:\nLibrary version: 1.02.167 (2019-11-30)\nDriver version: 4.45.0"
time="2022-05-21T12:21:29.767577192Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1
time="2022-05-21T12:21:29.767755988Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1
time="2022-05-21T12:21:29.767802526Z" level=info msg="metadata content store policy set" policy=shared
time="2022-05-21T12:21:29.768630692Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1
time="2022-05-21T12:21:29.768653124Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1
time="2022-05-21T12:21:29.768685215Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1
time="2022-05-21T12:21:29.768703098Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1
time="2022-05-21T12:21:29.768714820Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1
time="2022-05-21T12:21:29.768725631Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1
time="2022-05-21T12:21:29.768737894Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1
time="2022-05-21T12:21:29.768749065Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1
time="2022-05-21T12:21:29.768762360Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1
time="2022-05-21T12:21:29.768773090Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1
time="2022-05-21T12:21:29.768783509Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1
time="2022-05-21T12:21:29.768854944Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2
time="2022-05-21T12:21:29.768914405Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1
time="2022-05-21T12:21:29.769205653Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1
time="2022-05-21T12:21:29.769224659Z" level=info msg="loading plugin \"io.containerd.service.v1.fc-control\"..." type=io.containerd.service.v1
time="2022-05-21T12:21:29.769236962Z" level=debug msg="initializing fc-control plugin (root: \"/var/lib/firecracker-containerd/containerd/io.containerd.service.v1.fc-control\")"
time="2022-05-21T12:21:29.787087539Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1
time="2022-05-21T12:21:29.787154725Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787173691Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787190302Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787206202Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787221761Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787238152Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787254453Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787273940Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787289529Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1
time="2022-05-21T12:21:29.787350123Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787370240Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787386521Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787401710Z" level=info msg="loading plugin \"io.containerd.grpc.v1.fc-control-service\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787419032Z" level=debug msg="initializing fc-control-service plugin"
time="2022-05-21T12:21:29.787436225Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1
time="2022-05-21T12:21:29.787618237Z" level=info msg=serving... address=/run/firecracker-containerd/containerd.sock.ttrpc
time="2022-05-21T12:21:29.787663662Z" level=info msg=serving... address=/run/firecracker-containerd/containerd.sock
time="2022-05-21T12:21:29.787677859Z" level=debug msg="sd notification" error="<nil>" notified=false state="READY=1"
time="2022-05-21T12:21:29.787692446Z" level=info msg="containerd successfully booted in 0.042626s"
time="2022-05-21T12:21:29.869377712Z" level=debug msg="garbage collected" d="453.583µs"
time="2022-05-21T12:24:38.435253405Z" level=debug msg="(*service).Write started" expected="sha256:6a11e6dbd88b1ce1ebb284c769b52e3fdb66a0f37b392bded5612045ff2cae61" ref="manifest-sha256:6a11e6dbd88b1ce1ebb284c769b52e3fdb66a0f37b392bded5612045ff2cae61" total=1996
time="2022-05-21T12:24:38.732027372Z" level=debug msg="(*service).Write started" expected="sha256:8a5cab1e2faec39c2e1215778ed65d63584e279a82685d77a447c7c7d36a4b17" ref="config-sha256:8a5cab1e2faec39c2e1215778ed65d63584e279a82685d77a447c7c7d36a4b17" total=8868
time="2022-05-21T12:24:39.415921874Z" level=debug msg="stat snapshot" key="sha256:cd7100a72410606589a54b932cabd804a17f9ae5b42a1882bd56d263e02b6215"
time="2022-05-21T12:24:39.416202714Z" level=debug msg="prepare snapshot" key="extract-416064543-vxuX sha256:cd7100a72410606589a54b932cabd804a17f9ae5b42a1882bd56d263e02b6215" parent=
time="2022-05-21T12:24:39.416536273Z" level=debug msg=prepare key="firecracker-containerd/1/extract-416064543-vxuX sha256:cd7100a72410606589a54b932cabd804a17f9ae5b42a1882bd56d263e02b6215" parent=
time="2022-05-21T12:24:39.416957517Z" level=debug msg="creating new thin device 'fc-dev-thinpool-snap-1'"
time="2022-05-21T12:24:39.485662564Z" level=debug msg="mkfs.ext4 -E nodiscard,lazy_itable_init=0,lazy_journal_init=0 /dev/mapper/fc-dev-thinpool-snap-1"
time="2022-05-21T12:24:39.749567331Z" level=debug msg="(*service).Write started" expected="sha256:75d39d67fbb3ca85eb89ece0b38e24ab7dadb2fccf9576a00cd87588aad7c460" ref="manifest-sha256:75d39d67fbb3ca85eb89ece0b38e24ab7dadb2fccf9576a00cd87588aad7c460" total=1998
time="2022-05-21T12:24:40.366557876Z" level=debug msg="mkfs:\nmke2fs 1.45.5 (07-Jan-2020)\nCreating filesystem with 2621440 4k blocks and 655360 inodes\nFilesystem UUID: 2874b42c-f116-4a8c-98bd-e6fb6a227caa\nSuperblock backups stored on blocks: \n\t32768, 98304, 163840, 229376, 294912, 819200, 884736, 1605632\n\nAllocating group tables: 0/80\b\b\b\b\b \b\b\b\b\bdone \nWriting inode tables: 0/80\b\b\b\b\b \b\b\b\b\bdone \nCreating journal (16384 blocks): done\nWriting superblocks and filesystem accounting information: 0/80\b\b\b\b\b \b\b\b\b\bdone\n\n"
time="2022-05-21T12:24:40.383084820Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:40.383528527Z" level=debug msg="(*service).Write started" expected="sha256:dfd5ae2430bfdaa3eabe80a09fef72b7b1b34a9b5ffe7690b3822cdad290cba5" ref="layer-sha256:dfd5ae2430bfdaa3eabe80a09fef72b7b1b34a9b5ffe7690b3822cdad290cba5" total=55723509
time="2022-05-21T12:24:40.383542864Z" level=debug msg="(*service).Write started" expected="sha256:614456ff946738237eb1d5e7ddb9b3b9578292cd2de96317aa37d76ea0a4eea9" ref="layer-sha256:614456ff946738237eb1d5e7ddb9b3b9578292cd2de96317aa37d76ea0a4eea9" total=185738
time="2022-05-21T12:24:40.383631712Z" level=debug msg="(*service).Write started" expected="sha256:72c1fa02b2c870da7fd4c4a0af11b837cd448185b4ff31f3ced4c1e11199d743" ref="layer-sha256:72c1fa02b2c870da7fd4c4a0af11b837cd448185b4ff31f3ced4c1e11199d743" total=248074790
time="2022-05-21T12:24:40.383655687Z" level=debug msg="(*service).Write started" expected="sha256:ff3a5c916c92643ff77519ffa742d3ec61b7f591b6b7504599d95a4a41134e28" ref="layer-sha256:ff3a5c916c92643ff77519ffa742d3ec61b7f591b6b7504599d95a4a41134e28" total=2065537
time="2022-05-21T12:24:40.383684742Z" level=debug msg="(*service).Write started" expected="sha256:466a9644be5453fb0268d102159dd91b988e5d24f84431d0a5a57ee7ff21de2b" ref="layer-sha256:466a9644be5453fb0268d102159dd91b988e5d24f84431d0a5a57ee7ff21de2b" total=3742
time="2022-05-21T12:24:40.383715099Z" level=debug msg="(*service).Write started" expected="sha256:964f5a9ea2070018f381d9c968d435bc4576497232bd7d3e79121b180ef2169a" ref="layer-sha256:964f5a9ea2070018f381d9c968d435bc4576497232bd7d3e79121b180ef2169a" total=125
time="2022-05-21T12:24:40.383724968Z" level=debug msg="(*service).Write started" expected="sha256:95853ec29c67ccc835034ef04f5765d1064b835ffb476e2a073dbb8e7b3d7cf3" ref="layer-sha256:95853ec29c67ccc835034ef04f5765d1064b835ffb476e2a073dbb8e7b3d7cf3" total=87932348
time="2022-05-21T12:24:40.383854722Z" level=debug msg="(*service).Write started" expected="sha256:f6993a2cb9082ebcb2d8d151f19a1137ebbe7c642e8a3c41aac38f816c15c4c7" ref="layer-sha256:f6993a2cb9082ebcb2d8d151f19a1137ebbe7c642e8a3c41aac38f816c15c4c7" total=98
time="2022-05-21T12:24:40.541633938Z" level=debug msg="(*service).Write started" expected="sha256:3691e79f01ef2ba64a855ef7621b04b3dbb0b4c689d27ebaa8644d4cb1a7e28f" ref="config-sha256:3691e79f01ef2ba64a855ef7621b04b3dbb0b4c689d27ebaa8644d4cb1a7e28f" total=8312
time="2022-05-21T12:24:41.315518049Z" level=debug msg="stat snapshot" key="sha256:5216338b40a7b96416b8b9858974bbe4acc3096ee60acbc4dfb1ee02aecceb10"
time="2022-05-21T12:24:41.315847150Z" level=debug msg="prepare snapshot" key="extract-315698480-u59L sha256:5216338b40a7b96416b8b9858974bbe4acc3096ee60acbc4dfb1ee02aecceb10" parent=
time="2022-05-21T12:24:41.316209544Z" level=debug msg=prepare key="firecracker-containerd/2/extract-315698480-u59L sha256:5216338b40a7b96416b8b9858974bbe4acc3096ee60acbc4dfb1ee02aecceb10" parent=
time="2022-05-21T12:24:41.316272633Z" level=debug msg="creating new thin device 'fc-dev-thinpool-snap-2'"
time="2022-05-21T12:24:41.380191672Z" level=debug msg="mkfs.ext4 -E nodiscard,lazy_itable_init=0,lazy_journal_init=0 /dev/mapper/fc-dev-thinpool-snap-2"
time="2022-05-21T12:24:41.446436168Z" level=debug msg="using pigz for decompression"
time="2022-05-21T12:24:42.197733236Z" level=debug msg="mkfs:\nmke2fs 1.45.5 (07-Jan-2020)\nCreating filesystem with 2621440 4k blocks and 655360 inodes\nFilesystem UUID: ad64cd7c-02b2-40d9-b823-7e1ce89b8078\nSuperblock backups stored on blocks: \n\t32768, 98304, 163840, 229376, 294912, 819200, 884736, 1605632\n\nAllocating group tables: 0/80\b\b\b\b\b \b\b\b\b\bdone \nWriting inode tables: 0/80\b\b\b\b\b \b\b\b\b\bdone \nCreating journal (16384 blocks): done\nWriting superblocks and filesystem accounting information: 0/80\b\b\b\b\b \b\b\b\b\bdone\n\n"
time="2022-05-21T12:24:42.218491271Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:42.218951870Z" level=debug msg="(*service).Write started" expected="sha256:d02232cde789f60bfefd38a38c22df68cb75f0b4a6e17f11876650bc1845acaf" ref="layer-sha256:d02232cde789f60bfefd38a38c22df68cb75f0b4a6e17f11876650bc1845acaf" total=637463
time="2022-05-21T12:24:42.218985574Z" level=debug msg="(*service).Write started" expected="sha256:00a47c8ade3f6bcd1061541f4387e56d3fcba420f67b4234ade01d51635572f4" ref="layer-sha256:00a47c8ade3f6bcd1061541f4387e56d3fcba420f67b4234ade01d51635572f4" total=2435
time="2022-05-21T12:24:42.219006212Z" level=debug msg="(*service).Write started" expected="sha256:407da27a03363f6b9d368ec6e131f7f2db7c8cb2a149160d913d6f3698905a5d" ref="layer-sha256:407da27a03363f6b9d368ec6e131f7f2db7c8cb2a149160d913d6f3698905a5d" total=1887472
time="2022-05-21T12:24:42.219067338Z" level=debug msg="(*service).Write started" expected="sha256:61614c1a5710c76af6b2a9c7170a81eb0dd76ccf90e921abc0c9dcc1d5ed490e" ref="layer-sha256:61614c1a5710c76af6b2a9c7170a81eb0dd76ccf90e921abc0c9dcc1d5ed490e" total=31506523
time="2022-05-21T12:24:42.219040487Z" level=debug msg="(*service).Write started" expected="sha256:c9b1b535fdd91a9855fb7f82348177e5f019329a58c53c47272962dd60f71fc9" ref="layer-sha256:c9b1b535fdd91a9855fb7f82348177e5f019329a58c53c47272962dd60f71fc9" total=2802957
time="2022-05-21T12:24:42.219047070Z" level=debug msg="(*service).Write started" expected="sha256:2cc5ad85d9abaadf23d5ae53c3f32e7ccb2df1956869980bfd2491ff396d348a" ref="layer-sha256:2cc5ad85d9abaadf23d5ae53c3f32e7ccb2df1956869980bfd2491ff396d348a" total=301261
time="2022-05-21T12:24:42.219080272Z" level=debug msg="(*service).Write started" expected="sha256:0522d30cde10ac29ae2c555b9bde76c2b50aafc7ef7435bbc7e19de706bcadcd" ref="layer-sha256:0522d30cde10ac29ae2c555b9bde76c2b50aafc7ef7435bbc7e19de706bcadcd" total=230
time="2022-05-21T12:24:42.219102684Z" level=debug msg="(*service).Write started" expected="sha256:adc08e00a651383f0333647c65bedddc8826225b3a3d8da06c4f8e678f935b71" ref="layer-sha256:adc08e00a651383f0333647c65bedddc8826225b3a3d8da06c4f8e678f935b71" total=20418883
time="2022-05-21T12:24:42.266679448Z" level=debug msg="diff applied" d=820.460239ms digest="sha256:ff3a5c916c92643ff77519ffa742d3ec61b7f591b6b7504599d95a4a41134e28" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=2065537
time="2022-05-21T12:24:42.266997488Z" level=debug msg="commit snapshot" key="extract-416064543-vxuX sha256:cd7100a72410606589a54b932cabd804a17f9ae5b42a1882bd56d263e02b6215" name="sha256:cd7100a72410606589a54b932cabd804a17f9ae5b42a1882bd56d263e02b6215"
time="2022-05-21T12:24:42.267152089Z" level=debug msg=commit key="firecracker-containerd/1/extract-416064543-vxuX sha256:cd7100a72410606589a54b932cabd804a17f9ae5b42a1882bd56d263e02b6215" name="firecracker-containerd/3/sha256:cd7100a72410606589a54b932cabd804a17f9ae5b42a1882bd56d263e02b6215"
time="2022-05-21T12:24:42.369148974Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:42.370149522Z" level=debug msg="stat snapshot" key="sha256:e4e4f6845ea6130dbe3b08e769e3bbc16a9f0dfe037f0380c5123e9b0d9a34d6"
time="2022-05-21T12:24:42.370489393Z" level=debug msg="prepare snapshot" key="extract-370322448-tHLo sha256:e4e4f6845ea6130dbe3b08e769e3bbc16a9f0dfe037f0380c5123e9b0d9a34d6" parent="sha256:cd7100a72410606589a54b932cabd804a17f9ae5b42a1882bd56d263e02b6215"
time="2022-05-21T12:24:42.370884388Z" level=debug msg=prepare key="firecracker-containerd/4/extract-370322448-tHLo sha256:e4e4f6845ea6130dbe3b08e769e3bbc16a9f0dfe037f0380c5123e9b0d9a34d6" parent="firecracker-containerd/3/sha256:cd7100a72410606589a54b932cabd804a17f9ae5b42a1882bd56d263e02b6215"
time="2022-05-21T12:24:42.370939652Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-3' from 'fc-dev-thinpool-snap-1'"
time="2022-05-21T12:24:42.448962736Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:43.291468091Z" level=debug msg="diff applied" d=90.261239ms digest="sha256:c9b1b535fdd91a9855fb7f82348177e5f019329a58c53c47272962dd60f71fc9" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=2802957
time="2022-05-21T12:24:43.291974076Z" level=debug msg="commit snapshot" key="extract-315698480-u59L sha256:5216338b40a7b96416b8b9858974bbe4acc3096ee60acbc4dfb1ee02aecceb10" name="sha256:5216338b40a7b96416b8b9858974bbe4acc3096ee60acbc4dfb1ee02aecceb10"
time="2022-05-21T12:24:43.292098861Z" level=debug msg=commit key="firecracker-containerd/2/extract-315698480-u59L sha256:5216338b40a7b96416b8b9858974bbe4acc3096ee60acbc4dfb1ee02aecceb10" name="firecracker-containerd/5/sha256:5216338b40a7b96416b8b9858974bbe4acc3096ee60acbc4dfb1ee02aecceb10"
time="2022-05-21T12:24:43.378639782Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:43.380558221Z" level=debug msg="stat snapshot" key="sha256:5edfa66f961ecdadabac1b15441c567a06631fd4cb8a197a2f0399644a3c18d5"
time="2022-05-21T12:24:43.381242842Z" level=debug msg="prepare snapshot" key="extract-380898302-n1sU sha256:5edfa66f961ecdadabac1b15441c567a06631fd4cb8a197a2f0399644a3c18d5" parent="sha256:5216338b40a7b96416b8b9858974bbe4acc3096ee60acbc4dfb1ee02aecceb10"
time="2022-05-21T12:24:43.381640893Z" level=debug msg=prepare key="firecracker-containerd/6/extract-380898302-n1sU sha256:5edfa66f961ecdadabac1b15441c567a06631fd4cb8a197a2f0399644a3c18d5" parent="firecracker-containerd/5/sha256:5216338b40a7b96416b8b9858974bbe4acc3096ee60acbc4dfb1ee02aecceb10"
time="2022-05-21T12:24:43.381710374Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-4' from 'fc-dev-thinpool-snap-2'"
time="2022-05-21T12:24:43.455687766Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:43.517770432Z" level=debug msg="diff applied" d=61.720533ms digest="sha256:2cc5ad85d9abaadf23d5ae53c3f32e7ccb2df1956869980bfd2491ff396d348a" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=301261
time="2022-05-21T12:24:43.518169405Z" level=debug msg="commit snapshot" key="extract-380898302-n1sU sha256:5edfa66f961ecdadabac1b15441c567a06631fd4cb8a197a2f0399644a3c18d5" name="sha256:5edfa66f961ecdadabac1b15441c567a06631fd4cb8a197a2f0399644a3c18d5"
time="2022-05-21T12:24:43.518363731Z" level=debug msg=commit key="firecracker-containerd/6/extract-380898302-n1sU sha256:5edfa66f961ecdadabac1b15441c567a06631fd4cb8a197a2f0399644a3c18d5" name="firecracker-containerd/7/sha256:5edfa66f961ecdadabac1b15441c567a06631fd4cb8a197a2f0399644a3c18d5"
time="2022-05-21T12:24:43.585493207Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:43.587100829Z" level=debug msg="stat snapshot" key="sha256:483d3bedbf62615f629bcfd167c0f2b45df7afe3ecb6b3f7ffa7ebe2dab70faa"
time="2022-05-21T12:24:43.587699268Z" level=debug msg="prepare snapshot" key="extract-587398761-hPJS sha256:483d3bedbf62615f629bcfd167c0f2b45df7afe3ecb6b3f7ffa7ebe2dab70faa" parent="sha256:5edfa66f961ecdadabac1b15441c567a06631fd4cb8a197a2f0399644a3c18d5"
time="2022-05-21T12:24:43.588068184Z" level=debug msg=prepare key="firecracker-containerd/8/extract-587398761-hPJS sha256:483d3bedbf62615f629bcfd167c0f2b45df7afe3ecb6b3f7ffa7ebe2dab70faa" parent="firecracker-containerd/7/sha256:5edfa66f961ecdadabac1b15441c567a06631fd4cb8a197a2f0399644a3c18d5"
time="2022-05-21T12:24:43.588130952Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-5' from 'fc-dev-thinpool-snap-4'"
time="2022-05-21T12:24:43.645274792Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:46.176828369Z" level=debug msg="diff applied" d=1.479437422s digest="sha256:61614c1a5710c76af6b2a9c7170a81eb0dd76ccf90e921abc0c9dcc1d5ed490e" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=31506523
time="2022-05-21T12:24:46.177281684Z" level=debug msg="commit snapshot" key="extract-587398761-hPJS sha256:483d3bedbf62615f629bcfd167c0f2b45df7afe3ecb6b3f7ffa7ebe2dab70faa" name="sha256:483d3bedbf62615f629bcfd167c0f2b45df7afe3ecb6b3f7ffa7ebe2dab70faa"
time="2022-05-21T12:24:46.177570749Z" level=debug msg=commit key="firecracker-containerd/8/extract-587398761-hPJS sha256:483d3bedbf62615f629bcfd167c0f2b45df7afe3ecb6b3f7ffa7ebe2dab70faa" name="firecracker-containerd/9/sha256:483d3bedbf62615f629bcfd167c0f2b45df7afe3ecb6b3f7ffa7ebe2dab70faa"
time="2022-05-21T12:24:46.237554115Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:46.238576995Z" level=debug msg="stat snapshot" key="sha256:3adc9ed0c3e4c8e4fb839cc55b92b0fefb26459b3d03edcbe139e0bd097fca0f"
time="2022-05-21T12:24:46.238895235Z" level=debug msg="prepare snapshot" key="extract-238771492-CcnZ sha256:3adc9ed0c3e4c8e4fb839cc55b92b0fefb26459b3d03edcbe139e0bd097fca0f" parent="sha256:483d3bedbf62615f629bcfd167c0f2b45df7afe3ecb6b3f7ffa7ebe2dab70faa"
time="2022-05-21T12:24:46.239264402Z" level=debug msg=prepare key="firecracker-containerd/10/extract-238771492-CcnZ sha256:3adc9ed0c3e4c8e4fb839cc55b92b0fefb26459b3d03edcbe139e0bd097fca0f" parent="firecracker-containerd/9/sha256:483d3bedbf62615f629bcfd167c0f2b45df7afe3ecb6b3f7ffa7ebe2dab70faa"
time="2022-05-21T12:24:46.239337229Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-6' from 'fc-dev-thinpool-snap-5'"
time="2022-05-21T12:24:46.324387579Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:46.335863566Z" level=debug msg="diff applied" d=10.46531ms digest="sha256:0522d30cde10ac29ae2c555b9bde76c2b50aafc7ef7435bbc7e19de706bcadcd" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=230
time="2022-05-21T12:24:46.336144857Z" level=debug msg="commit snapshot" key="extract-238771492-CcnZ sha256:3adc9ed0c3e4c8e4fb839cc55b92b0fefb26459b3d03edcbe139e0bd097fca0f" name="sha256:3adc9ed0c3e4c8e4fb839cc55b92b0fefb26459b3d03edcbe139e0bd097fca0f"
time="2022-05-21T12:24:46.336294279Z" level=debug msg=commit key="firecracker-containerd/10/extract-238771492-CcnZ sha256:3adc9ed0c3e4c8e4fb839cc55b92b0fefb26459b3d03edcbe139e0bd097fca0f" name="firecracker-containerd/11/sha256:3adc9ed0c3e4c8e4fb839cc55b92b0fefb26459b3d03edcbe139e0bd097fca0f"
time="2022-05-21T12:24:46.422005696Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:46.423114728Z" level=debug msg="stat snapshot" key="sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270"
time="2022-05-21T12:24:46.423647052Z" level=debug msg="prepare snapshot" key="extract-423391750-ND6S sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270" parent="sha256:3adc9ed0c3e4c8e4fb839cc55b92b0fefb26459b3d03edcbe139e0bd097fca0f"
time="2022-05-21T12:24:46.424063217Z" level=debug msg=prepare key="firecracker-containerd/12/extract-423391750-ND6S sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270" parent="firecracker-containerd/11/sha256:3adc9ed0c3e4c8e4fb839cc55b92b0fefb26459b3d03edcbe139e0bd097fca0f"
time="2022-05-21T12:24:46.424146594Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-7' from 'fc-dev-thinpool-snap-6'"
time="2022-05-21T12:24:46.512145636Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:46.664512796Z" level=debug msg="diff applied" d=151.860434ms digest="sha256:407da27a03363f6b9d368ec6e131f7f2db7c8cb2a149160d913d6f3698905a5d" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=1887472
time="2022-05-21T12:24:46.666326848Z" level=debug msg="commit snapshot" key="extract-423391750-ND6S sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270" name="sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270"
time="2022-05-21T12:24:46.666727544Z" level=debug msg=commit key="firecracker-containerd/12/extract-423391750-ND6S sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270" name="firecracker-containerd/13/sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270"
time="2022-05-21T12:24:46.729776543Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:46.730786668Z" level=debug msg="stat snapshot" key="sha256:cf144147664338bb0a5b647411ad7e6c14b7c87d243cc114550f3f1c07d80edc"
time="2022-05-21T12:24:46.731153410Z" level=debug msg="prepare snapshot" key="extract-730983379-ugnd sha256:cf144147664338bb0a5b647411ad7e6c14b7c87d243cc114550f3f1c07d80edc" parent="sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270"
time="2022-05-21T12:24:46.731532916Z" level=debug msg=prepare key="firecracker-containerd/14/extract-730983379-ugnd sha256:cf144147664338bb0a5b647411ad7e6c14b7c87d243cc114550f3f1c07d80edc" parent="firecracker-containerd/13/sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270"
time="2022-05-21T12:24:46.731620090Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-8' from 'fc-dev-thinpool-snap-7'"
time="2022-05-21T12:24:46.800469801Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:46.810384333Z" level=debug msg="diff applied" d=9.564301ms digest="sha256:00a47c8ade3f6bcd1061541f4387e56d3fcba420f67b4234ade01d51635572f4" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=2435
time="2022-05-21T12:24:46.810671014Z" level=debug msg="commit snapshot" key="extract-730983379-ugnd sha256:cf144147664338bb0a5b647411ad7e6c14b7c87d243cc114550f3f1c07d80edc" name="sha256:cf144147664338bb0a5b647411ad7e6c14b7c87d243cc114550f3f1c07d80edc"
time="2022-05-21T12:24:46.810856303Z" level=debug msg=commit key="firecracker-containerd/14/extract-730983379-ugnd sha256:cf144147664338bb0a5b647411ad7e6c14b7c87d243cc114550f3f1c07d80edc" name="firecracker-containerd/15/sha256:cf144147664338bb0a5b647411ad7e6c14b7c87d243cc114550f3f1c07d80edc"
time="2022-05-21T12:24:46.878248575Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:46.879355713Z" level=debug msg="stat snapshot" key="sha256:66a1cfe2f05ed05d825c22c11c95890bcedc47070570e361a2afbd6fbade7ea2"
time="2022-05-21T12:24:46.879695234Z" level=debug msg="prepare snapshot" key="extract-879544840-15tN sha256:66a1cfe2f05ed05d825c22c11c95890bcedc47070570e361a2afbd6fbade7ea2" parent="sha256:cf144147664338bb0a5b647411ad7e6c14b7c87d243cc114550f3f1c07d80edc"
time="2022-05-21T12:24:46.880134622Z" level=debug msg=prepare key="firecracker-containerd/16/extract-879544840-15tN sha256:66a1cfe2f05ed05d825c22c11c95890bcedc47070570e361a2afbd6fbade7ea2" parent="firecracker-containerd/15/sha256:cf144147664338bb0a5b647411ad7e6c14b7c87d243cc114550f3f1c07d80edc"
time="2022-05-21T12:24:46.880213501Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-9' from 'fc-dev-thinpool-snap-8'"
time="2022-05-21T12:24:46.963029707Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:47.571625831Z" level=debug msg="(*service).Write started" expected="sha256:12dc6715ed1a8306f246ceaf7742c09e38a52a79d17421e4a50d7e0e09fdbc25" ref="manifest-sha256:12dc6715ed1a8306f246ceaf7742c09e38a52a79d17421e4a50d7e0e09fdbc25" total=1998
time="2022-05-21T12:24:47.575516079Z" level=debug msg="diff applied" d=612.204631ms digest="sha256:adc08e00a651383f0333647c65bedddc8826225b3a3d8da06c4f8e678f935b71" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=20418883
time="2022-05-21T12:24:47.575876970Z" level=debug msg="commit snapshot" key="extract-879544840-15tN sha256:66a1cfe2f05ed05d825c22c11c95890bcedc47070570e361a2afbd6fbade7ea2" name="sha256:66a1cfe2f05ed05d825c22c11c95890bcedc47070570e361a2afbd6fbade7ea2"
time="2022-05-21T12:24:47.576081836Z" level=debug msg=commit key="firecracker-containerd/16/extract-879544840-15tN sha256:66a1cfe2f05ed05d825c22c11c95890bcedc47070570e361a2afbd6fbade7ea2" name="firecracker-containerd/17/sha256:66a1cfe2f05ed05d825c22c11c95890bcedc47070570e361a2afbd6fbade7ea2"
time="2022-05-21T12:24:47.646240867Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:47.647235584Z" level=debug msg="stat snapshot" key="sha256:e172bf8795d813920c658a6772bb108238dc2cf13f1fc1ee1ac5c595c37da14a"
time="2022-05-21T12:24:47.647640107Z" level=debug msg="prepare snapshot" key="extract-647453625-dkKb sha256:e172bf8795d813920c658a6772bb108238dc2cf13f1fc1ee1ac5c595c37da14a" parent="sha256:66a1cfe2f05ed05d825c22c11c95890bcedc47070570e361a2afbd6fbade7ea2"
time="2022-05-21T12:24:47.648136062Z" level=debug msg=prepare key="firecracker-containerd/18/extract-647453625-dkKb sha256:e172bf8795d813920c658a6772bb108238dc2cf13f1fc1ee1ac5c595c37da14a" parent="firecracker-containerd/17/sha256:66a1cfe2f05ed05d825c22c11c95890bcedc47070570e361a2afbd6fbade7ea2"
time="2022-05-21T12:24:47.648230160Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-10' from 'fc-dev-thinpool-snap-9'"
time="2022-05-21T12:24:47.728820467Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:47.761914041Z" level=debug msg="diff applied" d=32.634048ms digest="sha256:d02232cde789f60bfefd38a38c22df68cb75f0b4a6e17f11876650bc1845acaf" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=637463
time="2022-05-21T12:24:47.762254243Z" level=debug msg="commit snapshot" key="extract-647453625-dkKb sha256:e172bf8795d813920c658a6772bb108238dc2cf13f1fc1ee1ac5c595c37da14a" name="sha256:e172bf8795d813920c658a6772bb108238dc2cf13f1fc1ee1ac5c595c37da14a"
time="2022-05-21T12:24:47.762417060Z" level=debug msg=commit key="firecracker-containerd/18/extract-647453625-dkKb sha256:e172bf8795d813920c658a6772bb108238dc2cf13f1fc1ee1ac5c595c37da14a" name="firecracker-containerd/19/sha256:e172bf8795d813920c658a6772bb108238dc2cf13f1fc1ee1ac5c595c37da14a"
time="2022-05-21T12:24:47.833540320Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:47.835257118Z" level=debug msg="create image" name="ghcr.io/ease-lab/pyaes:var_workload" target="sha256:75d39d67fbb3ca85eb89ece0b38e24ab7dadb2fccf9576a00cd87588aad7c460"
time="2022-05-21T12:24:47.835625083Z" level=debug msg="event published" ns=firecracker-containerd topic=/images/create type=containerd.services.images.v1.ImageCreate
time="2022-05-21T12:24:47.856726997Z" level=debug msg="garbage collected" d="794.93µs"
time="2022-05-21T12:24:47.881668745Z" level=debug msg="create VM request: VMID:\"2\" MachineCfg:<MemSizeMib:256 VcpuCount:1 > KernelArgs:\"ro noapic reboot=k panic=1 pci=off nomodules systemd.log_color=false systemd.unit=firecracker.target init=/sbin/overlay-init tsc=reliable quiet 8250.nr_uarts=0 ipv6.disable=1\" NetworkInterfaces:<StaticConfig:<MacAddress:\"02:FC:00:00:00:01\" HostDevName:\"2_tap\" IPConfig:<PrimaryAddr:\"190.128.0.3/10\" GatewayAddr:\"190.128.0.1\" Nameservers:\"8.8.8.8\" > > > TimeoutSeconds:100 OffloadEnabled:true "
time="2022-05-21T12:24:47.881743926Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:24:47.881729820Z" level=debug msg="create VM request: VMID:\"3\" MachineCfg:<MemSizeMib:256 VcpuCount:1 > KernelArgs:\"ro noapic reboot=k panic=1 pci=off nomodules systemd.log_color=false systemd.unit=firecracker.target init=/sbin/overlay-init tsc=reliable quiet 8250.nr_uarts=0 ipv6.disable=1\" NetworkInterfaces:<StaticConfig:<MacAddress:\"02:FC:00:00:00:02\" HostDevName:\"3_tap\" IPConfig:<PrimaryAddr:\"190.128.0.4/10\" GatewayAddr:\"190.128.0.1\" Nameservers:\"8.8.8.8\" > > > TimeoutSeconds:100 OffloadEnabled:true "
time="2022-05-21T12:24:47.881768353Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:24:47.882022292Z" level=debug msg="starting containerd-shim-aws-firecracker" vmID=2
time="2022-05-21T12:24:47.882041147Z" level=debug msg="starting containerd-shim-aws-firecracker" vmID=3
time="2022-05-21T12:24:47.885870150Z" level=debug msg="(*service).Write started" expected="sha256:a87533385b75fd1d476d8a1112a0c5db953e3d5b44d4f9db814b1e2e6abb8734" ref="config-sha256:a87533385b75fd1d476d8a1112a0c5db953e3d5b44d4f9db814b1e2e6abb8734" total=8312
time="2022-05-21T12:24:47.931247184Z" level=info msg="starting signal loop" namespace=firecracker-containerd path="/var/lib/firecracker-containerd/shim-base/firecracker-containerd#2" pid=25750
time="2022-05-21T12:24:47.931611973Z" level=info msg="creating new VM" runtime=aws.firecracker vmID=2
time="2022-05-21T12:24:47.931874027Z" level=info msg="Called startVMM(), setting up a VMM on firecracker.sock" runtime=aws.firecracker
time="2022-05-21T12:24:47.935192767Z" level=info msg="starting signal loop" namespace=firecracker-containerd path="/var/lib/firecracker-containerd/shim-base/firecracker-containerd#3" pid=25751
time="2022-05-21T12:24:47.935557806Z" level=info msg="creating new VM" runtime=aws.firecracker vmID=3
time="2022-05-21T12:24:47.935861148Z" level=info msg="Called startVMM(), setting up a VMM on firecracker.sock" runtime=aws.firecracker
time="2022-05-21T12:24:47.943351868Z" level=info msg="refreshMachineConfiguration: [GET /machine-config][200] getMachineConfigurationOK &{CPUTemplate: HtEnabled:0xc000581a63 MemSizeMib:0xc000581a58 TrackDirtyPages:false VcpuCount:0xc000581a50}" runtime=aws.firecracker
time="2022-05-21T12:24:47.943512210Z" level=info msg="PutGuestBootSource: [PUT /boot-source][204] putGuestBootSourceNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:47.943534622Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/runtime/default-rootfs.img, slot root_drive, root true." runtime=aws.firecracker
time="2022-05-21T12:24:47.943785506Z" level=info msg="Attached drive /var/lib/firecracker-containerd/runtime/default-rootfs.img: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:47.943802327Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#2/ctrstub0, slot MN2HE43UOVRDA, root false." runtime=aws.firecracker
time="2022-05-21T12:24:47.943925149Z" level=info msg="Attached drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#2/ctrstub0: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:47.943939957Z" level=info msg="Attaching NIC 2_tap (hwaddr 02:FC:00:00:00:01) at index 1" runtime=aws.firecracker
time="2022-05-21T12:24:47.947339560Z" level=info msg="refreshMachineConfiguration: [GET /machine-config][200] getMachineConfigurationOK &{CPUTemplate: HtEnabled:0xc000a08f63 MemSizeMib:0xc000a08f58 TrackDirtyPages:false VcpuCount:0xc000a08f50}" runtime=aws.firecracker
time="2022-05-21T12:24:47.947481367Z" level=info msg="PutGuestBootSource: [PUT /boot-source][204] putGuestBootSourceNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:47.947500443Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/runtime/default-rootfs.img, slot root_drive, root true." runtime=aws.firecracker
time="2022-05-21T12:24:47.947727291Z" level=info msg="Attached drive /var/lib/firecracker-containerd/runtime/default-rootfs.img: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:47.947753110Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#3/ctrstub0, slot MN2HE43UOVRDA, root false." runtime=aws.firecracker
time="2022-05-21T12:24:47.947890158Z" level=info msg="Attached drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#3/ctrstub0: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:47.947904044Z" level=info msg="Attaching NIC 3_tap (hwaddr 02:FC:00:00:00:02) at index 1" runtime=aws.firecracker
time="2022-05-21T12:24:47.955223391Z" level=info msg="startInstance successful: [PUT /actions][204] createSyncActionNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:47.955240503Z" level=info msg="calling agent" runtime=aws.firecracker vmID=2
time="2022-05-21T12:24:47.960050847Z" level=info msg="startInstance successful: [PUT /actions][204] createSyncActionNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:47.960068099Z" level=info msg="calling agent" runtime=aws.firecracker vmID=3
time="2022-05-21T12:24:48.498068019Z" level=debug msg="stat snapshot" key="sha256:5216338b40a7b96416b8b9858974bbe4acc3096ee60acbc4dfb1ee02aecceb10"
time="2022-05-21T12:24:48.498152759Z" level=debug msg=stat key="firecracker-containerd/5/sha256:5216338b40a7b96416b8b9858974bbe4acc3096ee60acbc4dfb1ee02aecceb10"
time="2022-05-21T12:24:48.498425353Z" level=debug msg="stat snapshot" key="sha256:5edfa66f961ecdadabac1b15441c567a06631fd4cb8a197a2f0399644a3c18d5"
time="2022-05-21T12:24:48.498453577Z" level=debug msg=stat key="firecracker-containerd/7/sha256:5edfa66f961ecdadabac1b15441c567a06631fd4cb8a197a2f0399644a3c18d5"
time="2022-05-21T12:24:48.498719518Z" level=debug msg="stat snapshot" key="sha256:483d3bedbf62615f629bcfd167c0f2b45df7afe3ecb6b3f7ffa7ebe2dab70faa"
time="2022-05-21T12:24:48.498802194Z" level=debug msg=stat key="firecracker-containerd/9/sha256:483d3bedbf62615f629bcfd167c0f2b45df7afe3ecb6b3f7ffa7ebe2dab70faa"
time="2022-05-21T12:24:48.499064048Z" level=debug msg="stat snapshot" key="sha256:3adc9ed0c3e4c8e4fb839cc55b92b0fefb26459b3d03edcbe139e0bd097fca0f"
time="2022-05-21T12:24:48.499105556Z" level=debug msg=stat key="firecracker-containerd/11/sha256:3adc9ed0c3e4c8e4fb839cc55b92b0fefb26459b3d03edcbe139e0bd097fca0f"
time="2022-05-21T12:24:48.499398219Z" level=debug msg="stat snapshot" key="sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270"
time="2022-05-21T12:24:48.499427574Z" level=debug msg=stat key="firecracker-containerd/13/sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270"
time="2022-05-21T12:24:48.499703344Z" level=debug msg="stat snapshot" key="sha256:e4226272b56cdeb334d6c3377374a6760ceae26f704fddedd5ff871e52d19784"
time="2022-05-21T12:24:48.500052072Z" level=debug msg="prepare snapshot" key="extract-499911517-YXoM sha256:e4226272b56cdeb334d6c3377374a6760ceae26f704fddedd5ff871e52d19784" parent="sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270"
time="2022-05-21T12:24:48.500455102Z" level=debug msg=prepare key="firecracker-containerd/20/extract-499911517-YXoM sha256:e4226272b56cdeb334d6c3377374a6760ceae26f704fddedd5ff871e52d19784" parent="firecracker-containerd/13/sha256:7893763263404b202f7a20649fdab037cf7200d38d5a5e6fc35ced2c072df270"
time="2022-05-21T12:24:48.500545703Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-11' from 'fc-dev-thinpool-snap-7'"
time="2022-05-21T12:24:48.576685656Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:48.577123051Z" level=debug msg="(*service).Write started" expected="sha256:0d776ee02572ee50935002da7bd7fcda4a60be4a48c4ace5dd3216c327e6767a" ref="layer-sha256:0d776ee02572ee50935002da7bd7fcda4a60be4a48c4ace5dd3216c327e6767a" total=20365161
time="2022-05-21T12:24:48.577132068Z" level=debug msg="(*service).Write started" expected="sha256:c09d5cdb7367fff0d581bb8003e3520dc3e8bd78811dfb51c92df81e58c3a50d" ref="layer-sha256:c09d5cdb7367fff0d581bb8003e3520dc3e8bd78811dfb51c92df81e58c3a50d" total=637463
time="2022-05-21T12:24:48.577150192Z" level=debug msg="(*service).Write started" expected="sha256:92b614cff45fafd028cec952e0cb2584e8d931bf0321e7d14bfafdf7a50ac3fa" ref="layer-sha256:92b614cff45fafd028cec952e0cb2584e8d931bf0321e7d14bfafdf7a50ac3fa" total=2233
time="2022-05-21T12:24:48.655506586Z" level=info msg="successfully started the VM" runtime=aws.firecracker vmID=2
time="2022-05-21T12:24:48.655959580Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/start type=VMStart
time="2022-05-21T12:24:48.659591121Z" level=debug msg="prepare snapshot" key=2 parent="sha256:e172bf8795d813920c658a6772bb108238dc2cf13f1fc1ee1ac5c595c37da14a"
time="2022-05-21T12:24:48.660053503Z" level=debug msg=prepare key=firecracker-containerd/21/2 parent="firecracker-containerd/19/sha256:e172bf8795d813920c658a6772bb108238dc2cf13f1fc1ee1ac5c595c37da14a"
time="2022-05-21T12:24:48.660137451Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-12' from 'fc-dev-thinpool-snap-10'"
time="2022-05-21T12:24:48.660288676Z" level=info msg="successfully started the VM" runtime=aws.firecracker vmID=3
time="2022-05-21T12:24:48.660603911Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/start type=VMStart
time="2022-05-21T12:24:48.663921018Z" level=debug msg="prepare snapshot" key=3 parent="sha256:e172bf8795d813920c658a6772bb108238dc2cf13f1fc1ee1ac5c595c37da14a"
time="2022-05-21T12:24:48.664368713Z" level=debug msg=prepare key=firecracker-containerd/22/3 parent="firecracker-containerd/19/sha256:e172bf8795d813920c658a6772bb108238dc2cf13f1fc1ee1ac5c595c37da14a"
time="2022-05-21T12:24:48.745766203Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-13' from 'fc-dev-thinpool-snap-10'"
time="2022-05-21T12:24:48.746100884Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:48.748570453Z" level=debug msg="get snapshot mounts" key=2
time="2022-05-21T12:24:48.748648289Z" level=debug msg=mounts key=firecracker-containerd/21/2
time="2022-05-21T12:24:48.758402008Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/create type=containerd.events.ContainerCreate
time="2022-05-21T12:24:48.760085894Z" level=debug msg="get snapshot mounts" key=2
time="2022-05-21T12:24:48.760127803Z" level=debug msg=mounts key=firecracker-containerd/21/2
time="2022-05-21T12:24:48.779856177Z" level=debug msg="garbage collected" d="769.191µs"
time="2022-05-21T12:24:48.807328789Z" level=debug msg=StartShim runtime=aws.firecracker task_id=2
time="2022-05-21T12:24:48.807982271Z" level=debug msg="create VM request: VMID:\"2\" "
time="2022-05-21T12:24:48.808019231Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:24:48.808404187Z" level=info msg="successfully started shim (git commit: 19c96c059d7a95e8eb7f27b4e2847c4a84898698)." runtime=aws.firecracker task_id=2 vmID=2
time="2022-05-21T12:24:48.811229016Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:24:48.822032224Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:48.825356394Z" level=debug msg="get snapshot mounts" key=3
time="2022-05-21T12:24:48.825405937Z" level=debug msg=mounts key=firecracker-containerd/22/3
time="2022-05-21T12:24:48.832160168Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/create type=containerd.events.ContainerCreate
time="2022-05-21T12:24:48.833470099Z" level=debug msg="get snapshot mounts" key=3
time="2022-05-21T12:24:48.833522718Z" level=debug msg=mounts key=firecracker-containerd/22/3
time="2022-05-21T12:24:48.847583772Z" level=debug msg="garbage collected" d="862.737µs"
time="2022-05-21T12:24:48.875450717Z" level=debug msg=StartShim runtime=aws.firecracker task_id=3
time="2022-05-21T12:24:48.876062441Z" level=debug msg="create VM request: VMID:\"3\" "
time="2022-05-21T12:24:48.876086607Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:24:48.876296012Z" level=info msg="successfully started shim (git commit: 19c96c059d7a95e8eb7f27b4e2847c4a84898698)." runtime=aws.firecracker task_id=3 vmID=3
time="2022-05-21T12:24:48.878901356Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:24:48.912138521Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/create type=containerd.events.TaskCreate
time="2022-05-21T12:24:48.928601586Z" level=info msg="successfully created task" ExecID= TaskID=2 pid_in_vm=720 runtime=aws.firecracker vmID=2
time="2022-05-21T12:24:48.934226826Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/start type=containerd.events.TaskStart
time="2022-05-21T12:24:48.975449154Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/create type=containerd.events.TaskCreate
time="2022-05-21T12:24:48.994231583Z" level=info msg="successfully created task" ExecID= TaskID=3 pid_in_vm=719 runtime=aws.firecracker vmID=3
time="2022-05-21T12:24:49.000235839Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/start type=containerd.events.TaskStart
time="2022-05-21T12:24:49.237655906Z" level=debug msg="diff applied" d=19.129675ms digest="sha256:92b614cff45fafd028cec952e0cb2584e8d931bf0321e7d14bfafdf7a50ac3fa" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=2233
time="2022-05-21T12:24:49.238000666Z" level=debug msg="commit snapshot" key="extract-499911517-YXoM sha256:e4226272b56cdeb334d6c3377374a6760ceae26f704fddedd5ff871e52d19784" name="sha256:e4226272b56cdeb334d6c3377374a6760ceae26f704fddedd5ff871e52d19784"
time="2022-05-21T12:24:49.238177780Z" level=debug msg=commit key="firecracker-containerd/20/extract-499911517-YXoM sha256:e4226272b56cdeb334d6c3377374a6760ceae26f704fddedd5ff871e52d19784" name="firecracker-containerd/23/sha256:e4226272b56cdeb334d6c3377374a6760ceae26f704fddedd5ff871e52d19784"
time="2022-05-21T12:24:49.313224341Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:49.317903387Z" level=debug msg="stat snapshot" key="sha256:046f313f5e6a5160dd5c71cdf40aed78bb816bce9030b7b1617dc499a623dac8"
time="2022-05-21T12:24:49.319639702Z" level=debug msg="prepare snapshot" key="extract-318339209-DMDW sha256:046f313f5e6a5160dd5c71cdf40aed78bb816bce9030b7b1617dc499a623dac8" parent="sha256:e4226272b56cdeb334d6c3377374a6760ceae26f704fddedd5ff871e52d19784"
time="2022-05-21T12:24:49.320082337Z" level=debug msg=prepare key="firecracker-containerd/24/extract-318339209-DMDW sha256:046f313f5e6a5160dd5c71cdf40aed78bb816bce9030b7b1617dc499a623dac8" parent="firecracker-containerd/23/sha256:e4226272b56cdeb334d6c3377374a6760ceae26f704fddedd5ff871e52d19784"
time="2022-05-21T12:24:49.320166776Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-14' from 'fc-dev-thinpool-snap-11'"
time="2022-05-21T12:24:49.411564581Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:51.009333002Z" level=debug msg="diff applied" d=607.995066ms digest="sha256:0d776ee02572ee50935002da7bd7fcda4a60be4a48c4ace5dd3216c327e6767a" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=20365161
time="2022-05-21T12:24:51.009737976Z" level=debug msg="commit snapshot" key="extract-318339209-DMDW sha256:046f313f5e6a5160dd5c71cdf40aed78bb816bce9030b7b1617dc499a623dac8" name="sha256:046f313f5e6a5160dd5c71cdf40aed78bb816bce9030b7b1617dc499a623dac8"
time="2022-05-21T12:24:51.009891575Z" level=debug msg=commit key="firecracker-containerd/24/extract-318339209-DMDW sha256:046f313f5e6a5160dd5c71cdf40aed78bb816bce9030b7b1617dc499a623dac8" name="firecracker-containerd/25/sha256:046f313f5e6a5160dd5c71cdf40aed78bb816bce9030b7b1617dc499a623dac8"
time="2022-05-21T12:24:51.085627988Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:51.086715970Z" level=debug msg="stat snapshot" key="sha256:21d1c32ff456f90efb6f819dc94981414274877bd50caee6b7866a3f65280253"
time="2022-05-21T12:24:51.087120152Z" level=debug msg="prepare snapshot" key="extract-86935985-FEyI sha256:21d1c32ff456f90efb6f819dc94981414274877bd50caee6b7866a3f65280253" parent="sha256:046f313f5e6a5160dd5c71cdf40aed78bb816bce9030b7b1617dc499a623dac8"
time="2022-05-21T12:24:51.087529244Z" level=debug msg=prepare key="firecracker-containerd/26/extract-86935985-FEyI sha256:21d1c32ff456f90efb6f819dc94981414274877bd50caee6b7866a3f65280253" parent="firecracker-containerd/25/sha256:046f313f5e6a5160dd5c71cdf40aed78bb816bce9030b7b1617dc499a623dac8"
time="2022-05-21T12:24:51.087606750Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-15' from 'fc-dev-thinpool-snap-14'"
time="2022-05-21T12:24:51.184203071Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:51.216163338Z" level=debug msg="diff applied" d=31.496652ms digest="sha256:c09d5cdb7367fff0d581bb8003e3520dc3e8bd78811dfb51c92df81e58c3a50d" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=637463
time="2022-05-21T12:24:51.216566308Z" level=debug msg="commit snapshot" key="extract-86935985-FEyI sha256:21d1c32ff456f90efb6f819dc94981414274877bd50caee6b7866a3f65280253" name="sha256:21d1c32ff456f90efb6f819dc94981414274877bd50caee6b7866a3f65280253"
time="2022-05-21T12:24:51.216717884Z" level=debug msg=commit key="firecracker-containerd/26/extract-86935985-FEyI sha256:21d1c32ff456f90efb6f819dc94981414274877bd50caee6b7866a3f65280253" name="firecracker-containerd/27/sha256:21d1c32ff456f90efb6f819dc94981414274877bd50caee6b7866a3f65280253"
time="2022-05-21T12:24:51.297623808Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:24:51.299635522Z" level=debug msg="create image" name="ghcr.io/ease-lab/helloworld:var_workload" target="sha256:12dc6715ed1a8306f246ceaf7742c09e38a52a79d17421e4a50d7e0e09fdbc25"
time="2022-05-21T12:24:51.300012674Z" level=debug msg="event published" ns=firecracker-containerd topic=/images/create type=containerd.services.images.v1.ImageCreate
time="2022-05-21T12:24:51.334027246Z" level=debug msg="garbage collected" d="817.562µs"
time="2022-05-21T12:24:51.338659313Z" level=debug msg="create VM request: VMID:\"5\" MachineCfg:<MemSizeMib:256 VcpuCount:1 > KernelArgs:\"ro noapic reboot=k panic=1 pci=off nomodules systemd.log_color=false systemd.unit=firecracker.target init=/sbin/overlay-init tsc=reliable quiet 8250.nr_uarts=0 ipv6.disable=1\" NetworkInterfaces:<StaticConfig:<MacAddress:\"02:FC:00:00:00:04\" HostDevName:\"5_tap\" IPConfig:<PrimaryAddr:\"190.128.0.6/10\" GatewayAddr:\"190.128.0.1\" Nameservers:\"8.8.8.8\" > > > TimeoutSeconds:100 OffloadEnabled:true "
time="2022-05-21T12:24:51.338710590Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:24:51.339036134Z" level=debug msg="starting containerd-shim-aws-firecracker" vmID=5
time="2022-05-21T12:24:51.387511334Z" level=info msg="starting signal loop" namespace=firecracker-containerd path="/var/lib/firecracker-containerd/shim-base/firecracker-containerd#5" pid=26240
time="2022-05-21T12:24:51.387996819Z" level=info msg="creating new VM" runtime=aws.firecracker vmID=5
time="2022-05-21T12:24:51.388381435Z" level=info msg="Called startVMM(), setting up a VMM on firecracker.sock" runtime=aws.firecracker
time="2022-05-21T12:24:51.400447155Z" level=info msg="refreshMachineConfiguration: [GET /machine-config][200] getMachineConfigurationOK &{CPUTemplate: HtEnabled:0xc000713313 MemSizeMib:0xc000713308 TrackDirtyPages:false VcpuCount:0xc000713300}" runtime=aws.firecracker
time="2022-05-21T12:24:51.400619800Z" level=info msg="PutGuestBootSource: [PUT /boot-source][204] putGuestBootSourceNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:51.400655678Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/runtime/default-rootfs.img, slot root_drive, root true." runtime=aws.firecracker
time="2022-05-21T12:24:51.400959661Z" level=info msg="Attached drive /var/lib/firecracker-containerd/runtime/default-rootfs.img: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:51.400997262Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#5/ctrstub0, slot MN2HE43UOVRDA, root false." runtime=aws.firecracker
time="2022-05-21T12:24:51.401149920Z" level=info msg="Attached drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#5/ctrstub0: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:51.401175088Z" level=info msg="Attaching NIC 5_tap (hwaddr 02:FC:00:00:00:04) at index 1" runtime=aws.firecracker
time="2022-05-21T12:24:51.418241991Z" level=info msg="startInstance successful: [PUT /actions][204] createSyncActionNoContent " runtime=aws.firecracker
time="2022-05-21T12:24:51.418258011Z" level=info msg="calling agent" runtime=aws.firecracker vmID=5
time="2022-05-21T12:24:52.118525353Z" level=info msg="successfully started the VM" runtime=aws.firecracker vmID=5
time="2022-05-21T12:24:52.119064731Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/start type=VMStart
time="2022-05-21T12:24:52.125840722Z" level=debug msg="prepare snapshot" key=5 parent="sha256:21d1c32ff456f90efb6f819dc94981414274877bd50caee6b7866a3f65280253"
time="2022-05-21T12:24:52.126585177Z" level=debug msg=prepare key=firecracker-containerd/28/5 parent="firecracker-containerd/27/sha256:21d1c32ff456f90efb6f819dc94981414274877bd50caee6b7866a3f65280253"
time="2022-05-21T12:24:52.126690075Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-16' from 'fc-dev-thinpool-snap-15'"
time="2022-05-21T12:24:52.197118965Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:24:52.201176990Z" level=debug msg="get snapshot mounts" key=5
time="2022-05-21T12:24:52.201443713Z" level=debug msg=mounts key=firecracker-containerd/28/5
time="2022-05-21T12:24:52.209478770Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/create type=containerd.events.ContainerCreate
time="2022-05-21T12:24:52.211163929Z" level=debug msg="get snapshot mounts" key=5
time="2022-05-21T12:24:52.211244871Z" level=debug msg=mounts key=firecracker-containerd/28/5
time="2022-05-21T12:24:52.243509282Z" level=debug msg="garbage collected" d="693.518µs"
time="2022-05-21T12:24:52.255266430Z" level=debug msg=StartShim runtime=aws.firecracker task_id=5
time="2022-05-21T12:24:52.255873806Z" level=debug msg="create VM request: VMID:\"5\" "
time="2022-05-21T12:24:52.255902810Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:24:52.256187858Z" level=info msg="successfully started shim (git commit: 19c96c059d7a95e8eb7f27b4e2847c4a84898698)." runtime=aws.firecracker task_id=5 vmID=5
time="2022-05-21T12:24:52.258642338Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:24:52.358603596Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/create type=containerd.events.TaskCreate
time="2022-05-21T12:24:52.371484614Z" level=info msg="successfully created task" ExecID= TaskID=5 pid_in_vm=719 runtime=aws.firecracker vmID=5
time="2022-05-21T12:24:52.377259267Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/start type=containerd.events.TaskStart
time="2022-05-21T12:25:05.465954028Z" level=debug msg="diff applied" d=9.911367773s digest="sha256:72c1fa02b2c870da7fd4c4a0af11b837cd448185b4ff31f3ced4c1e11199d743" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=248074790
time="2022-05-21T12:25:05.466983841Z" level=debug msg="commit snapshot" key="extract-370322448-tHLo sha256:e4e4f6845ea6130dbe3b08e769e3bbc16a9f0dfe037f0380c5123e9b0d9a34d6" name="sha256:e4e4f6845ea6130dbe3b08e769e3bbc16a9f0dfe037f0380c5123e9b0d9a34d6"
time="2022-05-21T12:25:05.467261284Z" level=debug msg=commit key="firecracker-containerd/4/extract-370322448-tHLo sha256:e4e4f6845ea6130dbe3b08e769e3bbc16a9f0dfe037f0380c5123e9b0d9a34d6" name="firecracker-containerd/29/sha256:e4e4f6845ea6130dbe3b08e769e3bbc16a9f0dfe037f0380c5123e9b0d9a34d6"
time="2022-05-21T12:25:05.532879944Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:25:05.533905458Z" level=debug msg="stat snapshot" key="sha256:761a2f6827609b34a6ef9b77073d68620598e904e91c550f9f0760a5da0246ff"
time="2022-05-21T12:25:05.534309160Z" level=debug msg="prepare snapshot" key="extract-534135171-Gsvg sha256:761a2f6827609b34a6ef9b77073d68620598e904e91c550f9f0760a5da0246ff" parent="sha256:e4e4f6845ea6130dbe3b08e769e3bbc16a9f0dfe037f0380c5123e9b0d9a34d6"
time="2022-05-21T12:25:05.534652628Z" level=debug msg=prepare key="firecracker-containerd/30/extract-534135171-Gsvg sha256:761a2f6827609b34a6ef9b77073d68620598e904e91c550f9f0760a5da0246ff" parent="firecracker-containerd/29/sha256:e4e4f6845ea6130dbe3b08e769e3bbc16a9f0dfe037f0380c5123e9b0d9a34d6"
time="2022-05-21T12:25:05.534714715Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-17' from 'fc-dev-thinpool-snap-3'"
time="2022-05-21T12:25:05.624614447Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:25:07.800504585Z" level=debug msg="diff applied" d=2.175628435s digest="sha256:dfd5ae2430bfdaa3eabe80a09fef72b7b1b34a9b5ffe7690b3822cdad290cba5" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=55723509
time="2022-05-21T12:25:07.800953933Z" level=debug msg="commit snapshot" key="extract-534135171-Gsvg sha256:761a2f6827609b34a6ef9b77073d68620598e904e91c550f9f0760a5da0246ff" name="sha256:761a2f6827609b34a6ef9b77073d68620598e904e91c550f9f0760a5da0246ff"
time="2022-05-21T12:25:07.801127119Z" level=debug msg=commit key="firecracker-containerd/30/extract-534135171-Gsvg sha256:761a2f6827609b34a6ef9b77073d68620598e904e91c550f9f0760a5da0246ff" name="firecracker-containerd/31/sha256:761a2f6827609b34a6ef9b77073d68620598e904e91c550f9f0760a5da0246ff"
time="2022-05-21T12:25:07.873595380Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:25:07.874943213Z" level=debug msg="stat snapshot" key="sha256:23ee52c98db3041bb984ad19b6b1791ec210dbeb9ee80380a78cd9214f54f442"
time="2022-05-21T12:25:07.875273877Z" level=debug msg="prepare snapshot" key="extract-875142168-cJsH sha256:23ee52c98db3041bb984ad19b6b1791ec210dbeb9ee80380a78cd9214f54f442" parent="sha256:761a2f6827609b34a6ef9b77073d68620598e904e91c550f9f0760a5da0246ff"
time="2022-05-21T12:25:07.875651739Z" level=debug msg=prepare key="firecracker-containerd/32/extract-875142168-cJsH sha256:23ee52c98db3041bb984ad19b6b1791ec210dbeb9ee80380a78cd9214f54f442" parent="firecracker-containerd/31/sha256:761a2f6827609b34a6ef9b77073d68620598e904e91c550f9f0760a5da0246ff"
time="2022-05-21T12:25:07.875742481Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-18' from 'fc-dev-thinpool-snap-17'"
time="2022-05-21T12:25:07.965547635Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:25:07.975635214Z" level=debug msg="diff applied" d=9.750623ms digest="sha256:f6993a2cb9082ebcb2d8d151f19a1137ebbe7c642e8a3c41aac38f816c15c4c7" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=98
time="2022-05-21T12:25:07.976126410Z" level=debug msg="commit snapshot" key="extract-875142168-cJsH sha256:23ee52c98db3041bb984ad19b6b1791ec210dbeb9ee80380a78cd9214f54f442" name="sha256:23ee52c98db3041bb984ad19b6b1791ec210dbeb9ee80380a78cd9214f54f442"
time="2022-05-21T12:25:07.976418922Z" level=debug msg=commit key="firecracker-containerd/32/extract-875142168-cJsH sha256:23ee52c98db3041bb984ad19b6b1791ec210dbeb9ee80380a78cd9214f54f442" name="firecracker-containerd/33/sha256:23ee52c98db3041bb984ad19b6b1791ec210dbeb9ee80380a78cd9214f54f442"
time="2022-05-21T12:25:08.054021408Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:25:08.055162621Z" level=debug msg="stat snapshot" key="sha256:323a98f7b4e8d733d95c457ac33ab2230a84bd409f7091912e007056fcee664c"
time="2022-05-21T12:25:08.055483496Z" level=debug msg="prepare snapshot" key="extract-55342741-WAg_ sha256:323a98f7b4e8d733d95c457ac33ab2230a84bd409f7091912e007056fcee664c" parent="sha256:23ee52c98db3041bb984ad19b6b1791ec210dbeb9ee80380a78cd9214f54f442"
time="2022-05-21T12:25:08.055931842Z" level=debug msg=prepare key="firecracker-containerd/34/extract-55342741-WAg_ sha256:323a98f7b4e8d733d95c457ac33ab2230a84bd409f7091912e007056fcee664c" parent="firecracker-containerd/33/sha256:23ee52c98db3041bb984ad19b6b1791ec210dbeb9ee80380a78cd9214f54f442"
time="2022-05-21T12:25:08.056025238Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-19' from 'fc-dev-thinpool-snap-18'"
time="2022-05-21T12:25:08.143518692Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:25:08.152801643Z" level=debug msg="diff applied" d=8.977835ms digest="sha256:964f5a9ea2070018f381d9c968d435bc4576497232bd7d3e79121b180ef2169a" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=125
time="2022-05-21T12:25:08.153114002Z" level=debug msg="commit snapshot" key="extract-55342741-WAg_ sha256:323a98f7b4e8d733d95c457ac33ab2230a84bd409f7091912e007056fcee664c" name="sha256:323a98f7b4e8d733d95c457ac33ab2230a84bd409f7091912e007056fcee664c"
time="2022-05-21T12:25:08.153267552Z" level=debug msg=commit key="firecracker-containerd/34/extract-55342741-WAg_ sha256:323a98f7b4e8d733d95c457ac33ab2230a84bd409f7091912e007056fcee664c" name="firecracker-containerd/35/sha256:323a98f7b4e8d733d95c457ac33ab2230a84bd409f7091912e007056fcee664c"
time="2022-05-21T12:25:08.229321577Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:25:08.230220162Z" level=debug msg="stat snapshot" key="sha256:a5dcf1102c49ed5747aa91058a2f763bb7ce8577819ad128c1b2d94c0d306c29"
time="2022-05-21T12:25:08.230507724Z" level=debug msg="prepare snapshot" key="extract-230381326-YdN1 sha256:a5dcf1102c49ed5747aa91058a2f763bb7ce8577819ad128c1b2d94c0d306c29" parent="sha256:323a98f7b4e8d733d95c457ac33ab2230a84bd409f7091912e007056fcee664c"
time="2022-05-21T12:25:08.230851603Z" level=debug msg=prepare key="firecracker-containerd/36/extract-230381326-YdN1 sha256:a5dcf1102c49ed5747aa91058a2f763bb7ce8577819ad128c1b2d94c0d306c29" parent="firecracker-containerd/35/sha256:323a98f7b4e8d733d95c457ac33ab2230a84bd409f7091912e007056fcee664c"
time="2022-05-21T12:25:08.230924581Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-20' from 'fc-dev-thinpool-snap-19'"
time="2022-05-21T12:25:08.286350614Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:25:10.949871578Z" level=debug msg="diff applied" d=2.663278636s digest="sha256:95853ec29c67ccc835034ef04f5765d1064b835ffb476e2a073dbb8e7b3d7cf3" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=87932348
time="2022-05-21T12:25:10.950602266Z" level=debug msg="commit snapshot" key="extract-230381326-YdN1 sha256:a5dcf1102c49ed5747aa91058a2f763bb7ce8577819ad128c1b2d94c0d306c29" name="sha256:a5dcf1102c49ed5747aa91058a2f763bb7ce8577819ad128c1b2d94c0d306c29"
time="2022-05-21T12:25:10.950969719Z" level=debug msg=commit key="firecracker-containerd/36/extract-230381326-YdN1 sha256:a5dcf1102c49ed5747aa91058a2f763bb7ce8577819ad128c1b2d94c0d306c29" name="firecracker-containerd/37/sha256:a5dcf1102c49ed5747aa91058a2f763bb7ce8577819ad128c1b2d94c0d306c29"
time="2022-05-21T12:25:11.006100937Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:25:11.007345605Z" level=debug msg="stat snapshot" key="sha256:4774be23f537f9f04d55671855857893520107cc5afbe9c2aaf9d0396d10dfaf"
time="2022-05-21T12:25:11.007767130Z" level=debug msg="prepare snapshot" key="extract-7616776-vAK0 sha256:4774be23f537f9f04d55671855857893520107cc5afbe9c2aaf9d0396d10dfaf" parent="sha256:a5dcf1102c49ed5747aa91058a2f763bb7ce8577819ad128c1b2d94c0d306c29"
time="2022-05-21T12:25:11.008122881Z" level=debug msg=prepare key="firecracker-containerd/38/extract-7616776-vAK0 sha256:4774be23f537f9f04d55671855857893520107cc5afbe9c2aaf9d0396d10dfaf" parent="firecracker-containerd/37/sha256:a5dcf1102c49ed5747aa91058a2f763bb7ce8577819ad128c1b2d94c0d306c29"
time="2022-05-21T12:25:11.008209033Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-21' from 'fc-dev-thinpool-snap-20'"
time="2022-05-21T12:25:11.081823557Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:25:11.092862250Z" level=debug msg="diff applied" d=10.610355ms digest="sha256:466a9644be5453fb0268d102159dd91b988e5d24f84431d0a5a57ee7ff21de2b" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=3742
time="2022-05-21T12:25:11.093137680Z" level=debug msg="commit snapshot" key="extract-7616776-vAK0 sha256:4774be23f537f9f04d55671855857893520107cc5afbe9c2aaf9d0396d10dfaf" name="sha256:4774be23f537f9f04d55671855857893520107cc5afbe9c2aaf9d0396d10dfaf"
time="2022-05-21T12:25:11.093358777Z" level=debug msg=commit key="firecracker-containerd/38/extract-7616776-vAK0 sha256:4774be23f537f9f04d55671855857893520107cc5afbe9c2aaf9d0396d10dfaf" name="firecracker-containerd/39/sha256:4774be23f537f9f04d55671855857893520107cc5afbe9c2aaf9d0396d10dfaf"
time="2022-05-21T12:25:11.141590750Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:25:11.142603100Z" level=debug msg="stat snapshot" key="sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb"
time="2022-05-21T12:25:11.142989929Z" level=debug msg="prepare snapshot" key="extract-142791475-HfT5 sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb" parent="sha256:4774be23f537f9f04d55671855857893520107cc5afbe9c2aaf9d0396d10dfaf"
time="2022-05-21T12:25:11.143465817Z" level=debug msg=prepare key="firecracker-containerd/40/extract-142791475-HfT5 sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb" parent="firecracker-containerd/39/sha256:4774be23f537f9f04d55671855857893520107cc5afbe9c2aaf9d0396d10dfaf"
time="2022-05-21T12:25:11.143588078Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-22' from 'fc-dev-thinpool-snap-21'"
time="2022-05-21T12:25:11.217970620Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:25:11.230484336Z" level=debug msg="diff applied" d=12.25088ms digest="sha256:614456ff946738237eb1d5e7ddb9b3b9578292cd2de96317aa37d76ea0a4eea9" media=application/vnd.docker.image.rootfs.diff.tar.gzip size=185738
time="2022-05-21T12:25:11.230799661Z" level=debug msg="commit snapshot" key="extract-142791475-HfT5 sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb" name="sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb"
time="2022-05-21T12:25:11.230985772Z" level=debug msg=commit key="firecracker-containerd/40/extract-142791475-HfT5 sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb" name="firecracker-containerd/41/sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb"
time="2022-05-21T12:25:11.281575282Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/commit type=containerd.events.SnapshotCommit
time="2022-05-21T12:25:11.283471699Z" level=debug msg="create image" name="docker.io/vhiveease/rnn_serving:var_workload" target="sha256:6a11e6dbd88b1ce1ebb284c769b52e3fdb66a0f37b392bded5612045ff2cae61"
time="2022-05-21T12:25:11.283876082Z" level=debug msg="event published" ns=firecracker-containerd topic=/images/create type=containerd.services.images.v1.ImageCreate
time="2022-05-21T12:25:11.301886687Z" level=debug msg="garbage collected" d="742.38µs"
time="2022-05-21T12:25:11.326557575Z" level=debug msg="create VM request: VMID:\"1\" MachineCfg:<MemSizeMib:256 VcpuCount:1 > KernelArgs:\"ro noapic reboot=k panic=1 pci=off nomodules systemd.log_color=false systemd.unit=firecracker.target init=/sbin/overlay-init tsc=reliable quiet 8250.nr_uarts=0 ipv6.disable=1\" NetworkInterfaces:<StaticConfig:<MacAddress:\"02:FC:00:00:00:00\" HostDevName:\"1_tap\" IPConfig:<PrimaryAddr:\"190.128.0.2/10\" GatewayAddr:\"190.128.0.1\" Nameservers:\"8.8.8.8\" > > > TimeoutSeconds:100 OffloadEnabled:true "
time="2022-05-21T12:25:11.326598613Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:25:11.326842783Z" level=debug msg="starting containerd-shim-aws-firecracker" vmID=1
time="2022-05-21T12:25:11.328074196Z" level=debug msg="create VM request: VMID:\"4\" MachineCfg:<MemSizeMib:256 VcpuCount:1 > KernelArgs:\"ro noapic reboot=k panic=1 pci=off nomodules systemd.log_color=false systemd.unit=firecracker.target init=/sbin/overlay-init tsc=reliable quiet 8250.nr_uarts=0 ipv6.disable=1\" NetworkInterfaces:<StaticConfig:<MacAddress:\"02:FC:00:00:00:03\" HostDevName:\"4_tap\" IPConfig:<PrimaryAddr:\"190.128.0.5/10\" GatewayAddr:\"190.128.0.1\" Nameservers:\"8.8.8.8\" > > > TimeoutSeconds:100 OffloadEnabled:true "
time="2022-05-21T12:25:11.328115444Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:25:11.328280325Z" level=debug msg="starting containerd-shim-aws-firecracker" vmID=4
time="2022-05-21T12:25:11.375283410Z" level=info msg="starting signal loop" namespace=firecracker-containerd path="/var/lib/firecracker-containerd/shim-base/firecracker-containerd#4" pid=26819
time="2022-05-21T12:25:11.375582144Z" level=info msg="creating new VM" runtime=aws.firecracker vmID=4
time="2022-05-21T12:25:11.375829260Z" level=info msg="Called startVMM(), setting up a VMM on firecracker.sock" runtime=aws.firecracker
time="2022-05-21T12:25:11.379299967Z" level=info msg="starting signal loop" namespace=firecracker-containerd path="/var/lib/firecracker-containerd/shim-base/firecracker-containerd#1" pid=26818
time="2022-05-21T12:25:11.379573272Z" level=info msg="creating new VM" runtime=aws.firecracker vmID=1
time="2022-05-21T12:25:11.379801954Z" level=info msg="Called startVMM(), setting up a VMM on firecracker.sock" runtime=aws.firecracker
time="2022-05-21T12:25:11.387199077Z" level=info msg="refreshMachineConfiguration: [GET /machine-config][200] getMachineConfigurationOK &{CPUTemplate: HtEnabled:0xc00093a873 MemSizeMib:0xc00093a868 TrackDirtyPages:false VcpuCount:0xc00093a860}" runtime=aws.firecracker
time="2022-05-21T12:25:11.387356214Z" level=info msg="PutGuestBootSource: [PUT /boot-source][204] putGuestBootSourceNoContent " runtime=aws.firecracker
time="2022-05-21T12:25:11.387386721Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/runtime/default-rootfs.img, slot root_drive, root true." runtime=aws.firecracker
time="2022-05-21T12:25:11.387637324Z" level=info msg="Attached drive /var/lib/firecracker-containerd/runtime/default-rootfs.img: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:25:11.387667281Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#4/ctrstub0, slot MN2HE43UOVRDA, root false." runtime=aws.firecracker
time="2022-05-21T12:25:11.387785674Z" level=info msg="Attached drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#4/ctrstub0: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:25:11.387801133Z" level=info msg="Attaching NIC 4_tap (hwaddr 02:FC:00:00:00:03) at index 1" runtime=aws.firecracker
time="2022-05-21T12:25:11.391162384Z" level=info msg="refreshMachineConfiguration: [GET /machine-config][200] getMachineConfigurationOK &{CPUTemplate: HtEnabled:0xc000d0cf53 MemSizeMib:0xc000d0cf48 TrackDirtyPages:false VcpuCount:0xc000d0cf40}" runtime=aws.firecracker
time="2022-05-21T12:25:11.391331583Z" level=info msg="PutGuestBootSource: [PUT /boot-source][204] putGuestBootSourceNoContent " runtime=aws.firecracker
time="2022-05-21T12:25:11.391349527Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/runtime/default-rootfs.img, slot root_drive, root true." runtime=aws.firecracker
time="2022-05-21T12:25:11.391574912Z" level=info msg="Attached drive /var/lib/firecracker-containerd/runtime/default-rootfs.img: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:25:11.391592405Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#1/ctrstub0, slot MN2HE43UOVRDA, root false." runtime=aws.firecracker
time="2022-05-21T12:25:11.391722781Z" level=info msg="Attached drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#1/ctrstub0: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:25:11.391744702Z" level=info msg="Attaching NIC 1_tap (hwaddr 02:FC:00:00:00:00) at index 1" runtime=aws.firecracker
time="2022-05-21T12:25:11.399102942Z" level=info msg="startInstance successful: [PUT /actions][204] createSyncActionNoContent " runtime=aws.firecracker
time="2022-05-21T12:25:11.399119804Z" level=info msg="calling agent" runtime=aws.firecracker vmID=4
time="2022-05-21T12:25:11.402984144Z" level=info msg="startInstance successful: [PUT /actions][204] createSyncActionNoContent " runtime=aws.firecracker
time="2022-05-21T12:25:11.402997218Z" level=info msg="calling agent" runtime=aws.firecracker vmID=1
time="2022-05-21T12:25:12.099358622Z" level=info msg="successfully started the VM" runtime=aws.firecracker vmID=4
time="2022-05-21T12:25:12.099812858Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/start type=VMStart
time="2022-05-21T12:25:12.103224644Z" level=info msg="successfully started the VM" runtime=aws.firecracker vmID=1
time="2022-05-21T12:25:12.103299526Z" level=debug msg="prepare snapshot" key=4 parent="sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb"
time="2022-05-21T12:25:12.103626583Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/start type=VMStart
time="2022-05-21T12:25:12.103741279Z" level=debug msg=prepare key=firecracker-containerd/42/4 parent="firecracker-containerd/41/sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb"
time="2022-05-21T12:25:12.103853942Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-23' from 'fc-dev-thinpool-snap-22'"
time="2022-05-21T12:25:12.112161052Z" level=debug msg="prepare snapshot" key=1 parent="sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb"
time="2022-05-21T12:25:12.112740785Z" level=debug msg=prepare key=firecracker-containerd/43/1 parent="firecracker-containerd/41/sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb"
time="2022-05-21T12:25:12.172554586Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-24' from 'fc-dev-thinpool-snap-22'"
time="2022-05-21T12:25:12.172825998Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:25:12.175410644Z" level=debug msg="get snapshot mounts" key=4
time="2022-05-21T12:25:12.175462131Z" level=debug msg=mounts key=firecracker-containerd/42/4
time="2022-05-21T12:25:12.184505791Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/create type=containerd.events.ContainerCreate
time="2022-05-21T12:25:12.185769685Z" level=debug msg="get snapshot mounts" key=4
time="2022-05-21T12:25:12.185831982Z" level=debug msg=mounts key=firecracker-containerd/42/4
time="2022-05-21T12:25:12.204733808Z" level=debug msg="garbage collected" d="671.056µs"
time="2022-05-21T12:25:12.227284136Z" level=debug msg=StartShim runtime=aws.firecracker task_id=4
time="2022-05-21T12:25:12.227854773Z" level=debug msg="create VM request: VMID:\"4\" "
time="2022-05-21T12:25:12.227882195Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:25:12.228234149Z" level=info msg="successfully started shim (git commit: 19c96c059d7a95e8eb7f27b4e2847c4a84898698)." runtime=aws.firecracker task_id=4 vmID=4
time="2022-05-21T12:25:12.230708767Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:25:12.245934257Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:25:12.248037234Z" level=debug msg="get snapshot mounts" key=1
time="2022-05-21T12:25:12.248084524Z" level=debug msg=mounts key=firecracker-containerd/43/1
time="2022-05-21T12:25:12.255745936Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/create type=containerd.events.ContainerCreate
time="2022-05-21T12:25:12.257099700Z" level=debug msg="get snapshot mounts" key=1
time="2022-05-21T12:25:12.257167918Z" level=debug msg=mounts key=firecracker-containerd/43/1
time="2022-05-21T12:25:12.276758834Z" level=debug msg="garbage collected" d="755.656µs"
time="2022-05-21T12:25:12.299323429Z" level=debug msg=StartShim runtime=aws.firecracker task_id=1
time="2022-05-21T12:25:12.299736608Z" level=debug msg="create VM request: VMID:\"1\" "
time="2022-05-21T12:25:12.299759562Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:25:12.300034340Z" level=info msg="successfully started shim (git commit: 19c96c059d7a95e8eb7f27b4e2847c4a84898698)." runtime=aws.firecracker task_id=1 vmID=1
time="2022-05-21T12:25:12.302584190Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:25:12.331164844Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/create type=containerd.events.TaskCreate
time="2022-05-21T12:25:12.342487943Z" level=info msg="successfully created task" ExecID= TaskID=4 pid_in_vm=720 runtime=aws.firecracker vmID=4
time="2022-05-21T12:25:12.348046769Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/start type=containerd.events.TaskStart
time="2022-05-21T12:25:12.403047171Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/create type=containerd.events.TaskCreate
time="2022-05-21T12:25:12.418334398Z" level=info msg="successfully created task" ExecID= TaskID=1 pid_in_vm=719 runtime=aws.firecracker vmID=1
time="2022-05-21T12:25:12.424258152Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/start type=containerd.events.TaskStart
time="2022-05-21T12:34:36.120304941Z" level=debug msg="create VM request: VMID:\"6\" MachineCfg:<MemSizeMib:256 VcpuCount:1 > KernelArgs:\"ro noapic reboot=k panic=1 pci=off nomodules systemd.log_color=false systemd.unit=firecracker.target init=/sbin/overlay-init tsc=reliable quiet 8250.nr_uarts=0 ipv6.disable=1\" NetworkInterfaces:<StaticConfig:<MacAddress:\"02:FC:00:00:00:05\" HostDevName:\"6_tap\" IPConfig:<PrimaryAddr:\"190.128.0.7/10\" GatewayAddr:\"190.128.0.1\" Nameservers:\"8.8.8.8\" > > > TimeoutSeconds:100 OffloadEnabled:true "
time="2022-05-21T12:34:36.120395703Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:34:36.120770030Z" level=debug msg="starting containerd-shim-aws-firecracker" vmID=6
time="2022-05-21T12:34:36.171393829Z" level=info msg="starting signal loop" namespace=firecracker-containerd path="/var/lib/firecracker-containerd/shim-base/firecracker-containerd#6" pid=31898
time="2022-05-21T12:34:36.171806810Z" level=info msg="creating new VM" runtime=aws.firecracker vmID=6
time="2022-05-21T12:34:36.172126354Z" level=info msg="Called startVMM(), setting up a VMM on firecracker.sock" runtime=aws.firecracker
time="2022-05-21T12:34:36.184094713Z" level=info msg="refreshMachineConfiguration: [GET /machine-config][200] getMachineConfigurationOK &{CPUTemplate: HtEnabled:0xc00038f113 MemSizeMib:0xc00038f0c8 TrackDirtyPages:false VcpuCount:0xc00038f0c0}" runtime=aws.firecracker
time="2022-05-21T12:34:36.184324107Z" level=info msg="PutGuestBootSource: [PUT /boot-source][204] putGuestBootSourceNoContent " runtime=aws.firecracker
time="2022-05-21T12:34:36.184367880Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/runtime/default-rootfs.img, slot root_drive, root true." runtime=aws.firecracker
time="2022-05-21T12:34:36.184666274Z" level=info msg="Attached drive /var/lib/firecracker-containerd/runtime/default-rootfs.img: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:34:36.184689628Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#6/ctrstub0, slot MN2HE43UOVRDA, root false." runtime=aws.firecracker
time="2022-05-21T12:34:36.184835213Z" level=info msg="Attached drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#6/ctrstub0: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-21T12:34:36.184853899Z" level=info msg="Attaching NIC 6_tap (hwaddr 02:FC:00:00:00:05) at index 1" runtime=aws.firecracker
time="2022-05-21T12:34:36.202232338Z" level=info msg="startInstance successful: [PUT /actions][204] createSyncActionNoContent " runtime=aws.firecracker
time="2022-05-21T12:34:36.202275319Z" level=info msg="calling agent" runtime=aws.firecracker vmID=6
time="2022-05-21T12:34:36.902597734Z" level=info msg="successfully started the VM" runtime=aws.firecracker vmID=6
time="2022-05-21T12:34:36.903160368Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/start type=VMStart
time="2022-05-21T12:34:36.906864862Z" level=debug msg="prepare snapshot" key=6 parent="sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb"
time="2022-05-21T12:34:36.907205697Z" level=debug msg=prepare key=firecracker-containerd/44/6 parent="firecracker-containerd/41/sha256:d7dccd214b2b808d39d11264689977e780b9e10662398cdae4fdc734fd008cdb"
time="2022-05-21T12:34:36.907301538Z" level=debug msg="creating snapshot device 'fc-dev-thinpool-snap-25' from 'fc-dev-thinpool-snap-22'"
time="2022-05-21T12:34:36.989814281Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare
time="2022-05-21T12:34:36.991850331Z" level=debug msg="get snapshot mounts" key=6
time="2022-05-21T12:34:36.991915985Z" level=debug msg=mounts key=firecracker-containerd/44/6
time="2022-05-21T12:34:36.999427017Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/create type=containerd.events.ContainerCreate
time="2022-05-21T12:34:37.000632887Z" level=debug msg="get snapshot mounts" key=6
time="2022-05-21T12:34:37.000692700Z" level=debug msg=mounts key=firecracker-containerd/44/6
time="2022-05-21T12:34:37.027685728Z" level=debug msg="garbage collected" d="831.743µs"
time="2022-05-21T12:34:37.043377407Z" level=debug msg=StartShim runtime=aws.firecracker task_id=6
time="2022-05-21T12:34:37.043960820Z" level=debug msg="create VM request: VMID:\"6\" "
time="2022-05-21T12:34:37.043999754Z" level=debug msg="using namespace: firecracker-containerd"
time="2022-05-21T12:34:37.044295763Z" level=info msg="successfully started shim (git commit: 19c96c059d7a95e8eb7f27b4e2847c4a84898698)." runtime=aws.firecracker task_id=6 vmID=6
time="2022-05-21T12:34:37.046486375Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:34:37.145872412Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/create type=containerd.events.TaskCreate
time="2022-05-21T12:34:37.163495884Z" level=info msg="successfully created task" ExecID= TaskID=6 pid_in_vm=720 runtime=aws.firecracker vmID=6
time="2022-05-21T12:34:37.169413153Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/start type=containerd.events.TaskStart
time="2022-05-21T12:35:33.515152090Z" level=info msg=exited ExecID= TaskID=2 exit_status=137 exited_at="2022-05-21 12:35:33.499631242 +0000 UTC" runtime=aws.firecracker vmID=2
time="2022-05-21T12:35:33.515258822Z" level=info msg="connection was closed: read /proc/self/fd/14: file already closed" ExecID= TaskID=2 runtime=aws.firecracker stream=stdin vmID=2
time="2022-05-21T12:35:33.515247701Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/exit type=containerd.events.TaskExit
time="2022-05-21T12:35:33.515314667Z" level=error msg="error closing io stream" ExecID= TaskID=2 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdin vmID=2
time="2022-05-21T12:35:34.023171185Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/delete type=containerd.events.TaskDelete
time="2022-05-21T12:35:34.029031355Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:35:34.029539656Z" level=info msg="shim disconnected" id=2
time="2022-05-21T12:35:34.029643773Z" level=warning msg="cleaning up after shim disconnected" id=2 namespace=firecracker-containerd
time="2022-05-21T12:35:34.029659402Z" level=info msg="cleaning up dead shim"
time="2022-05-21T12:35:34.525450707Z" level=info msg=exited ExecID= TaskID=4 exit_status=137 exited_at="2022-05-21 12:35:34.510419647 +0000 UTC" runtime=aws.firecracker vmID=4
time="2022-05-21T12:35:34.525547219Z" level=info msg="connection was closed: read /proc/self/fd/14: file already closed" ExecID= TaskID=4 runtime=aws.firecracker stream=stdin vmID=4
time="2022-05-21T12:35:34.525598526Z" level=error msg="error closing io stream" ExecID= TaskID=4 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdin vmID=4
time="2022-05-21T12:35:34.525646237Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/exit type=containerd.events.TaskExit
time="2022-05-21T12:35:34.536096345Z" level=info msg=exited ExecID= TaskID=5 exit_status=137 exited_at="2022-05-21 12:35:34.52122283 +0000 UTC" runtime=aws.firecracker vmID=5
time="2022-05-21T12:35:34.536129748Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/exit type=containerd.events.TaskExit
time="2022-05-21T12:35:34.536205812Z" level=info msg="connection was closed: read /proc/self/fd/14: file already closed" ExecID= TaskID=5 runtime=aws.firecracker stream=stdin vmID=5
time="2022-05-21T12:35:34.536276846Z" level=error msg="error closing io stream" ExecID= TaskID=5 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdin vmID=5
time="2022-05-21T12:35:35.034033817Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/delete type=containerd.events.TaskDelete
time="2022-05-21T12:35:35.043916212Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/delete type=containerd.events.TaskDelete
time="2022-05-21T12:35:35.044094730Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:35:35.044436526Z" level=info msg="shim disconnected" id=4
time="2022-05-21T12:35:35.044473245Z" level=warning msg="cleaning up after shim disconnected" id=4 namespace=firecracker-containerd
time="2022-05-21T12:35:35.044485418Z" level=info msg="cleaning up dead shim"
time="2022-05-21T12:35:35.050067843Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:35:35.050438454Z" level=info msg="shim disconnected" id=5
time="2022-05-21T12:35:35.050492075Z" level=warning msg="cleaning up after shim disconnected" id=5 namespace=firecracker-containerd
time="2022-05-21T12:35:35.050505430Z" level=info msg="cleaning up dead shim"
time="2022-05-21T12:35:37.072952376Z" level=info msg=exited ExecID= TaskID=1 exit_status=137 exited_at="2022-05-21 12:35:37.057779967 +0000 UTC" runtime=aws.firecracker vmID=1
time="2022-05-21T12:35:37.073011828Z" level=info msg="connection was closed: read /proc/self/fd/14: file already closed" ExecID= TaskID=1 runtime=aws.firecracker stream=stdin vmID=1
time="2022-05-21T12:35:37.073058046Z" level=error msg="error closing io stream" ExecID= TaskID=1 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdin vmID=1
time="2022-05-21T12:35:37.073077032Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/exit type=containerd.events.TaskExit
time="2022-05-21T12:35:37.082337741Z" level=info msg=exited ExecID= TaskID=3 exit_status=137 exited_at="2022-05-21 12:35:37.066925921 +0000 UTC" runtime=aws.firecracker vmID=3
time="2022-05-21T12:35:37.082409727Z" level=info msg="connection was closed: read /proc/self/fd/14: file already closed" ExecID= TaskID=3 runtime=aws.firecracker stream=stdin vmID=3
time="2022-05-21T12:35:37.082431398Z" level=error msg="error closing io stream" ExecID= TaskID=3 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdin vmID=3
time="2022-05-21T12:35:37.082454582Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/exit type=containerd.events.TaskExit
time="2022-05-21T12:35:37.581306291Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/delete type=containerd.events.TaskDelete
time="2022-05-21T12:35:37.589649115Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/delete type=containerd.events.TaskDelete
time="2022-05-21T12:35:37.591723026Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:35:37.592051978Z" level=info msg="shim disconnected" id=1
time="2022-05-21T12:35:37.592108655Z" level=warning msg="cleaning up after shim disconnected" id=1 namespace=firecracker-containerd
time="2022-05-21T12:35:37.592118524Z" level=info msg="cleaning up dead shim"
time="2022-05-21T12:35:37.596143814Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:35:37.596375042Z" level=info msg="shim disconnected" id=3
time="2022-05-21T12:35:37.596425156Z" level=warning msg="cleaning up after shim disconnected" id=3 namespace=firecracker-containerd
time="2022-05-21T12:35:37.596432901Z" level=info msg="cleaning up dead shim"
time="2022-05-21T12:35:38.515280086Z" level=error msg="error closing io stream" ExecID= TaskID=2 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stderr vmID=2
time="2022-05-21T12:35:38.515270889Z" level=error msg="error closing io stream" ExecID= TaskID=2 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdout vmID=2
time="2022-05-21T12:35:39.077965525Z" level=error msg="failed to delete" cmd="/usr/local/bin/containerd-shim-aws-firecracker -namespace firecracker-containerd -address /run/firecracker-containerd/containerd.sock -publish-binary /usr/local/bin/firecracker-containerd -id 2 -bundle /run/firecracker-containerd/io.containerd.runtime.v2.task/firecracker-containerd/2 delete" error="exit status 1"
time="2022-05-21T12:35:39.078094950Z" level=warning msg="failed to clean up after shim disconnected" error="aws.firecracker: rpc error: code = DeadlineExceeded desc = timed out waiting for VM start\n: exit status 1" id=2 namespace=firecracker-containerd
time="2022-05-21T12:35:39.079736584Z" level=debug msg="remove snapshot" key=2
time="2022-05-21T12:35:39.080298125Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/remove type=containerd.events.SnapshotRemove
time="2022-05-21T12:35:39.080821836Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/delete type=containerd.events.ContainerDelete
time="2022-05-21T12:35:39.081206864Z" level=debug msg="stop VM: VMID:\"2\" "
time="2022-05-21T12:35:39.081553950Z" level=info msg="stopping the VM" runtime=aws.firecracker vmID=2
time="2022-05-21T12:35:39.090562622Z" level=debug msg="schedule snapshotter cleanup" snapshotter=devmapper
time="2022-05-21T12:35:39.090651170Z" level=debug msg=walk
time="2022-05-21T12:35:39.090852350Z" level=debug msg=remove key=firecracker-containerd/21/2
time="2022-05-21T12:35:39.193934417Z" level=debug msg="removed snapshot" key=firecracker-containerd/21/2 snapshotter=devmapper
time="2022-05-21T12:35:39.193956018Z" level=debug msg=cleanup
time="2022-05-21T12:35:39.193964594Z" level=debug msg="snapshot garbage collected" d=103.357858ms snapshotter=devmapper
time="2022-05-21T12:35:39.193983350Z" level=debug msg="garbage collected" d="883.059µs"
time="2022-05-21T12:35:39.499060469Z" level=info msg="firecracker exited: status=0" runtime=aws.firecracker
time="2022-05-21T12:35:39.499663128Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/stop type=VMStop
time="2022-05-21T12:35:39.525661724Z" level=error msg="error closing io stream" ExecID= TaskID=4 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdout vmID=4
time="2022-05-21T12:35:39.525652867Z" level=error msg="error closing io stream" ExecID= TaskID=4 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stderr vmID=4
time="2022-05-21T12:35:39.536259771Z" level=error msg="error closing io stream" ExecID= TaskID=5 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stderr vmID=5
time="2022-05-21T12:35:39.536259741Z" level=error msg="error closing io stream" ExecID= TaskID=5 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdout vmID=5
time="2022-05-21T12:35:40.141582173Z" level=error msg="failed to delete" cmd="/usr/local/bin/containerd-shim-aws-firecracker -namespace firecracker-containerd -address /run/firecracker-containerd/containerd.sock -publish-binary /usr/local/bin/firecracker-containerd -id 4 -bundle /run/firecracker-containerd/io.containerd.runtime.v2.task/firecracker-containerd/4 delete" error="exit status 1"
time="2022-05-21T12:35:40.141716307Z" level=warning msg="failed to clean up after shim disconnected" error="aws.firecracker: rpc error: code = DeadlineExceeded desc = timed out waiting for VM start\n: exit status 1" id=4 namespace=firecracker-containerd
time="2022-05-21T12:35:40.141557086Z" level=error msg="failed to delete" cmd="/usr/local/bin/containerd-shim-aws-firecracker -namespace firecracker-containerd -address /run/firecracker-containerd/containerd.sock -publish-binary /usr/local/bin/firecracker-containerd -id 5 -bundle /run/firecracker-containerd/io.containerd.runtime.v2.task/firecracker-containerd/5 delete" error="exit status 1"
time="2022-05-21T12:35:40.141825333Z" level=warning msg="failed to clean up after shim disconnected" error="aws.firecracker: rpc error: code = DeadlineExceeded desc = timed out waiting for VM start\n: exit status 1" id=5 namespace=firecracker-containerd
time="2022-05-21T12:35:40.143440216Z" level=debug msg="remove snapshot" key=4
time="2022-05-21T12:35:40.143464943Z" level=debug msg="remove snapshot" key=5
time="2022-05-21T12:35:40.144007869Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/remove type=containerd.events.SnapshotRemove
time="2022-05-21T12:35:40.144447070Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/remove type=containerd.events.SnapshotRemove
time="2022-05-21T12:35:40.144839111Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/delete type=containerd.events.ContainerDelete
time="2022-05-21T12:35:40.145152784Z" level=debug msg="stop VM: VMID:\"4\" "
time="2022-05-21T12:35:40.145205604Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/delete type=containerd.events.ContainerDelete
time="2022-05-21T12:35:40.145438494Z" level=debug msg="stop VM: VMID:\"5\" "
time="2022-05-21T12:35:40.145482488Z" level=info msg="stopping the VM" runtime=aws.firecracker vmID=4
time="2022-05-21T12:35:40.145822941Z" level=info msg="stopping the VM" runtime=aws.firecracker vmID=5
time="2022-05-21T12:35:40.155076697Z" level=debug msg="schedule snapshotter cleanup" snapshotter=devmapper
time="2022-05-21T12:35:40.155138284Z" level=debug msg=walk
time="2022-05-21T12:35:40.155298086Z" level=debug msg=remove key=firecracker-containerd/42/4
time="2022-05-21T12:35:40.265315476Z" level=debug msg="removed snapshot" key=firecracker-containerd/42/4 snapshotter=devmapper
time="2022-05-21T12:35:40.265379196Z" level=debug msg=remove key=firecracker-containerd/28/5
time="2022-05-21T12:35:40.384238663Z" level=debug msg="removed snapshot" key=firecracker-containerd/28/5 snapshotter=devmapper
time="2022-05-21T12:35:40.384263490Z" level=debug msg=cleanup
time="2022-05-21T12:35:40.384276064Z" level=debug msg="snapshot garbage collected" d=229.16377ms snapshotter=devmapper
time="2022-05-21T12:35:40.384304889Z" level=debug msg="garbage collected" d="768.834µs"
time="2022-05-21T12:35:40.533405823Z" level=info msg="firecracker exited: status=0" runtime=aws.firecracker
time="2022-05-21T12:35:40.534026236Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/stop type=VMStop
time="2022-05-21T12:35:40.594976715Z" level=info msg="firecracker exited: status=0" runtime=aws.firecracker
time="2022-05-21T12:35:40.595667792Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/stop type=VMStop
time="2022-05-21T12:35:42.073176402Z" level=error msg="error closing io stream" ExecID= TaskID=1 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stderr vmID=1
time="2022-05-21T12:35:42.073200047Z" level=error msg="error closing io stream" ExecID= TaskID=1 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdout vmID=1
time="2022-05-21T12:35:42.082455386Z" level=error msg="error closing io stream" ExecID= TaskID=3 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdout vmID=3
time="2022-05-21T12:35:42.082481695Z" level=error msg="error closing io stream" ExecID= TaskID=3 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stderr vmID=3
time="2022-05-21T12:35:42.641916005Z" level=error msg="failed to delete" cmd="/usr/local/bin/containerd-shim-aws-firecracker -namespace firecracker-containerd -address /run/firecracker-containerd/containerd.sock -publish-binary /usr/local/bin/firecracker-containerd -id 1 -bundle /run/firecracker-containerd/io.containerd.runtime.v2.task/firecracker-containerd/1 delete" error="exit status 1"
time="2022-05-21T12:35:42.642036283Z" level=warning msg="failed to clean up after shim disconnected" error="aws.firecracker: rpc error: code = DeadlineExceeded desc = timed out waiting for VM start\n: exit status 1" id=1 namespace=firecracker-containerd
time="2022-05-21T12:35:42.643382057Z" level=debug msg="remove snapshot" key=1
time="2022-05-21T12:35:42.643884527Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/remove type=containerd.events.SnapshotRemove
time="2022-05-21T12:35:42.644369685Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/delete type=containerd.events.ContainerDelete
time="2022-05-21T12:35:42.644636419Z" level=debug msg="stop VM: VMID:\"1\" "
time="2022-05-21T12:35:42.644896030Z" level=info msg="stopping the VM" runtime=aws.firecracker vmID=1
time="2022-05-21T12:35:42.645825477Z" level=error msg="failed to delete" cmd="/usr/local/bin/containerd-shim-aws-firecracker -namespace firecracker-containerd -address /run/firecracker-containerd/containerd.sock -publish-binary /usr/local/bin/firecracker-containerd -id 3 -bundle /run/firecracker-containerd/io.containerd.runtime.v2.task/firecracker-containerd/3 delete" error="exit status 1"
time="2022-05-21T12:35:42.645872436Z" level=warning msg="failed to clean up after shim disconnected" error="aws.firecracker: rpc error: code = DeadlineExceeded desc = timed out waiting for VM start\n: exit status 1" id=3 namespace=firecracker-containerd
time="2022-05-21T12:35:42.647077524Z" level=debug msg="remove snapshot" key=3
time="2022-05-21T12:35:42.647643103Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/remove type=containerd.events.SnapshotRemove
time="2022-05-21T12:35:42.648095619Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/delete type=containerd.events.ContainerDelete
time="2022-05-21T12:35:42.648346724Z" level=debug msg="stop VM: VMID:\"3\" "
time="2022-05-21T12:35:42.648660918Z" level=info msg="stopping the VM" runtime=aws.firecracker vmID=3
time="2022-05-21T12:35:42.679480228Z" level=debug msg="schedule snapshotter cleanup" snapshotter=devmapper
time="2022-05-21T12:35:42.679547655Z" level=debug msg=walk
time="2022-05-21T12:35:42.679798159Z" level=debug msg=remove key=firecracker-containerd/43/1
time="2022-05-21T12:35:42.781310628Z" level=debug msg="removed snapshot" key=firecracker-containerd/43/1 snapshotter=devmapper
time="2022-05-21T12:35:42.781358118Z" level=debug msg=remove key=firecracker-containerd/22/3
time="2022-05-21T12:35:42.896503142Z" level=debug msg="removed snapshot" key=firecracker-containerd/22/3 snapshotter=devmapper
time="2022-05-21T12:35:42.896522358Z" level=debug msg=cleanup
time="2022-05-21T12:35:42.896532037Z" level=debug msg="snapshot garbage collected" d=217.018606ms snapshotter=devmapper
time="2022-05-21T12:35:42.896558006Z" level=debug msg="garbage collected" d="837.133µs"
time="2022-05-21T12:35:43.038832794Z" level=info msg="firecracker exited: status=0" runtime=aws.firecracker
time="2022-05-21T12:35:43.039421898Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/stop type=VMStop
time="2022-05-21T12:35:43.094980772Z" level=info msg="firecracker exited: status=0" runtime=aws.firecracker
time="2022-05-21T12:35:43.095515483Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/stop type=VMStop
time="2022-05-21T12:35:44.499752285Z" level=error msg="aws.firecracker: publisher not closed" shim_stream=stderr vmID=2
time="2022-05-21T12:35:44.502671604Z" level=debug msg="shim has been terminated" error="exit status 1" vmID=2
time="2022-05-21T12:35:45.533993126Z" level=error msg="aws.firecracker: publisher not closed" shim_stream=stderr vmID=4
time="2022-05-21T12:35:45.536593061Z" level=debug msg="shim has been terminated" error="exit status 1" vmID=4
time="2022-05-21T12:35:45.595866359Z" level=error msg="aws.firecracker: publisher not closed" shim_stream=stderr vmID=5
time="2022-05-21T12:35:45.598999713Z" level=debug msg="shim has been terminated" error="exit status 1" vmID=5
time="2022-05-21T12:35:48.039466018Z" level=error msg="aws.firecracker: publisher not closed" shim_stream=stderr vmID=1
time="2022-05-21T12:35:48.042676518Z" level=debug msg="shim has been terminated" error="exit status 1" vmID=1
time="2022-05-21T12:35:48.095532793Z" level=error msg="aws.firecracker: publisher not closed" shim_stream=stderr vmID=3
time="2022-05-21T12:35:48.101212842Z" level=debug msg="shim has been terminated" error="exit status 1" vmID=3
time="2022-05-21T12:45:33.967395092Z" level=info msg=exited ExecID= TaskID=6 exit_status=137 exited_at="2022-05-21 12:45:33.95509659 +0000 UTC" runtime=aws.firecracker vmID=6
time="2022-05-21T12:45:33.967606401Z" level=info msg="connection was closed: read /proc/self/fd/14: file already closed" ExecID= TaskID=6 runtime=aws.firecracker stream=stdin vmID=6
time="2022-05-21T12:45:33.967681954Z" level=error msg="error closing io stream" ExecID= TaskID=6 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdin vmID=6
time="2022-05-21T12:45:33.967674190Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/exit type=containerd.events.TaskExit
time="2022-05-21T12:45:34.475414760Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/tasks/delete type=containerd.events.TaskDelete
time="2022-05-21T12:45:34.487400333Z" level=info msg="PatchGuestDrive successful" runtime=aws.firecracker
time="2022-05-21T12:45:34.487860152Z" level=info msg="shim disconnected" id=6
time="2022-05-21T12:45:34.487923412Z" level=warning msg="cleaning up after shim disconnected" id=6 namespace=firecracker-containerd
time="2022-05-21T12:45:34.487937459Z" level=info msg="cleaning up dead shim"
time="2022-05-21T12:45:38.967669723Z" level=error msg="error closing io stream" ExecID= TaskID=6 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdout vmID=6
time="2022-05-21T12:45:38.967669984Z" level=error msg="error closing io stream" ExecID= TaskID=6 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stderr vmID=6
time="2022-05-21T12:45:39.537759581Z" level=error msg="failed to delete" cmd="/usr/local/bin/containerd-shim-aws-firecracker -namespace firecracker-containerd -address /run/firecracker-containerd/containerd.sock -publish-binary /usr/local/bin/firecracker-containerd -id 6 -bundle /run/firecracker-containerd/io.containerd.runtime.v2.task/firecracker-containerd/6 delete" error="exit status 1"
time="2022-05-21T12:45:39.537904966Z" level=warning msg="failed to clean up after shim disconnected" error="aws.firecracker: rpc error: code = DeadlineExceeded desc = timed out waiting for VM start\n: exit status 1" id=6 namespace=firecracker-containerd
time="2022-05-21T12:45:39.539368444Z" level=debug msg="remove snapshot" key=6
time="2022-05-21T12:45:39.539953000Z" level=debug msg="event published" ns=firecracker-containerd topic=/snapshot/remove type=containerd.events.SnapshotRemove
time="2022-05-21T12:45:39.540717957Z" level=debug msg="event published" ns=firecracker-containerd topic=/containers/delete type=containerd.events.ContainerDelete
time="2022-05-21T12:45:39.541061777Z" level=debug msg="stop VM: VMID:\"6\" "
time="2022-05-21T12:45:39.541464309Z" level=info msg="stopping the VM" runtime=aws.firecracker vmID=6
time="2022-05-21T12:45:39.566609475Z" level=debug msg="schedule snapshotter cleanup" snapshotter=devmapper
time="2022-05-21T12:45:39.566676362Z" level=debug msg=walk
time="2022-05-21T12:45:39.566867153Z" level=debug msg=remove key=firecracker-containerd/44/6
time="2022-05-21T12:45:39.665208864Z" level=debug msg="removed snapshot" key=firecracker-containerd/44/6 snapshotter=devmapper
time="2022-05-21T12:45:39.665227930Z" level=debug msg=cleanup
time="2022-05-21T12:45:39.665235825Z" level=debug msg="snapshot garbage collected" d=98.590682ms snapshotter=devmapper
time="2022-05-21T12:45:39.665255131Z" level=debug msg="garbage collected" d="699.814µs"
time="2022-05-21T12:45:39.949359456Z" level=info msg="firecracker exited: status=0" runtime=aws.firecracker
time="2022-05-21T12:45:39.950229812Z" level=debug msg="event forwarded" ns=firecracker-containerd topic=/firecracker-vm/stop type=VMStop
time="2022-05-21T12:45:44.950425067Z" level=error msg="aws.firecracker: publisher not closed" shim_stream=stderr vmID=6
time="2022-05-21T12:45:44.953747532Z" level=debug msg="shim has been terminated" error="exit status 1" vmID=6
the logs don't show any problems. what does kubectl get pods
show? please reproduce the issue with just helloworld function
Hi @ustiugov, the output of kubectl get pods
is No resources found in default namespace.
The output of kubectl get pods -A
is:
NAMESPACE NAME READY STATUS RESTARTS AGE
istio-system cluster-local-gateway-74c4558686-7w54r 1/1 Running 0 15m
istio-system istio-ingressgateway-f5b59cc7c-jb6dm 1/1 Running 0 15m
istio-system istiod-54bbfb4d85-4ldbk 1/1 Running 0 15m
knative-eventing eventing-controller-59475d565c-79xg9 1/1 Running 0 15m
knative-eventing eventing-webhook-74cbb75cb-nrrx6 1/1 Running 0 15m
knative-eventing imc-controller-84c7f75c67-45jg6 1/1 Running 0 15m
knative-eventing imc-dispatcher-7786967556-ld7rw 1/1 Running 0 15m
knative-eventing mt-broker-controller-65bb965bf9-64fq6 1/1 Running 0 15m
knative-eventing mt-broker-filter-8496c9765-w925x 1/1 Running 0 15m
knative-eventing mt-broker-ingress-67959dc68f-4fp5x 1/1 Running 0 15m
knative-serving activator-7f7865c9f5-7f8mf 1/1 Running 0 15m
knative-serving autoscaler-5f795f4cb7-ljpn6 1/1 Running 0 15m
knative-serving controller-5b7545f6f5-8sd7f 1/1 Running 0 15m
knative-serving default-domain-m5jmz 1/1 Running 0 15m
knative-serving domain-mapping-9f9784f9b-pwhlb 1/1 Running 0 15m
knative-serving domainmapping-webhook-67896589f6-m6c4q 1/1 Running 0 15m
knative-serving net-istio-controller-6b84bc75d6-nxszc 1/1 Running 0 15m
knative-serving net-istio-webhook-f96dbffb4-j7w5q 1/1 Running 0 15m
knative-serving webhook-557f4b554d-xrmpp 1/1 Running 0 15m
kube-system calico-kube-controllers-644b84fc59-57d2z 1/1 Running 0 15m
kube-system canal-tjcsv 2/2 Running 0 15m
kube-system coredns-64897985d-mmpp7 1/1 Running 0 15m
kube-system coredns-64897985d-n4qzt 1/1 Running 0 15m
kube-system etcd 1/1 Running 2 16m
kube-system kube-apiserver 1/1 Running 2 16m
kube-system kube-controller-manager 1/1 Running 16 15m
kube-system kube-proxy-zx2tn 1/1 Running 0 15m
kube-system kube-scheduler 1/1 Running 29 15m
metallb-system controller-557988499-fgdhz 1/1 Running 0 15m
metallb-system speaker-n8cqs 1/1 Running 0 15m
registry docker-registry-pod-p4b6h 1/1 Running 0 15m
registry registry-etc-hosts-update-7s4hx 1/1 Running 0 15m
kubectl get pods -A
shows no deployed functions.
Yes, I presume that's because of no ready revisions. But I am not sure about why that it happening.
Output of kn services list
:
NAME URL LATEST AGE CONDITIONS READY REASON
helloworld-0 http://helloworld-0.default.example.com 19m 0 OK / 3 False RevisionMissing : Configuration "helloworld-0" does not have any ready Revision.
Output of kubectl get revisions
NAME CONFIG NAME K8S SERVICE NAME GENERATION READY REASON ACTUAL REPLICAS DESIRED REPLICAS
helloworld-0-00001 helloworld-0 1 False ProgressDeadlineExceeded 0
Output of kubectl describe revision/helloworld-0-00001
Name: helloworld-0-00001
Namespace: default
Labels: serving.knative.dev/configuration=helloworld-0
serving.knative.dev/configurationGeneration=1
serving.knative.dev/configurationUID=1cbbfe51-92c5-4cf0-a91f-a5d0fb28859c
serving.knative.dev/routingState=active
serving.knative.dev/service=helloworld-0
serving.knative.dev/serviceUID=20b5d456-e512-4c66-8a16-b2509f56e7b7
Annotations: autoscaling.knative.dev/target: 1
serving.knative.dev/creator: kubernetes-admin
serving.knative.dev/routes: helloworld-0
serving.knative.dev/routingStateModified: 2022-05-21T13:36:46Z
API Version: serving.knative.dev/v1
Kind: Revision
Metadata:
Creation Timestamp: 2022-05-21T13:36:46Z
Generation: 1
Managed Fields:
API Version: serving.knative.dev/v1
Fields Type: FieldsV1
fieldsV1:
f:metadata:
f:annotations:
.:
f:autoscaling.knative.dev/target:
f:serving.knative.dev/creator:
f:serving.knative.dev/routes:
f:serving.knative.dev/routingStateModified:
f:labels:
.:
f:serving.knative.dev/configuration:
f:serving.knative.dev/configurationGeneration:
f:serving.knative.dev/configurationUID:
f:serving.knative.dev/routingState:
f:serving.knative.dev/service:
f:serving.knative.dev/serviceUID:
f:ownerReferences:
.:
k:{"uid":"1cbbfe51-92c5-4cf0-a91f-a5d0fb28859c"}:
f:spec:
.:
f:containerConcurrency:
f:containers:
f:enableServiceLinks:
f:timeoutSeconds:
Manager: Go-http-client
Operation: Update
Time: 2022-05-21T13:36:46Z
API Version: serving.knative.dev/v1
Fields Type: FieldsV1
fieldsV1:
f:status:
.:
f:actualReplicas:
f:conditions:
f:containerStatuses:
f:observedGeneration:
Manager: Go-http-client
Operation: Update
Subresource: status
Time: 2022-05-21T13:47:16Z
Owner References:
API Version: serving.knative.dev/v1
Block Owner Deletion: true
Controller: true
Kind: Configuration
Name: helloworld-0
UID: 1cbbfe51-92c5-4cf0-a91f-a5d0fb28859c
Resource Version: 7100
UID: 608c7b88-edc1-4492-9bf8-fe231d47f7ad
Spec:
Container Concurrency: 0
Containers:
Env:
Name: GUEST_PORT
Value: 50051
Name: GUEST_IMAGE
Value: ghcr.io/ease-lab/helloworld:var_workload
Image: crccheck/hello-world:latest
Name: user-container
Ports:
Container Port: 50051
Name: h2c
Protocol: TCP
Readiness Probe:
Success Threshold: 1
Tcp Socket:
Port: 0
Resources:
Enable Service Links: false
Timeout Seconds: 300
Status:
Actual Replicas: 0
Conditions:
Last Transition Time: 2022-05-21T13:47:16Z
Message: The target is not receiving traffic.
Reason: NoTraffic
Severity: Info
Status: False
Type: Active
Last Transition Time: 2022-05-21T13:46:47Z
Message: Container failed with: container exited with no error
Reason: ExitCode0
Status: False
Type: ContainerHealthy
Last Transition Time: 2022-05-21T13:47:16Z
Message: Initial scale was never achieved
Reason: ProgressDeadlineExceeded
Status: False
Type: Ready
Last Transition Time: 2022-05-21T13:47:16Z
Message: Initial scale was never achieved
Reason: ProgressDeadlineExceeded
Status: False
Type: ResourcesAvailable
Container Statuses:
Name: user-container
Observed Generation: 1
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning InternalError 8m18s revision-controller failed to update deployment "helloworld-0-00001-deployment": Operation cannot be fulfilled on deployments.apps "helloworld-0-00001-deployment": the object has been modified; please apply your changes to the latest version and try again
after you started using the other branch, have you started with a new clean node or kept using the old one?
I have been using the old node. But I have cleared all previous files (starting with an empty filesystem), and cloned the new branch, then started the process.
I suggest using a fresh node
Sure, I'll try that. That may take a few days however. Do you propose any way to clean up the current node in a way to use it for the fresh branch ?
Hi @ustiugov, I tried using a fresh AWS ec2 instance running AMD and Ubuntu 20.04 for this. I used the new branch (#465) and also applied the change of (#481) locally. However, I ran into the exactly the same issue once again (same output for kubectl get pods
, kubectl get revisions
etc).
vhive.stdout
time="2022-05-22T06:48:17.906469735Z" level=warning msg="Using google dns 8.8.8.8\n"
time="2022-05-22T06:48:18.182771404Z" level=error msg="VM config for pod a816c49fc0c057c63746107ce10b2a119973b3536f6dd4c0bfe0d29d5fac762e does not exist"
time="2022-05-22T06:48:18.182800714Z" level=error error="VM config for pod does not exist"
time="2022-05-22T06:48:18.198215955Z" level=error msg="coordinator failed to start VM" error="failed to create the microVM in firecracker-containerd: rpc error: code = Unkno
wn desc = failed to create VM: failed to start the VM: Put \"http://localhost/actions\": EOF" image="vhiveease/rnn_serving:var_workload" vmID=165
time="2022-05-22T06:48:18.198257295Z" level=error msg="failed to start VM" error="failed to create the microVM in firecracker-containerd: rpc error: code = Unknown desc = fa
iled to create VM: failed to start the VM: Put \"http://localhost/actions\": EOF"
time="2022-05-22T06:48:18.204791658Z" level=error msg="VM config for pod 2c9008a36f3f462efaebef6e22905a6fb1d7fc21f52b67124c2e43bb410c1c33 does not exist"
time="2022-05-22T06:48:18.204810928Z" level=error error="VM config for pod does not exist"
time="2022-05-22T06:48:28.178727997Z" level=error msg="VM config for pod 93d79475aa67f0fc707f0b5e554185f8a1bc804b47a7427ba590c6e295b319a1 does not exist"
time="2022-05-22T06:48:28.178749627Z" level=error error="VM config for pod does not exist"
time="2022-05-22T06:48:29.178473676Z" level=error msg="VM config for pod dd3b7e9755d55deadd08aab43b9ebc631c9694530a237adf53446796d6a13a91 does not exist"
time="2022-05-22T06:48:29.178500227Z" level=error error="VM config for pod does not exist"
time="2022-05-22T06:48:29.949906752Z" level=warning msg="Failed to Fetch k8s dns clusterIP exit status 1\nThe connection to the server localhost:8080 was refused - did you s
pecify the right host or port?\n\n"
time="2022-05-22T06:48:29.949939242Z" level=warning msg="Using google dns 8.8.8.8\n"
time="2022-05-22T06:48:30.180305088Z" level=error msg="VM config for pod a816c49fc0c057c63746107ce10b2a119973b3536f6dd4c0bfe0d29d5fac762e does not exist"
time="2022-05-22T06:48:30.180337768Z" level=error error="VM config for pod does not exist"
time="2022-05-22T06:48:30.258215038Z" level=error msg="coordinator failed to start VM" error="failed to create the microVM in firecracker-containerd: rpc error: code = Unkno
wn desc = failed to create VM: failed to start the VM: Put \"http://localhost/actions\": EOF" image="ghcr.io/ease-lab/pyaes:var_workload" vmID=166
time="2022-05-22T06:48:30.258258959Z" level=error msg="failed to start VM" error="failed to create the microVM in firecracker-containerd: rpc error: code = Unknown desc = fa
iled to create VM: failed to start the VM: Put \"http://localhost/actions\": EOF"
time="2022-05-22T06:48:30.260811597Z" level=error msg="VM config for pod 8b136b626a162d8954c178f9052b9cd164c69754058388eb9eec8cc5d656ac30 does not exist"
time="2022-05-22T06:48:30.260828387Z" level=error error="VM config for pod does not exist"
time="2022-05-22T06:48:33.183440545Z" level=error msg="VM config for pod 2c9008a36f3f462efaebef6e22905a6fb1d7fc21f52b67124c2e43bb410c1c33 does not exist"
time="2022-05-22T06:48:33.183504665Z" level=error error="VM config for pod does not exist"
time="2022-05-22T06:48:37.477829630Z" level=info msg="HEARTBEAT: number of active VMs: 0"
time="2022-05-22T06:48:37.486898090Z" level=info msg="FuncPool heartbeat: ==== Stats by cold functions ====\nfID, #started, #served\n==================================="
time="2022-05-22T06:48:39.880526375Z" level=warning msg="Failed to Fetch k8s dns clusterIP exit status 1\nThe connection to the server localhost:8080 was refused - did you s
pecify the right host or port?\n\n"
time="2022-05-22T06:48:39.880558325Z" level=warning msg="Using google dns 8.8.8.8\n"
time="2022-05-22T06:48:40.182251734Z" level=error msg="coordinator failed to start VM" error="failed to create the microVM in firecracker-containerd: rpc error: code = Unkno
wn desc = failed to create VM: failed to start the VM: Put \"http://localhost/actions\": EOF" image="ghcr.io/ease-lab/helloworld:var_workload" vmID=167
time="2022-05-22T06:48:40.182304234Z" level=error msg="failed to start VM" error="failed to create the microVM in firecracker-containerd: rpc error: code = Unknown desc = fa
iled to create VM: failed to start the VM: Put \"http://localhost/actions\": EOF"
time="2022-05-22T06:48:40.185197387Z" level=error msg="VM config for pod 93d79475aa67f0fc707f0b5e554185f8a1bc804b47a7427ba590c6e295b319a1 does not exist"
time="2022-05-22T06:48:40.185220167Z" level=error error="VM config for pod does not exist"
firecracker.stderr
M(), setting up a VMM on firecracker.sock" runtime=aws.firecracker
time="2022-05-22T06:49:32.970316862Z" level=info msg="refreshMachineConfiguration: [GET /machine-config][200] getMachineConfigurationOK &{CPUTemplate: HtEnabled:0xc00053e5e
3 MemSizeMib:0xc00053e588 TrackDirtyPages:false VcpuCount:0xc00053e580}" runtime=aws.firecracker
time="2022-05-22T06:49:32.970531724Z" level=info msg="PutGuestBootSource: [PUT /boot-source][204] putGuestBootSourceNoContent " runtime=aws.firecracker
time="2022-05-22T06:49:32.970552044Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/runtime/default-rootfs.img, slot root_drive, root true." runtime=aws.fi
recracker
time="2022-05-22T06:49:32.970900508Z" level=info msg="Attached drive /var/lib/firecracker-containerd/runtime/default-rootfs.img: [PUT /drives/{drive_id}][204] putGuestDriveB
yIdNoContent " runtime=aws.firecracker
time="2022-05-22T06:49:32.970920308Z" level=info msg="Attaching drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#171/ctrstub0, slot MN2HE43UOVRDA, root
false." runtime=aws.firecracker
time="2022-05-22T06:49:32.971099140Z" level=info msg="Attached drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd#171/ctrstub0: [PUT /drives/{drive_id}][
204] putGuestDriveByIdNoContent " runtime=aws.firecracker
time="2022-05-22T06:49:32.971116620Z" level=info msg="Attaching NIC 171_tap (hwaddr 02:FC:00:00:00:AA) at index 1" runtime=aws.firecracker
time="2022-05-22T06:49:33.156858045Z" level=error msg="Starting instance: Put \"http://localhost/actions\": EOF" runtime=aws.firecracker
time="2022-05-22T06:49:33.156946386Z" level=error msg="failed to create VM" error="failed to start the VM: Put \"http://localhost/actions\": EOF" runtime=aws.firecracker vmI
D=171
time="2022-05-22T06:49:33.157153618Z" level=warning msg="firecracker exited: signal: aborted (core dumped)" runtime=aws.firecracker
time="2022-05-22T06:49:33.162219583Z" level=error msg="shim CreateVM returned error" error="rpc error: code = Unknown desc = failed to create VM: failed to start the VM: Put
\"http://localhost/actions\": EOF"
time="2022-05-22T06:49:38.157281139Z" level=error msg="aws.firecracker: publisher not closed" shim_stream=stderr vmID=171
time="2022-05-22T06:49:38.158392301Z" level=debug msg="shim has been terminated" error="exit status 1" vmID=171
Let me know if you observe something, or need more detailed logs.
Note All this is one a single-node cluster
Hi @ustiugov, I managed to get the problem fixed by setting up on a new machine, and by enabling KVM, and ensuring that this script worked okay.
The functions are now getting deployed properly now. However, I am having an issue with the istio set-up. This is similar to #475.
NAMESPACE NAME READY STATUS RESTARTS AGE
istio-system cluster-local-gateway-74c4558686-8g9zs 0/1 CrashLoopBackOff 41 (4m34s ago) 3h9m
istio-system istio-ingressgateway-f5b59cc7c-qqgrr 0/1 CrashLoopBackOff 41 (4m52s ago) 3h9m
Output of kubectl describe pod cluster-local-gateway-74c4558686-8g9zs -n istio-system
Name: cluster-local-gateway-74c4558686-8g9zs
Namespace: istio-system
Priority: 0
Start Time: Mon, 23 May 2022 13:50:03 +0000
Labels: app=cluster-local-gateway
chart=gateways
heritage=Tiller
install.operator.istio.io/owning-resource=unknown
istio=cluster-local-gateway
istio.io/rev=default
operator.istio.io/component=IngressGateways
pod-template-hash=74c4558686
release=istio
service.istio.io/canonical-name=cluster-local-gateway
service.istio.io/canonical-revision=latest
sidecar.istio.io/inject=false
Annotations: cni.projectcalico.org/podIP: 192.168.0.34/32
cni.projectcalico.org/podIPs: 192.168.0.34/32
prometheus.io/path: /stats/prometheus
prometheus.io/port: 15020
prometheus.io/scrape: true
sidecar.istio.io/inject: false
Status: Running
IP: 192.168.0.34
IPs:
IP: 192.168.0.34
Controlled By: ReplicaSet/cluster-local-gateway-74c4558686
Containers:
istio-proxy:
Container ID: containerd://079fe320e2704ab386383d25917b927a84a58301a26c76dc02bc09c5c3be988a
Image: docker.io/istio/proxyv2:1.12.5
Image ID: docker.io/istio/proxyv2@sha256:780f49744311374e0905e5d15a4bd251bbc48284cb653ca9d609ac3894558462
Ports: 15020/TCP, 8080/TCP, 8443/TCP, 15090/TCP
Host Ports: 0/TCP, 0/TCP, 0/TCP, 0/TCP
Args:
proxy
router
--domain
$(POD_NAMESPACE).svc.cluster.local
--proxyLogLevel=warning
--proxyComponentLogLevel=misc:error
--log_output_level=default:info
State: Waiting
Reason: CrashLoopBackOff
Last State: Terminated
Reason: Completed
Exit Code: 0
Started: Mon, 23 May 2022 17:00:15 +0000
Finished: Mon, 23 May 2022 17:00:16 +0000
Ready: False
Restart Count: 42
Limits:
cpu: 2
memory: 1Gi
Requests:
cpu: 100m
memory: 128Mi
Readiness: http-get http://:15021/healthz/ready delay=1s timeout=1s period=2s #success=1 #failure=30
Environment:
JWT_POLICY: first-party-jwt
PILOT_CERT_PROVIDER: istiod
CA_ADDR: istiod.istio-system.svc:15012
NODE_NAME: (v1:spec.nodeName)
POD_NAME: cluster-local-gateway-74c4558686-8g9zs (v1:metadata.name)
POD_NAMESPACE: istio-system (v1:metadata.namespace)
INSTANCE_IP: (v1:status.podIP)
HOST_IP: (v1:status.hostIP)
SERVICE_ACCOUNT: (v1:spec.serviceAccountName)
ISTIO_META_WORKLOAD_NAME: cluster-local-gateway
ISTIO_META_OWNER: kubernetes://apis/apps/v1/namespaces/istio-system/deployments/cluster-local-gateway
ISTIO_META_MESH_ID: cluster.local
TRUST_DOMAIN: cluster.local
ISTIO_META_UNPRIVILEGED_POD: true
ISTIO_META_CLUSTER_ID: Kubernetes
Mounts:
/etc/istio/config from config-volume (rw)
/etc/istio/ingressgateway-ca-certs from ingressgateway-ca-certs (ro)
/etc/istio/ingressgateway-certs from ingressgateway-certs (ro)
/etc/istio/pod from podinfo (rw)
/etc/istio/proxy from istio-envoy (rw)
/var/lib/istio/data from istio-data (rw)
/var/run/secrets/istio from istiod-ca-cert (rw)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-tcr6n (ro)
Conditions:
Type Status
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
istiod-ca-cert:
Type: ConfigMap (a volume populated by a ConfigMap)
Name: istio-ca-root-cert
Optional: false
podinfo:
Type: DownwardAPI (a volume populated by information about the pod)
Items:
metadata.labels -> labels
metadata.annotations -> annotations
istio-envoy:
Type: EmptyDir (a temporary directory that shares a pod's lifetime)
Medium:
SizeLimit: <unset>
istio-data:
Type: EmptyDir (a temporary directory that shares a pod's lifetime)
Medium:
SizeLimit: <unset>
config-volume:
Type: ConfigMap (a volume populated by a ConfigMap)
Name: istio
Optional: true
ingressgateway-certs:
Type: Secret (a volume populated by a Secret)
SecretName: istio-ingressgateway-certs
Optional: true
ingressgateway-ca-certs:
Type: Secret (a volume populated by a Secret)
SecretName: istio-ingressgateway-ca-certs
Optional: true
kube-api-access-tcr6n:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
QoS Class: Burstable
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning BackOff 12m (x865 over 3h12m) kubelet Back-off restarting failed container
Normal Pulled 2m40s (x43 over 3h12m) kubelet Container image "docker.io/istio/proxyv2:1.12.5" already present on machine
Output of kubectl logs cluster-local-gateway-74c4558686-8g9zs -n istio-system
2022-05-23T17:00:15.709727Z info FLAG: --concurrency="0"
2022-05-23T17:00:15.709824Z info FLAG: --domain="istio-system.svc.cluster.local"
2022-05-23T17:00:15.709833Z info FLAG: --help="false"
2022-05-23T17:00:15.709838Z info FLAG: --log_as_json="false"
2022-05-23T17:00:15.709842Z info FLAG: --log_caller=""
2022-05-23T17:00:15.709846Z info FLAG: --log_output_level="default:info"
2022-05-23T17:00:15.709850Z info FLAG: --log_rotate=""
2022-05-23T17:00:15.709853Z info FLAG: --log_rotate_max_age="30"
2022-05-23T17:00:15.709857Z info FLAG: --log_rotate_max_backups="1000"
2022-05-23T17:00:15.709861Z info FLAG: --log_rotate_max_size="104857600"
2022-05-23T17:00:15.709866Z info FLAG: --log_stacktrace_level="default:none"
2022-05-23T17:00:15.709873Z info FLAG: --log_target="[stdout]"
2022-05-23T17:00:15.709878Z info FLAG: --meshConfig="./etc/istio/config/mesh"
2022-05-23T17:00:15.709882Z info FLAG: --outlierLogPath=""
2022-05-23T17:00:15.709886Z info FLAG: --proxyComponentLogLevel="misc:error"
2022-05-23T17:00:15.709890Z info FLAG: --proxyLogLevel="warning"
2022-05-23T17:00:15.709895Z info FLAG: --serviceCluster="istio-proxy"
2022-05-23T17:00:15.709899Z info FLAG: --stsPort="0"
2022-05-23T17:00:15.709903Z info FLAG: --templateFile=""
2022-05-23T17:00:15.709908Z info FLAG: --tokenManagerPlugin="GoogleTokenExchange"
2022-05-23T17:00:15.709913Z info FLAG: --vklog="0"
2022-05-23T17:00:15.709918Z info Version 1.12.5-6332f0901f96ca97cf114d57b466d4bcd055b08c-Clean
2022-05-23T17:00:15.710614Z info Proxy role ips=[192.168.0.34 fe80::82b:46ff:feca:51d3] type=router id=cluster-local-gateway-74c4558686-8g9zs.istio-system domain=istio-system.svc.cluster.local
2022-05-23T17:00:15.710740Z info Apply mesh config from file defaultConfig:
discoveryAddress: istiod.istio-system.svc:15012
proxyMetadata: {}
tracing:
zipkin:
address: zipkin.istio-system:9411
enablePrometheusMerge: true
rootNamespace: istio-system
trustDomain: cluster.local
2022-05-23T17:00:15.712776Z info Effective config: binaryPath: /usr/local/bin/envoy
configPath: ./etc/istio/proxy
controlPlaneAuthPolicy: MUTUAL_TLS
discoveryAddress: istiod.istio-system.svc:15012
drainDuration: 45s
parentShutdownDuration: 60s
proxyAdminPort: 15000
proxyMetadata: {}
serviceCluster: istio-proxy
statNameLength: 189
statusPort: 15020
terminationDrainDuration: 5s
tracing:
zipkin:
address: zipkin.istio-system:9411
2022-05-23T17:00:15.712806Z info JWT policy is first-party-jwt
2022-05-23T17:00:15.718815Z info CA Endpoint istiod.istio-system.svc:15012, provider Citadel
2022-05-23T17:00:15.718859Z info Opening status port 15020
2022-05-23T17:00:15.718911Z info Using CA istiod.istio-system.svc:15012 cert with certs: var/run/secrets/istio/root-cert.pem
2022-05-23T17:00:15.719073Z info citadelclient Citadel client using custom root cert: istiod.istio-system.svc:15012
2022-05-23T17:00:15.741484Z info ads All caches have been synced up in 35.750914ms, marking server ready
2022-05-23T17:00:15.741820Z info sds SDS server for workload certificates started, listening on "etc/istio/proxy/SDS"
2022-05-23T17:00:15.741850Z info xdsproxy Initializing with upstream address "istiod.istio-system.svc:15012" and cluster "Kubernetes"
2022-05-23T17:00:15.741944Z info sds Starting SDS grpc server
2022-05-23T17:00:15.742273Z info Pilot SAN: [istiod.istio-system.svc]
2022-05-23T17:00:15.742287Z info starting Http service at 127.0.0.1:15004
2022-05-23T17:00:15.743775Z info Pilot SAN: [istiod.istio-system.svc]
2022-05-23T17:00:15.745484Z info Starting proxy agent
2022-05-23T17:00:15.745533Z info Epoch 0 starting
2022-05-23T17:00:15.745555Z info Envoy command: [-c etc/istio/proxy/envoy-rev0.json --restart-epoch 0 --drain-time-s 45 --drain-strategy immediate --parent-shutdown-time-s 60 --local-address-ip-version v4 --file-flush-interval-msec 1000 --disable-hot-restart --log-format %Y-%m-%dT%T.%fZ %l envoy %n %v -l warning --component-log-level misc:error]
[warn] evutil_make_internal_pipe_: pipe: Too many open files
[warn] event_base_new_with_config: Unable to make base notifiable.
2022-05-23T17:00:15.959603Z critical envoy assert assert failure: event_base != nullptr. Details: Failed to initialize libevent event_base
2022-05-23T17:00:15.959717Z critical envoy backtrace Caught Aborted, suspect faulting address 0x5390000002c
2022-05-23T17:00:15.959725Z critical envoy backtrace Backtrace (use tools/stack_decode.py to get line numbers):
2022-05-23T17:00:15.959729Z critical envoy backtrace Envoy version: eb9f894bff4c135904eb83513c795db899c838d1/1.20.3-dev/Clean/RELEASE/BoringSSL
2022-05-23T17:00:15.959989Z critical envoy backtrace #0: __restore_rt [0x7f328064f3c0]
2022-05-23T17:00:15.968871Z critical envoy backtrace #1: Envoy::Event::DispatcherImpl::DispatcherImpl() [0x55f8ff655fc1]
2022-05-23T17:00:15.975578Z critical envoy backtrace #2: Envoy::Api::Impl::allocateDispatcher() [0x55f8ff07a740]
2022-05-23T17:00:15.976570Z info cache generated new workload certificate latency=234.644688ms ttl=23h59m59.023453911s
2022-05-23T17:00:15.976604Z info cache Root cert has changed, start rotating root cert
2022-05-23T17:00:15.976639Z info ads XDS: Incremental Pushing:0 ConnectedEndpoints:0 Version:
2022-05-23T17:00:15.976716Z info cache returned workload trust anchor from cache ttl=23h59m59.023288508s
2022-05-23T17:00:15.979510Z critical envoy backtrace #3: Envoy::Server::ProdWorkerFactory::createWorker() [0x55f8ff075566]
2022-05-23T17:00:15.983446Z critical envoy backtrace #4: Envoy::Server::ListenerManagerImpl::ListenerManagerImpl() [0x55f8ff36dd05]
2022-05-23T17:00:15.987376Z critical envoy backtrace #5: Envoy::Server::InstanceImpl::initialize() [0x55f8ff05bbda]
2022-05-23T17:00:15.991313Z critical envoy backtrace #6: Envoy::Server::InstanceImpl::InstanceImpl() [0x55f8ff0578e4]
2022-05-23T17:00:15.995219Z critical envoy backtrace #7: std::__1::make_unique<>() [0x55f8fd698af4]
2022-05-23T17:00:15.999153Z critical envoy backtrace #8: Envoy::MainCommonBase::MainCommonBase() [0x55f8fd697e59]
2022-05-23T17:00:16.003037Z critical envoy backtrace #9: Envoy::MainCommon::MainCommon() [0x55f8fd699507]
2022-05-23T17:00:16.006932Z critical envoy backtrace #10: Envoy::MainCommon::main() [0x55f8fd69969c]
2022-05-23T17:00:16.010816Z critical envoy backtrace #11: main [0x55f8fd69590c]
2022-05-23T17:00:16.010876Z critical envoy backtrace #12: __libc_start_main [0x7f328046d0b3]
2022-05-23T17:00:16.163848Z error Epoch 0 exited with error: signal: aborted (core dumped)
2022-05-23T17:00:16.163874Z info No more active epochs, terminating
Output of kubectl logs istio-ingressgateway-f5b59cc7c-qqgrr -n istio-system
2022-05-23T17:05:05.697445Z info FLAG: --concurrency="0"
2022-05-23T17:05:05.697522Z info FLAG: --domain="istio-system.svc.cluster.local"
2022-05-23T17:05:05.697529Z info FLAG: --help="false"
2022-05-23T17:05:05.697535Z info FLAG: --log_as_json="false"
2022-05-23T17:05:05.697539Z info FLAG: --log_caller=""
2022-05-23T17:05:05.697543Z info FLAG: --log_output_level="default:info"
2022-05-23T17:05:05.697547Z info FLAG: --log_rotate=""
2022-05-23T17:05:05.697554Z info FLAG: --log_rotate_max_age="30"
2022-05-23T17:05:05.697558Z info FLAG: --log_rotate_max_backups="1000"
2022-05-23T17:05:05.697562Z info FLAG: --log_rotate_max_size="104857600"
2022-05-23T17:05:05.697567Z info FLAG: --log_stacktrace_level="default:none"
2022-05-23T17:05:05.697581Z info FLAG: --log_target="[stdout]"
2022-05-23T17:05:05.697589Z info FLAG: --meshConfig="./etc/istio/config/mesh"
2022-05-23T17:05:05.697593Z info FLAG: --outlierLogPath=""
2022-05-23T17:05:05.697599Z info FLAG: --proxyComponentLogLevel="misc:error"
2022-05-23T17:05:05.697603Z info FLAG: --proxyLogLevel="warning"
2022-05-23T17:05:05.697609Z info FLAG: --serviceCluster="istio-proxy"
2022-05-23T17:05:05.697614Z info FLAG: --stsPort="0"
2022-05-23T17:05:05.697618Z info FLAG: --templateFile=""
2022-05-23T17:05:05.697623Z info FLAG: --tokenManagerPlugin="GoogleTokenExchange"
2022-05-23T17:05:05.697633Z info FLAG: --vklog="0"
2022-05-23T17:05:05.697638Z info Version 1.12.5-6332f0901f96ca97cf114d57b466d4bcd055b08c-Clean
2022-05-23T17:05:05.698147Z info Proxy role ips=[192.168.0.36 fe80::c0e4:b1ff:fe16:c34d] type=router id=istio-ingressgateway-f5b59cc7c-qqgrr.istio-system domain=istio-system.svc.cluster.local
2022-05-23T17:05:05.698236Z info Apply mesh config from file defaultConfig:
discoveryAddress: istiod.istio-system.svc:15012
proxyMetadata: {}
tracing:
zipkin:
address: zipkin.istio-system:9411
enablePrometheusMerge: true
rootNamespace: istio-system
trustDomain: cluster.local
2022-05-23T17:05:05.700145Z info Effective config: binaryPath: /usr/local/bin/envoy
configPath: ./etc/istio/proxy
controlPlaneAuthPolicy: MUTUAL_TLS
discoveryAddress: istiod.istio-system.svc:15012
drainDuration: 45s
parentShutdownDuration: 60s
proxyAdminPort: 15000
proxyMetadata: {}
serviceCluster: istio-proxy
statNameLength: 189
statusPort: 15020
terminationDrainDuration: 5s
tracing:
zipkin:
address: zipkin.istio-system:9411
2022-05-23T17:05:05.700165Z info JWT policy is first-party-jwt
2022-05-23T17:05:05.709326Z info CA Endpoint istiod.istio-system.svc:15012, provider Citadel
2022-05-23T17:05:05.709363Z info Using CA istiod.istio-system.svc:15012 cert with certs: var/run/secrets/istio/root-cert.pem
2022-05-23T17:05:05.709380Z info Opening status port 15020
2022-05-23T17:05:05.709476Z info citadelclient Citadel client using custom root cert: istiod.istio-system.svc:15012
2022-05-23T17:05:05.731186Z info ads All caches have been synced up in 37.532188ms, marking server ready
2022-05-23T17:05:05.731633Z info sds SDS server for workload certificates started, listening on "etc/istio/proxy/SDS"
2022-05-23T17:05:05.731660Z info xdsproxy Initializing with upstream address "istiod.istio-system.svc:15012" and cluster "Kubernetes"
2022-05-23T17:05:05.731736Z info sds Starting SDS grpc server
2022-05-23T17:05:05.732257Z info Pilot SAN: [istiod.istio-system.svc]
2022-05-23T17:05:05.732359Z info starting Http service at 127.0.0.1:15004
2022-05-23T17:05:05.733798Z info Pilot SAN: [istiod.istio-system.svc]
2022-05-23T17:05:05.735114Z info Starting proxy agent
2022-05-23T17:05:05.735138Z info Epoch 0 starting
2022-05-23T17:05:05.735154Z info Envoy command: [-c etc/istio/proxy/envoy-rev0.json --restart-epoch 0 --drain-time-s 45 --drain-strategy immediate --parent-shutdown-time-s 60 --local-address-ip-version v4 --file-flush-interval-msec 1000 --disable-hot-restart --log-format %Y-%m-%dT%T.%fZ %l envoy %n %v -l warning --component-log-level misc:error]
[warn] evutil_make_internal_pipe_: pipe: Too many open files
[warn] event_base_new_with_config: Unable to make base notifiable.
2022-05-23T17:05:05.923436Z critical envoy assert assert failure: event_base != nullptr. Details: Failed to initialize libevent event_base
2022-05-23T17:05:05.923489Z critical envoy backtrace Caught Aborted, suspect faulting address 0x53900000027
2022-05-23T17:05:05.923531Z critical envoy backtrace Backtrace (use tools/stack_decode.py to get line numbers):
2022-05-23T17:05:05.923535Z critical envoy backtrace Envoy version: eb9f894bff4c135904eb83513c795db899c838d1/1.20.3-dev/Clean/RELEASE/BoringSSL
2022-05-23T17:05:05.923785Z critical envoy backtrace #0: __restore_rt [0x7eff08eaf3c0]
2022-05-23T17:05:05.932570Z critical envoy backtrace #1: Envoy::Event::DispatcherImpl::DispatcherImpl() [0x55f9fadb0fc1]
2022-05-23T17:05:05.939399Z critical envoy backtrace #2: Envoy::Api::Impl::allocateDispatcher() [0x55f9fa7d5740]
2022-05-23T17:05:05.943302Z critical envoy backtrace #3: Envoy::Server::ProdWorkerFactory::createWorker() [0x55f9fa7d0566]
2022-05-23T17:05:05.947208Z critical envoy backtrace #4: Envoy::Server::ListenerManagerImpl::ListenerManagerImpl() [0x55f9faac8d05]
2022-05-23T17:05:05.951100Z critical envoy backtrace #5: Envoy::Server::InstanceImpl::initialize() [0x55f9fa7b6bda]
2022-05-23T17:05:05.954994Z critical envoy backtrace #6: Envoy::Server::InstanceImpl::InstanceImpl() [0x55f9fa7b28e4]
2022-05-23T17:05:05.958846Z critical envoy backtrace #7: std::__1::make_unique<>() [0x55f9f8df3af4]
2022-05-23T17:05:05.962695Z critical envoy backtrace #8: Envoy::MainCommonBase::MainCommonBase() [0x55f9f8df2e59]
2022-05-23T17:05:05.966553Z critical envoy backtrace #9: Envoy::MainCommon::MainCommon() [0x55f9f8df4507]
2022-05-23T17:05:05.970403Z critical envoy backtrace #10: Envoy::MainCommon::main() [0x55f9f8df469c]
2022-05-23T17:05:05.974252Z critical envoy backtrace #11: main [0x55f9f8df090c]
2022-05-23T17:05:05.974308Z critical envoy backtrace #12: __libc_start_main [0x7eff08ccd0b3]
2022-05-23T17:05:06.128876Z error Epoch 0 exited with error: signal: aborted (core dumped)
2022-05-23T17:05:06.129021Z info No more active epochs, terminating
Output of kubectl describe pod istio-ingressgateway-f5b59cc7c-qqgrr -n istio-system
Name: istio-ingressgateway-f5b59cc7c-qqgrr
Namespace: istio-system
Priority: 0
Start Time: Mon, 23 May 2022 13:50:03 +0000
Labels: app=istio-ingressgateway
chart=gateways
heritage=Tiller
install.operator.istio.io/owning-resource=unknown
istio=ingressgateway
istio.io/rev=default
operator.istio.io/component=IngressGateways
pod-template-hash=f5b59cc7c
release=istio
service.istio.io/canonical-name=istio-ingressgateway
service.istio.io/canonical-revision=latest
sidecar.istio.io/inject=false
Annotations: cni.projectcalico.org/podIP: 192.168.0.36/32
cni.projectcalico.org/podIPs: 192.168.0.36/32
prometheus.io/path: /stats/prometheus
prometheus.io/port: 15020
prometheus.io/scrape: true
sidecar.istio.io/inject: false
Status: Running
IP: 192.168.0.36
IPs:
IP: 192.168.0.36
Controlled By: ReplicaSet/istio-ingressgateway-f5b59cc7c
Containers:
istio-proxy:
Container ID: containerd://d64d7ee4594da3fe65ba4775e590f5244323eda1cbfc839cfcda93e6935e057e
Image: docker.io/istio/proxyv2:1.12.5
Image ID: docker.io/istio/proxyv2@sha256:780f49744311374e0905e5d15a4bd251bbc48284cb653ca9d609ac3894558462
Ports: 15021/TCP, 8080/TCP, 8443/TCP, 15090/TCP
Host Ports: 0/TCP, 0/TCP, 0/TCP, 0/TCP
Args:
proxy
router
--domain
$(POD_NAMESPACE).svc.cluster.local
--proxyLogLevel=warning
--proxyComponentLogLevel=misc:error
--log_output_level=default:info
State: Waiting
Reason: CrashLoopBackOff
Last State: Terminated
Reason: Completed
Exit Code: 0
Started: Mon, 23 May 2022 17:05:05 +0000
Finished: Mon, 23 May 2022 17:05:06 +0000
Ready: False
Restart Count: 43
Limits:
cpu: 2
memory: 1Gi
Requests:
cpu: 100m
memory: 128Mi
Readiness: http-get http://:15021/healthz/ready delay=1s timeout=1s period=2s #success=1 #failure=30
Environment:
JWT_POLICY: first-party-jwt
PILOT_CERT_PROVIDER: istiod
CA_ADDR: istiod.istio-system.svc:15012
NODE_NAME: (v1:spec.nodeName)
POD_NAME: istio-ingressgateway-f5b59cc7c-qqgrr (v1:metadata.name)
POD_NAMESPACE: istio-system (v1:metadata.namespace)
INSTANCE_IP: (v1:status.podIP)
HOST_IP: (v1:status.hostIP)
SERVICE_ACCOUNT: (v1:spec.serviceAccountName)
ISTIO_META_WORKLOAD_NAME: istio-ingressgateway
ISTIO_META_OWNER: kubernetes://apis/apps/v1/namespaces/istio-system/deployments/istio-ingressgateway
ISTIO_META_MESH_ID: cluster.local
TRUST_DOMAIN: cluster.local
ISTIO_META_UNPRIVILEGED_POD: true
ISTIO_META_CLUSTER_ID: Kubernetes
Mounts:
/etc/istio/config from config-volume (rw)
/etc/istio/ingressgateway-ca-certs from ingressgateway-ca-certs (ro)
/etc/istio/ingressgateway-certs from ingressgateway-certs (ro)
/etc/istio/pod from podinfo (rw)
/etc/istio/proxy from istio-envoy (rw)
/var/lib/istio/data from istio-data (rw)
/var/run/secrets/istio from istiod-ca-cert (rw)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-4m2xw (ro)
Conditions:
Type Status
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
istiod-ca-cert:
Type: ConfigMap (a volume populated by a ConfigMap)
Name: istio-ca-root-cert
Optional: false
podinfo:
Type: DownwardAPI (a volume populated by information about the pod)
Items:
metadata.labels -> labels
metadata.annotations -> annotations
istio-envoy:
Type: EmptyDir (a temporary directory that shares a pod's lifetime)
Medium:
SizeLimit: <unset>
istio-data:
Type: EmptyDir (a temporary directory that shares a pod's lifetime)
Medium:
SizeLimit: <unset>
config-volume:
Type: ConfigMap (a volume populated by a ConfigMap)
Name: istio
Optional: true
ingressgateway-certs:
Type: Secret (a volume populated by a Secret)
SecretName: istio-ingressgateway-certs
Optional: true
ingressgateway-ca-certs:
Type: Secret (a volume populated by a Secret)
SecretName: istio-ingressgateway-ca-certs
Optional: true
kube-api-access-4m2xw:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
QoS Class: Burstable
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning BackOff 6m11s (x915 over 3h16m) kubelet Back-off restarting failed container
Normal Pulled 74s (x44 over 3h16m) kubelet Container image "docker.io/istio/proxyv2:1.12.5" already present on machine
Istio Error during set-up
- Processing resources for Istio core.
✔ Istio core installed
- Processing resources for Istiod.
- Processing resources for Istiod. Waiting for Deployment/istio-system/istiod
✔ Istiod installed
- Processing resources for Ingress gateways.
- Processing resources for Ingress gateways. Waiting for Deployment/istio-system/cluster-local-gateway, Deployment/istio-system/istio-ingressgateway
✘ Ingress gateways encountered an error: failed to wait for resource: resources not ready after 5m0s: timed out waiting for the condition
Deployment/istio-system/cluster-local-gateway (container failed to start: CrashLoopBackOff: back-off 2m40s restarting failed container=istio-proxy pod=cluster-local-gateway-74c4558686-ncbjb_istio-system(67dc64df-0d90-4d43-aa1e-e4ed458f1f90))
Deployment/istio-system/istio-ingressgateway (container failed to start: CrashLoopBackOff: back-off 2m40s restarting failed container=istio-proxy pod=istio-ingressgateway-f5b59cc7c-bj9mc_istio-system(06717413-71bf-409a-b6e0-98309676c0c3))
- Pruning removed resourcesError: failed to install manifests: errors occurred during operation
@aditya2803 ok, pre-requisites is a good catch. I don't quite get how functions can be deployed properly without istio installation being successful. Can you collect a complete log of the bash scripts that set up the cluster?
Here it is:
//KVM
$ [ -r /dev/kvm ] && [ -w /dev/kvm ] && echo "OK" || echo "FAIL"
$ sudo apt -y install bridge-utils cpu-checker libvirt-clients libvirt-daemon qemu qemu-kvm
$ kvm-ok
$ err=""; [ "$(uname) $(uname -m)" = "Linux x86_64" ] || err="ERROR: your system is not Linux x86_64."; [ -r /dev/kvm ] && [ -w /dev/kvm ] || err="$err\nERROR: /dev/kvm is innaccessible."; (( $(uname -r | cut -d. -f1)*1000 + $(uname -r | cut -d. -f2) >= 4014 )) || err="$err\nERROR: your kernel version ($(uname -r)) is too old."; dmesg | grep -i "hypervisor detected" && echo "WARNING: you are running in a virtual machine. Firecracker is not well tested under nested virtualization."; [ -z "$err" ] && echo "Your system looks ready for Firecracker!" || echo -e "$err"
$ ls -al /dev/kvm
$ grep kvm /etc/group
$ sudo adduser aditya kvm
$ grep kvm /etc/group
//GIT
$ gh repo clone ease-lab/vhive
$ cd vhive
$ gh pr checkout 465
//CLUSTER
$ ./scripts/cloudlab/setup_node.sh > >(tee -a /tmp/vhive-logs/setup_worker_kubelet.stdout) 2> >(tee -a /tmp/vhive-logs/setup_worker_kubelet.stderr >&2)
$ sudo screen -dmS containerd bash -c "containerd > >(tee -a /tmp/vhive-logs/containerd.stdout) 2> >(tee -a /tmp/vhive-logs/containerd.stderr >&2)"
$ sleep 5;
$ sudo PATH=$PATH screen -dmS firecracker bash -c "/usr/local/bin/firecracker-containerd --config /etc/firecracker-containerd/config.toml > >(tee -a /tmp/vhive-logs/firecracker.stdout) 2> >(tee -a /tmp/vhive-logs/firecracker.stderr >&2)"
$ sleep 5;
$ source /etc/profile && go build
$ sudo screen -dmS vhive bash -c "./vhive > >(tee -a /tmp/vhive-logs/vhive.stdout) 2> >(tee -a /tmp/vhive-logs/vhive.stderr >&2)"
$ sleep 5;
$ ./scripts/cluster/create_one_node_cluster.sh > >(tee -a /tmp/vhive-logs/create_singlenode_cluster.stdout) 2> >(tee -a /tmp/vhive-logs/create_singlenode_cluster.stderr >&2)
Also, for my understanding, isn't Istio just used for serving the function endpoints ? Isn't the deployment of the functions independent of it ? I understand why function invocation won't work without it, but not clear about the deploying part.
Debugging function deployment in a failed knative cluster is not a good strategy. Let us focus on Istio first.
Please provide /tmp/vhive-logs/setup_worker_kubelet.*
and /tmp/vhive-logs/create_singlenode_cluster.*
.
Sure. I tried running the cleanup and again starting the cluster a couple of times, so the logs are appended with that. Sorry about it.
/tmp/vhive-logs/setup_worker_kubelet.stdout
logs
APT::Periodic::Update-Package-Lists "0";
APT::Periodic::Download-Upgradeable-Packages "0";
APT::Periodic::AutocleanInterval "0";
APT::Periodic::Unattended-Upgrade "0";
APT::Periodic::Update-Package-Lists "0";
APT::Periodic::Download-Upgradeable-Packages "0";
APT::Periodic::AutocleanInterval "0";
APT::Periodic::Unattended-Upgrade "0";
containerd github.com/containerd/containerd v1.6.2 de8046a5501db9e0e478e1c10cbcfb21af4c6b2d
OK
🚧 Compile
net.bridge.bridge-nf-call-iptables = 1
net.ipv4.ip_forward = 1
net.bridge.bridge-nf-call-ip6tables = 1
0 209715200 thin-pool /dev/loop4 /dev/loop3 128 32768 1 skip_block_zeroing
APT::Periodic::Update-Package-Lists "0";
APT::Periodic::Download-Upgradeable-Packages "0";
APT::Periodic::AutocleanInterval "0";
APT::Periodic::Unattended-Upgrade "0";
containerd github.com/containerd/containerd v1.6.2 de8046a5501db9e0e478e1c10cbcfb21af4c6b2d
OK
🚧 Compile
net.bridge.bridge-nf-call-iptables = 1
net.ipv4.ip_forward = 1
net.bridge.bridge-nf-call-ip6tables = 1
0 209715200 thin-pool /dev/loop6 /dev/loop5 128 32768 1 skip_block_zeroing
APT::Periodic::Update-Package-Lists "0";
APT::Periodic::Download-Upgradeable-Packages "0";
APT::Periodic::AutocleanInterval "0";
APT::Periodic::Unattended-Upgrade "0";
containerd github.com/containerd/containerd v1.6.2 de8046a5501db9e0e478e1c10cbcfb21af4c6b2d
OK
🚧 Compile
net.bridge.bridge-nf-call-iptables = 1
net.ipv4.ip_forward = 1
net.bridge.bridge-nf-call-ip6tables = 1
0 209715200 thin-pool /dev/loop12 /dev/loop11 128 32768 1 skip_block_zeroing
/tmp/vhive-logs/setup_worker_kubelet.stderr
logs
Created symlink /etc/systemd/system/apt-daily.service → /dev/null.
Created symlink /etc/systemd/system/apt-daily-upgrade.service → /dev/null.
E: Unable to locate package skopeo
sysctl: setting key "net.ipv4.conf.all.promote_secondaries": Invalid argument
sudo: route: command not found
sudo: nft: command not found
sudo: nft: command not found
sudo: nft: command not found
sudo: nft: command not found
sudo: nft: command not found
sudo: nft: command not found
E: Unable to locate package skopeo
sysctl: setting key "net.ipv4.conf.all.promote_secondaries": Invalid argument
Warning: apt-key output should not be parsed (stdout is not a terminal)
sysctl: setting key "net.ipv4.conf.all.promote_secondaries": Invalid argument
device-mapper: reload ioctl on fc-dev-thinpool failed: No such device or address
Command failed.
E: Unable to locate package skopeo
sysctl: setting key "net.ipv4.conf.all.promote_secondaries": Invalid argument
Warning: apt-key output should not be parsed (stdout is not a terminal)
fatal: destination path '/home/aditya/client' already exists and is not an empty directory.
sysctl: setting key "net.ipv4.conf.all.promote_secondaries": Invalid argument
E: Unable to locate package skopeo
sysctl: setting key "net.ipv4.conf.all.promote_secondaries": Invalid argument
Warning: apt-key output should not be parsed (stdout is not a terminal)
fatal: destination path '/home/aditya/client' already exists and is not an empty directory.
sysctl: setting key "net.ipv4.conf.all.promote_secondaries": Invalid argument
/tmp/vhive-logs/create_singlenode_cluster.stdout
logs
[init] Using Kubernetes version: v1.23.6
[preflight] Running pre-flight checks
[preflight] Pulling images required for setting up a Kubernetes cluster
[preflight] This might take a minute or two, depending on the speed of your internet connection
[preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
[certs] Using certificateDir folder "/etc/kubernetes/pki"
[certs] Generating "ca" certificate and key
[certs] Generating "apiserver" certificate and key
[certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local spgblr-dyt-09] and IPs [10.96.0.1 10.138.143.25]
[certs] Generating "apiserver-kubelet-client" certificate and key
[certs] Generating "front-proxy-ca" certificate and key
[certs] Generating "front-proxy-client" certificate and key
[certs] Generating "etcd/ca" certificate and key
[certs] Generating "etcd/server" certificate and key
[certs] etcd/server serving cert is signed for DNS names [localhost spgblr-dyt-09] and IPs [10.138.143.25 127.0.0.1 ::1]
[certs] Generating "etcd/peer" certificate and key
[certs] etcd/peer serving cert is signed for DNS names [localhost spgblr-dyt-09] and IPs [10.138.143.25 127.0.0.1 ::1]
[certs] Generating "etcd/healthcheck-client" certificate and key
[certs] Generating "apiserver-etcd-client" certificate and key
[certs] Generating "sa" key and public key
[kubeconfig] Using kubeconfig folder "/etc/kubernetes"
[kubeconfig] Writing "admin.conf" kubeconfig file
[kubeconfig] Writing "kubelet.conf" kubeconfig file
[kubeconfig] Writing "controller-manager.conf" kubeconfig file
[kubeconfig] Writing "scheduler.conf" kubeconfig file
[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
[kubelet-start] Starting the kubelet
[control-plane] Using manifest folder "/etc/kubernetes/manifests"
[control-plane] Creating static Pod manifest for "kube-apiserver"
[control-plane] Creating static Pod manifest for "kube-controller-manager"
[control-plane] Creating static Pod manifest for "kube-scheduler"
[etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
[apiclient] All control plane components are healthy after 11.503210 seconds
[upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
[kubelet] Creating a ConfigMap "kubelet-config-1.23" in namespace kube-system with the configuration for the kubelets in the cluster
NOTE: The "kubelet-config-1.23" naming of the kubelet ConfigMap is deprecated. Once the UnversionedKubeletConfigMap feature gate graduates to Beta the default name will become just "kubelet-config". Kubeadm upgrade will handle this transition transparently.
[upload-certs] Skipping phase. Please see --upload-certs
[mark-control-plane] Marking the node spgblr-dyt-09 as control-plane by adding the labels: [node-role.kubernetes.io/master(deprecated) node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
[mark-control-plane] Marking the node spgblr-dyt-09 as control-plane by adding the taints [node-role.kubernetes.io/master:NoSchedule]
[bootstrap-token] Using token: as63qq.en6z0w4fi61kpsvr
[bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to get nodes
[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
[bootstrap-token] configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
[bootstrap-token] configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
[bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
[kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
[addons] Applied essential addon: CoreDNS
[addons] Applied essential addon: kube-proxy
Your Kubernetes control-plane has initialized successfully!
To start using your cluster, you need to run the following as a regular user:
mkdir -p $HOME/.kube
sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
sudo chown $(id -u):$(id -g) $HOME/.kube/config
Alternatively, if you are the root user, you can run:
export KUBECONFIG=/etc/kubernetes/admin.conf
You should now deploy a pod network to the cluster.
Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
https://kubernetes.io/docs/concepts/cluster-administration/addons/
Then you can join any number of worker nodes by running the following on each as root:
kubeadm join 10.138.143.25:6443 --token as63qq.en6z0w4fi61kpsvr \
--discovery-token-ca-cert-hash sha256:66e83512c0e30b74d197786cf28a0f4eba4fc7cc153d09cf9c1eab12b326256e
node/spgblr-dyt-09 untainted
configmap/canal-config created
customresourcedefinition.apiextensions.k8s.io/bgpconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/bgppeers.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/blockaffinities.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/clusterinformations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/felixconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/globalnetworkpolicies.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/globalnetworksets.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/hostendpoints.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamblocks.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamconfigs.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamhandles.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ippools.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/kubecontrollersconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/networkpolicies.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/networksets.crd.projectcalico.org created
clusterrole.rbac.authorization.k8s.io/calico-kube-controllers created
clusterrolebinding.rbac.authorization.k8s.io/calico-kube-controllers created
clusterrole.rbac.authorization.k8s.io/calico-node created
clusterrole.rbac.authorization.k8s.io/flannel created
clusterrolebinding.rbac.authorization.k8s.io/canal-flannel created
clusterrolebinding.rbac.authorization.k8s.io/canal-calico created
daemonset.apps/canal created
serviceaccount/canal created
deployment.apps/calico-kube-controllers created
serviceaccount/calico-kube-controllers created
poddisruptionbudget.policy/calico-kube-controllers created
configmap/kube-proxy configured
namespace/metallb-system created
podsecuritypolicy.policy/controller created
podsecuritypolicy.policy/speaker created
serviceaccount/controller created
serviceaccount/speaker created
clusterrole.rbac.authorization.k8s.io/metallb-system:controller created
clusterrole.rbac.authorization.k8s.io/metallb-system:speaker created
role.rbac.authorization.k8s.io/config-watcher created
role.rbac.authorization.k8s.io/pod-lister created
clusterrolebinding.rbac.authorization.k8s.io/metallb-system:controller created
clusterrolebinding.rbac.authorization.k8s.io/metallb-system:speaker created
rolebinding.rbac.authorization.k8s.io/config-watcher created
rolebinding.rbac.authorization.k8s.io/pod-lister created
daemonset.apps/speaker created
deployment.apps/controller created
secret/memberlist created
configmap/config created
Downloading istio-1.12.5 from https://github.com/istio/istio/releases/download/1.12.5/istio-1.12.5-linux-amd64.tar.gz ...
Istio 1.12.5 Download Complete!
Istio has been successfully downloaded into the istio-1.12.5 folder on your system.
Next Steps:
See https://istio.io/latest/docs/setup/install/ to add Istio to your Kubernetes cluster.
To configure the istioctl client tool for your workstation,
add the /home/aditya/vhive/istio-1.12.5/bin directory to your environment path variable with:
export PATH="$PATH:/home/aditya/vhive/istio-1.12.5/bin"
Begin the Istio pre-installation check by running:
istioctl x precheck
Need more information? Visit https://istio.io/latest/docs/setup/install/
customresourcedefinition.apiextensions.k8s.io/certificates.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/configurations.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/clusterdomainclaims.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/domainmappings.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/ingresses.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/metrics.autoscaling.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/podautoscalers.autoscaling.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/revisions.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/routes.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/serverlessservices.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/services.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/images.caching.internal.knative.dev created
namespace/knative-serving created
clusterrole.rbac.authorization.k8s.io/knative-serving-aggregated-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/knative-serving-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/knative-serving-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-serving-namespaced-edit created
clusterrole.rbac.authorization.k8s.io/knative-serving-namespaced-view created
clusterrole.rbac.authorization.k8s.io/knative-serving-core created
clusterrole.rbac.authorization.k8s.io/knative-serving-podspecable-binding created
serviceaccount/controller created
clusterrole.rbac.authorization.k8s.io/knative-serving-admin created
clusterrolebinding.rbac.authorization.k8s.io/knative-serving-controller-admin created
clusterrolebinding.rbac.authorization.k8s.io/knative-serving-controller-addressable-resolver created
customresourcedefinition.apiextensions.k8s.io/images.caching.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/certificates.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/configurations.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/clusterdomainclaims.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/domainmappings.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/ingresses.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/metrics.autoscaling.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/podautoscalers.autoscaling.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/revisions.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/routes.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/serverlessservices.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/services.serving.knative.dev unchanged
image.caching.internal.knative.dev/queue-proxy created
configmap/config-autoscaler created
configmap/config-defaults created
configmap/config-deployment created
configmap/config-domain created
configmap/config-features created
configmap/config-gc created
configmap/config-leader-election created
configmap/config-logging created
configmap/config-network created
configmap/config-observability created
configmap/config-tracing created
horizontalpodautoscaler.autoscaling/activator created
poddisruptionbudget.policy/activator-pdb created
deployment.apps/activator created
service/activator-service created
deployment.apps/autoscaler created
service/autoscaler created
deployment.apps/controller created
service/controller created
deployment.apps/domain-mapping created
deployment.apps/domainmapping-webhook created
service/domainmapping-webhook created
horizontalpodautoscaler.autoscaling/webhook created
poddisruptionbudget.policy/webhook-pdb created
deployment.apps/webhook created
service/webhook created
validatingwebhookconfiguration.admissionregistration.k8s.io/config.webhook.serving.knative.dev created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.serving.knative.dev created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.domainmapping.serving.knative.dev created
secret/domainmapping-webhook-certs created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.webhook.domainmapping.serving.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.webhook.serving.knative.dev created
secret/webhook-certs created
namespace/registry created
persistentvolume/docker-repo-pv created
persistentvolumeclaim/docker-repo-pvc created
replicaset.apps/docker-registry-pod created
service/docker-registry created
daemonset.apps/registry-etc-hosts-update created
job.batch/default-domain created
service/default-domain-service created
clusterrole.rbac.authorization.k8s.io/knative-serving-istio created
gateway.networking.istio.io/knative-ingress-gateway created
gateway.networking.istio.io/knative-local-gateway created
service/knative-local-gateway created
configmap/config-istio created
peerauthentication.security.istio.io/webhook created
peerauthentication.security.istio.io/domainmapping-webhook created
peerauthentication.security.istio.io/net-istio-webhook created
deployment.apps/net-istio-controller created
deployment.apps/net-istio-webhook created
secret/net-istio-webhook-certs created
service/net-istio-webhook created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.istio.networking.internal.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/config.webhook.istio.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/apiserversources.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/brokers.eventing.knative.dev created
customresourcedefinition.apiextensions.k8s.io/channels.messaging.knative.dev created
customresourcedefinition.apiextensions.k8s.io/containersources.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/eventtypes.eventing.knative.dev created
customresourcedefinition.apiextensions.k8s.io/parallels.flows.knative.dev created
customresourcedefinition.apiextensions.k8s.io/pingsources.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/sequences.flows.knative.dev created
customresourcedefinition.apiextensions.k8s.io/sinkbindings.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/subscriptions.messaging.knative.dev created
customresourcedefinition.apiextensions.k8s.io/triggers.eventing.knative.dev created
namespace/knative-eventing created
serviceaccount/eventing-controller created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-resolver created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-source-observer created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-sources-controller created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-manipulator created
serviceaccount/pingsource-mt-adapter created
clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-pingsource-mt-adapter created
serviceaccount/eventing-webhook created
clusterrolebinding.rbac.authorization.k8s.io/eventing-webhook created
rolebinding.rbac.authorization.k8s.io/eventing-webhook created
clusterrolebinding.rbac.authorization.k8s.io/eventing-webhook-resolver created
clusterrolebinding.rbac.authorization.k8s.io/eventing-webhook-podspecable-binding created
configmap/config-br-default-channel created
configmap/config-br-defaults created
configmap/default-ch-webhook created
configmap/config-ping-defaults created
configmap/config-features created
configmap/config-kreference-mapping created
configmap/config-leader-election created
configmap/config-logging created
configmap/config-observability created
configmap/config-tracing created
deployment.apps/eventing-controller created
deployment.apps/pingsource-mt-adapter created
horizontalpodautoscaler.autoscaling/eventing-webhook created
poddisruptionbudget.policy/eventing-webhook created
deployment.apps/eventing-webhook created
service/eventing-webhook created
customresourcedefinition.apiextensions.k8s.io/apiserversources.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/brokers.eventing.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/channels.messaging.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/containersources.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/eventtypes.eventing.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/parallels.flows.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/pingsources.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/sequences.flows.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/sinkbindings.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/subscriptions.messaging.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/triggers.eventing.knative.dev unchanged
clusterrole.rbac.authorization.k8s.io/addressable-resolver created
clusterrole.rbac.authorization.k8s.io/service-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/serving-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/channel-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/broker-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/flows-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/eventing-broker-filter created
clusterrole.rbac.authorization.k8s.io/eventing-broker-ingress created
clusterrole.rbac.authorization.k8s.io/eventing-config-reader created
clusterrole.rbac.authorization.k8s.io/channelable-manipulator created
clusterrole.rbac.authorization.k8s.io/meta-channelable-manipulator created
clusterrole.rbac.authorization.k8s.io/knative-eventing-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-messaging-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-flows-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-sources-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-bindings-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-eventing-namespaced-edit created
clusterrole.rbac.authorization.k8s.io/knative-eventing-namespaced-view created
clusterrole.rbac.authorization.k8s.io/knative-eventing-controller created
clusterrole.rbac.authorization.k8s.io/knative-eventing-pingsource-mt-adapter created
clusterrole.rbac.authorization.k8s.io/podspecable-binding created
clusterrole.rbac.authorization.k8s.io/builtin-podspecable-binding created
clusterrole.rbac.authorization.k8s.io/source-observer created
clusterrole.rbac.authorization.k8s.io/eventing-sources-source-observer created
clusterrole.rbac.authorization.k8s.io/knative-eventing-sources-controller created
clusterrole.rbac.authorization.k8s.io/knative-eventing-webhook created
role.rbac.authorization.k8s.io/knative-eventing-webhook created
validatingwebhookconfiguration.admissionregistration.k8s.io/config.webhook.eventing.knative.dev created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.eventing.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.webhook.eventing.knative.dev created
secret/eventing-webhook-certs created
mutatingwebhookconfiguration.admissionregistration.k8s.io/sinkbindings.webhook.sources.knative.dev created
namespace/knative-eventing unchanged
serviceaccount/imc-controller created
clusterrolebinding.rbac.authorization.k8s.io/imc-controller created
rolebinding.rbac.authorization.k8s.io/imc-controller created
clusterrolebinding.rbac.authorization.k8s.io/imc-controller-resolver created
serviceaccount/imc-dispatcher created
clusterrolebinding.rbac.authorization.k8s.io/imc-dispatcher created
configmap/config-imc-event-dispatcher created
configmap/config-observability unchanged
configmap/config-tracing configured
deployment.apps/imc-controller created
service/inmemorychannel-webhook created
service/imc-dispatcher created
deployment.apps/imc-dispatcher created
customresourcedefinition.apiextensions.k8s.io/inmemorychannels.messaging.knative.dev created
clusterrole.rbac.authorization.k8s.io/imc-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/imc-channelable-manipulator created
clusterrole.rbac.authorization.k8s.io/imc-controller created
clusterrole.rbac.authorization.k8s.io/imc-dispatcher created
role.rbac.authorization.k8s.io/knative-inmemorychannel-webhook created
mutatingwebhookconfiguration.admissionregistration.k8s.io/inmemorychannel.eventing.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.inmemorychannel.eventing.knative.dev created
secret/inmemorychannel-webhook-certs created
clusterrole.rbac.authorization.k8s.io/knative-eventing-mt-channel-broker-controller created
clusterrole.rbac.authorization.k8s.io/knative-eventing-mt-broker-filter created
serviceaccount/mt-broker-filter created
clusterrole.rbac.authorization.k8s.io/knative-eventing-mt-broker-ingress created
serviceaccount/mt-broker-ingress created
clusterrolebinding.rbac.authorization.k8s.io/eventing-mt-channel-broker-controller created
clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-mt-broker-filter created
clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-mt-broker-ingress created
deployment.apps/mt-broker-filter created
service/broker-filter created
deployment.apps/mt-broker-ingress created
service/broker-ingress created
deployment.apps/mt-broker-controller created
horizontalpodautoscaler.autoscaling/broker-ingress-hpa created
horizontalpodautoscaler.autoscaling/broker-filter-hpa created
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
istio-ingressgateway LoadBalancer 10.110.46.166 192.168.1.240 15021:31995/TCP,80:32106/TCP,443:31066/TCP 5m18s
[init] Using Kubernetes version: v1.23.6
[preflight] Running pre-flight checks
[preflight] Pulling images required for setting up a Kubernetes cluster
[preflight] This might take a minute or two, depending on the speed of your internet connection
[preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
[certs] Using certificateDir folder "/etc/kubernetes/pki"
[certs] Generating "ca" certificate and key
[certs] Generating "apiserver" certificate and key
[certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local spgblr-dyt-09] and IPs [10.96.0.1 10.138.143.25]
[certs] Generating "apiserver-kubelet-client" certificate and key
[certs] Generating "front-proxy-ca" certificate and key
[certs] Generating "front-proxy-client" certificate and key
[certs] Generating "etcd/ca" certificate and key
[certs] Generating "etcd/server" certificate and key
[certs] etcd/server serving cert is signed for DNS names [localhost spgblr-dyt-09] and IPs [10.138.143.25 127.0.0.1 ::1]
[certs] Generating "etcd/peer" certificate and key
[certs] etcd/peer serving cert is signed for DNS names [localhost spgblr-dyt-09] and IPs [10.138.143.25 127.0.0.1 ::1]
[certs] Generating "etcd/healthcheck-client" certificate and key
[certs] Generating "apiserver-etcd-client" certificate and key
[certs] Generating "sa" key and public key
[kubeconfig] Using kubeconfig folder "/etc/kubernetes"
[kubeconfig] Writing "admin.conf" kubeconfig file
[kubeconfig] Writing "kubelet.conf" kubeconfig file
[kubeconfig] Writing "controller-manager.conf" kubeconfig file
[kubeconfig] Writing "scheduler.conf" kubeconfig file
[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
[kubelet-start] Starting the kubelet
[control-plane] Using manifest folder "/etc/kubernetes/manifests"
[control-plane] Creating static Pod manifest for "kube-apiserver"
[control-plane] Creating static Pod manifest for "kube-controller-manager"
[control-plane] Creating static Pod manifest for "kube-scheduler"
[etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
[apiclient] All control plane components are healthy after 11.003029 seconds
[upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
[kubelet] Creating a ConfigMap "kubelet-config-1.23" in namespace kube-system with the configuration for the kubelets in the cluster
NOTE: The "kubelet-config-1.23" naming of the kubelet ConfigMap is deprecated. Once the UnversionedKubeletConfigMap feature gate graduates to Beta the default name will become just "kubelet-config". Kubeadm upgrade will handle this transition transparently.
[upload-certs] Skipping phase. Please see --upload-certs
[mark-control-plane] Marking the node spgblr-dyt-09 as control-plane by adding the labels: [node-role.kubernetes.io/master(deprecated) node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
[mark-control-plane] Marking the node spgblr-dyt-09 as control-plane by adding the taints [node-role.kubernetes.io/master:NoSchedule]
[bootstrap-token] Using token: 6pe1hk.zrosfdtcheqvjew7
[bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to get nodes
[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
[bootstrap-token] configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
[bootstrap-token] configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
[bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
[kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
[addons] Applied essential addon: CoreDNS
[addons] Applied essential addon: kube-proxy
Your Kubernetes control-plane has initialized successfully!
To start using your cluster, you need to run the following as a regular user:
mkdir -p $HOME/.kube
sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
sudo chown $(id -u):$(id -g) $HOME/.kube/config
Alternatively, if you are the root user, you can run:
export KUBECONFIG=/etc/kubernetes/admin.conf
You should now deploy a pod network to the cluster.
Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
https://kubernetes.io/docs/concepts/cluster-administration/addons/
Then you can join any number of worker nodes by running the following on each as root:
kubeadm join 10.138.143.25:6443 --token 6pe1hk.zrosfdtcheqvjew7 \
--discovery-token-ca-cert-hash sha256:874d8674e33941a21a4eb43813c30dee409fbfdf72a5e27db5a88c3ba8e312fd
node/spgblr-dyt-09 untainted
configmap/canal-config created
customresourcedefinition.apiextensions.k8s.io/bgpconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/bgppeers.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/blockaffinities.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/clusterinformations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/felixconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/globalnetworkpolicies.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/globalnetworksets.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/hostendpoints.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamblocks.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamconfigs.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamhandles.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ippools.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/kubecontrollersconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/networkpolicies.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/networksets.crd.projectcalico.org created
clusterrole.rbac.authorization.k8s.io/calico-kube-controllers created
clusterrolebinding.rbac.authorization.k8s.io/calico-kube-controllers created
clusterrole.rbac.authorization.k8s.io/calico-node created
clusterrole.rbac.authorization.k8s.io/flannel created
clusterrolebinding.rbac.authorization.k8s.io/canal-flannel created
clusterrolebinding.rbac.authorization.k8s.io/canal-calico created
daemonset.apps/canal created
serviceaccount/canal created
deployment.apps/calico-kube-controllers created
serviceaccount/calico-kube-controllers created
poddisruptionbudget.policy/calico-kube-controllers created
configmap/kube-proxy configured
namespace/metallb-system created
podsecuritypolicy.policy/controller created
podsecuritypolicy.policy/speaker created
serviceaccount/controller created
serviceaccount/speaker created
clusterrole.rbac.authorization.k8s.io/metallb-system:controller created
clusterrole.rbac.authorization.k8s.io/metallb-system:speaker created
role.rbac.authorization.k8s.io/config-watcher created
role.rbac.authorization.k8s.io/pod-lister created
clusterrolebinding.rbac.authorization.k8s.io/metallb-system:controller created
clusterrolebinding.rbac.authorization.k8s.io/metallb-system:speaker created
rolebinding.rbac.authorization.k8s.io/config-watcher created
rolebinding.rbac.authorization.k8s.io/pod-lister created
daemonset.apps/speaker created
deployment.apps/controller created
secret/memberlist created
configmap/config created
Downloading istio-1.12.5 from https://github.com/istio/istio/releases/download/1.12.5/istio-1.12.5-linux-amd64.tar.gz ...
Istio 1.12.5 Download Complete!
Istio has been successfully downloaded into the istio-1.12.5 folder on your system.
Next Steps:
See https://istio.io/latest/docs/setup/install/ to add Istio to your Kubernetes cluster.
To configure the istioctl client tool for your workstation,
add the /home/aditya/vhive/istio-1.12.5/bin directory to your environment path variable with:
export PATH="$PATH:/home/aditya/vhive/istio-1.12.5/bin"
Begin the Istio pre-installation check by running:
istioctl x precheck
Need more information? Visit https://istio.io/latest/docs/setup/install/
customresourcedefinition.apiextensions.k8s.io/certificates.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/configurations.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/clusterdomainclaims.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/domainmappings.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/ingresses.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/metrics.autoscaling.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/podautoscalers.autoscaling.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/revisions.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/routes.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/serverlessservices.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/services.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/images.caching.internal.knative.dev created
namespace/knative-serving created
clusterrole.rbac.authorization.k8s.io/knative-serving-aggregated-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/knative-serving-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/knative-serving-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-serving-namespaced-edit created
clusterrole.rbac.authorization.k8s.io/knative-serving-namespaced-view created
clusterrole.rbac.authorization.k8s.io/knative-serving-core created
clusterrole.rbac.authorization.k8s.io/knative-serving-podspecable-binding created
serviceaccount/controller created
clusterrole.rbac.authorization.k8s.io/knative-serving-admin created
clusterrolebinding.rbac.authorization.k8s.io/knative-serving-controller-admin created
clusterrolebinding.rbac.authorization.k8s.io/knative-serving-controller-addressable-resolver created
customresourcedefinition.apiextensions.k8s.io/images.caching.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/certificates.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/configurations.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/clusterdomainclaims.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/domainmappings.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/ingresses.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/metrics.autoscaling.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/podautoscalers.autoscaling.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/revisions.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/routes.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/serverlessservices.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/services.serving.knative.dev unchanged
image.caching.internal.knative.dev/queue-proxy created
configmap/config-autoscaler created
configmap/config-defaults created
configmap/config-deployment created
configmap/config-domain created
configmap/config-features created
configmap/config-gc created
configmap/config-leader-election created
configmap/config-logging created
configmap/config-network created
configmap/config-observability created
configmap/config-tracing created
horizontalpodautoscaler.autoscaling/activator created
poddisruptionbudget.policy/activator-pdb created
deployment.apps/activator created
service/activator-service created
deployment.apps/autoscaler created
service/autoscaler created
deployment.apps/controller created
service/controller created
deployment.apps/domain-mapping created
deployment.apps/domainmapping-webhook created
service/domainmapping-webhook created
horizontalpodautoscaler.autoscaling/webhook created
poddisruptionbudget.policy/webhook-pdb created
deployment.apps/webhook created
service/webhook created
validatingwebhookconfiguration.admissionregistration.k8s.io/config.webhook.serving.knative.dev created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.serving.knative.dev created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.domainmapping.serving.knative.dev created
secret/domainmapping-webhook-certs created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.webhook.domainmapping.serving.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.webhook.serving.knative.dev created
secret/webhook-certs created
namespace/registry created
persistentvolume/docker-repo-pv created
persistentvolumeclaim/docker-repo-pvc created
replicaset.apps/docker-registry-pod created
service/docker-registry created
daemonset.apps/registry-etc-hosts-update created
job.batch/default-domain created
service/default-domain-service created
clusterrole.rbac.authorization.k8s.io/knative-serving-istio created
gateway.networking.istio.io/knative-ingress-gateway created
gateway.networking.istio.io/knative-local-gateway created
service/knative-local-gateway created
configmap/config-istio created
peerauthentication.security.istio.io/webhook created
peerauthentication.security.istio.io/domainmapping-webhook created
peerauthentication.security.istio.io/net-istio-webhook created
deployment.apps/net-istio-controller created
deployment.apps/net-istio-webhook created
secret/net-istio-webhook-certs created
service/net-istio-webhook created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.istio.networking.internal.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/config.webhook.istio.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/apiserversources.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/brokers.eventing.knative.dev created
customresourcedefinition.apiextensions.k8s.io/channels.messaging.knative.dev created
customresourcedefinition.apiextensions.k8s.io/containersources.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/eventtypes.eventing.knative.dev created
customresourcedefinition.apiextensions.k8s.io/parallels.flows.knative.dev created
customresourcedefinition.apiextensions.k8s.io/pingsources.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/sequences.flows.knative.dev created
customresourcedefinition.apiextensions.k8s.io/sinkbindings.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/subscriptions.messaging.knative.dev created
customresourcedefinition.apiextensions.k8s.io/triggers.eventing.knative.dev created
namespace/knative-eventing created
serviceaccount/eventing-controller created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-resolver created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-source-observer created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-sources-controller created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-manipulator created
serviceaccount/pingsource-mt-adapter created
clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-pingsource-mt-adapter created
serviceaccount/eventing-webhook created
clusterrolebinding.rbac.authorization.k8s.io/eventing-webhook created
rolebinding.rbac.authorization.k8s.io/eventing-webhook created
clusterrolebinding.rbac.authorization.k8s.io/eventing-webhook-resolver created
clusterrolebinding.rbac.authorization.k8s.io/eventing-webhook-podspecable-binding created
configmap/config-br-default-channel created
configmap/config-br-defaults created
configmap/default-ch-webhook created
configmap/config-ping-defaults created
configmap/config-features created
configmap/config-kreference-mapping created
configmap/config-leader-election created
configmap/config-logging created
configmap/config-observability created
configmap/config-tracing created
deployment.apps/eventing-controller created
deployment.apps/pingsource-mt-adapter created
horizontalpodautoscaler.autoscaling/eventing-webhook created
poddisruptionbudget.policy/eventing-webhook created
deployment.apps/eventing-webhook created
service/eventing-webhook created
customresourcedefinition.apiextensions.k8s.io/apiserversources.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/brokers.eventing.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/channels.messaging.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/containersources.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/eventtypes.eventing.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/parallels.flows.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/pingsources.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/sequences.flows.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/sinkbindings.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/subscriptions.messaging.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/triggers.eventing.knative.dev unchanged
clusterrole.rbac.authorization.k8s.io/addressable-resolver created
clusterrole.rbac.authorization.k8s.io/service-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/serving-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/channel-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/broker-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/flows-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/eventing-broker-filter created
clusterrole.rbac.authorization.k8s.io/eventing-broker-ingress created
clusterrole.rbac.authorization.k8s.io/eventing-config-reader created
clusterrole.rbac.authorization.k8s.io/channelable-manipulator created
clusterrole.rbac.authorization.k8s.io/meta-channelable-manipulator created
clusterrole.rbac.authorization.k8s.io/knative-eventing-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-messaging-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-flows-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-sources-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-bindings-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-eventing-namespaced-edit created
clusterrole.rbac.authorization.k8s.io/knative-eventing-namespaced-view created
clusterrole.rbac.authorization.k8s.io/knative-eventing-controller created
clusterrole.rbac.authorization.k8s.io/knative-eventing-pingsource-mt-adapter created
clusterrole.rbac.authorization.k8s.io/podspecable-binding created
clusterrole.rbac.authorization.k8s.io/builtin-podspecable-binding created
clusterrole.rbac.authorization.k8s.io/source-observer created
clusterrole.rbac.authorization.k8s.io/eventing-sources-source-observer created
clusterrole.rbac.authorization.k8s.io/knative-eventing-sources-controller created
clusterrole.rbac.authorization.k8s.io/knative-eventing-webhook created
role.rbac.authorization.k8s.io/knative-eventing-webhook created
validatingwebhookconfiguration.admissionregistration.k8s.io/config.webhook.eventing.knative.dev created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.eventing.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.webhook.eventing.knative.dev created
secret/eventing-webhook-certs created
mutatingwebhookconfiguration.admissionregistration.k8s.io/sinkbindings.webhook.sources.knative.dev created
namespace/knative-eventing unchanged
serviceaccount/imc-controller created
clusterrolebinding.rbac.authorization.k8s.io/imc-controller created
rolebinding.rbac.authorization.k8s.io/imc-controller created
clusterrolebinding.rbac.authorization.k8s.io/imc-controller-resolver created
serviceaccount/imc-dispatcher created
clusterrolebinding.rbac.authorization.k8s.io/imc-dispatcher created
configmap/config-imc-event-dispatcher created
configmap/config-observability unchanged
configmap/config-tracing configured
deployment.apps/imc-controller created
service/inmemorychannel-webhook created
service/imc-dispatcher created
deployment.apps/imc-dispatcher created
customresourcedefinition.apiextensions.k8s.io/inmemorychannels.messaging.knative.dev created
clusterrole.rbac.authorization.k8s.io/imc-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/imc-channelable-manipulator created
clusterrole.rbac.authorization.k8s.io/imc-controller created
clusterrole.rbac.authorization.k8s.io/imc-dispatcher created
role.rbac.authorization.k8s.io/knative-inmemorychannel-webhook created
mutatingwebhookconfiguration.admissionregistration.k8s.io/inmemorychannel.eventing.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.inmemorychannel.eventing.knative.dev created
secret/inmemorychannel-webhook-certs created
clusterrole.rbac.authorization.k8s.io/knative-eventing-mt-channel-broker-controller created
clusterrole.rbac.authorization.k8s.io/knative-eventing-mt-broker-filter created
serviceaccount/mt-broker-filter created
clusterrole.rbac.authorization.k8s.io/knative-eventing-mt-broker-ingress created
serviceaccount/mt-broker-ingress created
clusterrolebinding.rbac.authorization.k8s.io/eventing-mt-channel-broker-controller created
clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-mt-broker-filter created
clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-mt-broker-ingress created
deployment.apps/mt-broker-filter created
service/broker-filter created
deployment.apps/mt-broker-ingress created
service/broker-ingress created
deployment.apps/mt-broker-controller created
horizontalpodautoscaler.autoscaling/broker-ingress-hpa created
horizontalpodautoscaler.autoscaling/broker-filter-hpa created
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
istio-ingressgateway LoadBalancer 10.100.129.155 192.168.1.240 15021:32503/TCP,80:30550/TCP,443:30395/TCP 5m22s
[init] Using Kubernetes version: v1.23.6
[preflight] Running pre-flight checks
[preflight] Pulling images required for setting up a Kubernetes cluster
[preflight] This might take a minute or two, depending on the speed of your internet connection
[preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
[certs] Using certificateDir folder "/etc/kubernetes/pki"
[certs] Generating "ca" certificate and key
[certs] Generating "apiserver" certificate and key
[certs] apiserver serving cert is signed for DNS names [kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local spgblr-dyt-09] and IPs [10.96.0.1 10.138.143.25]
[certs] Generating "apiserver-kubelet-client" certificate and key
[certs] Generating "front-proxy-ca" certificate and key
[certs] Generating "front-proxy-client" certificate and key
[certs] Generating "etcd/ca" certificate and key
[certs] Generating "etcd/server" certificate and key
[certs] etcd/server serving cert is signed for DNS names [localhost spgblr-dyt-09] and IPs [10.138.143.25 127.0.0.1 ::1]
[certs] Generating "etcd/peer" certificate and key
[certs] etcd/peer serving cert is signed for DNS names [localhost spgblr-dyt-09] and IPs [10.138.143.25 127.0.0.1 ::1]
[certs] Generating "etcd/healthcheck-client" certificate and key
[certs] Generating "apiserver-etcd-client" certificate and key
[certs] Generating "sa" key and public key
[kubeconfig] Using kubeconfig folder "/etc/kubernetes"
[kubeconfig] Writing "admin.conf" kubeconfig file
[kubeconfig] Writing "kubelet.conf" kubeconfig file
[kubeconfig] Writing "controller-manager.conf" kubeconfig file
[kubeconfig] Writing "scheduler.conf" kubeconfig file
[kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
[kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
[kubelet-start] Starting the kubelet
[control-plane] Using manifest folder "/etc/kubernetes/manifests"
[control-plane] Creating static Pod manifest for "kube-apiserver"
[control-plane] Creating static Pod manifest for "kube-controller-manager"
[control-plane] Creating static Pod manifest for "kube-scheduler"
[etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
[wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
[apiclient] All control plane components are healthy after 11.503265 seconds
[upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
[kubelet] Creating a ConfigMap "kubelet-config-1.23" in namespace kube-system with the configuration for the kubelets in the cluster
NOTE: The "kubelet-config-1.23" naming of the kubelet ConfigMap is deprecated. Once the UnversionedKubeletConfigMap feature gate graduates to Beta the default name will become just "kubelet-config". Kubeadm upgrade will handle this transition transparently.
[upload-certs] Skipping phase. Please see --upload-certs
[mark-control-plane] Marking the node spgblr-dyt-09 as control-plane by adding the labels: [node-role.kubernetes.io/master(deprecated) node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
[mark-control-plane] Marking the node spgblr-dyt-09 as control-plane by adding the taints [node-role.kubernetes.io/master:NoSchedule]
[bootstrap-token] Using token: 6kix0t.4r4r5vhxogz2ywqo
[bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to get nodes
[bootstrap-token] configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
[bootstrap-token] configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
[bootstrap-token] configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
[bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
[kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
[addons] Applied essential addon: CoreDNS
[addons] Applied essential addon: kube-proxy
Your Kubernetes control-plane has initialized successfully!
To start using your cluster, you need to run the following as a regular user:
mkdir -p $HOME/.kube
sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
sudo chown $(id -u):$(id -g) $HOME/.kube/config
Alternatively, if you are the root user, you can run:
export KUBECONFIG=/etc/kubernetes/admin.conf
You should now deploy a pod network to the cluster.
Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
https://kubernetes.io/docs/concepts/cluster-administration/addons/
Then you can join any number of worker nodes by running the following on each as root:
kubeadm join 10.138.143.25:6443 --token 6kix0t.4r4r5vhxogz2ywqo \
--discovery-token-ca-cert-hash sha256:5b43e41d27f166c17df05144564a09955b1a06f5d89b5a0a245d868542f58fc3
node/spgblr-dyt-09 untainted
configmap/canal-config created
customresourcedefinition.apiextensions.k8s.io/bgpconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/bgppeers.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/blockaffinities.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/clusterinformations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/felixconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/globalnetworkpolicies.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/globalnetworksets.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/hostendpoints.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamblocks.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamconfigs.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ipamhandles.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/ippools.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/kubecontrollersconfigurations.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/networkpolicies.crd.projectcalico.org created
customresourcedefinition.apiextensions.k8s.io/networksets.crd.projectcalico.org created
clusterrole.rbac.authorization.k8s.io/calico-kube-controllers created
clusterrolebinding.rbac.authorization.k8s.io/calico-kube-controllers created
clusterrole.rbac.authorization.k8s.io/calico-node created
clusterrole.rbac.authorization.k8s.io/flannel created
clusterrolebinding.rbac.authorization.k8s.io/canal-flannel created
clusterrolebinding.rbac.authorization.k8s.io/canal-calico created
daemonset.apps/canal created
serviceaccount/canal created
deployment.apps/calico-kube-controllers created
serviceaccount/calico-kube-controllers created
poddisruptionbudget.policy/calico-kube-controllers created
configmap/kube-proxy configured
namespace/metallb-system created
podsecuritypolicy.policy/controller created
podsecuritypolicy.policy/speaker created
serviceaccount/controller created
serviceaccount/speaker created
clusterrole.rbac.authorization.k8s.io/metallb-system:controller created
clusterrole.rbac.authorization.k8s.io/metallb-system:speaker created
role.rbac.authorization.k8s.io/config-watcher created
role.rbac.authorization.k8s.io/pod-lister created
clusterrolebinding.rbac.authorization.k8s.io/metallb-system:controller created
clusterrolebinding.rbac.authorization.k8s.io/metallb-system:speaker created
rolebinding.rbac.authorization.k8s.io/config-watcher created
rolebinding.rbac.authorization.k8s.io/pod-lister created
daemonset.apps/speaker created
deployment.apps/controller created
secret/memberlist created
configmap/config created
Downloading istio-1.12.5 from https://github.com/istio/istio/releases/download/1.12.5/istio-1.12.5-linux-amd64.tar.gz ...
Istio 1.12.5 Download Complete!
Istio has been successfully downloaded into the istio-1.12.5 folder on your system.
Next Steps:
See https://istio.io/latest/docs/setup/install/ to add Istio to your Kubernetes cluster.
To configure the istioctl client tool for your workstation,
add the /home/aditya/vhive/istio-1.12.5/bin directory to your environment path variable with:
export PATH="$PATH:/home/aditya/vhive/istio-1.12.5/bin"
Begin the Istio pre-installation check by running:
istioctl x precheck
Need more information? Visit https://istio.io/latest/docs/setup/install/
customresourcedefinition.apiextensions.k8s.io/certificates.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/configurations.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/clusterdomainclaims.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/domainmappings.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/ingresses.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/metrics.autoscaling.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/podautoscalers.autoscaling.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/revisions.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/routes.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/serverlessservices.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/services.serving.knative.dev created
customresourcedefinition.apiextensions.k8s.io/images.caching.internal.knative.dev created
namespace/knative-serving created
clusterrole.rbac.authorization.k8s.io/knative-serving-aggregated-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/knative-serving-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/knative-serving-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-serving-namespaced-edit created
clusterrole.rbac.authorization.k8s.io/knative-serving-namespaced-view created
clusterrole.rbac.authorization.k8s.io/knative-serving-core created
clusterrole.rbac.authorization.k8s.io/knative-serving-podspecable-binding created
serviceaccount/controller created
clusterrole.rbac.authorization.k8s.io/knative-serving-admin created
clusterrolebinding.rbac.authorization.k8s.io/knative-serving-controller-admin created
clusterrolebinding.rbac.authorization.k8s.io/knative-serving-controller-addressable-resolver created
customresourcedefinition.apiextensions.k8s.io/images.caching.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/certificates.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/configurations.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/clusterdomainclaims.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/domainmappings.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/ingresses.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/metrics.autoscaling.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/podautoscalers.autoscaling.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/revisions.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/routes.serving.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/serverlessservices.networking.internal.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/services.serving.knative.dev unchanged
image.caching.internal.knative.dev/queue-proxy created
configmap/config-autoscaler created
configmap/config-defaults created
configmap/config-deployment created
configmap/config-domain created
configmap/config-features created
configmap/config-gc created
configmap/config-leader-election created
configmap/config-logging created
configmap/config-network created
configmap/config-observability created
configmap/config-tracing created
horizontalpodautoscaler.autoscaling/activator created
poddisruptionbudget.policy/activator-pdb created
deployment.apps/activator created
service/activator-service created
deployment.apps/autoscaler created
service/autoscaler created
deployment.apps/controller created
service/controller created
deployment.apps/domain-mapping created
deployment.apps/domainmapping-webhook created
service/domainmapping-webhook created
horizontalpodautoscaler.autoscaling/webhook created
poddisruptionbudget.policy/webhook-pdb created
deployment.apps/webhook created
service/webhook created
validatingwebhookconfiguration.admissionregistration.k8s.io/config.webhook.serving.knative.dev created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.serving.knative.dev created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.domainmapping.serving.knative.dev created
secret/domainmapping-webhook-certs created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.webhook.domainmapping.serving.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.webhook.serving.knative.dev created
secret/webhook-certs created
namespace/registry created
persistentvolume/docker-repo-pv created
persistentvolumeclaim/docker-repo-pvc created
replicaset.apps/docker-registry-pod created
service/docker-registry created
daemonset.apps/registry-etc-hosts-update created
job.batch/default-domain created
service/default-domain-service created
clusterrole.rbac.authorization.k8s.io/knative-serving-istio created
gateway.networking.istio.io/knative-ingress-gateway created
gateway.networking.istio.io/knative-local-gateway created
service/knative-local-gateway created
configmap/config-istio created
peerauthentication.security.istio.io/webhook created
peerauthentication.security.istio.io/domainmapping-webhook created
peerauthentication.security.istio.io/net-istio-webhook created
deployment.apps/net-istio-controller created
deployment.apps/net-istio-webhook created
secret/net-istio-webhook-certs created
service/net-istio-webhook created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.istio.networking.internal.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/config.webhook.istio.networking.internal.knative.dev created
customresourcedefinition.apiextensions.k8s.io/apiserversources.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/brokers.eventing.knative.dev created
customresourcedefinition.apiextensions.k8s.io/channels.messaging.knative.dev created
customresourcedefinition.apiextensions.k8s.io/containersources.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/eventtypes.eventing.knative.dev created
customresourcedefinition.apiextensions.k8s.io/parallels.flows.knative.dev created
customresourcedefinition.apiextensions.k8s.io/pingsources.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/sequences.flows.knative.dev created
customresourcedefinition.apiextensions.k8s.io/sinkbindings.sources.knative.dev created
customresourcedefinition.apiextensions.k8s.io/subscriptions.messaging.knative.dev created
customresourcedefinition.apiextensions.k8s.io/triggers.eventing.knative.dev created
namespace/knative-eventing created
serviceaccount/eventing-controller created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-resolver created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-source-observer created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-sources-controller created
clusterrolebinding.rbac.authorization.k8s.io/eventing-controller-manipulator created
serviceaccount/pingsource-mt-adapter created
clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-pingsource-mt-adapter created
serviceaccount/eventing-webhook created
clusterrolebinding.rbac.authorization.k8s.io/eventing-webhook created
rolebinding.rbac.authorization.k8s.io/eventing-webhook created
clusterrolebinding.rbac.authorization.k8s.io/eventing-webhook-resolver created
clusterrolebinding.rbac.authorization.k8s.io/eventing-webhook-podspecable-binding created
configmap/config-br-default-channel created
configmap/config-br-defaults created
configmap/default-ch-webhook created
configmap/config-ping-defaults created
configmap/config-features created
configmap/config-kreference-mapping created
configmap/config-leader-election created
configmap/config-logging created
configmap/config-observability created
configmap/config-tracing created
deployment.apps/eventing-controller created
deployment.apps/pingsource-mt-adapter created
horizontalpodautoscaler.autoscaling/eventing-webhook created
poddisruptionbudget.policy/eventing-webhook created
deployment.apps/eventing-webhook created
service/eventing-webhook created
customresourcedefinition.apiextensions.k8s.io/apiserversources.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/brokers.eventing.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/channels.messaging.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/containersources.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/eventtypes.eventing.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/parallels.flows.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/pingsources.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/sequences.flows.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/sinkbindings.sources.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/subscriptions.messaging.knative.dev unchanged
customresourcedefinition.apiextensions.k8s.io/triggers.eventing.knative.dev unchanged
clusterrole.rbac.authorization.k8s.io/addressable-resolver created
clusterrole.rbac.authorization.k8s.io/service-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/serving-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/channel-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/broker-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/flows-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/eventing-broker-filter created
clusterrole.rbac.authorization.k8s.io/eventing-broker-ingress created
clusterrole.rbac.authorization.k8s.io/eventing-config-reader created
clusterrole.rbac.authorization.k8s.io/channelable-manipulator created
clusterrole.rbac.authorization.k8s.io/meta-channelable-manipulator created
clusterrole.rbac.authorization.k8s.io/knative-eventing-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-messaging-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-flows-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-sources-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-bindings-namespaced-admin created
clusterrole.rbac.authorization.k8s.io/knative-eventing-namespaced-edit created
clusterrole.rbac.authorization.k8s.io/knative-eventing-namespaced-view created
clusterrole.rbac.authorization.k8s.io/knative-eventing-controller created
clusterrole.rbac.authorization.k8s.io/knative-eventing-pingsource-mt-adapter created
clusterrole.rbac.authorization.k8s.io/podspecable-binding created
clusterrole.rbac.authorization.k8s.io/builtin-podspecable-binding created
clusterrole.rbac.authorization.k8s.io/source-observer created
clusterrole.rbac.authorization.k8s.io/eventing-sources-source-observer created
clusterrole.rbac.authorization.k8s.io/knative-eventing-sources-controller created
clusterrole.rbac.authorization.k8s.io/knative-eventing-webhook created
role.rbac.authorization.k8s.io/knative-eventing-webhook created
validatingwebhookconfiguration.admissionregistration.k8s.io/config.webhook.eventing.knative.dev created
mutatingwebhookconfiguration.admissionregistration.k8s.io/webhook.eventing.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.webhook.eventing.knative.dev created
secret/eventing-webhook-certs created
mutatingwebhookconfiguration.admissionregistration.k8s.io/sinkbindings.webhook.sources.knative.dev created
namespace/knative-eventing unchanged
serviceaccount/imc-controller created
clusterrolebinding.rbac.authorization.k8s.io/imc-controller created
rolebinding.rbac.authorization.k8s.io/imc-controller created
clusterrolebinding.rbac.authorization.k8s.io/imc-controller-resolver created
serviceaccount/imc-dispatcher created
clusterrolebinding.rbac.authorization.k8s.io/imc-dispatcher created
configmap/config-imc-event-dispatcher created
configmap/config-observability unchanged
configmap/config-tracing configured
deployment.apps/imc-controller created
service/inmemorychannel-webhook created
service/imc-dispatcher created
deployment.apps/imc-dispatcher created
customresourcedefinition.apiextensions.k8s.io/inmemorychannels.messaging.knative.dev created
clusterrole.rbac.authorization.k8s.io/imc-addressable-resolver created
clusterrole.rbac.authorization.k8s.io/imc-channelable-manipulator created
clusterrole.rbac.authorization.k8s.io/imc-controller created
clusterrole.rbac.authorization.k8s.io/imc-dispatcher created
role.rbac.authorization.k8s.io/knative-inmemorychannel-webhook created
mutatingwebhookconfiguration.admissionregistration.k8s.io/inmemorychannel.eventing.knative.dev created
validatingwebhookconfiguration.admissionregistration.k8s.io/validation.inmemorychannel.eventing.knative.dev created
secret/inmemorychannel-webhook-certs created
clusterrole.rbac.authorization.k8s.io/knative-eventing-mt-channel-broker-controller created
clusterrole.rbac.authorization.k8s.io/knative-eventing-mt-broker-filter created
serviceaccount/mt-broker-filter created
clusterrole.rbac.authorization.k8s.io/knative-eventing-mt-broker-ingress created
serviceaccount/mt-broker-ingress created
clusterrolebinding.rbac.authorization.k8s.io/eventing-mt-channel-broker-controller created
clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-mt-broker-filter created
clusterrolebinding.rbac.authorization.k8s.io/knative-eventing-mt-broker-ingress created
deployment.apps/mt-broker-filter created
service/broker-filter created
deployment.apps/mt-broker-ingress created
service/broker-ingress created
deployment.apps/mt-broker-controller created
horizontalpodautoscaler.autoscaling/broker-ingress-hpa created
horizontalpodautoscaler.autoscaling/broker-filter-hpa created
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
istio-ingressgateway LoadBalancer 10.106.207.86 192.168.1.240 15021:32734/TCP,80:31659/TCP,443:31125/TCP 5m20s
/tmp/vhive-logs/create_singlenode_cluster.stderr
logs
+++ dirname ./scripts/cluster/create_one_node_cluster.sh
++ cd ./scripts/cluster
++ pwd
+ DIR=/home/aditya/vhive/scripts/cluster
++ cd /home/aditya/vhive/scripts/cluster
++ cd ..
++ cd ..
++ pwd
+ ROOT=/home/aditya/vhive
+ STOCK_CONTAINERD=
+ /home/aditya/vhive/scripts/cluster/setup_worker_kubelet.sh
+ '[' '' == stock-only ']'
+ CRI_SOCK=/etc/vhive-cri/vhive-cri.sock
+++ cat /proc/1/cpuset
++ basename /
+ CONTAINERID=/
+ '[' 64 -eq 1 ']'
+ sudo kubeadm init --ignore-preflight-errors=all --cri-socket /etc/vhive-cri/vhive-cri.sock --pod-network-cidr=192.168.0.0/16
I0523 12:45:28.989075 60812 version.go:255] remote version is much newer: v1.24.0; falling back to: stable-1.23
+ mkdir -p /home/aditya/.kube
+ sudo cp -i /etc/kubernetes/admin.conf /home/aditya/.kube/config
++ id -u
++ id -g
+ sudo chown 1001:27 /home/aditya/.kube/config
+ '[' 1001 -eq 0 ']'
+ kubectl taint nodes --all node-role.kubernetes.io/master-
+ /home/aditya/vhive/scripts/cluster/setup_master_node.sh
Warning: policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
Warning: resource configmaps/kube-proxy is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically.
Warning: policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+
Warning: spec.template.spec.nodeSelector[beta.kubernetes.io/os]: deprecated since v1.14; use "kubernetes.io/os" instead
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 101 100 101 0 0 349 0 --:--:-- --:--:-- --:--:-- 349
100 4926 100 4926 0 0 6928 0 --:--:-- --:--:-- --:--:-- 6928
! values.global.jwtPolicy is deprecated; use Values.global.jwtPolicy=third-party-jwt. See http://istio.io/latest/docs/ops/best-practices/security/#configure-third-party-service-account-tokens for more information instead
- Processing resources for Istio core.
✔ Istio core installed
- Processing resources for Istiod.
- Processing resources for Istiod. Waiting for Deployment/istio-system/istiod
✔ Istiod installed
- Processing resources for Ingress gateways.
- Processing resources for Ingress gateways. Waiting for Deployment/istio-system/cluster-local-gateway, Deployment/istio-system/istio-ingressgateway
✘ Ingress gateways encountered an error: failed to wait for resource: resources not ready after 5m0s: timed out waiting for the condition
Deployment/istio-system/cluster-local-gateway (container failed to start: CrashLoopBackOff: back-off 2m40s restarting failed container=istio-proxy pod=cluster-local-gateway-74c4558686-ncbjb_istio-system(67dc64df-0d90-4d43-aa1e-e4ed458f1f90))
Deployment/istio-system/istio-ingressgateway (container failed to start: CrashLoopBackOff: back-off 2m40s restarting failed container=istio-proxy pod=istio-ingressgateway-f5b59cc7c-bj9mc_istio-system(06717413-71bf-409a-b6e0-98309676c0c3))
- Pruning removed resourcesError: failed to install manifests: errors occurred during operation
Warning: autoscaling/v2beta2 HorizontalPodAutoscaler is deprecated in v1.23+, unavailable in v1.26+; use autoscaling/v2 HorizontalPodAutoscaler
Warning: autoscaling/v2beta2 HorizontalPodAutoscaler is deprecated in v1.23+, unavailable in v1.26+; use autoscaling/v2 HorizontalPodAutoscaler
Warning: policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
Warning: autoscaling/v2beta2 HorizontalPodAutoscaler is deprecated in v1.23+, unavailable in v1.26+; use autoscaling/v2 HorizontalPodAutoscaler
+++ dirname ./scripts/cluster/create_one_node_cluster.sh
++ cd ./scripts/cluster
++ pwd
+ DIR=/home/aditya/vhive/scripts/cluster
++ cd /home/aditya/vhive/scripts/cluster
++ cd ..
++ cd ..
++ pwd
+ ROOT=/home/aditya/vhive
+ STOCK_CONTAINERD=
+ /home/aditya/vhive/scripts/cluster/setup_worker_kubelet.sh
+ '[' '' == stock-only ']'
+ CRI_SOCK=/etc/vhive-cri/vhive-cri.sock
+++ cat /proc/1/cpuset
++ basename /
+ CONTAINERID=/
+ '[' 64 -eq 1 ']'
+ sudo kubeadm init --ignore-preflight-errors=all --cri-socket /etc/vhive-cri/vhive-cri.sock --pod-network-cidr=192.168.0.0/16
I0523 13:07:32.510735 97167 version.go:255] remote version is much newer: v1.24.0; falling back to: stable-1.23
+ mkdir -p /home/aditya/.kube
+ sudo cp -i /etc/kubernetes/admin.conf /home/aditya/.kube/config
++ id -u
++ id -g
+ sudo chown 1001:27 /home/aditya/.kube/config
+ '[' 1001 -eq 0 ']'
+ kubectl taint nodes --all node-role.kubernetes.io/master-
+ /home/aditya/vhive/scripts/cluster/setup_master_node.sh
Warning: policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
Warning: resource configmaps/kube-proxy is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically.
Warning: policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+
Warning: spec.template.spec.nodeSelector[beta.kubernetes.io/os]: deprecated since v1.14; use "kubernetes.io/os" instead
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 101 100 101 0 0 353 0 --:--:-- --:--:-- --:--:-- 354
100 4926 100 4926 0 0 6889 0 --:--:-- --:--:-- --:--:-- 6889
! values.global.jwtPolicy is deprecated; use Values.global.jwtPolicy=third-party-jwt. See http://istio.io/latest/docs/ops/best-practices/security/#configure-third-party-service-account-tokens for more information instead
- Processing resources for Istio core.
✔ Istio core installed
- Processing resources for Istiod.
- Processing resources for Istiod. Waiting for Deployment/istio-system/istiod
✔ Istiod installed
- Processing resources for Ingress gateways.
- Processing resources for Ingress gateways. Waiting for Deployment/istio-system/cluster-local-gateway, Deployment/istio-system/istio-ingressgateway
✘ Ingress gateways encountered an error: failed to wait for resource: resources not ready after 5m0s: timed out waiting for the condition
Deployment/istio-system/cluster-local-gateway (container failed to start: CrashLoopBackOff: back-off 2m40s restarting failed container=istio-proxy pod=cluster-local-gateway-74c4558686-frksv_istio-system(25eb9504-13a3-410a-b986-6bcfc0170331))
Deployment/istio-system/istio-ingressgateway (container failed to start: CrashLoopBackOff: back-off 2m40s restarting failed container=istio-proxy pod=istio-ingressgateway-f5b59cc7c-l824g_istio-system(3aeafca5-0555-4429-8d23-05e3cddc49f4))
- Pruning removed resourcesError: failed to install manifests: errors occurred during operation
Warning: autoscaling/v2beta2 HorizontalPodAutoscaler is deprecated in v1.23+, unavailable in v1.26+; use autoscaling/v2 HorizontalPodAutoscaler
Warning: autoscaling/v2beta2 HorizontalPodAutoscaler is deprecated in v1.23+, unavailable in v1.26+; use autoscaling/v2 HorizontalPodAutoscaler
Warning: policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
Warning: autoscaling/v2beta2 HorizontalPodAutoscaler is deprecated in v1.23+, unavailable in v1.26+; use autoscaling/v2 HorizontalPodAutoscaler
+++ dirname ./scripts/cluster/create_one_node_cluster.sh
++ cd ./scripts/cluster
++ pwd
+ DIR=/home/aditya/vhive/scripts/cluster
++ cd /home/aditya/vhive/scripts/cluster
++ cd ..
++ cd ..
++ pwd
+ ROOT=/home/aditya/vhive
+ STOCK_CONTAINERD=
+ /home/aditya/vhive/scripts/cluster/setup_worker_kubelet.sh
+ '[' '' == stock-only ']'
+ CRI_SOCK=/etc/vhive-cri/vhive-cri.sock
+++ cat /proc/1/cpuset
++ basename /
+ CONTAINERID=/
+ '[' 64 -eq 1 ']'
+ sudo kubeadm init --ignore-preflight-errors=all --cri-socket /etc/vhive-cri/vhive-cri.sock --pod-network-cidr=192.168.0.0/16
I0523 13:23:36.520089 122821 version.go:255] remote version is much newer: v1.24.0; falling back to: stable-1.23
+ mkdir -p /home/aditya/.kube
+ sudo cp -i /etc/kubernetes/admin.conf /home/aditya/.kube/config
++ id -u
++ id -g
+ sudo chown 1001:27 /home/aditya/.kube/config
+ '[' 1001 -eq 0 ']'
+ kubectl taint nodes --all node-role.kubernetes.io/master-
+ /home/aditya/vhive/scripts/cluster/setup_master_node.sh
Warning: policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
Warning: resource configmaps/kube-proxy is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically.
Warning: policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+
Warning: spec.template.spec.nodeSelector[beta.kubernetes.io/os]: deprecated since v1.14; use "kubernetes.io/os" instead
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 101 100 101 0 0 351 0 --:--:-- --:--:-- --:--:-- 351
100 4926 100 4926 0 0 6851 0 --:--:-- --:--:-- --:--:-- 16474
! values.global.jwtPolicy is deprecated; use Values.global.jwtPolicy=third-party-jwt. See http://istio.io/latest/docs/ops/best-practices/security/#configure-third-party-service-account-tokens for more information instead
- Processing resources for Istio core.
✔ Istio core installed
- Processing resources for Istiod.
- Processing resources for Istiod. Waiting for Deployment/istio-system/istiod
✔ Istiod installed
- Processing resources for Ingress gateways.
- Processing resources for Ingress gateways. Waiting for Deployment/istio-system/cluster-local-gateway, Deployment/istio-system/istio-ingressgateway
✘ Ingress gateways encountered an error: failed to wait for resource: resources not ready after 5m0s: timed out waiting for the condition
Deployment/istio-system/cluster-local-gateway (container failed to start: CrashLoopBackOff: back-off 2m40s restarting failed container=istio-proxy pod=cluster-local-gateway-74c4558686-r6h8b_istio-system(ee188c65-566e-470e-a91b-e672bc193fb3))
Deployment/istio-system/istio-ingressgateway (container failed to start: CrashLoopBackOff: back-off 2m40s restarting failed container=istio-proxy pod=istio-ingressgateway-f5b59cc7c-h9qcr_istio-system(868240ef-8d43-464e-93a9-d73d9d458c3a))
- Pruning removed resourcesError: failed to install manifests: errors occurred during operation
Warning: autoscaling/v2beta2 HorizontalPodAutoscaler is deprecated in v1.23+, unavailable in v1.26+; use autoscaling/v2 HorizontalPodAutoscaler
Warning: autoscaling/v2beta2 HorizontalPodAutoscaler is deprecated in v1.23+, unavailable in v1.26+; use autoscaling/v2 HorizontalPodAutoscaler
Warning: policy/v1beta1 PodDisruptionBudget is deprecated in v1.21+, unavailable in v1.25+; use policy/v1 PodDisruptionBudget
Warning: autoscaling/v2beta2 HorizontalPodAutoscaler is deprecated in v1.23+, unavailable in v1.26+; use autoscaling/v2 HorizontalPodAutoscaler
I suggest to clean up the node again with ./scripts/github_runner/clean_cri_runner.sh
and delete the logs. Then, try to run the setup script again and see if Istio gets installed. If not, this is some issue with the node setup, which is not vHive-specific and probably should be reported to Knative and/or Istio maintainers.
Please crop the logs in future and attach full logs as files if necessary, otherwise the issue thread becomes unmanageably long.
Hi @ustiugov, apologies for the full log files, I'll keep it in mind next time onward.
After running the clean* script, and redeploying the cluster, istio suddenly was deployed perfectly :) I did try this a few hours earlier, but somehow did not work that time. Anyways, all pods are running okay now.
Also, functions are getting deployed and invoked normally now, and I am getting the final output in the rps1.00_lat.csv
file.
Thanks a lot for your support. Also, shall I submit a PR for the changes in the set-up guide, including the KVM check-up etc ?
@aditya2803 glad to hear! we always welcome improvements from the community 👍 please close the Issue if it's resolved
Description
I am trying to set up vHive on a single node cluster, and get it working by deploying and then invoking the functions, as described in the guide here. I am able to follow through the steps manually, and all the kubernetes pods are running as desired. However, when deploying functions using this link, I ran into some errors.
System Configuration
lscpu output:
cat /etc/os-release output:
Logs
vHive logs:
(I get the same issue as #476 initially. I then used the solution proposed on the ticket. Above logs are post application of the solution.)
Notes There is a similar issue mentioned here. This seems to be a firecracker-containerd issue for non-Intel vendors, which they seem to have fixed later (as per the issue). I am not sure whether the firecracker-containerd binary used in vHive is the latest one. When I clone the latest firecracker-containerd repo, install it, and replace the /vhive/bin/firecracker-containerd binary with the one I built, the vHive error log gets reduced to:
I have also gone through #525 and have access to /dev/kvm. Also, I am running on a bare-metal x86_64 amd server running Ubuntu 20.04.
Expected Behavior Functions should be deployed normally.
Steps to reproduce Simply follow the start-up guide provided to set up an one-node cluster & then run the deployer.