Closed 316953425 closed 7 months ago
@316953425 Thanks for the report! Could you please deploy it again and share your cluster dump?
kubectl cluster-info dump -n nsm-system --output-directory nsm-dump
kubectl cluster-info dump -n ns-kernel2ip2kernel --output-directory nsm-dump
kubectl cluster-info dump -n nsm-system --output-directory nsm-dump
hi, @glazychev-art kubectl cluster-info dump -n nsm-system --output-directory nsm-dump nsm-dump.zip kubectl cluster-info dump -n ns-kernel2ip2kernel-dual-stack --output-directory nsm-dump nsm-dump.zip kubectl cluster-info dump -n ns-kernel2ip2kernel --output-directory nsm-dump nsm-dump.zip
Is the content of the file what you want?
Thank you. that is interesting.
Could you please change the log level from INFO
to DEBUG
for NSM components?
https://github.com/networkservicemesh/deployments-k8s/blob/v1.11.0/apps/nsmgr/nsmgr.yaml#L33-L34
https://github.com/networkservicemesh/deployments-k8s/blob/v1.11.0/apps/forwarder-vpp/forwarder.yaml#L30-L31
https://github.com/networkservicemesh/deployments-k8s/blob/v1.11.0/apps/nse-kernel/nse.yaml#L31-L32
After that, redeploy basic and Kernel2IP2Kernel
kubectl cluster-info dump -n nsm-system --output-directory nsm-dump
hi,@glazychev-art kubectl cluster-info dump -n nsm-system --output-directory nsm-dump
kubectl cluster-info dump -n ns-kernel2ip2kernel --output-directory nsm-dump
I'm sorry, my mistake.
Let's please use the TRACE
level
kubectl cluster-info dump -n nsm-system --output-directory nsm-dump
hi @glazychev-art kubectl cluster-info dump -n nsm-system --output-directory nsm-dump nsm-dump.zip
kubectl cluster-info dump -n ns-kernel2ip2kernel --output-directory nsm-dump nsm-dump.zip
Thanks, A few questions: Does the Kernel2Ethernet2Kernel work for you?
Is the connection between two forwarder-vpp pods fully functional?
172.16.102.12 <------> 172.16.102.13
Perhaps some packet filtering rules are used?
Does the Kernel2Ethernet2Kernel work for you?
Hi @glazychev-art
No packet filtering rules applyed 172.16.102.12 can ping 172.16.102.13
Kernel2Ethernet2Kernel I also tried it, alpine can start but it no interface for nsm nsm-dump.zip
It looks like there is a problem in the connection between the pods. Perhaps there are some special parameters for cilium?
Could you please share kubectl get pods -A -o wide
?
kubectl get pods -A -o wide
hi, @glazychev-art yes
[root@CNCP-MS-01 use-cases]# kubectl get pods -A -o wide
NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
cncp-production dhcp4-679f4549f4-4xwd7 1/1 Running 0 2d20h 10.32.2.99 cncp-ms-02
hi @glazychev-art If the above one doesn't look clear, you can look at this file. env.txt
Got it, thanks
Let's check this, could you remove hostNetwork: true
for forwarder and redeploy it?
https://github.com/networkservicemesh/deployments-k8s/blob/v1.11.0/apps/forwarder-vpp/forwarder.yaml#L19C1-L19C24
Got it, thanks Let's check this, could you remove
hostNetwork: true
for forwarder and redeploy it? https://github.com/networkservicemesh/deployments-k8s/blob/v1.11.0/apps/forwarder-vpp/forwarder.yaml#L19C1-L19C24
hi @glazychev-art Also doesn't work well
The part inside the red box will cause this problem ?
No, it is not a problem, I think
Could you share the same output from any of forwarder-vpp?
kubectl cluster-info dump -n nsm-system --output-directory nsm-dump
hi @glazychev-art yes nsm-dump.zip
looks interesting
Can you please deploy pods on different nodes and try to check availability of pod addresses? Without NSM.
To check the connection you can use ping
, nc
for both side
For example, we are interested in
hi @glazychev-art I'm not sure what you mean, are you suspecting that this address is unavailable? - name: NSM_CIDR_PREFIX value: 10.234.1.1/24
No, I suspect there are some restrictions between pods. Therefore I suggest:
alpine
) on different nodesnc -l -u -p 51820
, on the other nc -u <IP-of-the-firts-pod> 51820
hi @glazychev-art @glazychev-art
it looks like work well
hi @glazychev-art Do you have any other suggestions? Thank you very much.
Could you please try https://github.com/networkservicemesh/deployments-k8s/tree/v1.11.1 instead of v1.11.0 ?
Could you please try https://github.com/networkservicemesh/deployments-k8s/tree/v1.11.1 instead of v1.11.0 ? hi @glazychev-art I tried master, the error message is almost the same as 1.11.0 thanks for you reply
When you tried it? The current master may not be entirely stable
When you tried it? The current master may not be entirely stable
ok i will try v1.11.1
When you tried it? The current master may not be entirely stable
hi @glazychev-art I just deployed the v1.11.1 version, but it still doesn't work. The phenomenon is the same as v1.11.0.
Hi @316953425 , Ok, got it, thanks. Then I suggest diving into a more detailed search for the problem.
- name: NSM_ENVS
value: NSM_LOG_LEVEL=TRACE,NSM_LIVENESSCHECKENABLED=false
INFO
to TRACE
for NSM components:
https://github.com/networkservicemesh/deployments-k8s/blob/v1.11.1/apps/nsmgr/nsmgr.yaml#L33-L34
https://github.com/networkservicemesh/deployments-k8s/blob/v1.11.1/apps/forwarder-vpp/forwarder.yaml#L30-L31
https://github.com/networkservicemesh/deployments-k8s/blob/v1.11.1/apps/nse-kernel/nse.yaml#L31-L32 hostNetwork: true
for forwarder-vppping -i 0.01 <NSE-IP-addr>
NSC_NODE=$(kubectl get pods -l app=alpine -n ns-kernel2ethernet2kernel --template '{{range .items}}{{.spec.nodeName}}{{"\n"}}{{end}}')
FORWARDER_NSC=$(kubectl get pods -l app=forwarder-vpp --field-selector spec.nodeName==${NSC_NODE} -n nsm-system --template '{{range .items}}{{.metadata.name}}{{"\n"}}{{end}}')
NSE_NODE=$(kubectl get pods -l app=nse-kernel -n ns-kernel2ethernet2kernel --template '{{range .items}}{{.spec.nodeName}}{{"\n"}}{{end}}')
FORWARDER_NSE=$(kubectl get pods -l app=forwarder-vpp --field-selector spec.nodeName==${NSE_NODE} -n nsm-system --template '{{range .items}}{{.metadata.name}}{{"\n"}}{{end}}')
kubectl exec -it -n nsm-system ${FORWARDER_NSC} vppctl
Please attach the output of the following commands from vpp:
show int
show int addr
show hardware-interface
show errors
show acl-plugin acl
show ip neighbor
show ip fib
Then let's collect vpp traces:
kubectl exec -it -n nsm-system ${FORWARDER_NSC} vppctl trace add af-packet-input 1000
sleep 20 # wait a few seconds here
kubectl exec -it -n nsm-system ${FORWARDER_NSC} vppctl show trace max 1000 > ${FORWARDER_NSC}_trace.log
_Please do the same but for ${FORWARDERNSE} !
Please show
kubectl get pods -A -o wide
Please attach the cluster-info. You can run 2 commands at once, everything should be saved correctly in nsm-dump:
kubectl cluster-info dump -n nsm-system --output-directory nsm-dump
kubectl cluster-info dump -n ns-kernel2ethernet2kernel --output-directory nsm-dump
We really appreciate your help! Thanks!
hi @glazychev-art I seem to have found a method that you can reproduce locally. You can try it. You install k8s locally and use cilium as the cni component, and then deploy our Kernel2Ethernet2Kernel. There are no other components in my environment. Thank you very much.
My installation steps are as follows:
1、kubectl apply -k https://github.com/networkservicemesh/deployments-k8s/examples/spire/single_cluster?ref=v1.11.1 2、kubectl apply -f https://raw.githubusercontent.com/networkservicemesh/deployments-k8s/v1.11.1/examples/spire/single_cluster/clusterspiffeid-template.yaml 3、kubectl apply -k https://github.com/networkservicemesh/deployments-k8s/examples/basic?ref=v1.11.1 4、kubectl apply -k https://github.com/networkservicemesh/deployments-k8s/examples/use-cases/Kernel2IP2Kernel?ref=v1.11.1
I installed it according to the official website. Is there any problem with my installation steps? Another thing to note is that my three nodes are both master and node at the same time. Will this have any impact?
Thanks a lot!
Your installation steps look good.
Question:
Is it possible to get the following output from nodes:
ip a
ip r
ip neigh
hi @glazychev-art I feel that there may be a conflict between cilium and nsm, causing nsm to not be able to support functions. What do you think?
kubectl exec -it -n nsm-system ${FORWARDER_NSC} vppctl
hi @glazychev-art
kubectl exec -it -n nsm-system ${FORWARDER_NSC} vppctl show int show int addr show hardware-interface show errors show acl-plugin acl
Result: NSC.txt
kubectl exec -it -n nsm-system ${FORWARDER_NSC} vppctl trace add af-packet-input 1000 sleep 20 # wait a few seconds here kubectl exec -it -n nsm-system ${FORWARDER_NSC} vppctl show trace max 1000 > ${FORWARDER_NSC}_trace.log
Result: forwarder-vpp-tm64s_trace.log
NSE Result:
NSE.txt forwarder-vpp-nr7qr_trace.log
kubectl get pods -A -o wide pod.txt
kubectl cluster-info dump -n nsm-system --output-directory nsm-dump nsm-dump.zip
kubectl cluster-info dump -n ns-kernel2ethernet2kernel --output-directory nsm-dump
kubectl logs alpine -n ns-kernel2ethernet2kernel -c cmd-nsc > log cmd-nsc-log.txt
ping from nsc to nse it not work like this:
ping from nse to nsc it not work too ,like this:
I change patch-nse.yaml NSM_CIDR_PREFIX value only:
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: nse-kernel
spec:
template:
spec:
containers:
- name: nse
env:
- name: NSM_CIDR_PREFIX
#value: 172.16.1.100/31
value: 10.16.3.100/31
- name: NSM_SERVICE_NAMES
value: "kernel2ethernet2kernel"
- name: NSM_REGISTER_SERVICE
value: "false"
hi @glazychev-art If you feel that the vpp-forward information above is lacking, you can see this attachment. forwarder-vpp-tm64s_trace.log forwarder-vpp-nr7qr_trace.log
Hi @316953425 , Thank you very much, this is very useful information!
I feel that there may be a conflict between cilium and nsm, causing nsm to not be able to support functions. What do you think?
You are right. More specifically, I think that VPP doesn't work with the cilium setup. I found a workaround locally, but there is one more thing:
Still unclear why the forwarder doesn't work if hostNetwork: true
is set?
If you don't mind, could you please repeat these steps, but without point 3?
I slightly modified the instructions, added more VPP commands, and also indicated that ping
should be started with the -i
flag
Thanks! You really help a lot!
9. kubectl get pods -A -o wide
hi @glazychev-art, I repeat these steps and without point 3
nsc: NSC.txt forwarder-vpp-jfk7c_trace.log
nse:
NSE.txt forwarder-vpp-4wqjm_trace.log
pod: pod.txt
kubectl cluster-info dump -n nsm-system --output-directory nsm-dump nsm-dump.zip
kubectl cluster-info dump -n ns-kernel2ethernet2kernel --output-directory nsm-dump nsm-dump.zip
If you don't mind, could you tell me the local workaround you are looking for?
Thank you, if you don’t mind, a few more traces (ping should still be running):
kubectl exec -it -n nsm-system ${FORWARDER_NSC} vppctl clear trace
kubectl exec -it -n nsm-system ${FORWARDER_NSC} vppctl trace add virtio-input 1000
sleep 20 # wait a few seconds here
kubectl exec -it -n nsm-system ${FORWARDER_NSC} vppctl show trace max 1000 > ${FORWARDER_NSC}_virtio_trace.log
I've rebuilt the forwarder-vpp image locally, and it works for me without hostNetwork: true
If you want, you can test it:
artgl/cmd-forwarder-vpp:cilium_test
kubectl exec -it -n nsm-system ${FORWARDER_NSC} vppctl show trace max 1000 > ${FORWARDER_NSC}_virtio_trace.log
hi @glazychev-art
nsc: forwarder-vpp-jfk7c_virtio_trace.log
nse: forwarder-vpp-4wqjm_virtio_trace.log
Is this ok? thanks
artgl/cmd-forwarder-vpp:cilium_test
hi @glazychev-art
I tried it locally. Whether it is Kernel2IP2Kernel or Kernel2Ethernet2Kernel, I can ping it, but the pod is always in PodInitializing.
Kernel2IP2Kernel:
Kernel2Ethernet2Kernel:
Could you please share
kubectl describe pod -n ns-kernel2ethernet2kernel alpine
kubectl describe pod -n ns-kernel2ethernet2kernel alpine
hi @glazychev-art , thanks apl.txt
use the artgl/cmd-forwarder-vpp:cilium_test image
As far as I can see, the pod now has a Running state.
It took some time to download the image ghcr.io/networkservicemesh/cmd-nsc:v1.11.1
@316953425
But I think we haven't solved all the problems yet.
Previously, we tried to use nс
for alpine pods:
https://github.com/networkservicemesh/deployments-k8s/issues/10581#issuecomment-1827211733
https://github.com/networkservicemesh/deployments-k8s/issues/10581#issuecomment-1827282829
Can we do the same thing, but on nodes?
You can either go directly to the nodes (if possible), or deploy alpine pods on different nodes with hostNetwork: true
and do nc
hi @glazychev-art
see this for alpine always in PodInitializing
kubectl describe pod -n ns-kernel2ip2kernel alpine
Name: alpine
Namespace: ns-kernel2ip2kernel
Priority: 0
Node: cncp-ms-03/172.16.102.13
Start Time: Wed, 29 Nov 2023 15:47:38 +0800
Labels: app=alpine
spiffe.io/spiffe-id=true
Annotations: k8s.v1.cni.cncf.io/network-status:
[{
"name": "cilium",
"interface": "eth0",
"ips": [
"fd00:1032:0:1::ac1b",
"10.32.1.242"
],
"mac": "de:8a:52:8a:47:82",
"default": true,
"dns": {},
"gateway": [
"fd00:1032:0:1::b494",
"10.32.1.224"
]
}]
networkservicemesh.io: kernel://kernel2ip2kernel/nsm-1
Status: Pending
IP: 10.32.1.242
IPs:
IP: 10.32.1.242
IP: fd00:1032:0:1::ac1b
Init Containers:
cmd-nsc-init:
Container ID: docker://ac85aae24e1084d1d00c5e1d0ec514a3fc8a35484d9b2c68be68486f7c54ec18
Image: ghcr.io/networkservicemesh/cmd-nsc-init:v1.11.1
Image ID: docker-pullable://ghcr.io/networkservicemesh/cmd-nsc-init@sha256:ab1d77fa389b0d39ba781c1acedfba141e4cdc1a72d959910056a35d682f5427
Port: <none>
Host Port: <none>
State: Terminated
Reason: Completed
Exit Code: 0
Started: Wed, 29 Nov 2023 15:47:40 +0800
Finished: Wed, 29 Nov 2023 15:47:49 +0800
Ready: True
Restart Count: 0
Limits:
cpu: 200m
memory: 80Mi
Requests:
cpu: 100m
memory: 40Mi
Environment:
NSM_LOG_LEVEL: INFO
SPIFFE_ENDPOINT_SOCKET: unix:///run/spire/sockets/agent.sock
POD_NAME: alpine (v1:metadata.name)
NSM_NETWORK_SERVICES: kernel://kernel2ip2kernel/nsm-1
NSM_NAME: $(POD_NAME)-4b24758c-d8e6-4fd2-b25a-cd175e41c082
Mounts:
/run/spire/sockets from spire-agent-socket (ro)
/var/lib/networkservicemesh from nsm-socket (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-2hv7h (ro)
Containers:
alpine:
Container ID:
Image: alpine:3.15.0
Image ID:
Port: <none>
Host Port: <none>
Command:
/bin/sh
-c
trap : TERM INT; sleep infinity & wait
State: Waiting
Reason: PodInitializing
Ready: False
Restart Count: 0
Environment: <none>
Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-2hv7h (ro)
cmd-nsc:
Container ID:
Image: ghcr.io/networkservicemesh/cmd-nsc:v1.11.1
Image ID:
Port: <none>
Host Port: <none>
State: Waiting
Reason: PodInitializing
Ready: False
Restart Count: 0
Limits:
cpu: 200m
memory: 80Mi
Requests:
cpu: 100m
memory: 40Mi
Environment:
NSM_LOG_LEVEL: INFO
SPIFFE_ENDPOINT_SOCKET: unix:///run/spire/sockets/agent.sock
POD_NAME: alpine (v1:metadata.name)
NSM_NETWORK_SERVICES: kernel://kernel2ip2kernel/nsm-1
NSM_NAME: $(POD_NAME)-4b24758c-d8e6-4fd2-b25a-cd175e41c082
Mounts:
/run/spire/sockets from spire-agent-socket (ro)
/var/lib/networkservicemesh from nsm-socket (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-2hv7h (ro)
Conditions:
Type Status
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-2hv7h:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
spire-agent-socket:
Type: HostPath (bare host directory volume)
Path: /run/spire/sockets
HostPathType: Directory
nsm-socket:
Type: HostPath (bare host directory volume)
Path: /var/lib/networkservicemesh
HostPathType: Directory
QoS Class: Burstable
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 67s default-scheduler Successfully assigned ns-kernel2ip2kernel/alpine to cncp-ms-03
Normal AddedInterface 66s multus Add eth0 [fd00:1032:0:1::ac1b/128 10.32.1.242/32] from cilium
Normal Pulled 66s kubelet Container image "ghcr.io/networkservicemesh/cmd-nsc-init:v1.11.1" already present on machine
Normal Created 65s kubelet Created container cmd-nsc-init
Normal Started 65s kubelet Started container cmd-nsc-init
Normal Pulled 55s kubelet Container image "alpine:3.15.0" already present on machine
Normal Created 55s kubelet Created container alpine
Normal Started 55s kubelet Started container alpine
Normal Pulling 55s kubelet Pulling image "ghcr.io/networkservicemesh/cmd-nsc:v1.11.1"
kubectl describe pod -n ns-kernel2ethernet2kernel alpine
Name: alpine
Namespace: ns-kernel2ethernet2kernel
Priority: 0
Node: cncp-ms-03/172.16.102.13
Start Time: Wed, 29 Nov 2023 15:51:26 +0800
Labels: app=alpine
spiffe.io/spiffe-id=true
Annotations: k8s.v1.cni.cncf.io/network-status:
[{
"name": "cilium",
"interface": "eth0",
"ips": [
"fd00:1032:0:1::4dae",
"10.32.1.222"
],
"mac": "d2:af:f5:c2:3d:2d",
"default": true,
"dns": {},
"gateway": [
"fd00:1032:0:1::b494",
"10.32.1.224"
]
}]
networkservicemesh.io: kernel://kernel2ethernet2kernel/nsm-1
Status: Pending
IP: 10.32.1.222
IPs:
IP: 10.32.1.222
IP: fd00:1032:0:1::4dae
Init Containers:
cmd-nsc-init:
Container ID: docker://83aad383e15915fc64cd4e113c7ec021d08df014623574b367b427f9ce497698
Image: ghcr.io/networkservicemesh/cmd-nsc-init:v1.11.1
Image ID: docker-pullable://ghcr.io/networkservicemesh/cmd-nsc-init@sha256:ab1d77fa389b0d39ba781c1acedfba141e4cdc1a72d959910056a35d682f5427
Port: <none>
Host Port: <none>
State: Terminated
Reason: Completed
Exit Code: 0
Started: Wed, 29 Nov 2023 15:51:28 +0800
Finished: Wed, 29 Nov 2023 15:51:36 +0800
Ready: True
Restart Count: 0
Limits:
cpu: 200m
memory: 80Mi
Requests:
cpu: 100m
memory: 40Mi
Environment:
NSM_LOG_LEVEL: INFO
SPIFFE_ENDPOINT_SOCKET: unix:///run/spire/sockets/agent.sock
POD_NAME: alpine (v1:metadata.name)
NSM_NETWORK_SERVICES: kernel://kernel2ethernet2kernel/nsm-1
NSM_NAME: $(POD_NAME)-f58167f0-fbc9-4673-b2be-41815d911c22
Mounts:
/run/spire/sockets from spire-agent-socket (ro)
/var/lib/networkservicemesh from nsm-socket (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-gdbgh (ro)
Containers:
alpine:
Container ID:
Image: alpine:3.15.0
Image ID:
Port: <none>
Host Port: <none>
Command:
/bin/sh
-c
trap : TERM INT; sleep infinity & wait
State: Waiting
Reason: PodInitializing
Ready: False
Restart Count: 0
Environment: <none>
Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-gdbgh (ro)
cmd-nsc:
Container ID:
Image: ghcr.io/networkservicemesh/cmd-nsc:v1.11.1
Image ID:
Port: <none>
Host Port: <none>
State: Waiting
Reason: PodInitializing
Ready: False
Restart Count: 0
Limits:
cpu: 200m
memory: 80Mi
Requests:
cpu: 100m
memory: 40Mi
Environment:
NSM_LOG_LEVEL: INFO
SPIFFE_ENDPOINT_SOCKET: unix:///run/spire/sockets/agent.sock
POD_NAME: alpine (v1:metadata.name)
NSM_NETWORK_SERVICES: kernel://kernel2ethernet2kernel/nsm-1
NSM_NAME: $(POD_NAME)-f58167f0-fbc9-4673-b2be-41815d911c22
Mounts:
/run/spire/sockets from spire-agent-socket (ro)
/var/lib/networkservicemesh from nsm-socket (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-gdbgh (ro)
Conditions:
Type Status
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-gdbgh:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
spire-agent-socket:
Type: HostPath (bare host directory volume)
Path: /run/spire/sockets
HostPathType: Directory
nsm-socket:
Type: HostPath (bare host directory volume)
Path: /var/lib/networkservicemesh
HostPathType: Directory
QoS Class: Burstable
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 49s default-scheduler Successfully assigned ns-kernel2ethernet2kernel/alpine to cncp-ms-03
Normal AddedInterface 48s multus Add eth0 [fd00:1032:0:1::4dae/128 10.32.1.222/32] from cilium
Normal Pulled 48s kubelet Container image "ghcr.io/networkservicemesh/cmd-nsc-init:v1.11.1" already present on machine
Normal Created 47s kubelet Created container cmd-nsc-init
Normal Started 47s kubelet Started container cmd-nsc-init
Normal Pulled 38s kubelet Container image "alpine:3.15.0" already present on machine
Normal Created 38s kubelet Created container alpine
Normal Started 37s kubelet Started container alpine
Normal Pulling 37s kubelet Pulling image "ghcr.io/networkservicemesh/cmd-nsc:v1.11.1"
Can we do the same thing, but on nodes? You can either go directly to the nodes (if possible), or deploy alpine pods on different nodes with
hostNetwork: true
and donc
hi @glazychev-art
I understand what you mean:
1、deploy nsm with hostNetwork: true
2、In one node exec : nc -l -u -p 51820
3、In another node exec :nc -u
like this:
172.16.102.11 is the node ip
Am I understanding correctly?
As far as I can see, the pod now has a Running state. It took some time to download the image
ghcr.io/networkservicemesh/cmd-nsc:v1.11.1
hi @glazychev-art After a long time ,alpine automatically changed to running, I didn’t do anything I think this time is a bit long, maybe there are some problems
172.16.102.11 is the node ip
Am I understanding correctly?
Yes, you are.
This looks very strange.
Question: does Kernel2IP2Kernel work with the image artgl/cmd-forwarder-vpp:cilium_test
and hostNetwork: true
for forwarder?
After a long time ,alpine automatically changed to running, I didn’t do anything I think this time is a bit long, maybe there are some problems
There is no problem here, downloading the docker image may take some time
172.16.102.11 is the node ip Am I understanding correctly?
Yes, you are. This looks very strange. Question: does Kernel2IP2Kernel work with the image
artgl/cmd-forwarder-vpp:cilium_test
andhostNetwork: true
for forwarder?After a long time ,alpine automatically changed to running, I didn’t do anything I think this time is a bit long, maybe there are some problems
There is no problem here, downloading the docker image may take some time
hi @glazychev-art
Kernel2IP2Kernel work with the image artgl/cmd-forwarder-vpp:cilium_test but remove hostNetwork: true for forwarder
If with hostNetwork: true for forwarder spire agent and server can not start
If with hostNetwork: true for forwarder spire agent and server can not start
This shouldn't happen I think. Could you try to redeploy everything (including spire), but wait until spire-agents and spire-server are ready before NSM deploying
If with hostNetwork: true for forwarder spire agent and server can not start
This shouldn't happen I think. Could you try to redeploy everything (including spire), but wait until spire-agents and spire-server are ready before NSM deploying
hi @glazychev-art with the image artgl/cmd-forwarder-vpp:cilium_test with hostNetwork: true
deploy spire delploy nsm delploy kernel2ethernet2kernel
is right? cat apps/forwarder-vpp/forwarder.yaml
spire agent and server can not start
hi @glazychev-art with the image artgl/cmd-forwarder-vpp:cilium_test not with hostNetwork: true
deploy spire delploy nsm delploy kernel2ethernet2kernel
spire agent and server work well
Try just deploying spire
. Don't deploy NSM yet.
Wait until spire
is Running
hi @glazychev-art and other developers
I really like the nsm project, it feels like an exciting project and I'm in production this week, I deploy deployments-k8s version 1.11.0, and then deploy the officially provided Kernel2IP2Kernel example, but the alpine pod cannot be started. The error message is as follows:
My environment is roughly as follows. The namespace starting with cncp is developed by ourselves, and the others are open source.cncp should have no impact on our nsm