Closed buptliuwei closed 6 years ago
I re-compiled the TX2 kernel and loaded some kernel modules needed for netfliter and weave. The problem is solved.
@buptliuwei Which kernel modules did you add? I have the same issue with the TX2 but cannot pinpoint the kernel modules for weave.
@martwetzels Hi, this is my modules after re-complied
nf_conntrack_netlink 24755 0
xt_nat 2320 5
xt_recent 10058 2
ipt_REJECT 1951 1
nf_reject_ipv4 3438 1 ipt_REJECT
ip_set 33915 0
nfnetlink 7318 2 ip_set,nf_conntrack_netlink
xt_comment 1348 32
xt_mark 1663 5
fuse 83099 2
ipt_MASQUERADE 2115 2
nf_nat_masquerade_ipv4 2931 1 ipt_MASQUERADE
iptable_nat 2285 1
nf_nat_ipv4 6554 1 iptable_nat
xt_addrtype 3298 3
iptable_filter 2119 1
ip_tables 18322 2 iptable_filter,iptable_nat
xt_conntrack 3551 3
nf_nat 16285 3 nf_nat_ipv4,xt_nat,nf_nat_masquerade_ipv4
br_netfilter 13923 0
overlay 33899 2
openvswitch 85585 2
bcmdhd 7447670 0
pci_tegra 60337 0
bluedroid_pm 11195 0
and you can read my gist . Hope it works.
Thanks! I just found the requirement for openvswitch and vxlan on a weave documentation page. Your gist is very useful, too bad it already took me a few hours to reinvent the wheel this week. Do you mind if I do a detailed write-up on Medium?
Btw, did you also manage to get the GPU capabilities visible on the node within the cluster from the TX2?
@martwetzels Never mind. We are working hard to get the GPU capabilities visible.
@buptliuwei hi,brother,where is your gist,I can not find it,can you give me the url?
@StupidYe hi this is my gist: https://gist.github.com/buptliuwei/8a340cc151507cb48a071cda04e1f882
@StupidYe 你好,gist没有邮件提醒,现在才看到,不好意思啊。首先proxy的manifest,我是这样做的,通过编辑器打开原来的kube-proxy文件,然后复制到自己新建的文本中,改成kube-proxy-arm,然后做一些修改,主要是nodeselector。关于flannel部署失败的问题,从报的error上看也是不能路由到service ip。很大关系和kube-proxy有关,我不知道你arm上装的系统是什么,ubuntu的话,可能是有些内核模块没打开,flannel也是overlay的解决方案,需要openvswitch等内核模块的支持,你可以lsmod下,看看模块是否都加载了。最后,kubeadm的文档里面建议在arm上部署网络插件的话,weave的兼容性最好。所以你也可以考虑下weave.
@buptliuwei 非常谢谢你的回复,我会尝试使用weave。Thanks
Thanks! I just found the requirement for openvswitch and vxlan on a weave documentation page. Your gist is very useful, too bad it already took me a few hours to reinvent the wheel this week. Do you mind if I do a detailed write-up on Medium?
Btw, did you also manage to get the GPU capabilities visible on the node within the cluster from the TX2?
hi buddy! Have you made any progress on getting the GPU capabilities visible on the node within the cluster from the TX2? Beacause of the lack of official support ,I think it is a hard work to manage the GPU on TX2 nodes by k8s master. I am getting stuck in , do you mind giving me some instructions ?
@yeliuang I did not proceed with getting the GPUs visible in K8s because @buptliuwei said he was working on it; it already cost me quite some time. To finish up the project we used a different approach, but I am still interested in getting this to work.
Just for the record (as I've been struggling quite some time to get this running) here's my working config:
CONFIG_NETFILTER_XT_SET=m
CONFIG_NETFILTER_XT_MATCH_MULTIPORT=m
CONFIG_NETFILTER_XT_MATCH_PHYSDEV=m
CONFIG_NETFILTER_XT_MATCH_RECENT=m
CONFIG_NETFILTER_XT_TARGET_REDIRECT=m
CONFIG_IP_SET=m
CONFIG_IP_SET_MAX=256
CONFIG_IP_SET_HASH_IP=m
CONFIG_IP_SET_HASH_NET=m
CONFIG_NF_NAT_REDIRECT=m
CONFIG_IP_NF_TARGET_REDIRECT=m
I cannot say for sure if all of those modules are necessary, my issue (non accessible services running on the same minion) has been resolved with adding xt_physdev (after some tedious iptables debugging).
Is this a BUG REPORT or FEATURE REQUEST?
Choose one: BUG REPORT or FEATURE REQUEST /kind bug
Versions
kubeadm version (use
kubeadm version
):1.10.0Environment:
kubectl version
):1.10.0uname -a
): Linux ubuntu 4.13.0-39-generic # 44~16.04.1-Ubuntu SMP Thu Apr 5 16:43:10 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux for amd64 && Linux tegra-ubuntu 4.4.38 # 1 SMP PREEMPT Sun Apr 22 02:51:59 UTC 2018 aarch64 aarch64 aarch64 GNU/Linux for armWhat happened?
my iptables-save shows that kube-proxy did not set up a rule for 10.96.0.1 in arm node. But the kube-proxy pod in arm node is running.
and i checked the logs for kube-proxy . The output is:
What you expected to happen?
the iptables output should be:
How to reproduce it (as minimally and precisely as possible)?
I create a kube-proxy for arm nodes using "node selector". such as Multiplatform (amd64 and arm) Kubernetes cluster setup
Anything else we need to know?
Because of this, i could not deploy weave.