akash-network / community

Starting point for joining and contributing to building Akash Network
MIT License
55 stars 28 forks source link

[Provider Audit]: provider.dzan.top #608

Closed lmlmvxi9527 closed 4 months ago

lmlmvxi9527 commented 5 months ago

Prerequisite Steps:

1. Make sure your provider has community provider attributes and your contact details (email, website):


  $ provider-services query provider get akash1wq8e6n6dxwqdprh569t9dfnam37775a9yl99mz -o text
  ...
  attributes:
  ...
  - key: host
  value: akash
  - key: tier
  value: community
  info:
info:
  email: lmlmvxi@gmail.com
  website: https://ingress.dzan.top:8085/api/v1/computing/cp

done

2. Make sure your provider *.ingress resolves to your provider IP (ideally worker node IP)

$  host anything.ingress.dzan.top
anything.ingress.dzan.top has address 38.32.189.85

done

3. Please make sure your Akash provider doesn't block any Akash specific ports.

done

Audit Steps:

1. Title the issue: " [Provider Audit]: Provider Address" (e.g. "[Provider Audit]: provider.europlots.com")

2. Wait for response via comments. If no issues during provider Audit, process will be complete, provider should start bidding on leases, and Audit ticket will be closed.

3. If there are issues during the provider Audit, debug those issues, and Audit will be complete.

4. Audit Issue will be closed by core team member.

Leave contact information (optional)

  1. Merlin
  2. Discord :meilin_89585
  3. email: lmlmvxi@gmail.com
shimpa1 commented 5 months ago

hi @lmlmvxi9527 ,

lmlmvxi9527 commented 5 months ago

Hi, I have checked and reopened port range 30000-32767/TCP

shimpa1 commented 5 months ago

good, it works now.

@andy108369 provider akash1wq8e6n6dxwqdprh569t9dfnam37775a9yl99mz ready to be signed

lmlmvxi9527 commented 5 months ago

tks

andy108369 commented 5 months ago

Hi @lmlmvxi9527 :wave: I've noticed your provider advertises rtx1660ti GPU as per its provider attributes:

$ provider-services query provider get akash1wq8e6n6dxwqdprh569t9dfnam37775a9yl99mz -o text
attributes:
...
- key: capabilities/gpu
  value: nvidia
- key: capabilities/gpu/vendor/nvidia/model/rtx1660ti
  value: "true"
- key: capabilities/gpu/vendor/nvidia/model/rtx1660ti/ram/6Gi
  value: "true"
- key: capabilities/gpu/vendor/nvidia/model/rtx1660ti/ram/6Gi/interface/pcie
  value: "true"
- key: capabilities/gpu/vendor/nvidia/model/rtx1660ti/interface/pcie
  value: "true"
...

However, we do not have this GPU in the list for the inventory-operator to discover https://github.com/akash-network/provider-configs/blob/main/devices/pcie/gpus.json

Mind please sharing your output for this command?

kubectl get nodes -o wide --show-labels

Also, have you tried deploying GPU deployment on your provider using https://console.akash.network (or CLI) ? Before I'd sign your provider, we need to make sure it can actually bid on the GPU deployment order requests and without the updated gpus.json file, I doubt it can.. yet

lmlmvxi9527 commented 5 months ago

kubectl get nodes -o wide --show-labels NAME STATUS ROLES AGE VERSION INTERNAL-IP EXTERNAL-IP OS-IMAGE KERNEL-VERSION CONTAINER-RUNTIME LABELS node1 Ready control-plane 130d v1.27.7 10.8.102.131 Ubuntu 22.04.4 LTS 5.15.0-94-generic containerd://1.7.5 AMD=true,NVIDIA-GTX-1660-TI=true,akash.network=true,beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=node1,kubernetes.io/os=linux,node-role.kubernetes.io/control-plane=,node.kubernetes.io/exclude-from-external-load-balancers= node10 Ready 13d v1.27.7 10.8.102.148 Ubuntu 22.04.3 LTS 5.15.0-100-generic containerd://1.7.5 AMD=true,NVIDIA-GTX-1660-TI=true,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.interface.pcie=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.ram.6Gi=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti=1,akash.network=true,allow-nvdp=true,beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=node10,kubernetes.io/os=linux node11 Ready 13d v1.27.7 10.8.102.139 Ubuntu 22.04.4 LTS 5.15.0-112-generic containerd://1.7.5 AMD=true,NVIDIA-GTX-1660-TI=true,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.interface.pcie=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.ram.6Gi=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti=1,akash.network=true,allow-nvdp=true,beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=node11,kubernetes.io/os=linux,nvidia.com/mps.capable=true node2 Ready control-plane 130d v1.27.7 10.8.102.135 Ubuntu 22.04.4 LTS 5.15.0-94-generic containerd://1.7.5 AMD=true,NVIDIA-GTX-1660-TI=true,akash.network=true,beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=node2,kubernetes.io/os=linux,node-role.kubernetes.io/control-plane=,node.kubernetes.io/exclude-from-external-load-balancers= node3 Ready control-plane 130d v1.27.7 10.8.102.136 Ubuntu 22.04.4 LTS 5.15.0-94-generic containerd://1.7.5 AMD=true,NVIDIA-GTX-1660-TI=true,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.interface.pcie=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.ram.6Gi=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti=1,akash.network=true,allow-nvdp=true,beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=node3,kubernetes.io/os=linux,node-role.kubernetes.io/control-plane=,node.kubernetes.io/exclude-from-external-load-balancers= node4 Ready 13d v1.27.7 10.8.102.143 Ubuntu 22.04.4 LTS 5.15.0-112-generic containerd://1.7.5 AMD=true,NVIDIA-GTX-1660-TI=true,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.interface.pcie=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.ram.6Gi=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti=1,akash.network=true,allow-nvdp=true,beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=node4,kubernetes.io/os=linux node5 Ready 13d v1.27.7 10.8.100.40 Ubuntu 22.04.4 LTS 5.15.0-105-generic containerd://1.7.5 AMD=true,NVIDIA-GTX-1660-TI=true,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.interface.pcie=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.ram.6Gi=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti=1,akash.network=true,allow-nvdp=true,beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=node5,kubernetes.io/os=linux node6 Ready 13d v1.27.7 10.8.100.47 Ubuntu 22.04.4 LTS 5.15.0-105-generic containerd://1.7.5 AMD=true,NVIDIA-GTX-1660-TI=true,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.interface.pcie=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.ram.6Gi=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti=1,akash.network=true,allow-nvdp=true,beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=node6,kubernetes.io/os=linux node7 Ready 13d v1.27.7 10.8.100.48 Ubuntu 22.04.4 LTS 5.15.0-105-generic containerd://1.7.5 AMD=true,NVIDIA-GTX-1660-TI=true,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.interface.pcie=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.ram.6Gi=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti=1,akash.network=true,allow-nvdp=true,beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=node7,kubernetes.io/os=linux node8 Ready 13d v1.27.7 10.8.100.53 Ubuntu 22.04.4 LTS 5.15.0-105-generic containerd://1.7.5 AMD=true,NVIDIA-GTX-1660-TI=true,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.interface.pcie=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.ram.6Gi=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti=1,akash.network=true,allow-nvdp=true,beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=node8,kubernetes.io/os=linux node9 Ready 13d v1.27.7 10.8.102.142 Ubuntu 22.04.4 LTS 5.15.0-97-generic containerd://1.7.5 AMD=true,NVIDIA-GTX-1660-TI=true,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.interface.pcie=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti.ram.6Gi=1,akash.network/capabilities.gpu.vendor.nvidia.model.rtx1660ti=1,akash.network=true,allow-nvdp=true,beta.kubernetes.io/arch=amd64,beta.kubernetes.io/os=linux,kubernetes.io/arch=amd64,kubernetes.io/hostname=node9,kubernetes.io/os=linux

lmlmvxi9527 commented 5 months ago

I confirmed that my rtx1660ti gpus is not in the https://github.com/akash-network/provider-configs/blob/main/devices/pcie/gpus.json list. Is this model not supported yet? What should I do? Can I cancel it in the attributes?

lmlmvxi9527 commented 5 months ago

Sir, I have cancelled the gpu related keys in attributes

shimpa1 commented 5 months ago

@lmlmvxi9527 Please follow this: https://akash.network/docs/providers/provider-feature-discovery-gpu-configuration-integration-guide/ The team will add your GPU type and it should be available soon

andy108369 commented 4 months ago

@lmlmvxi9527 adding to what @shimpa1 said above:

I've noticed your ingress isn't configured properly, i.e. when deploying a web server deployment, say nginx, it then should be accessible.

However, the issue is that in your case it attempts redirecting the HTTP request to HTTPS and HTTPS fails. Normally, neither should fail nor there should be HTTP -> HTTPS redirect set.

$ curl -LI 83ir0r0i3hecjbde8mgffdd568.ingress.dzan.top
HTTP/1.1 301 Moved Permanently
Server: nginx/1.18.0 (Ubuntu)
Date: Mon, 08 Jul 2024 11:57:22 GMT
Content-Type: text/html
Content-Length: 178
Connection: keep-alive
Location: https://83ir0r0i3hecjbde8mgffdd568.ingress.dzan.top/
$ curl -k -L https://83ir0r0i3hecjbde8mgffdd568.ingress.dzan.top
<html>
<head><title>504 Gateway Time-out</title></head>
<body>
<center><h1>504 Gateway Time-out</h1></center>
<hr><center>nginx</center>
</body>
</html>
andy108369 commented 4 months ago

@lmlmvxi9527 closing due to inactivity. Please feel free to reopen when ready.