Open Itime-ren opened 1 week ago
@Itime-ren Try adding --disable-ipv6
to your llama stack run
command.
- (llamastack-stack-3.2-1B) root@720:~/.llama/checkpoints# llama stack run stack-3.2-1B --port 5000 --disable-ipv6 It doesn't work,no difference
Can you share the full run.yaml? Seems to be in this location /root/.llama/builds/conda/stack-3.2-1B-run.yaml
Can you share the full run.yaml? Seems to be in this location
/root/.llama/builds/conda/stack-3.2-1B-run.yaml
I have read the source code and saw that the default port is 5000. I ran “ llama stack run stack-3.2-1B --port 5000 -- disable-ipv6”. My system's 5000 port was not occupied, but the result was still dynamically assigned ports and IPv6 was enabled by default, The detailed system output is as described above
/root/.llama/builds/conda/stack-3.2-1B-run.yaml,as follow
version: v1
built_at: '2024-10-06T02:16:39.007013'
image_name: stack-3.2-1B
docker_image: null
conda_env: stack-3.2-1B
apis_to_serve:
- memory
- inference
- safety
- models
- agents
- memory_banks
- shields
api_providers:
inference:
providers:
- meta-reference
safety:
providers:
- meta-reference
agents:
provider_type: meta-reference
config:
persistence_store:
namespace: null
type: sqlite
db_path: /root/.llama/runtime/kvstore.db
memory:
providers:
- meta-reference
telemetry:
provider_type: meta-reference
config: {}
routing_table:
inference:
- provider_type: meta-reference
config:
model: Llama3.2-1B-Instruct
quantization: null
torch_seed: null
max_seq_len: 4096
max_batch_size: 1
routing_key: Llama3.2-1B-Instruct
safety:
- provider_type: meta-reference
config:
llama_guard_shield: null
enable_prompt_guard: false
routing_key:
- llama_guard
- code_scanner_guard
- injection_shield
- jailbreak_shield
memory:
- provider_type: meta-reference
config: {}
routing_key: vector
downloaded the 1B model from Huggingface and encountered an error while running it. The following is the configuration process, and I am puzzled as to why I need to link it to the address [: ffff: 0.0.2.208]: 48461