Closed Kanupriyagoyal closed 1 week ago
When on the CLI the command is --collect-cpu-metrics
(dashes instead of underscores).
#805 Questions regarding config search
In this 2nd point. Always getting instance as kind_cpu. In my case its happening reverse as I am getting 'KIND_GPU', which parameter i can pass in command line to make sure it will create KIND_CPU instance
found some PR https://github.com/triton-inference-server/model_analyzer/pull/806
Running model analyzer with:
model-analyzer profile --triton-launch-mode remote --client-protocol grpc --triton-grpc-endpoint {my_ip}:{port} --triton-metrics-url http://{my_ip}:{port}/metrics --profile-models snapml_model --gpus [''] --output-model-repository-path out_models --override-output-model-repository
**[Model Analyzer] No GPUs requested**
[Model Analyzer] Creating model config: snapml_model_config_0
[Model Analyzer] Setting instance_group to [{'count': 1, 'kind': 'KIND_GPU'}]
[Model Analyzer] Setting max_batch_size to 1
[Model Analyzer] Enabling dynamic_batching
[Model Analyzer]
[Model Analyzer] Model snapml_model_config_0 load failed: [StatusCode.INVALID_ARGUMENT] load failed for model 'snapml_model': version 1 is at UNAVAILABLE state: Invalid argument: instance group snapml_model_0 of model snapml_model has kind KIND_GPU but server does not support GPUs;
[Model Analyzer] No changes made to analyzer data, no checkpoint saved.
[Model Analyzer] Creating model config: snapml_model_config_1
[Model Analyzer] Setting instance_group to [{'count': 2, 'kind': 'KIND_GPU'}]
[Model Analyzer] Setting max_batch_size to 1
[Model Analyzer] Enabling dynamic_batching
[Model Analyzer]
[Model Analyzer] Model snapml_model_config_1 load failed: [StatusCode.INVALID_ARGUMENT] load failed for model 'snapml_model': version 1 is at UNAVAILABLE state: Invalid argument: instance group snapml_model_0 of model snapml_model has kind KIND_GPU but server does not support GPUs;
[Model Analyzer] No changes made to analyzer data, no checkpoint saved.
[Model Analyzer] Creating model config: snapml_model_config_2
[Model Analyzer] Setting instance_group to [{'count': 3, 'kind': 'KIND_GPU'}]
[Model Analyzer] Setting max_batch_size to 1
[Model Analyzer] Enabling dynamic_batching
[Model Analyzer]
[Model Analyzer] Model snapml_model_config_2 load failed: [StatusCode.INVALID_ARGUMENT] load failed for model 'snapml_model': version 1 is at UNAVAILABLE state: Invalid argument: instance group snapml_model_0 of model snapml_model has kind KIND_GPU but server does not support GPUs;
[Model Analyzer] No changes made to analyzer data, no checkpoint saved.
[Model Analyzer] Creating model config: snapml_model_config_3
[Model Analyzer] Setting instance_group to [{'count': 4, 'kind': 'KIND_GPU'}]
[Model Analyzer] Setting max_batch_size to 1
[Model Analyzer] Enabling dynamic_batching
[Model Analyzer]
[Model Analyzer] Model snapml_model_config_3 load failed: [StatusCode.INVALID_ARGUMENT] load failed for model 'snapml_model': version 1 is at UNAVAILABLE state: Invalid argument: instance group snapml_model_0 of model snapml_model has kind KIND_GPU but server does not support GPUs;
[Model Analyzer] No changes made to analyzer data, no checkpoint saved.
[Model Analyzer] Creating model config: snapml_model_config_4
[Model Analyzer] Setting instance_group to [{'count': 5, 'kind': 'KIND_GPU'}]
[Model Analyzer] Setting max_batch_size to 1
[Model Analyzer] Enabling dynamic_batching
[Model Analyzer]
[Model Analyzer] Model snapml_model_config_4 load failed: [StatusCode.INVALID_ARGUMENT] load failed for model 'snapml_model': version 1 is at UNAVAILABLE state: Invalid argument: instance group snapml_model_0 of model snapml_model has kind KIND_GPU but server does not support GPUs;
[Model Analyzer] No changes made to analyzer data, no checkpoint saved.
[Model Analyzer]
[Model Analyzer] Done with brute mode search.
@nv-braf is there any option to pass in command line so that it can take KIND_CPU instance only!!
Not on the command line, but you can specify cpu_only
as a flag to the model in the YAML file. Please see our documentation for an example on how to do this: https://github.com/triton-inference-server/model_analyzer/blob/main/docs/config.md#cpu_only
@nv-braf Thanks cpu_only flag i had tried and working fine. I was looking for command line option.
I am using model analyzer 1.42 and server r24.07 1.https://github.com/triton-inference-server/model_analyzer/blob/main/docs/config.md#cli-and-yaml-config-options
List of composing models that should be profiled using CPU instances only
cpu_only_composing_models:
Not able to use as command line options - But able to use it in yaml
Specifies which metric(s) are to be collected.
[ collect_cpu_metrics: | default: false ]
How in quick search mode it will take KIND_CPU instance only. How to make sure that model runs only on cpu instance.
How to avoid gpu metrics warnings also