Skip to content

[Bug] add_sub example failed #120

@fumihwh

Description

@fumihwh

Problem:
add_sub example failed. log as below.
INFO[client.py:82] Model add_sub_i0 load failed: [StatusCode.INTERNAL] failed to load 'add_sub_i0', no version is available
It's an INFO, but finally I get empty metrices:

Server Only:
Model           GPU ID   Batch   Concurrency   Max GPU Memory Usage(MB)   Max GPU Memory Available(MB)   Max GPU Utilization(%)
triton-server   0        0       0             166.0                      14943.0                        0.0
triton-server   1        0       0             166.0                      14943.0                        0.0
triton-server   2        0       0             166.0                      14943.0                        0.0
triton-server   3        0       0             166.0                      14943.0                        0.0

Models (GPU Metrics):
Model   GPU ID   Batch   Concurrency   Model Config Path   Max GPU Memory Usage(MB)   Max GPU Memory Available(MB)   Max GPU Utilization(%)

Models (Inference):
Model   Batch   Concurrency   Model Config Path   Throughput(infer/sec)   Average Latency(us)   Max RAM Usage(MB)   Max RAM Available(MB)

Models (GPU Metrics - Failed Constraints):
Model   GPU ID   Batch   Concurrency   Model Config Path   Max GPU Memory Usage(MB)   Max GPU Memory Available(MB)   Max GPU Utilization(%)

Models (Inference - Failed Constraints):
Model   Batch   Concurrency   Model Config Path   Throughput(infer/sec)   Average Latency(us)   Max RAM Usage(MB)   Max RAM Available(MB)

All I've done is:

  • Pull image from ngc
    • nvcr.io/nvidia/tritonserver:21.03-py3-sdk as doc says
    • nvcr.io/nvidia/tritonserver:21.03-py3 for --triton-launch-mode=docker
  • Clone model_analyzer repo to $HOME
    • cd $HOME && git clone https://github.com/triton-inference-server/model_analyzer.git
  • Start docker container as doc says: docker run -it --rm --gpus all \ -v /var/run/docker.sock:/var/run/docker.sock \ -v $HOME/model_analyzer/examples/quick-start:/quick_start_repository \ --net=host --name model-analyzer \ nvcr.io/nvidia/tritonserver:21.03-py3-sdk /bin/bash
  • Under /workspace folder, run model-analyzer -m /quick_start_repository -n add_sub --triton-launch-mode=docker --triton-version=21.03-py3 --export-path=analysis_results --log-level=DEBUG --override-output-model-repository

Did I miss something?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions