You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to run the Pytorch Triton example in the repository. I have followed the README.md file and it seems to work fine. However, at step 4 in the Instruction section. I encounter this error after running the Docker error: creating server: Internal - failed to stat file ./model_repository
This is the whole output:
2023-08-23T07:42:13+0000 [INFO] [cli] Service loaded from Bento directory: bentoml.Service(tag="triton-integration-pytorch:withgpu", path="/home/bentoml/bento/")
2023-08-23T07:42:13+0000 [INFO] [cli] Environ for worker 0: set CPU thread count to 12
2023-08-23T07:42:13+0000 [INFO] [cli] Prometheus metrics for HTTP BentoServer from "/home/bentoml/bento" can be accessed at http://localhost:3000/metrics.
2023-08-23T07:42:14+0000 [INFO] [cli] Starting production HTTP BentoServer from "/home/bentoml/bento" listening on http://0.0.0.0:3000 (Press CTRL+C to quit)
W0823 07:42:14.602007 32 pinned_memory_manager.cc:236] Unable to allocate pinned system memory, pinned memory pool will not be available: CUDA driver version is insufficient for CUDA runtime version
I0823 07:42:14.603316 32 cuda_memory_manager.cc:115] CUDA memory pool disabled
Error: Failed to initialize NVML
W0823 07:42:14.633157 32 metrics.cc:785] DCGM unable to start: DCGM initialization error
I0823 07:42:14.633736 32 metrics.cc:757] Collecting CPU metrics
I0823 07:42:14.634028 32 tritonserver.cc:2264]
+----------------------------------+--------------------------------------------------------------------------------------------------------+
| Option | Value |
+----------------------------------+--------------------------------------------------------------------------------------------------------+
| server_id | triton |
| server_version | 2.29.0 |
| server_extensions | classification sequence model_repository model_repository(unload_dependents) schedule_policy model_con |
| | figuration system_shared_memory cuda_shared_memory binary_tensor_data statistics trace logging |
| model_repository_path[0] | ./model_repository |
| model_control_mode | MODE_EXPLICIT |
| startup_models_0 | torchscript_yolov5s |
| strict_model_config | 0 |
| rate_limit | OFF |
| pinned_memory_pool_byte_size | 268435456 |
| response_cache_byte_size | 0 |
| min_supported_compute_capability | 6.0 |
| strict_readiness | 1 |
| exit_timeout | 30 |
+----------------------------------+--------------------------------------------------------------------------------------------------------+
I0823 07:42:14.635074 32 server.cc:261] No server context available. Exiting immediately.
error: creating server: Internal - failed to stat file ./model_repository
Describe the bug
I am trying to run the Pytorch Triton example in the repository. I have followed the README.md file and it seems to work fine. However, at step 4 in the
Instruction
section. I encounter this error after running the Dockererror: creating server: Internal - failed to stat file ./model_repository
This is the whole output:
To reproduce
No response
Expected behavior
No response
Environment
Environment variable
System information
bentoml
: 1.1.1python
: 3.9.17platform
: Linux-6.2.0-26-generic-x86_64-with-glibc2.35uid_gid
: 1000:1000conda
: 23.5.2in_conda_env
: Trueconda_packages
The text was updated successfully, but these errors were encountered: