Open
Description
Describe the bug
This issue is related to the issue JVM bug 82 in sagemaker-inference-toolkit
To reproduce
Clone the SaeMaker example
Deploy the model using the same endpoint.
Check CloudWatch logs and the number of CPU cores detected will be like Number of CPUs: 1
JVM detect the CPU count as 1 when more CPUs are available for the container.
Expected behavior
The CPU count from CloudWatch should match the CPU count for the used instance. For example, 4 if the instance is ml.m4.xlarge
System information
Container: pytorch-inference:1.7-cpu-py3 and pytorch-inference:1.7-gpu-py3
SageMaker inference v1.1.2
Additional context
This clearly does not allow the usage of all CPUs on the instance for Sagemaker Inference
Metadata
Metadata
Assignees
Labels
No labels