InferX — Serverless GPU Inference Platform for Production Workloads

Snapshot ID Nodename State GPU Pageable (MB) Pinned (MB) Docker Image Name Build ID
ID Memory Size (MB)
public/ActionAnalytics/CR-70B/79#g8398d4 g8398d4 Ready
0
1
2
3
71202
71202
71202
71202
5460.0 4.328125 vllm/vllm-openai:v0.9.0 [19, 0, 139, 214, 177, 226, 45, 20, 180, 184, 142, 158, 226, 10, 61, 52, 155, 112, 81, 10]
public/Qwen/IntelliAsk-Qwen3-32B-450-Merged/91#g8398d4 g8398d4 Ready
0
1
60725
60725
3644.0 4.1640625 vllm/vllm-openai:v0.9.0 [19, 0, 139, 214, 177, 226, 45, 20, 180, 184, 142, 158, 226, 10, 61, 52, 155, 112, 81, 10]
public/Trial/Huihui-Qwen3-Next-80B-A3B-Thinking-abliterated/85#g8398d4 g8398d4 Ready
0
1
2
3
47135
47135
47135
47135
7266.0 33.87890625 vllm/vllm-openai:v0.15.0 [19, 0, 139, 214, 177, 226, 45, 20, 180, 184, 142, 158, 226, 10, 61, 52, 155, 112, 81, 10]
public/Trial/L3.3-70B-Loki-V2.0/94#g8398d4 g8398d4 Ready
0
1
74291
74291
3604.0 2.1640625 vllm/vllm-openai:v0.9.0 [19, 0, 139, 214, 177, 226, 45, 20, 180, 184, 142, 158, 226, 10, 61, 52, 155, 112, 81, 10]
public/Trial/translategemma-27b-it-FP8-Dynamic/112#g8398d4 g8398d4 Ready
0
32023
3188.0 0.20703125 vllm/vllm-openai:v0.9.0 [19, 0, 139, 214, 177, 226, 45, 20, 180, 184, 142, 158, 226, 10, 61, 52, 155, 112, 81, 10]