logo

InferX AI Function Platform (Lambda Function for Inference)

    --   Serve tens models in one box with ultra-fast (<2 sec) cold start (contact: support@inferx.net)

Action

Pods

tenant namespace pod name state Node name Req Gpu Count Req Gpu vRam Type Standby allocated GPU vRam (MB) allocated GPU Slots
gpu pageable pinned
public Qwen public/Qwen/Qwen2.5-1.5B/146/1152 Standby node2 1 8000 MB Restore Blob : 7588 MB Blob : 1314 MB Blob : 7168 MB 0 {}
public Qwen public/Qwen/Qwen2.5-7B-Instruct-1M/156/1213 Standby node2 2 13800 MB Restore Blob : 24228 MB Blob : 1586 MB Blob : 14336 MB 0 {}
public Qwen public/Qwen/Qwen2.5-7B-Instruct-GPTQ-Int8/144/1154 Standby node2 1 14200 MB Restore Blob : 13164 MB Blob : 1314 MB Blob : 7168 MB 0 {}
public Qwen public/Qwen/Qwen2.5-Coder-1.5B-Instruct/1208/1220 Standby node2 1 6000 MB Restore Blob : 5762 MB Blob : 1314 MB Blob : 7168 MB 0 {}
public Qwen public/Qwen/Qwen2.5-Coder-14B-Instruct-GPTQ-Int8/142/1228 Standby node2 2 13800 MB Restore Blob : 25420 MB Blob : 1626 MB Blob : 12288 MB 0 {}
public Qwen public/Qwen/Qwen2.5-Coder-3B/137/1222 Standby node2 1 10000 MB Restore Blob : 9622 MB Blob : 1310 MB Blob : 4608 MB 0 {}
public Qwen public/Qwen/Qwen2.5-Coder-7B-Instruct/140/1224 Standby node2 2 13800 MB Restore Blob : 24240 MB Blob : 1574 MB Blob : 14336 MB 0 {}
public Qwen public/Qwen/Qwen2.5-Math-1.5B-Instruct/825/1227 Standby node2 1 8000 MB Restore Blob : 7616 MB Blob : 1338 MB Blob : 7168 MB 0 {}
public Qwen public/Qwen/Qwen2.5-Math-1.5B/150/1215 Standby node2 1 8000 MB Restore Blob : 7620 MB Blob : 1314 MB Blob : 7168 MB 0 {}
public Qwen public/Qwen/Qwen2.5-Math-7B-Instruct/841/1161 Standby node2 2 13800 MB Restore Blob : 24218 MB Blob : 1574 MB Blob : 14336 MB 0 {}
public Qwen public/Qwen/Qwen2.5-Math-7B/154/1203 Standby node2 2 13800 MB Restore Blob : 24218 MB Blob : 1586 MB Blob : 14336 MB 0 {}