logo

InferX AI Function Platform (Lambda Function for Inference)

    --   Serve tens models in one box with ultra-fast (<2 sec) cold start (contact: support@inferx.net)

Action

Pods

tenant namespace pod name state Node name Req Gpu Count Req Gpu vRam Type Standby allocated GPU vRam (MB) allocated GPU Slots
gpu pageable pinned
public deepseek-ai public/deepseek-ai/DeepSeek-R1-Distill-Llama-8B/230/1178 Standby node2 2 13800 MB Restore Blob : 24436 MB Blob : 1684 MB Blob : 8192 MB 0 {}
public deepseek-ai public/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B/232/1179 Standby node2 1 13000 MB Restore Blob : 11954 MB Blob : 1362 MB Blob : 7168 MB 0 {}
public deepseek-ai public/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B/234/1223 Ready node2 2 13800 MB Restore Blob : 24270 MB Blob : 1630 MB Blob : 14336 MB 13824 {'0': {'contextCnt': 1, 'slotCnt': 54}, '1': {'contextCnt': 1, 'slotCnt': 54}}
public deepseek-ai public/deepseek-ai/deepseek-llm-7b-chat/226/1181 Standby node2 1 14600 MB Restore Blob : 14748 MB Blob : 1304 MB Blob : 7680 MB 0 {}
public deepseek-ai public/deepseek-ai/deepseek-llm-7b-chat_2gpu/228/1182 Standby node2 2 14200 MB Restore Blob : 25552 MB Blob : 1784 MB Blob : 15360 MB 0 {}
public deepseek-ai public/deepseek-ai/deepseek-math-7b-instruct/236/1195 Standby node2 2 13800 MB Restore Blob : 24594 MB Blob : 1888 MB Blob : 15360 MB 0 {}