logo

InferX AI Function Platform (Lambda Function for Inference)

    --   Serve tens models in one box with ultra-fast (<2 sec) cold start (contact: support@inferx.net)



Pods

tenant namespace pod name state Node name Req Gpu Count Req Gpu vRam Type Standby allocated GPU vRam (MB) allocated GPU Slots
gpu pageable pinned
public THUDM public/THUDM/chatglm3-6b-128k/148/932 Standby node3 1 13800 MB Restore Blob : 13714 MB Blob : 1260 MB Blob : 7168 MB 0 {}
public THUDM public/THUDM/chatglm3-6b-32k/146/933 Standby node3 1 13800 MB Restore Blob : 13742 MB Blob : 1098 MB Blob : 7168 MB 0 {}
public THUDM public/THUDM/chatglm3-6b/144/934 Standby node3 1 13800 MB Restore Blob : 13690 MB Blob : 1256 MB Blob : 7168 MB 0 {}