logo

InferX AI Function Platform (Lambda Function for Inference)

    --   Serve tens models in one box with ultra-fast (<2 sec) cold start (contact: support@inferx.net)

Action

Pods

tenant namespace pod name state Node name Req Gpu Count Req Gpu vRam Type Standby allocated GPU vRam (MB) allocated GPU Slots
gpu pageable pinned
public THUDM public/THUDM/chatglm3-6b-128k/164/1164 Standby node2 1 13800 MB Restore Blob : 13714 MB Blob : 1264 MB Blob : 7168 MB 0 {}
public THUDM public/THUDM/chatglm3-6b-32k/162/1165 Standby node2 1 13800 MB Restore Blob : 13742 MB Blob : 1102 MB Blob : 7168 MB 0 {}
public THUDM public/THUDM/chatglm3-6b/160/1201 Standby node2 1 13800 MB Restore Blob : 13690 MB Blob : 1264 MB Blob : 7168 MB 0 {}