logo

InferX AI Function Platform (Lambda Function for Inference)

    --   Serve tens models in one box with ultra-fast (<2 sec) cold start (contact: support@inferx.net)

Action

Pods

tenant namespace pod name state Node name Req Gpu Count Req Gpu vRam Type Standby allocated GPU vRam (MB) allocated GPU Slots
gpu pageable pinned
public microsoft public/microsoft/Phi-3-mini-128k-instruct/187/1185 Standby node2 1 13000 MB Restore Blob : 12120 MB Blob : 1254 MB Blob : 4096 MB 0 {}
public microsoft public/microsoft/Phi-3-mini-4k-instruct/185/1186 Standby node2 1 13000 MB Restore Blob : 12108 MB Blob : 1258 MB Blob : 4096 MB 0 {}