logo

InferX AI Function Platform (Lambda Function for Inference)

    --   Serve tens models in one box with ultra-fast (<2 sec) cold start (contact: support@inferx.net)



Pods

tenant namespace pod name state Node name Req Gpu Count Req Gpu vRam Type Standby allocated GPU vRam (MB) allocated GPU Slots
gpu pageable pinned
public TinyLlama public/TinyLlama/TinyLlama-1.1B-Chat-v1.0/100/935 Standby node3 1 6000 MB Restore Blob : 5694 MB Blob : 1254 MB Blob : 5632 MB 0 {}
public TinyLlama public/TinyLlama/TinyLlama-1.1B-Chat-v1.0_13GB/106/936 Standby node3 1 13800 MB Restore Blob : 12602 MB Blob : 1254 MB Blob : 5632 MB 0 {}
public TinyLlama public/TinyLlama/TinyLlama-1.1B-Chat-v1.0_2gpu/109/937 Standby node3 2 13800 MB Restore Blob : 24598 MB Mem : 1546 MB Blob : 11264 MB 0 {}