logo

InferX AI Function Platform (Lambda Function for Inference)

    --   Serve tens models in one box with ultra-fast (<2 sec) cold start (contact: support@inferx.net)

Action

Models

tenant namespace model name gpu count vram (GB) cpu memory (GB) standby state snapshot nodes revision
gpu pageable pinned
public openbmb MiniCPM-2B-dpo-bf16 1 13.8 12.0 28.0 Blob Blob Blob Normal ['node2'] 208
public openbmb MiniCPM-2B-sft-bf16 1 9.0 12.0 24.0 Blob Blob Blob Normal ['node2'] 210

Summary

Model Count

2

Required GPU Count

2

Required VRAM (GB)

22.8 GB

Required CPU Cores

24.0

Required Memory (GB)

52.0 GB