logo

InferX AI Function Platform (Lambda Function for Inference)

    --   Serve tens models in one box with ultra-fast (<2 sec) cold start (contact: support@inferx.net)

Action

Models

tenant namespace model name gpu count vram (GB) cpu memory (GB) standby state snapshot nodes revision
gpu pageable pinned
public allenai OLMo-1B-hf 1 14.6 12.0 50.0 Blob Blob Blob Normal ['node2'] 219
public allenai OLMo-1B-hf_2gpu 2 14.6 12.0 50.0 Blob Blob Blob Normal ['node2'] 221
public allenai OLMo-7B-hf 2 13.8 20.0 50.0 Blob Blob Blob Normal ['node2'] 223

Summary

Model Count

3

Required GPU Count

5

Required VRAM (GB)

71.4 GB

Required CPU Cores

44.0

Required Memory (GB)

150.0 GB