logo

InferX AI Function Platform (Lambda Function for Inference)

    --   Serve tens models in one box with ultra-fast (<2 sec) cold start (contact: support@inferx.net)

Action

Models

tenant namespace model name gpu count vram (GB) cpu memory (GB) standby state snapshot nodes revision
gpu pageable pinned
public deepseek-ai DeepSeek-R1-Distill-Llama-8B 2 13.8 20.0 60.0 Blob Blob Blob Normal ['node2'] 230
public deepseek-ai DeepSeek-R1-Distill-Qwen-1.5B 1 13.0 20.0 50.0 Blob Blob Blob Normal ['node2'] 232
public deepseek-ai DeepSeek-R1-Distill-Qwen-7B 2 13.8 20.0 60.0 Blob Blob Blob Normal ['node2'] 234
public deepseek-ai deepseek-llm-7b-chat 1 14.6 20.0 60.0 Blob Blob Blob Normal ['node2'] 226
public deepseek-ai deepseek-llm-7b-chat_2gpu 2 14.2 20.0 60.0 Blob Blob Blob Normal ['node2'] 228
public deepseek-ai deepseek-math-7b-instruct 2 13.8 20.0 60.0 Blob Blob Blob Normal ['node2'] 236

Summary

Model Count

6

Required GPU Count

10

Required VRAM (GB)

138.8 GB

Required CPU Cores

120.0

Required Memory (GB)

350.0 GB