Skip to content
@ModelCloud

ModelCloud.ai

Our mission is to give allow everyone, including bots, unlimited and free access to llm/ai models.

Pinned Loading

  1. GPTQModel GPTQModel Public

    Production ready LLM model compression/quantization toolkit with hw accelerated inference support for both cpu/gpu via HF, vLLM, and SGLang.

    Python 578 82

  2. Device-SMI Device-SMI Public

    Self-contained Python lib with zero-dependencies that give you a unified device properties for gpu, cpu, and npu. No more calling separate tools such as nvidia-smi or /proc/cpuinfo and parsing it y…

    Python 11 1

Repositories

Showing 10 of 12 repositories

Top languages

Loading…

Most used topics

Loading…