Skip to content

huggingface/inference-playground

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

title emoji colorFrom colorTo sdk pinned app_port
Inference Playground
🔋
blue
pink
docker
false
3000

Hugging Face Inference Playground

Build GitHub Contributor Covenant

This application provides a user interface to interact with various large language models, leveraging the @huggingface/inference library. It allows you to easily test and compare models hosted on Hugging Face, connect to different third-party Inference Providers, and even configure your own custom OpenAI-compatible endpoints.

Local Setup

TL;DR: After cloning, run pnpm i && pnpm run dev --open

Prerequisites

Before you begin, ensure you have the following installed:

  • Node.js: Version 20 or later is recommended.
  • pnpm: Install it globally via npm install -g pnpm.
  • Hugging Face Account & Token: You'll need a free Hugging Face account and an access token to interact with models. Generate a token with at least read permissions from hf.co/settings/tokens.

Follow these steps to get the Inference Playground running on your local machine:

  1. Clone the Repository:

    git clone https://github.com/huggingface/inference-playground.git
    cd inference-playground
  2. Install Dependencies:

    pnpm install
  3. Start the Development Server:

    pnpm run dev
  4. Access the Playground:

    • Open your web browser and navigate to http://localhost:5173 (or the port indicated in your terminal).

Features

  • Model Interaction: Chat with a wide range of models available through Hugging Face Inference.
  • Provider Support: Connect to various third-party inference providers (like Together, Fireworks, Replicate, etc.).
  • Custom Endpoints: Add and use your own OpenAI-compatible API endpoints.
  • Comparison View: Run prompts against two different models or configurations side-by-side.
  • Configuration: Adjust generation parameters like temperature, max tokens, and top-p.
  • Session Management: Save and load your conversation setups using Projects and Checkpoints.
  • Code Snippets: Generate code snippets for various languages to replicate your inference calls.

We hope you find the Inference Playground useful for exploring and experimenting with language models!

About

Resources

License

Code of conduct

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •