inference-playground
"--- title: Inference Playground emoji: 🔋 colorFrom: blue colorTo: pink sdk: docker pinned: false app_port: 3000 ---"
Best Scenarios
Technical Constraints
🕸️ Neural Graph Explorer
v15.13📚 Learn More
📈 Interest Trend
* Real-time activity index across HuggingFace, GitHub and Research citations.
🕸️ Neural Graph Explorer
v15.13📚 Learn More
📈 Interest Trend
* Real-time activity index across HuggingFace, GitHub and Research citations.
Benchmark integration for interactive spaces is in preview.
🔬Deep Dive
Expand Details [+]▾
🛠️ Technical Profile
⚡ Hardware & Scale
🌐 Cloud & Rights
🎮 Demo Preview
💻 Usage
docker pull inference-playground git clone https://huggingface.co/spaces/huggingface/inference-playground Space Overview
This application provides a user interface to interact with various large language models, leveraging the @huggingface/inference library. It allows you to easily test and compare models hosted on Hugging Face, connect to different third-party Inference Providers, and even configure your own custom OpenAI-compatible endpoints.
Local Setup
TL;DR: After cloning, run pnpm i && pnpm run dev --open
Prerequisites
Before you begin, ensure you have the following installed:
- Node.js: Version 20 or later is recommended.
- pnpm: Install it globally via
npm install -g pnpm. - Hugging Face Account & Token: You'll need a free Hugging Face account and an access token to interact with models. Generate a token with at least
readpermissions from [hf.co/settings/tokens](https:/
This application provides a user interface to interact with various large language models, leveraging the @huggingface/inference library. It allows you to easily test and compare models hosted on Hugging Face, connect to different third-party Inference Providers, and even configure your own custom OpenAI-compatible endpoints.
Local Setup
TL;DR: After cloning, run pnpm i && pnpm run dev --open
Prerequisites
Before you begin, ensure you have the following installed:
- Node.js: Version 20 or later is recommended.
- pnpm: Install it globally via
npm install -g pnpm. - Hugging Face Account & Token: You'll need a free Hugging Face account and an access token to interact with models. Generate a token with at least
readpermissions from hf.co/settings/tokens.
git clone https://github.com/huggingface/inference-playground.git
cd inference-playground
pnpm install
pnpm run dev
http://localhost:5173 (or the port indicated in your terminal).Features
- Model Interaction: Chat with a wide range of models available through Hugging Face Inference.
- Provider Support: Connect to various third-party inference providers (like Together, Fireworks, Replicate, etc.).
- Custom Endpoints: Add and use your own OpenAI-compatible API endpoints.
- Comparison View: Run prompts against two different models or configurations side-by-side.
- Configuration: Adjust generation parameters like temperature, max tokens, and top-p.
- Session Management: Save and load your conversation setups using Projects and Checkpoints.
- Code Snippets: Generate code snippets for various languages to replicate your inference calls.
- Organization Billing: Specify an organization to bill usage to for Team and Enterprise accounts.
Organization Billing
For Team and Enterprise Hugging Face Hub organizations, you can centralize billing for all users by specifying an organization to bill usage to. This feature allows:
- Centralized Billing: All inference requests can be billed to your organization instead of individual user accounts
- Usage Tracking: Track inference usage across your organization from the organization's billing page
- Spending Controls: Organization administrators can set spending limits and manage provider access
How to Use Organization Billing
X-HF-Bill-To header in API requests when an organization is specifiedRequirements
- You must be a member of a Team or Enterprise Hugging Face Hub organization
- The organization must have billing enabled
- You need appropriate permissions to bill usage to the organization
We hope you find the Inference Playground useful for exploring and experimenting with language models!
4,779 characters total