
TypingMind×Use GLM-5 on Hugging Face with your own API key on TypingMind
Hugging Face Inference is a managed service that gives you access to thousands of open-source AI models — including Meta Llama, Mistral, Qwen, DeepSeek, and more — through a single OpenAI-compatible API.
Here is how to use GLM-5 (zai-org/GLM-5) on Hugging Face with TypingMind.
GLM-5 Overview
| Model Name | GLM-5 |
| Provider | Hugging Face |
| Model ID | zai-org/GLM-5 |
| Release Date | Feb 11, 2026 |
| Last Updated | Feb 11, 2026 |
| Context Window | 202,752 tokens |
| Max Output | 131,072 tokens |
| Pricing /1M tokens | $1 input $3.2 output $0.2 cache read |
| Input Modalities | text |
| Output Modalities | text |
| Capabilities | Open WeightsReasoningTool CallingTemperature Control |
Complete Setup Guide
Create a Hugging Face Account
Go to huggingface.co and create a free account if you don't have one. Hugging Face provides access to thousands of open-source AI models through their Inference API — no complex setup required.
Generate an Access Token
Go to your Hugging Face Access Tokens settings and click "+ Create new token".
Select Fine-grained as the token type and give it a name (e.g. "typingmind"). Under the Inference section, enable the following permissions:
- ✅ Make calls to Inference Providers — required to use Hugging Face hosted models
- ☑️ Make calls to your Inference Endpoints — optional, for dedicated endpoints
- ☑️ Manage your Inference Endpoints — optional, for endpoint management

Click "Create token" and copy the generated token — you'll need it in the next step.
⚠️ Important: Keep your access token secure and never share it publicly. Store it safely as you'll need it to connect with TypingMind.
Add a Custom Model to TypingMind
Go to typingmind.com, click the Models tab on the left workspace bar, then click "Add Custom Model" in the top right corner.

Select "Create Manually" as the creation method.

Fill in the following fields:
| Field | Value |
|---|---|
| Name | HuggingFace GLM-5(or any name you prefer) |
| API Type | OpenAI Chat Completions API |
| Endpoint URL | https://router.huggingface.co/v1/chat/completions |
| Model ID | zai-org/GLM-5 |
| Context Length | 202,752for GLM-5 |
Scroll down to the Authentication section and set:
| Field | Value |
|---|---|
| Authentication Type | API Key via HTTP Header |
| Header Key | Authorization |
| Header Value | Bearer {{API key}} |

Under Model Capabilities, enable the features supported by your chosen model:
- ✅ Support System Role — recommended for most models
- ✅ Support Streaming Output — recommended for all models
- ☑️ Support Plugins — enable if the model supports function/tool calling
- ☑️ Support Image Input — enable for vision-capable models
- ☑️ Support Thinking — enable for reasoning/thinking models

After filling in all fields, click "Test". If everything is set up correctly, you will see the message "Nice, the endpoint is working!"
Click "+ Add Model" to finish.
Use the Model
Now you can select your newly added Hugging Face model from the model picker and start chatting!
- Select GLM-5 from the model dropdown menu
- Start typing your message in the chat input
- Enjoy faster responses and better features than the official interface
- Switch between different AI models as needed

💡 Pro tips for better results:
- Use specific, detailed prompts for better responses (How to use Prompt Library)
- Create AI agents with custom instructions for repeated tasks (How to create AI Agents)
- Use plugins to extend GLM-5 capabilities (How to use plugins)
Hugging Face FAQ
Why use GLM-5 on Hugging Face with TypingMind instead of the official Hugging Face interface?
Hugging Face's web interface is built for model exploration and research — not for everyday AI chat. TypingMind gives you a polished, full-featured chat interface on top of Hugging Face Inference: conversation history, prompt library, AI agents, plugin support, file uploads, and more. You bring your own HF access token and pay only for what you use, with no TypingMind subscription required.
Do I need a Hugging Face subscription to use GLM-5?
No subscription required for many models. Hugging Face offers a free tier with rate limits, and a PRO plan ($9/month) for higher usage. Many open-weight models are completely free to use. You only pay for inference on paid models. TypingMind is also a one-time purchase, not a subscription.
How much does it cost to use GLM-5 via Hugging Face?
GLM-5 on Hugging Face costs $1/1M input tokens and $3.2/1M output tokens (cache read: $0.2/1M tokens). Many open-weight models are free to use. You can monitor usage in your Hugging Face billing dashboard.
What is Hugging Face Inference and how does it work with TypingMind?
Hugging Face Inference is a managed service that lets you run thousands of open-source AI models via a simple OpenAI-compatible API. TypingMind lets you connect to Hugging Face using your own access token, so you can chat with models like GLM-5 directly — without any intermediary subscription.
Can I use other Hugging Face models besides GLM-5 in TypingMind?
Yes! Hugging Face gives you access to thousands of open-source models from Meta, Mistral, Qwen, DeepSeek, and more. You can add multiple models as custom models in TypingMind and switch between them instantly in any conversation.
Is my data private when using Hugging Face with TypingMind?
Yes. TypingMind stores your conversations locally — in your browser (web) or on your device (desktop app). Hugging Face processes inference requests through their secure infrastructure. Your data never passes through TypingMind servers.
What token permissions do I need for the Hugging Face access token?
You need a Fine-grained token with "Make calls to Inference Providers" permission enabled under the Inference section. Optionally, you can also enable "Make calls to your Inference Endpoints" if you use dedicated endpoints.
Can I use GLM-5 on Hugging Face for commercial projects?
It depends on the model's license. GLM-5 is an open-weight model — check its specific license on Hugging Face for commercial use terms. TypingMind supports commercial use.












