Model Management

The Model Management page allows users to view and manage a list of supported models for fine-tuning or inference purposes. Models can be downloaded directly from Hugging Face repositories. Before using the feature, ensure that your device is connected to the internet, as model downloads require access to external servers.

Additionally, on the first download attempt, users will be prompted to provide an Access Token from Hugging Face. This token can be obtained by following Hugging Face's official guidelines for token generation.

Features

  1. Model List:

    • Displays all available models, including their names, sizes, and current processing status.

    • Users can search for specific models using the search bar.

  2. Model Details:

    • Model Name: Specifies the name of the model (e.g., meta-llama/Llama-3.1-8B-Instruct).

    • Support: Use the Phison tag for full parameters and Unsloth for the LoRA mechanism. A model tagged with both supports two different LLM fine-tuning approaches.

    • Size: Shows the storage size of each model (e.g., 29.93 GiB, 262.86 GiB).

    • Process Status: Indicates whether the model is ready to use (✓ for ready).

    • Action: Allows users to manage the model (e.g., download, delete).

  3. Download Process:

    • Displays the download progress for models that are currently being fetched.

    • Provides a clear indication of the model size and estimated time remaining.

    • Supports pause and resumefunctionality for interrupted downloads.

  4. Access Token Requirement:

    • During the initial download, users are required to provide a Hugging Face Access Token for authentication.

    • For token generation, visit the Hugging Face website and follow the official instructions to create and retrieve your personal token.

Supported Models

  • Full Parameters (Phison Supported)

Model Family

Model

Deepseek-ai

DeepSeek-R1-Distill-Llama-70B

DeepSeek-R1-Distill-Qwen-32B

deepseek-llm-7b-chat

Google

gemma-3-27b-it

gemma-3-1b-it

gemma-2-9b-it

Meta-llama

Llama-3.3-70B-Instruct

Llama-3.2-3B-Instruct

Llama-3.1-405B-Instruct

Llama-3.1-8B-Instruct

Llama-3-Taiwan-70B-Instruct

Microsoft

Phi-4-reasoning

phi-4

Phi-3.5-mini-instruct

Mistral

Mistral-Small-3.1-24B-Instruct-2503

Mixtral-8x22B-Instruct-v0.1

Mixtral-8x7B-Instruct-v0.1

Qwen

Qwen3-32B

Qwen3-0.6B

QwQ-32B

Qwen2.5-72B-Instruct

Qwen2-7B

  • LoRA (Unsloth Supported)

Model Family

Model

Deepseek-ai

DeepSeek-R1-Distill-Llama-8B

DeepSeek-R1-Distill-Qwen-14B

DeepSeek-R1-Distill-Qwen-1.5B

DeepSeek-R1-Distill-Qwen-7B

Google

gemma-3-1b-it

gemma-3-4b-it

gemma-3-12b-it

Meta-llama

Llama-3.1-8B-Instruct

Llama-3.2-1B-Instruct

Llama-3.2-3B-Instruct

Mistralai

Mistral-7B-Instruct-v0.3

Mistral-Nemo-Instruct-2407

Qwen

Qwen3-0.6B

Qwen3-0.6B-Base

Qwen3-1.7B

Qwen3-1.7B-Base

Qwen3-4B

Qwen3-4B-Base

Qwen3-8B

Qwen3-8B-Base

Qwen3-14B

Qwen3-14B-Base

Microsoft

Phi-3-mini-4k-instruct

Phi-3-medium-4k-instruct

IBM-granite

granite-3.2-2b-instruct

granite-3.2-8b-instruct

Steps for Access Token

  1. Navigate to the Hugging Face website: https://huggingface.co.

  2. Log in to your Hugging Face account.

  3. Go to your account settings and select the Access Tokens tab.

  4. Generate a new token with the required permissions.

  5. Copy the token and provide it when prompted during the model download process in the application.

For more details, refer to the Hugging Face documentation.

Last updated