MaixPy MaixCAM Large Model User Guide
Update history
| Date | Version | Author | Update content |
|---|---|---|---|
| 2026-01-05 | 1.0.0 | lxowalle | Added large model usage documentation |
Introduction
This document provides an overview of how to obtain and use large models, including models for text-to-image, image-to-image, speech-to-text, text-to-speech, and chat.
List of supported large models:
| Supported Model | Supported Platform | Memory Requirement | Description |
|---|---|---|---|
| lcm-lora-sdv1-5-maixcam2 | MaixCAM2 | 4G | Text-to-Image / Image-to-Image |
| lcm-lora-sdv1-5-320x320-maixcam2 | MaixCAM2 | 4G | Text-to-Image / Image-to-Image |
| sensevoice-maixcam2 | MaixCAM2 | 1G | Speech-to-Text |
| whisper-basic-maixcam2 | MaixCAM2 | 1G | Speech-to-Text |
| melotts-maixcam2 | MaixCAM2 | 1G | Text-to-Speech |
| smolvlm-256m-instruct-maixcam2 | MaixCAM2 | 1G | Vision-Language Model |
| InternVL2.5-1B-maixcam2 | MaixCAM2 | 4G | Vision-Language Model |
| Qwen3-VL-2B-Instruct-GPTQ-Int4-AX630C-P320-CTX448-maixcam2 | MaixCAM2 | 4G | Vision-Language Model |
| deepseek-r1-distill-qwen-1.5B-maixcam2 | MaixCAM2 | 4G | Language Model |
| Qwen2.5-1.5B-Instruct-maixcam2 | MaixCAM2 | 4G | Language Model |
| Qwen2.5-0.5B-Instruct-maixcam2 | MaixCAM2 | 4G | Language Model |
Download Methods
Currently, two download methods are provided: cloud storage download and HuggingFace download.
Cloud Storage Download
Download Images (Baidu Netdisk) Extraction code:vjex
From the List of supported large models above, locate the required model and download it from the cloud storage.
For example, for the lcm-lora-sdv1-5-maixcam2 model, download a file similar to lcm-lora-sdv1-5-maixcam2-202601051759.zip, The suffix 202601051759 indicates the model packaging time.
HuggingFace Download
注:
- Downloading from HuggingFace requires a stable network environment. Poor connectivity may result in interrupted downloads
- The following methods can also be executed directly in the terminal of the target platform (e.g., MaixCAM2).
Download via Command Line
Install the download tool:
pip install huggingface_hubSet the download endpoint. The default is
https://huggingface.co
For users in China, it is recommended to usehttps://hf-mirror.com# Linux/MacOS export HF_ENDPOINT=https://hf-mirror.com # or 'https://huggingface.co' # Windows ## cmd set HF_ENDPOINT=https://hf-mirror.com ## PowerShell $env:HF_ENDPOINT = "https://hf-mirror.com"Example for downloading the
lcm-lora-sdv1-5-maixcam2model.
To download another model, replacelcm-lora-sdv1-5-maixcam2with the desired model name.# Download model (new) hf download sipeed/lcm-lora-sdv1-5-maixcam2 --local-dir /root/models # Download model (legacy) huggingface-cli download sipeed/lcm-lora-sdv1-5-maixcam2 --local-dir /root/modelsDownload Using Python
Install the
huggingface_hubpackage:pip install huggingface_hubExample for downloading the
lcm-lora-sdv1-5-maixcam2model. To download other models, replace themodel_namevariable accordingly.# This scripy is used to install models from huggingface # Only support MaixCAM2 platform import os os.environ['HF_ENDPOINT'] = 'https://hf-mirror.com' # or 'https://huggingface.co' from huggingface_hub import snapshot_download from huggingface_hub.utils import tqdm CAPTURE_PROGRESS=False class Tqdm(tqdm): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) if CAPTURE_PROGRESS: print(f"[INIT] {self.desc} | {self.n}/{self.total} ({self.n/self.total*100:.2f}%)") else: print('') def update(self, n=1): super().update(n) if CAPTURE_PROGRESS: print(f"[UPDATE] {self.desc} | {self.n}/{self.total} ({self.n/self.total*100:.2f}%)") else: print('') def close(self): super().close() if CAPTURE_PROGRESS: print(f"[CLOSE] {self.desc} | {self.n}/{self.total} ({self.n/self.total*100:.2f}%)") else: print('') model_name = 'lcm-lora-sdv1-5-maixcam2' repo_id = f'sipeed/{model_name}' local_dir = f'/root/models/{model_name}' snapshot_download( repo_id=repo_id, local_dir=local_dir, # allow_patterns="*.py", tqdm_class=Tqdm, )
Uploading Models to the Board
After downloading, you can upload the model to the board using the scp command. It is recommended to upload models to the /root/models directory.
Using lcm-lora-sdv1-5-maixcam2 as an example:
Confirm the Files to Upload
The directory structure of the
lcm-lora-sdv1-5-maixcam2model is:lcm-lora-sdv1-5-maixcam2 ├── lcm-lora-sdv1-5-maixcam2 # The actual model files to upload ├── README.md ├── README_ZH.md └── launcher.pyThe folder
lcm-lora-sdv1-5-maixcam2under the parent directory is the actual model directory that must be uploaded. Make sure to upload the entire model directory.Upload the Model to the Development Board
Example of uploading the
lcm-lora-sdv1-5-maixcam2model to the board usingscp(assuming the board IP address is 192.168.10.100):Notes:
- It is recommended to upload models via the USB network interface for higher transfer speed. For details on obtaining the USB network IP address, seeWired Connection
scp -r lcm-lora-sdv1-5-maixcam2/lcm-lora-sdv1-5-maixcam2 root@192.168.10.100:/root/models
Using the Models
For instructions on how to use each model, please refer to the corresponding documentation, for example:
Qwen Large Language Model
InternVL Vision-Language Model
SmolVLM Vision-Language Model