WeClone
| Entity Passport | |
| Registry ID | gh-model--xming521--weclone |
| License | AGPL-3.0 |
| Provider | github |
Cite this model
Academic & Research Attribution
@misc{gh_model__xming521__weclone,
author = {xming521},
title = {WeClone Model},
year = {2026},
howpublished = {\url{https://github.com/xming521/weclone}},
note = {Accessed via Free2AITools Knowledge Fortress}
} ๐ฌTechnical Deep Dive
Full Specifications [+]โพ
Quick Commands
git clone https://github.com/xming521/weclone โ๏ธ Nexus Index V2.0
๐ฌ Index Insight
FNI V2.0 for WeClone: Semantic (S:50), Authority (A:0), Popularity (P:76), Recency (R:97), Quality (Q:50).
Verification Authority
๐ What's Next?
Technical Deep Dive
๐ One-stop solution for creating your digital avatar from chat history ๐ก
็ฎไฝไธญๆ๏ฝ English๏ฝ Project Homepage ๏ฝ Documentation
[!IMPORTANT]
Telegram is now supported as a data source !
โจCore Features
- ๐ซ Complete end-to-end solution for creating digital avatars, including chat data export, preprocessing, model training, and deployment
- ๐ฌ Fine-tune LLM using chat history with support for image modal data, infusing it with that authentic "flavor"
- ๐ Integrate with Telegram, WhatsApp (coming soon) to create your own digital avatar
- ๐ก๏ธ Privacy information filtering with localized fine-tuning and deployment for secure and controllable data
๐Features & Notes
Data Source Platform Support
| Platform | Text | Images | Voice | Video | Animated Emojis/Stickers | Links (Sharing) | Quote | Forward | Location | Files |
|---|---|---|---|---|---|---|---|---|---|---|
| Telegram | โ | โ | โ | โ | โ ๏ธConvert to Emoji | โ | โ | โ | โ | โ |
| ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | |
| Discord | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง |
| Slack | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง | ๐ง |
Deployment Platform Support
| Platform | Deployment Support |
|---|---|
| Telegram | โ |
| ๐ง | |
| WeChat (Personal Account) | โ (Based on openclaw-weixin) |
| Discord | โ |
| Slack | โ |
[!IMPORTANT]
- WeClone is still in rapid iteration phase, current performance does not represent final results.
- LLM fine-tuning effectiveness largely depends on model size, quantity and quality of chat data. Theoretically, larger models with more data yield better results.
- The performance of the 7B model is average, while models with 14B or more parameters tend to deliver better results.
- Windows environment has not been rigorously tested. You can use WSL as the runtime environment.
Recent Updates
[25/07/10] Data source added Telegram
[25/06/05] Support for image modal data fine-tuning
Online Fine-Tuning
- Big Model Lab (Lab4AI) (with 50 CNY voucher): https://www.lab4ai.cn/project/detail?utm_source=weclone1&id=ab83d14684fa45d197f67eddb3d8316c&type=project
Hardware Requirements
The project uses Qwen2.5-VL-7B-Instruct model by default with LoRA method for SFT stage fine-tuning. You can also use other models and methods supported by LLaMA Factory.
Estimated VRAM requirements:
| Method | Precision | 7B | 14B | 30B | 70B | xB |
|---|---|---|---|---|---|---|
Full (bf16 or fp16) |
32 | 120GB | 240GB | 600GB | 1200GB | 18xGB |
Full (pure_bf16) |
16 | 60GB | 120GB | 300GB | 600GB | 8xGB |
| Freeze/LoRA/GaLore/APOLLO/BAdam | 16 | 16GB | 32GB | 64GB | 160GB | 2xGB |
| QLoRA | 8 | 10GB | 20GB | 40GB | 80GB | xGB |
| QLoRA | 4 | 6GB | 12GB | 24GB | 48GB | x/2GB |
| QLoRA | 2 | 4GB | 8GB | 16GB | 24GB | x/4GB |
Environment Setup
CUDA installation (skip if already installed, requires version 12.6 or above)
It is recommended to use uv to install dependencies, which is a very fast Python environment manager. After installing uv, you can use the following commands to create a new Python environment and install dependencies.
git clone https://github.com/xming521/WeClone.git && cd WeClone
uv venv .venv --python=3.12
source .venv/bin/activate # windows .venv\Scripts\activate
uv pip install --group main -e .
- Copy the configuration file template and rename it to
settings.jsonc, and make subsequent configuration changes in this file:
cp examples/tg.template.jsonc settings.jsonc
[!NOTE] Training and inference related configurations are unified in the file
settings.jsonc
- Use the following command to test whether the CUDA environment is correctly configured and can be recognized by PyTorch (not needed for Mac):
python -c "import torch; print('CUDA Available:', torch.cuda.is_available());"
- (Optional) Install FlashAttention to accelerate training and inference:
uv pip install flash-attn --no-build-isolation.
Model Download
It is recommended to use Hugging Face to download models, or use the following command:
git lfs install
git clone https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct models/Qwen2.5-VL-7B-Instruct
Data Preparation
Please use Telegram Desktop to export chat records. Click the top right corner in the chat interface, then click "Export chat history". Select Photos for message types and JSON for format. You can export multiple contacts (group chat records are not recommended), then place the exported ChatExport_* in the ./dataset/telegram directory, meaning put different people's chat record folders together in ./dataset/telegram.
Data Preprocessing
- First, modify the
language,platform, andinclude_typein the configuration file according to your needs. - If you use telegram, you need to modify the
telegram_args.my_idin the configuration file to your own telegram user ID. - By default, the project uses Microsoft Presidio to remove
phone numbers, email addresses, credit card numbers, IP addresses, geographic location names, international bank account numbers, cryptocurrency wallet addresses, age information, and generic ID numbersfrom the data, but it cannot guarantee 100% identification. - Therefore, a blocklist
blocked_wordsis provided insettings.jsonc, allowing users to manually add words or phrases they want to filter (the entire sentence containing blocked words will be removed by default).
[!IMPORTANT] ๐จ Please be sure to protect personal privacy and do not leak personal information!
- Execute the following command to process the data. You can modify the
make_dataset_argsin settings.jsonc according to your own chat style.
weclone-cli make-dataset
More Parameter Details: Data Preprocessing
Configure Parameters and Fine-tune Model
- (Optional) Modify
model_name_or_path,template,lora_targetinsettings.jsoncto select other locally downloaded models. - Modify
per_device_train_batch_sizeandgradient_accumulation_stepsto adjust VRAM usage. - You can modify parameters like
num_train_epochs,lora_rank,lora_dropoutintrain_sft_argsbased on your dataset's quantity and quality.
Single GPU Training
weclone-cli train-sft
Multi-GPU Training
Uncomment the deepspeed line in settings.jsonc and use the following command for multi-GPU training:
uv pip install "deepspeed<=0.16.9"
deepspeed --num_gpus=number_of_gpus weclone/train/train_sft.py
Simple Inference with Browser Demo
Test suitable temperature and top_p values, then modify infer_args in settings.jsonc for subsequent inference use.
weclone-cli webchat-demo
Inference Using API
weclone-cli server
Test with Common Chat Questions
Does not include questions asking for personal information, only daily conversation. Test results are in test_result-my.txt.
weclone-cli server
weclone-cli test-model
๐ผ๏ธ Results Showcase
[!TIP] We're looking for interesting examples of native English speakers chatting with WeClone! Feel free to share them with us on Twitter.
๐ค Deploy to Chat Bots
AstrBot
AstrBot is an easy-to-use multi-platform LLM chatbot and development framework โจ Supports Discord, Telegram, Slack, Feishu and other platforms.
Usage steps:
- Deploy AstrBot
- Deploy messaging platforms like Discord, Telegram, Slack in AstrBot
- Execute
weclone-cli serverto start the API service - Add a new service provider in AstrBot, select OpenAI type, fill in the API Base URL according to AstrBot's deployment method (e.g., for docker deployment it might be http://172.17.0.1:8005/v1), fill in the model as gpt-3.5-turbo, and enter any API Key
- Tool calling is not supported after fine-tuning, please turn off the default tools first by sending the command:
/tool off_allon the messaging platform, otherwise the fine-tuned effect won't be visible. - Set the system prompt in AstrBot according to the default_system used during fine-tuning.
[!IMPORTANT] Check the api_service logs to ensure that the large model service request parameters are consistent with those used during fine-tuning as much as possible, and turn off all tool plugin capabilities.
LangBot
LangBot is an easy-to-use open-source LLM chatbot platform suitable for various scenarios. It connects to various global instant messaging platforms. You can set up your IM bot in just 5 minutes.
- Deploy LangBot
- Add a bot (Discord, Telegram, Slack, Lark e.g.) in LangBot
- Execute
weclone-cli serverto start the WeClone API service - Add a new model in the model page, name it
gpt-3.5-turbo, select OpenAI as the provider, fill in the request URL as WeClone's address. For detailed connection methods, refer to the documentation, and enter any API Key.
- Select the model you just added in the pipeline configuration, or modify the prompt configuration
๐ Roadmap
- Support more data sources
- Richer context: including contextual conversations, chat participant information, time, etc.
- Memory support
- Multimodal support: image support already implemented
- Data augmentation
- GUI support
- COT (Chain of Thought) thinking support
Troubleshooting
[Official Documentation FAQ](https://docs.weclone.love/docs/introduce/FAQ.html)
It is also recommended to use DeepWiki for problem solving.
โค๏ธ Contributing
Any Issues/Pull Requests are welcome!
You can contribute by checking Issues or helping review PRs (Pull Requests). For new feature additions, please discuss through Issues first.
Development environment:
uv pip install --group dev -e .
pre-commit install
The project uses pytest for testing, pyright for type checking, and ruff for code formatting.
Before submitting your code, you should run pytest tests to ensure all tests pass.
๐ Acknowledgments
Thanks to the following code contributors and other community members for their contributions
This project also benefits from excellent open source projects such as PyWxDump, LLaMA-Factory, AstrBot, LangBot, and others.
This project is supported by:
โ ๏ธ Disclaimer
[!CAUTION] This project is for learning, research and experimental purposes only. There are significant risks in using it for production environments, please assess carefully. Do not use for illegal purposes, consequences are at your own risk.
[!IMPORTANT]
WeClone is currently not partnered with any platform and has not issued any cryptocurrency. The only official website is: [weclone.love](https://www.weclone.love). Beware of imitations.
Click to view disclaimer terms
1. Use at Your Own Risk
- Users should fully understand and bear all related risks when using this project
- The project authors are not responsible for any direct or indirect losses arising from the use of this project
- Including but not limited to: data loss, financial loss, legal disputes, personal reputation damage, social relationship impact, psychological trauma, career development obstacles, business reputation damage, etc.
2. Production Environment Risk Warning
- Use for commercial purposes or providing external services requires bearing all risks yourself
- All consequences that may result from production environment use (including but not limited to service interruption, data security issues, user complaints, legal liability, etc.) are entirely borne by the user
- It is recommended to conduct thorough testing, verification and risk assessment before using in production environments
3. Model Output Unreliability
- Fine-tuned models may produce inaccurate, harmful or misleading content
- Model outputs do not represent the views or intentions of real persons
- Users should conduct manual review and verification of model outputs
4. Data Security and Privacy
- Users should ensure that uploaded chat records and other data comply with relevant laws and regulations
- Users should obtain appropriate authorization from data-related persons
- This project is not responsible for data leakage or privacy infringement
5. Legal Compliance
- Users should ensure that using this project complies with local laws and regulations
- Involving artificial intelligence, data protection, intellectual property and other related laws
- Users bear the consequences of illegal use
6. Technical Support Limitations
- This project is provided "as is" without any express or implied warranties
- Authors do not promise to provide continuous technical support or maintenance
- No guarantee of project stability, reliability or applicability
Usage Recommendations
Mandatory Bot Identity Identification
When using digital avatars generated by this project, it is strongly recommended to:
- Clearly identify as "AI Bot" or "Digital Avatar" at the beginning of each conversation
- Prominently mark "AI-generated content" in the user interface
- Avoid letting users mistake it for real human conversation, which could cause risks
Risk Assessment Recommendations
If you must use in production environments, it is recommended to:
- Conduct comprehensive security testing
- Establish complete content review mechanisms
- Develop emergency response plans
- Purchase appropriate insurance coverage
- Consult legal professionals for advice
This disclaimer may be revised with project updates, users should regularly check the latest version. Continuing to use this project indicates agreement with the latest disclaimer terms.
Once you download, clone, modify, distribute or use the code or models of this project in any way, it indicates that you have fully read, understood and agreed to unconditionally accept all terms of this disclaimer.
Please carefully read and understand all contents of this disclaimer, ensuring strict compliance with relevant regulations when using this project.
โญ Star History
[!TIP] If this project is helpful to you, or if you are interested in the future development of this project, please give the project a Star, thank you
๐ Quick Start
git clone https://github.com/xming521/WeClone.git && cd WeClone
uv venv .venv --python=3.12
source .venv/bin/activate # windows .venv\Scripts\activate
uv pip install --group main -e .
โ ๏ธ Incomplete Data
Some information about this model is not available. Use with Caution - Verify details from the original source before relying on this data.
View Original Source โ๐ Limitations & Considerations
- โข Benchmark scores may vary based on evaluation methodology and hardware configuration.
- โข VRAM requirements are estimates; actual usage depends on quantization and batch size.
- โข FNI scores are relative rankings and may change as new models are added.
- โ License Unknown: Verify licensing terms before commercial use.
Social Proof
AI Summary: Based on GitHub metadata. Not a recommendation.
๐ก๏ธ Model Transparency Report
Technical metadata sourced from upstream repositories.
๐ Identity & Source
- id
- gh-model--xming521--weclone
- slug
- xming521--weclone
- source
- github
- author
- xming521
- license
- AGPL-3.0
- tags
- chat-history, digital-avatar, llm, qwen, telegram, python
โ๏ธ Technical Specs
- architecture
- null
- params billions
- null
- context length
- null
- pipeline tag
- text-generation
๐ Engagement & Metrics
- downloads
- 0
- stars
- 16,511
- forks
- 0
Data indexed from public sources. Updated daily.