# X-AnyLabeling **Repository Path**: opencvsdk/X-AnyLabeling ## Basic Information - **Project Name**: X-AnyLabeling - **Description**: X-AnyLabeling 是一款基于AI推理引擎和丰富功能特性于一体的强大辅助标注工具,其专注于实际应用,致力于为多模态数据工程师提供工业级的一站式解决方案,可自动快速进行各种复杂任务的标定。 - **Primary Language**: Unknown - **License**: GPL-3.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2026-04-09 - **Last Updated**: 2026-04-28 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README

X-AnyLabeling

[English](README.md) | [简体中文](README_zh-CN.md)

![](https://user-images.githubusercontent.com/18329471/234640541-a6a65fbc-d7a5-4ec3-9b65-55305b01a7aa.png)
Auto-Training
Auto-Labeling
Detect Anything
Segment Anything
Promptable Concept Grounding
VQA
Chatbot
Image Classifier
OCR
## 🥳 What's New - `2026-04-26`: Added client-side (local ONNX) support for [SAM 3](./examples/grounding/sam3/README.md) text-grounded segmentation. - `2026-04-25`: Added support for the [GeCo2](./examples/counting/geco2/README.md) object counting model. - `2026-04-19`: Added annotation check status workflow for reviewing labeled images. - `2026-04-19`: Added support for TensorRT backend inference for YOLO models. - `2026-04-18`: Added support exporting annotation visualizations as images or videos. - `2026-04-18`: Added support for [PaddleOCR](./docs/en/paddle_ocr.md) document parsing and intelligent text recognition annotation panel. - `2026-04-01`: Added support for Japanese and Korean UI languages (`ja_JP`, `ko_KR`). - `2026-03-22`: Added support for built-in Settings in the GUI for directly adjusting common options. - `2026-03-10`: Added support for 3D Cuboid shape annotation from rectangle. - `2026-03-01`: Completed the PyQt5-to-PyQt6 upgrade refactor (Beta release), along with additional feature fixes and optimizations. - For more details, please refer to the [CHANGELOG](./CHANGELOG.md) ## X-AnyLabeling **X-AnyLabeling** is a powerful annotation tool that integrates an AI engine for fast and automatic labeling. It's designed for multi-modal data engineers, offering industrial-grade solutions for complex tasks. Also, we highly recommend trying out [X-AnyLabeling-Server](https://github.com/CVHub520/X-AnyLabeling-Server), a simple, lightweight, and extensible framework that enables remote inference capabilities for X-AnyLabeling. ## Features - Supports remote inference service. - Processes both `images` and `videos`. - Supports flexible backends including `ONNX Runtime`, `TensorRT`, and `OpenCV DNN`. - Supports UI localization in `English`, `Chinese`, `Japanese`, and `Korean`. - Allows custom models and secondary development. - Supports one-click inference for all images in the current task. - Supports import/export for formats like `COCO`, `VOC`, `YOLO`, `DOTA`, `MOT`, `MASK`, `PPOCR`, `MMGD`, `VLM-R1`, `ShareGPT`. - Handles tasks like `classification`, `detection`, `segmentation`, `caption`, `rotation`, `tracking`, `estimation`, `ocr`, `vqa`, `grounding`, `document parsing` and so on. - Supports diverse annotation styles: `polygons`, `rectangles`, `cuboids`, `rotated boxes`, `quadrilaterals`, `circles`, `lines`, `line strips`, `points`, and annotations for `text detection`, `recognition`, and `KIE`. ### Model library | **Task Category** | **Supported Models** | | :--- | :--- | | 🖼️ Image Classification | YOLOv5-Cls, YOLOv8-Cls, YOLO11-Cls, InternImage, PULC | | 🎯 Object Detection | YOLOv5/6/7/8/9/10, YOLO11/12/26, YOLOX, YOLO-NAS, D-FINE, DAMO-YOLO, Gold_YOLO, RT-DETR, RF-DETR, DEIMv2 | | 🖌️ Instance Segmentation | YOLOv5-Seg, YOLOv8-Seg, YOLO11-Seg, YOLO26-Seg, Hyper-YOLO-Seg, RF-DETR-Seg | | 🏃 Pose Estimation | YOLOv8-Pose, YOLO11-Pose, YOLO26-Pose, DWPose, RTMO | | 👣 Tracking | Bot-SORT, ByteTrack, SAM2/3-Video | | 🔄 Rotated Object Detection | YOLOv5-Obb, YOLOv8-Obb, YOLO11-Obb, YOLO26-Obb | | 📏 Depth Estimation | Depth Anything | | 🧩 Segment Anything | SAM 1/2/3, SAM-HQ, SAM-Med2D, EdgeSAM, EfficientViT-SAM, MobileSAM | | ✂️ Image Matting | RMBG 1.4/2.0 | | 💡 Proposal | UPN | | 🏷️ Tagging | RAM, RAM++ | | 📄 OCR | PP-OCRv4, PP-OCRv5 | | 🧾 Layout Analysis | PP-DocLayoutV3 | | 📑 Document Parsing | PaddleOCR-VL, PaddleOCR-VL-1.5 | | 🗣️ Vision Foundation Models | Rex-Omni, Florence2 | | 👁️ Vision Language Models | Qwen3-VL, Gemini, ChatGPT, GLM | | 🛣️ Lane Detection | CLRNet | | 🔢 Object Counting | CountGD, GeCO, GeCo2 | | 📍 Grounding | Grounding DINO, YOLO-World, YOLOE | | 📚 Other | 👉 [model_zoo](./docs/en/model_zoo.md) 👈 | ## Docs 0. [Remote Inference Service](https://github.com/CVHub520/X-AnyLabeling-Server) 1. [Installation & Quickstart](./docs/en/get_started.md) 2. [Usage](./docs/en/user_guide.md) 3. [Command Line Interface](./docs/en/cli.md) 4. [Customize a model](./docs/en/custom_model.md) 5. [Chatbot](./docs/en/chatbot.md) 6. [VQA](./docs/en/vqa.md) 7. [Multi-class Image Classifier](./docs/en/image_classifier.md) 8. [Document Parsing and Intelligent Text Recognition](./docs/en/paddle_ocr.md) ## Examples - [Classification](./examples/classification/) - [Image-Level](./examples/classification/image-level/README.md) - [Shape-Level](./examples/classification/shape-level/README.md) - [Detection](./examples/detection/) - [HBB Object Detection](./examples/detection/hbb/README.md) - [OBB Object Detection](./examples/detection/obb/README.md) - [Segmentation](./examples/segmentation/README.md) - [Instance Segmentation](./examples/segmentation/instance_segmentation/) - [Binary Semantic Segmentation](./examples/segmentation/binary_semantic_segmentation/) - [Multiclass Semantic Segmentation](./examples/segmentation/multiclass_semantic_segmentation/) - [Description](./examples/description/) - [Tagging](./examples/description/tagging/README.md) - [Captioning](./examples/description/captioning/README.md) - [Estimation](./examples/estimation/) - [Pose Estimation](./examples/estimation/pose_estimation/README.md) - [Depth Estimation](./examples/estimation/depth_estimation/README.md) - [OCR](./examples/optical_character_recognition/) - [Text Recognition](./examples/optical_character_recognition/text_recognition/) - [Key Information Extraction](./examples/optical_character_recognition/key_information_extraction/README.md) - [MOT](./examples/multiple_object_tracking/README.md) - [Tracking by HBB Object Detection](./examples/multiple_object_tracking/README.md) - [Tracking by OBB Object Detection](./examples/multiple_object_tracking/README.md) - [Tracking by Instance Segmentation](./examples/multiple_object_tracking/README.md) - [Tracking by Pose Estimation](./examples/multiple_object_tracking/README.md) - [iVOS](./examples/interactive_video_object_segmentation) - [SAM2-Video](./examples/interactive_video_object_segmentation/sam2/README.md) - [SAM3-Video](./examples/interactive_video_object_segmentation/sam3/README.md) - [Matting](./examples/matting/) - [Image Matting](./examples/matting/image_matting/README.md) - [Vision-Language](./examples/vision_language/) - [Rex-Omni](./examples/vision_language/rexomni/README.md) - [Florence 2](./examples/vision_language/florence2/README.md) - [Counting](./examples/counting/) - [GeCo](./examples/counting/geco/README.md) - [GeCo2](./examples/counting/geco2/README.md) - [Grounding](./examples/grounding/) - [YOLOE](./examples/grounding/yoloe/README.md) - [SAM 3](./examples/grounding/sam3/README.md) - [Training](./examples/training/) - [Ultralytics](./examples/training/ultralytics/README.md) ## Contribute We believe in open collaboration! **X‑AnyLabeling** continues to grow with the support of the community. Whether you're fixing bugs, improving documentation, or adding new features, your contributions make a real impact. To get started, please read our [Contributing Guide](./CONTRIBUTING.md) and make sure to agree to the [Contributor License Agreement (CLA)](./CLA.md) before submitting a pull request. If you find this project helpful, please consider giving it a ⭐️ star! Have questions or suggestions? Open an [issue](https://github.com/CVHub520/X-AnyLabeling/issues) or email us at cv_hub@163.com. A huge thank you 🙏 to everyone helping to make X‑AnyLabeling better. ## License This project is licensed under the [GPL-3.0 license](./LICENSE) and is completely open source and free. The original intention is to enable more developers, researchers, and enterprises to conveniently use this AI application platform, promoting the development of the entire industry. We encourage everyone to use it freely (including commercial use), and you can also add features based on this project and commercialize it, but you must retain the brand identity and indicate the source project address. Additionally, to understand the ecosystem and usage of X-AnyLabeling, if you use this project for academic, research, teaching, or enterprise purposes, please fill out the [registration form](https://forms.gle/MZCKhU7UJ4TRSWxR7). This registration is only for statistical purposes and will not incur any fees. We will strictly keep all information confidential. X-AnyLabeling is independently developed and maintained by an individual. If this project has been helpful to you, we welcome your support through the donation links below to help sustain the project's continued development. Your support is the greatest encouragement! If you have any questions about the project or would like to collaborate, please feel free to contact via WeChat: ww10874 or email provided above. ## Sponsors - [buy-me-a-coffee](https://ko-fi.com/cvhub520) - [Wechat/Alipay](https://github.com/CVHub520/X-AnyLabeling/blob/main/README_zh-CN.md#%E8%B5%9E%E5%8A%A9) ## Acknowledgement I extend my heartfelt thanks to the developers and contributors of [AnyLabeling](https://github.com/vietanhdev/anylabeling), [LabelMe](https://github.com/wkentaro/labelme), [LabelImg](https://github.com/tzutalin/labelImg), [roLabelImg](https://github.com/cgvict/roLabelImg), [PPOCRLabel](https://github.com/PFCCLab/PPOCRLabel) and [CVAT](https://github.com/opencv/cvat), whose work has been crucial to the success of this project. ## Citing If you use this software in your research, please cite it as below: ``` @misc{X-AnyLabeling, year = {2023}, author = {Wei Wang}, publisher = {Github}, organization = {CVHub}, journal = {Github repository}, title = {Advanced Auto Labeling Solution with Added Features}, howpublished = {\url{https://github.com/CVHub520/X-AnyLabeling}} } ``` --- ![Star History Chart](https://api.star-history.com/svg?repos=CVHub520/X-AnyLabeling&type=Date)
🔝 Back to Top