(CapPic)
An image viewer and AI-assisted editing tool that helps with curating datasets for generative AI models, finetunes and LoRA.
-
Image Viewer: Display and navigate images
- Quick-starting desktop application built with Qt
- Runs smoothly with tens of thousands of images
- Modular interface that lets you place windows on different monitors
- Open multiple tabs
- Zoom/pan and fullscreen mode
- Gallery with thumbnails and optionally captions
- Semantic image sorting with text prompts
- Compare two images
- Measure size, area and pixel distances
- Slideshow
-
Image/Mask Editor: Prepare images for training
- Crop and save parts of images
- Scale images, optionally using AI upscale models
- Dynamic save paths with template variables
- Manually edit masks with multiple layers
- Support for pressure-sensitive drawing pens
- Record masking operations into macros
- Automated masking
-
Captioning: Describe images with text
- Edit captions manually with drag-and-drop support
- Multi-Edit Mode for editing captions of multiple images simultaneously
- Focus Mode where one key stroke adds a tag, saves the file and skips to the next image
- Tag grouping, merging, sorting, filtering and replacement rules
- Colored text highlighting
- CLIP Token Counter
- Automated captioning with support for grounding
- Prompt presets
- Multi-turn conversations with each answer saved to different entries in a
.jsonfile - Further refinement with LLMs
-
Stats/Filters: Summarize your data and get an overview
- List all tags, image resolutions, masked regions, or size of concept folders
- Filter images and create subsets
- Combine and chain filters
- Export the summaries as CSV
-
Batch Processing: Process whole folders at once
- Flexible batch captioning, tagging and transformation
- Batch scaling of images
- Batch masking with user-defined macros
- Batch cropping of images using your macros
- Copy and move files, create symlinks, ZIP captions for backups
-
AI Assistance:
- Support for state-of-the-art captioning and masking models
- Model and sampling settings, GPU acceleration with CPU offload support
- On-the-fly NF4 and INT8 quantization
- Run inference locally and/or on multiple remote machines over SSH
- Separate inference subprocess isolates potential crashes and allows complete VRAM cleanup
These are the supported architectures with links to the original models.
Find more specialized finetuned models on huggingface.co.
-
Tagging
Generate keyword captions for images. -
Captioning
Generate complete-sentence captions for images. -
LLM
Transform existing captions/tags.- Models in GGUF format with embedded chat template (llama-cpp backend).
-
Upscaling
Resize images to higher resolutions.- Model architectures supported by the spandrel backend.
- Find more models at openmodeldb.info.
-
Masking
Generate greyscale masks.- Box Detection
- YOLO/Adetailer detection models
- Search for YOLO models on huggingface.co.
- Florence-2
- Qwen2.5-VL
- YOLO/Adetailer detection models
- Segmentation / Background Removal
- InSPyReNet (Plus_Ultra)
- RMBG-2.0
- Florence-2
- Box Detection
-
Embedding
Sort images by their similarity to a prompt.- CLIP
- SigLIP
- SigLIP (ONNX), SigLIP2-giant-opt (ONNX)
(recommended: largest text model + fp16 vision model)
Requires Python 3.10 or later.
By default, prebuilt packages for CUDA 12.8 are installed. If you need a different CUDA version, change the URLs in requirements-pytorch.txt and requirements-flashattn.txt before running the setup script.
- Git clone or download this repository.
- Run
setup.shon Linux,setup.baton Windows.- Packages are installed into a virtual environment.
The setup script will ask you a couple of questions.
You can choose to install only the GUI and image processing packages without AI assistance. Or when installing on a headless server for remote inference, you can choose to install only the backend.
If the setup scripts didn't work for you, but you manually got it running, please share your solution and raise an issue.
- Linux:
run.sh - Windows:
run.batorrun-console.bat
You can open files or folders directly in qapyq by associating the file types with the respective run script in your OS.
For shortcuts, icons are available in the qapyq/res folder.
If git was used to clone the repository, simply use git pull to update.
If the repository was downloaded as a zip archive, download it again and replace the installed files.
To update the installed packages in the virtual environment, run the setup script again.
New dependencies may be added. If the program fails to start or crashes, run the setup script to install the missing packages.
More information is available in the Wiki.
Use the page index on the right side to find topics and navigate the Wiki.
How to setup and configure AI models: Model Setup
How to use qapyq: User Guide
How to caption with qapyq: Captioning
How to use qapyq's features in a workflow: Tips and Workflows
If you have questions, please ask in the Discussions.
- Natural sorting of files
- Gallery list view with captions
- Summary and stats of captions and tags
- Shortcuts and improved ease-of-use
- AI-assisted mask editing
- Overlays (difference image) for comparison tool
- Image resizing
- Run inference on remote machines
- Adapt new captioning and masking models
- Possibly a plugin system for new tools
- Integration with ComfyUI
- Docs, Screenshots, Video Guides

