¶
Build better computer vision models faster with self-supervised pre-training
LightlyTrain leverages self-supervised learning (SSL) and distillation from foundation models to train computer vision models on large datasets without labels. It provides simple Python, Command Line, and Docker interfaces to train models with popular pretraining methods such as SimCLR, DINO or distillation from DINOv2.
Why LightlyTrain?¶
🚀 Higher accuracy – Pretrained models generalize better and achieve higher performance.
💸 Cost efficiency – Make use of unlabeled data instead of letting it go to waste.
⚡ Faster convergence – Pretraining with LightlyTrain speeds up learning and reduces compute time.
🖼️ Better image embeddings – Extract more meaningful features than with supervised models.
🏗️ Stronger foundation – A great starting point for downstream tasks like:
Image classification
Object detection
Segmentation
🔄 Improved domain adaptation – Self-supervised learning enables better adaptability to data shifts.
Lightly are the experts in computer vision pretraining and developed LightlyTrain to simplify model training for any task and dataset.
How It Works¶
Install LightlyTrain:
pip install lightly-train
Then start pretraining with:
import lightly_train
if __name__ == "__main__":
lightly_train.train(
out="out/my_experiment", # Output directory
data="my_data_dir", # Directory with images
model="torchvision/resnet50", # Model to train
)
This will pretrain a Torchvision ResNet-50 model using unlabeled images from my_data_dir
.
All training logs, model exports, and checkpoints are saved to the output directory
at out/my_experiment
.
The final model is exported to out/my_experiment/exported_models/exported_last.pt
.
It can directly be used for fine-tuning. Follow the Quick Start
guide to learn how to fine-tune a pretrained model.
Follow the Embed guide to learn how to generate image embeddings with the pretrained model for clustering, retrieval, or visualization tasks.
Features¶
Train models on any image data without labels
Train models from popular libraries such as torchvision, TIMM, Ultralytics, SuperGradients, RT-DETR, and YOLOv12.
Train custom models with ease
No self-supervised learning expertise required
Automatic SSL method selection (coming soon!)
Python, Command Line, and Docker support
Built for high performance including multi-GPU and multi-node support
Export models for fine-tuning or inference
Generate and export image embeddings
Monitor training progress with TensorBoard, Weights & Biases, and more
Supported Models¶
ResNet
ConvNext
All models
YOLOv5
YOLOv6
YOLOv8
YOLO11
YOLO12
RT-DETR
YOLOv12
PP-LiteSeg
SSD
YOLO-NAS
See supported models for a detailed list of all supported models.
Contact us if you need support for additional models or libraries.
Supported SSL Methods¶
DINO
Distillation (recommended 🚀)
SimCLR
See methods for details.
License¶
LightlyTrain is available under an AGPL-3.0 and a commercial license. Please contact us at info@lightly.ai for more information.