In modern agriculture, leaf disease segmentation is crucial for crop disease management and yield improvement. Since most deep learning-based segmentation models require extensive annotations, pursuing unsupervised methods becomes a practical solution. Although advanced models like the Segment Anything Model (SAM) can produce precise class-agnostic masks without annotations, their reliance on human-interactive prompts limits their utility to unsupervised disease segmentation tasks. To address this limitation, we present UPLS, a novel three-stage framework that leverages unsupervised prompt learning to enable SAM for leaf disease segmentation. This method employs unsupervised contrastive learning to progressively segment leaf and disease areas and automatically generate disease prompts for SAM. Concretely, for fine and intricate lesions, we utilize an improved high-frequency attention mechanism to extract high-frequency features from the leaf area, and construct contrastive losses between global priors and foreground/background features. Furthermore, we devise a strategy to automatically generate disease-related prompts from the segmented leaf and initial lesion regions, enabling SAM to refine disease boundaries without human intervention. Experiments on the public leaf disease datasets show that UPLS outperforms existing non-fully supervised segmentation methods in accuracy and robustness.
git clone https://github.com/Tianluda/UPLS.git
cd UPLS
Download the Apple Leaf Disease Dataset and the Extra test dataset used in this paper. After downloading, place the datasets under the "./experiments/data" folder.
Follow CCAM to build the environment.
The dependencies are configured as follows:
- Python 3
- PyTorch 1.12.1
- OpenCV-Python
- Numpy
- Scipy
- MatplotLib
- Yaml
- Easydict
Download the following pretrained parameters and place them in the current directory. 'mocov2': here 'detco': here 'Plant': here
├── UPLS/
| ├── core
| ├—— ...
| ├—— moco_r50_v2-e3b0c442.pth
| └── detco_200ep.pth
| └── ResNet50-Plant-model-80.pth
First, train the UPLS-1 model with MOCO pretrained weights to extract the leaf foreground area.
nohup /bin/bash ./start_MOCO.sh >> ./cron_MOCO.log 2>&1
The code will create experiment folders for model checkpoints (./experiment/models), log files (.experiments/logs), visualization (./experiments/images/) and prediction results (./experiments/predictions/).
├── experiments/
| ├—— images
| ├—— logs
| ├—— models
| ├── predictions
The above code has generated a binary map of the leaf area. Next, use the first part of the code in './Process-Evaluation.ipynb' to generate the leaf foreground area (RGBA four-channel image).
In the second step, use the UCLSS-2 model with DETCO or Plant pretrained weights to extract the initial disease spot foreground area.
nohup /bin/bash ./start_DETCO.sh >> ./cron_DETCO.log 2>&1
nohup /bin/bash ./start_Plant.sh >> ./cron_Plant.log 2>&1
Next, use the Prompt-Guided Segmentation Refinement method based on SAM to refine the disease spot area. Copy the scripts in './segment-anything-main/notebooks' to the corresponding directory in the SAM source code, and execute the following command:
nohup ./segment-anything-main/notebooks/C2AM_SAM_Box.sh >> ./segment-anything-main/notebooks/C2AM_SAM_Box.log 2>&1
Use the second part of the code in './Process-Evaluation.ipynb' to evaluate the model performance on the test dataset.