Simple script using GPT-4 to parallelize download and extract files for SA-1B Dataset.
- Python >= 3.6
- requests >= 2.0
Install with:
pip install requests
The download.py script uses the sa1b_links.txt file by default as the input file for downloading and extracting images:
python download.py --processes 4 --input_file sa1b_links.txt --raw_dir raw --images_dir images --masks_dir annotations
The download.py script takes the following command-line arguments:
--processes: The number of processes to use for downloading and extracting files (default: 4)
--input_file: The path to the input file containing file names and URLs (default: 'sa1b.txt')
--raw_dir: The directory to store downloaded files (default: 'raw')
--images_dir: The directory to store extracted jpg files (default: 'images')
--masks_dir: The directory to store extracted json files (default: 'annotations')
--skip_existing: Skip extraction if the file has already been extracted (default: False)