Skip to content

a tool that help you find out all duplicated media from a specifc directory

Notifications You must be signed in to change notification settings

eawlot3000/uniqle_media

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

36 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

uniqle media 📁

A user-specific multi-threads supporting CLI tool that help you find out duplicated media (images/videos) from a specifc directory

Features

  • support user-specific multi-threads now!
  • create a duplicated folder to store your wasted so that you can review it and move on
  • any errors handling and also performance display

Performance

  1. I have tested a folder which has 683 images up to 835MB with different formats, including 19 duplicates. with MAX_THREADS = 64, it took 14 seconds to finish perfectly!
  2. For a updated 1.18GB folder that has 2438 images and vidoes includes 534 duplicates, took 23 seconds! ======= I have tested a folder which has 683 images up to 835MB with different formats, including 19 duplicates. with MAX_THREADS = 64, it only took 14 seconds to finish perfectly!

how to maximum the speed if you have better configs?

modify MAX_THREADS = [YOUR DESIRED THREADS] in fast.py file. You should try yourself for best performance because everyone's machine is different.

requirements

pip install -r requirements.txt

usage

python fast.py [FOLDER/]


TODO

  • time priority
  • exception data type support? [-args]
  • multi directories at the same time? [multi arguments]
  • multi threads to improve speed? ✅
  • in root? (great for clean your disk space yuh?)
  • GPU support? (tf needs python <= 3.9, NO arm mac)

tree view

.
├── README.md
├── copy_main.py
├── fast.py ==> search faster!
├── main.py
├── performance
├── requirements.txt

Releases

No releases published

Packages

No packages published

Languages