I am generating several million jpg images using a python script for a deep learning project. I am storing these on a 14TB hard drive on my Mac. The total size of the image folder containing the images is projected to be about 1.2TB.
I notice that as the script proceeds (gets up into the 400-500k range), it starts to slow and the images don't immediately show up on the HD. For example, I stopped the script at 8am this morning but the most recent image was saved to the HD at 03:46. If I relaunch finder (it's a Mac), more of the most recently saved images are shown, but not all.
Now I am finding when I open the HD and look in the folder, nothing is shown even though if I look at it via terminal I can see all the images.
This to me seems like a bandwidth issue with the HD saving the images and that it is still saving images well after I stopped the script i.e. trying to catch up.
I am no expert with this sort of thing but I wonder if someone could help
Pin point the issue (it may be simple...
Give me some advice on the best way to fix this problem. Is there a better way to format the HD so that one can store millions of small files?
Lastly, I should add that I have removed the HD from Spotlight on my mac (to prevent it from trying index all the files)...so I don't think this is the issue.