Sweet logic bro
So I accidentally a new imageboard
Furry detected...
Even if not, while it is an overgeneralization in most cases, it isn't in this case given all the other ridiculous shit from the furry community, including (but not limited to) plushie sex, vore, beastiality, diaperfurs, and so on...
It is a very safe bet to say that whole fandom is flaming trash.
Maybe off topic? But on the subject of...
I've played this game. Dir reads take a mega long time with tons and tons of files in them. I've found in my testing when I did something like.
#pseudo codesum = generate_sum(filename)dirname = s[0]s[1]/s[2]s[3]/mkdir dirnamemv filename dirname
And I don't know if you're interest in this but noting that visually identical (down to the pixel) images can have different md5sums due to so many different variables that simple cryptographic hashing isn't sufficient for something like this. You're going to need to do perceptual hashing if you care about that. At the very least you can turn the image greyscale and scale down to a reasonable size you can easily store and compute a hamming distance against other hashes. That will catch less false positives than you may think.
Here's a shitty little testing thing with hardly any functionality I did a couple of years ago that should illustrate what I mean. 0x0.st
It's what I used to generate 5.png from 4.jpg. 5.png just visually represents a 64byte hash that can be stored in a file name perhaps.
These links may hook you up.
phash.org
Meant to say 256
Try again, smartass.
The furry fandom is way too vast to be generalizing like this.
In a way, they kind of asked for it- with their constant attempts to absorb other fandoms and the like, as well as the rejection of some anti-bestiality people in the earlier days. Not going to judge them based on zoophiles hiding int heir ranks though.
It's just for naming the images, if I wanted to scrape tags from other sites I'd need to md5 them at some point. It's also a decent way of avoiding name collisions with uploaded material.
I did something like this back in the day. What I did is downscale the image to ~50 pixels of data while logging the ratio. After that I converted the image to a limited indexed palette I created. Then it was simple to store the images as strings and compare them. The one requirement was that the image have roughly the same aspect ratio, within a certain tolerance, for the algorithm to be computationally light on larger imagesets.
People bitching about your porn taste OP, but good for you for doing something productive. Was wondering how fast that textfile was over the SQL database? I know it'll be slower but I don't mind waiting a second or two for the sake of not having to set up a local server on my toaster of a machine. I've thought of doing something similar for my own image folders, but I pull images from so many places I'd have to manually tag like 20GB worth of shit and gave up.
suicide might also be a fix.>>980631