On Sat Nov 1, 2025 at 3:23 PM CET,
pavanag...@gmail.com wrote:
> Hi, I know this isn't stricly a FFF question but is there a way to find
> duplicate URLs? I mergd 2 libraries together and now there are a lot of
> duplicates floating around.
I have few Python/shell scripts in
https://git.sr.ht/~mcepl/epubgrep. Using epub_meta [1] library
it is really easy to wipe up simple scripts dealing with EPub
metadata.
knih_dedup is my attempt for the deduplication script, but it is
exactly wrong like you suggest such script should not be written
(it uses names, not URLs). epub_get_URL.py could be what you are
after.
Best,
Matěj
[1]
https://pypi.org/project/epub_meta/
--
http://matej.ceplovi.cz/blog/, @mc...@en.osm.town
GPG Finger: 3C76 A027 CA45 AD70 98B5 BC1D 7920 5802 880B C9D8
In political activity men sail a boundless and bottomless sea;
there is neither harbor for shelter nor floor for anchorage,
neither starting point nor appointed destination.
-- Michael Oakeshott: Rationalism in Politics