Remove Duplicate Files From Pc

0 views
Skip to first unread message

Marieta Reeks

unread,
Aug 5, 2024, 11:43:51 AM8/5/24
to eananmiemis
Thisis not really a space issue, I have lots, but I have nearly 1000 ebooks that I sometimes file in multiple places. (No, I don't use the dreaded Kindle thanks.) So I may download it in a "holiday reading" folder, then I may also download in the "Publishers folder" and at some point I may have reviewed it and moved it to the "author" folder and then maybe I have forgotten that I downloaded it in the "lost interest in this" folder. So there could be anywhere from 1-4 version of a book floating around.

If I search specifically by name it shows me all of the versions and then I can delete the ones that are in the "wrong" place, but I don't have the time or inclination (despite my love of procrastination by rearranging files) to search 1000 titles. So is there a simple Dropbox search function for duplicates, or do I have to download some kind of app that will do the job. Can you download file titles and locations to a spreadsheet? At least that would let me filter/sort and then go back and find the ones I want to delete.


If that's required, I don't want one that automatically deletes duplicates, I want to pick which ones are required, but if you have any recommendations of on that is easy to use and pretty much straight forward (nothing fancy required, just find the duplicates (or those that may have a similar name since sometimes they get saved with slightly different titles), I'd appreciate the help.


Duplicate Cleaner has enough features to satisfy even the most demanding power user: findduplicate folders, unique files, search inside zip files, advanced filtering, virtual folders,snapshot states and much more.Full feature list


Duplicate Cleaner is a tool for finding and removing duplicate files from your computeror network drives. It is intended to be used on user content - documents, photos,images, music, video but can be used to scan any type of files.


Free has the basic functionality, and is only for personal/home use - not for use in acommercial environment. Pro has lots more functions including similar image detection,finding duplicate folders and unique files, searching in zip files and advanced filtersand search methods.Full featurelist and comparison.


Duplicate Cleaner (Pro Edition) is licensed for Personal or Commercial use. The Licenseis perpetual and all updates to the Duplicate Cleaner 5 series are included in theprice.Includes technical support by via our support centre and the forum.


Duplicate files occupy unnecessary space in your computer hence slowing it down. These files are copies of already existing files on your device. You may have file duplicates of photos, videos, audios, archives, documents, etc.


The best way to deal with such files is to trace and remove them from wherever they are on your computer. There exist a couple of techniques and strategies to remove duplicate files on Windows 10. This article will discuss 3 ways, including the following:


This may create space to store more data and speed up your computer in the long run. The standard built-in tools to remove duplicate files on Windows 10 include File Explorer and Windows PowerShell. Below are steps to use each of these tools.


4. The above process will take some time to complete (depending on the number of files in the search folder). When it finishes, open the results file (.txt file) to identify the location of the duplicate files and manually delete them.


Using third-party software to remove duplicate files on Windows 10 is a great alternative to inbuilt tools. Third-party software is quite effective and efficient in removing all the duplicate files on your computer.


Step 1. Open EaseUS Dupfiles Cleaner and click Scan Now to start cleaning. EaseUS Dupfiles Cleaner will automatically select all data in all partitions. You can delete partitions you don't want to clean up by pressing the "-" sign in Scan Folders and choose the file types in Filename Pattern.


* You can click Smart Selections to further check which type of files you need to clean up and if you cannot identify the content from the file name, you can directly click the file name in the upper right corner to preview it.


Windows Explorer helps you manually view different forms of data stored on your computer. The views exist in the form of "Extra Large Icons," "Large Icons," "Medium Icons," "Small Icons," "List," "Details," "Tiles," and "Content."


This method works well for you if you already know the names of the duplicate files. You'll only need to open the drive and type the file's name on the search bar. You can also search for the known files using their respective extensions, i.e., .jpg, .mp3, .pdf, etc.


You can remove duplicate files on your Windows 10 computer by using built-in tools, third-party software, or manually. Common built-in tools to delete duplicate files include File Explorer and Windows PowerShell. The best third-party software to delete duplicate files is EaseUS DupFiles Cleaner.


This software automatically searches all files on your computer to identify duplicates. To manually delete duplicate files in your Windows 10 computer, you can use Windows Views, Windows Search, or sorting files. Among the three methods, EaseUS DupFiles Cleaner is the most preferred.


I've found a lot of duplicates but many are slightly different at the end of the file paths, but they may be the same size and have the same name. I've attached a snippet. How do I know if I can delete any of the duplicates since I have so many? As you can see, many duplicate files end in "_1". Can they be deleted?


I would strongly advise against searching for or even deleting duplicate files in the system folders, such as "Program Files" or "Windows". Many programs and also Windows unfortunately have the habit of creating files multiple times, however, in the vast majority of cases deleting these files leads to damage, so that a repair or reinstallation becomes necessary.


But when all is said and done only you can decide which files you want to keep and which you don't.

Nobody else can make that decision for you, and certainly an application can't decide it for you.


Taking note of the above I would suggest that setting the 'Match by', Ignore', and 'Include' similar to these screenshots is the safest (best) way to search for your own duplicates, you may want to 'add' more paths to search.

This first will search only user Steve's Documents, Downloads, and Pictures folders and report only any exact duplicates found there whatever they are named.




An alternative would be to do it like this and tell it to search all the sub-folders in 'Steve', this is of course searching more folders and so is likely to find more duplicates which may not be the ones you are looking for:




After searching you can output a list of what has been found to a textfile if you want to check them out further before deciding if you want to delete them or not.

You cannot import that textfile back into duplicate finder, things may have changed since you made that textfile so a new scan is needed each time you use duplicate finder.

Once you have decided which, if any, files to delete then you can use the tickboxes to delete one (or more) of the duplicates.


When using the tickboxes then you have to select them one by one, there is no 'Delete All' button - that would be just too dangerous and you could very easily delete (hundreds of) things that you want or need.


What I need is to remove all the repetitions while (preferably, but this can be sacrificed for a significant performance boost) maintaining the original sequence order. In the result each line is to be unique. If there were 100 equal lines (usually the duplicates are spread across the file and won't be neighbours) there is to be only one of the kind left.


UPDATE: the awk '!seen[$0]++' filename solution seemed working just fine for me as long as the files were near 2 GiB or smaller but now as I am to clean-up a 8 GiB file it doesn't work any more. It seems taking infinity on a Mac with 4 GiB RAM and a 64-bit Windows 7 PC with 4 GiB RAM and 6 GiB swap just runs out of memory. And I don't feel enthusiastic about trying it on Linux with 4 GiB RAM given this experience.


There's a simple (which is not to say obvious) method using standard utilities which doesn't require a large memory except to run sort, which in most implementations has specific optimizations for huge files (a good external sort algorithm). An advantage of this method is that it only loops over all the lines inside special-purpose utilities, never inside interpreted languages.


For a large amount of duplication, a method that only requires storing a single copy of each line in memory will perform better. With some interpretation overhead, there's a very concise awk script for that (already posted by enzotib):


Assuming you can afford to keep as much as the de-duplicated file in memory (if your data is indeed duplicated by a factor of 100, that should be about 20MiB + overhead), you can do this very easily with Perl.


That's fine as sort needs to have read all its input before it can start outputting anything (before it can tell which is the line that sorts first which could very well be the last line of the input).


sort will (intelligently) use temporary files so as to avoid loading the whole input in memory. You'll need enough space in $TMPDIR (or /tmp if that variable is not set). Some sort implementations can compress the temp files (like with --compress-program=lzop with GNU sort) which can help if you're short on disk space or have slow disks.


Typically one can delete duplicate files on the My Cloud the same way they do on their computer. Use an application/software that searches the drive(s)/folders you designate and display any duplicate files/folders found.


What My Cloud unit do you have? The single bay/single drive My Cloud units (the general subject of this subforum) do not officially support third party apps being installed like the multi-bay My Cloud units do.

3a8082e126
Reply all
Reply to author
Forward
0 new messages