Fedora Security :: Cleaning Duplicate Files ?

Dec 17, 2009

I was wondering if anyone knew about fdupes? What I would like to do is to delete duplicate copies of files that are not needed from my whole system.

View 8 Replies


ADVERTISEMENT

Fedora :: Cleaning The Orphaned Files Automatically?

Oct 26, 2009

If we update or remove some packages (in addition manually installed software), some files such as previous version dynamic lib files are left, so it may conflict with new stuffs sometimes. Is there any efficient way to remove these kind of orphaned files all, automatically?

View 2 Replies View Related

Ubuntu Security :: Antivirus System For Cleaning Windows Platform Viruses?

Mar 26, 2010

I would like to use my Ubuntu 9.10 Karmic system to scan and clean viruses from Windows based HDD. I have KlamAV, AVG, and AVAST. I couldn't get F-prot to work. I don't find a debian package of it and don't know how to install tar.gz and the tutorials don't help a lot. Are there other antivirus solutions which are good for cleaning Windows OS but work on Linux/WINE? I wonder if I can install antivirus software in WINE and scan with no problem the hard disks.

View 5 Replies View Related

Ubuntu :: HDD Deep Cleaning Of Deleted Files?

Jul 14, 2010

Is there a way to deep clean the supposedly "empty" areas of HDD. I've found "shred" and similar tools by googling, but they allow either deleting a file or complete wipeout of a HDD. What I'd like to do is clean up what's left of already deleted files (which can probably be still "undeleted") on a live HDD (with useful data, which doesn't need to be destroyed).

View 5 Replies View Related

Debian :: Cleaning Up Apt/dpkg After Deleting Installed Files?

Mar 9, 2010

I'd like to clean up apt/dpkg so that it doesn't think the package whose files I've deleted is still installed, but when I run:

Code:

apt-get remove --purge linux-image-2.6.32-2.slh.4

I get the following:

Code:

Reading package lists... Done
Building dependency tree
Reading state information... Done
E: Couldn't find package linux-image-2.6.32-2.slh.4

The package does show up in the list when I run the command:

Code:

dpkg -l "linux-image*"

with "ii" in the 1st column, whatever that means. Should I just rm the matching files in /var/lib/dpkg/info?

View 6 Replies View Related

Slackware :: System Cluttered With Unused Files - Cleaning Out Junk

Jun 16, 2011

When you unsuccessfully try to install software, does your Slackware system become cluttered with unused files? Is there a program, other than 'slackpkg clean-system', that can review your installation and identify any useless stuff? My tmp directory is set, in fstab, to delete its contents on boot. Is the tmp directory the only directory to be concerned with?

View 11 Replies View Related

Security :: How To Prevent Duplicate UIDs

Oct 8, 2010

I have been learning Linux for the past few months and just recently started with Bash programming. Using scripts it is possible to find users with duplicate UIDs but is there any way or script why which duplicate UIDs can be prevented altogether.

View 7 Replies View Related

Fedora :: Cleaning Up /var/tmp

Dec 19, 2009

Can i safelly clean up /var/tmp? it contains 730mb of files

View 12 Replies View Related

General :: Find Duplicate Files By Name ?

Jul 24, 2010

We have a huge amount of duplicate files in a folder and I would like some pointers on to writing a bash script to create a list of the duplicate files. I've seen examples that check for the md5 sum of files... but I dont need that, the file name is enough.

View 4 Replies View Related

Software :: Duplicate Files And A Basedirectory ?

Dec 28, 2010

I have two directories, I want to know which files in the second directory also appear in the first and delete the duplicate in the second directory. Filenames might be different (so that rules out diff).

My problem is that various programs (such as fdupes and freedup) are very capable of finding duplicate files but randomly delete (or link) files from the first or the second directory.

Here an example with fdupes:

Code:

As you can see, the file in the third pair is removed from dir1 instead of from dir2. My aim is to have only files deleted from the dir1. I know that fdupes can't do this, as I emailed with the author.

View 2 Replies View Related

Ubuntu :: Bash Script - Not Duplicate Files?

Sep 1, 2010

I am working through a collection of many thousand photographs in many collections, spanning decades. One of the folders is a "Selected_images" folder which should only contain copies of images from the collections.

How can I check that the files in "Selected_images" are all indeed copies and have an identical counterpart somewhere in the collections hierarchy, not necessarily with the same name?

At present I am scanning the output of fdupes -r collection, which is tedious (although, in fact, no photographs should ever be duplicated except in "Selected_images").

View 4 Replies View Related

Applications :: Find Duplicate Files In My Folders

Jul 2, 2010

I have found some duplicate files in my folders. Is there a way to clean them out?

View 2 Replies View Related

General :: Detect Duplicate Files Using Md5sum ?

Feb 21, 2011

What i am trying is to check the file duplication in a folder and remove a file if it is a duplicate of another file ie the contents are duplicate; but names may be same.

Basically i am using md5sum to calculate the md5sum values of each file and redirecting to a file. And i am thinking of comparing the md5sum values.But i am finding it hard to decide how to complete the code after redirecting the output of calculation of md5sum to a file.

View 3 Replies View Related

General :: Remove Duplicate Files From Two Folders?

Aug 19, 2010

I have two folders - Folder abc and Folder xyz which contains 1000's of files with few of them having the same file names. How can I remove the duplicates from Folder abc?

View 14 Replies View Related

Software :: Best Tool To Find Duplicate Files ?

May 16, 2010

I have a 1TB drive that has MANY duplicate files all over it. a good linux tool that can find duplicate files on such a large drive (almost full) drive?

View 1 Replies View Related

Software :: Find Duplicate Files - Program With Gui ?

Jan 29, 2011

I would like to check two folders for duplicate files (two pretty old backup instances).

MY folders are quite alike so I would like to stop the NON-duplicated files for that I want to be able to do some checks not only for the filename but alsofor the filesize (might be the case that two files have same name but not size).

The ideal would be to suggest me such a program with gui but if not I will try run any script code that is available outhere.

View 4 Replies View Related

Programming :: Script To Remove Duplicate Jpg Files

May 25, 2010

Thanks y'all for the great script and explanation. This helped a lot in my own project. I thought I'd share the efforts.The project is this: I've got lots of duplicate JPGs from all the family members who've named the same photo with different names. Since md5sum generates a "fingerprint" based on the file contents, not the name, I want to use the md5sum of each jpg to uniquely name each photo and also remove exact duplicates.

It has the following flaws:
0) it doesn't handle certain non-alphanumerics
1) it keeps both photo-shopped and unaltered photos (different md5s)
2) it (currently) doesn't preserve descriptive filenames.

(For me, removal of duplicates is more important than keeping the filenames. I may change that to concatenate the md5 and the filename.)Please note that the commented "rename" command should be used to strip non-aphanumerics from the file names, and the script should be launched with the commented "find" command.

View 1 Replies View Related

Fedora :: Cleaning Old Packages From Yum Cache?

Mar 12, 2010

I have enabled the option to keep the yum cache when updating Fedora 12 and I was wondering if there was a way in which I could remove old versions of the same package from the cache?

I have Fedora installed on more than one machine and I make them share a common update location for their updates to avoid downloading the same packages twice (bandwidth quota expensive in Australia).

how I could accomplish that? yum clean doesn't seem to have any options to only remove old packages!!!

Example: The yum update cache has:

wireshark-gnome-1.2.5-3.fc12.x86_64.rpm
wireshark-gnome-1.2.6-1.fc12.x86_64.rpm

I wish to only keep the most recent version that is wireshark-gnome-1.2.6-1.fc12.x86_64.rpm, but for all the packages.

View 4 Replies View Related

OpenSUSE :: Find Duplicate Files In Multiple Disks?

Jan 11, 2011

I am looking for an application (better kde one) that can search two external hard disks I have and find any duplicate files. I did some backups before to one disk which i copied few years ago to the other disk. Right now I would like some program to check files and tell me if there are the same.

View 9 Replies View Related

General :: How Do I Recursively Delete Duplicate Files Using Terminal

Mar 1, 2011

I'm using a mac, and just transferred a bunch o photos from another computer, and as it turns out, there is a bunch of duplicates.I'm not too familiar with the mac terminal, but if there is a solution for linux, it will probably work for the mac.Just need to be able to recursively scan all folders in my Pictures folder and then Delete them.

View 1 Replies View Related

Ubuntu :: Delete Duplicate Files Command Line?

Apr 14, 2010

Is there a way to remove duplicate files from a specific folder through SSH? I've uploaded a lot of flash games on my server and I can see in the Webmin's file manager that I have many duplicates. Their names are different, of course.

View 2 Replies View Related

General :: GUI Method To Copy (duplicate) Files / Media?

Sep 23, 2010

Is there a GUI method to copy files or is this always done with the Bash terminal.

View 6 Replies View Related

General :: Find Duplicate Files Using Bash Script ?

Jun 8, 2010

What kind of method to find the duplicates files on linux,

1.how to find just using the file name, sometimes i figure out people often to copy their files to another directory and i want to find out if there any same file name in the linux box.

2. what about if i want to find the duplicate files based on contents of the file, example is in picture file if users store picture files from digital camera first they just save the file name in default but when they want to give that picture to others they will rename it, i've been used method md5 for this situation in python script but it takes long time

I'm asking this question just to know to use bash script a lot in work and i want to test out fdupes at home, is fdupes use similar md5 scan to find duplicate files?

View 4 Replies View Related

General :: BASH: Find Duplicate Files (MAC/LINUX Compatible)?

Mar 18, 2011

I am looking for a bash script which is compatible with Mac, to find duplicate files in a directory.

View 2 Replies View Related

Ubuntu :: Write A Script To Go Through And Delete Bulk Duplicate Files?

Jul 7, 2010

I am using my Ubuntu machine to serve as a media server and network storage. The problem I have is iTunes on my desktop managed to make 2 copies of every song on the machine so instead of the 30GB I have its up to almost 100gb. I was wondering if there was a way to write a script to go through and delete the duplicates. The duplicates are the same filename as the original except a 1 or 2 following. Wasn't looking forward to deleting 12,000 files by hand.

View 1 Replies View Related

Ubuntu :: Searching For And Deleting Duplicate Files That Exist In Different Paths?

May 30, 2011

Is their (I assume their is) of searching for and deleting duplicate files that exist in different paths?

View 3 Replies View Related

General :: Find Duplicate Files And Delete All Except Most Recent Version

Aug 17, 2010

I have a directory containing a ton of photos, some of which are duplicates but just with different names. Is there any way in linux to find all the duplicates and remove all of them except the most recent version? I know on Windows there are utilities that will do this through a GUI, but I'm using Linux through the CLI only.

View 6 Replies View Related

General :: GUI Application For Finding / Deleting Duplicate Data Files

Jan 10, 2011

I copied a back up of my windows 'my documents' fold and all of its' sub folders into my linux (Mint Debian) Documents directory. I found that many of my files can be found in more that one directory so, what I want to do is to find all the dups and deal with them. Is there a good linux application to resolve this 'duplicates' problem. (I don't want to touch the linux system files.)

View 2 Replies View Related

Fedora :: Update 15 And Taking Too Long - Even Cleaning Of Packages

Aug 16, 2011

I tried to update fedora 15 and it took too long , even the cleaing of packages took too long At the end of the update got the status cleaning up packages that took even longer than the whole update.

View 3 Replies View Related

Debian :: Duplicate Files In User Library - Safe To Remove Outdated

Feb 8, 2011

I have just upped from lenny to squeeze. I didn't mean to, really, but the package manager was well into its stride by the time I realised what was happening. Mostly all went well, BUT /usr is now 100% full. I notice that there are duplicate files in /usr/lib, eg Oct 11 22:35 libgcj.so.10.0.0 and Sep 14 2008 libgcj.so.90.0.0 (I assume the latter has been replaced by the former?). Is it safe to remove the "outdated" lib files? Is there an elegant way of doing it?

View 4 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved