General :: Finding Files That Relates To Certain Applications
Jan 28, 2010
I'm wondering if you can share some tips in regards to finding .conf files in programs when installing using package managers. I'm scratching my head on the fact that when you install a program through yum/apt-get, I don't know what and where the software is being installed at. In Windows, I know that when it installs an application, it goes into the Program Files directory, it's that simple.I know Linux has predefined directories for applications but sometimes it installs configuration files in /etc or some other locations in /usr which I have a tough time sifting through.
Is there a way to trace what .conf or any files for that matter which relates to what software that needs it? It's just hard for me to understand what file relates to what application at the moment. As much as I would like to learn more about Linux, this process for me takes up alot of time. I hope you can help me out on this one.
I have a situation where a directory has about 1.5 million files in it. On an hourly basis, I want to be able to find any files that have changed in the last hour, compress them, encrypt them and then copy them to both a local backup machine and an off site backup.
Is there any kind of utility or kernel module that creates some type of log of modified files? I know I can use find, but the search for -mtime in this directory takes quite a while and will not suffice for an hourly backup.
i am a newbie in linux ,i am writing a bash script to identify the files which are exactly 7 days ( a week old) i tried this command find /var/backup -mtime +7 -exec ls -d {} ;but this gives me even the files which are older than 7 days
I am looking for this `struct messages_sdd_t` and I need to search through a lot of *.c files to find it.However, I can't seen to find a match as I want to exclude all the words 'struct' and 'messages_sdd_t'. As I want to search on this only 'struct messages_sdd_t' The reason for this is, as struct is used many times and I keep getting pages or search results.The directory I am searching in, has another directories so it will have to search recursively.I have been doing this without success:Code: find . -type f -name '*.c' | xargs grep 'struct messages_sdd_t'and thisCode: find . -type f -name '*.c' | xargs egrep -w 'struct|messages_sdd_t'
I know find can do what I am looking for, but I am wondering if there is an alternative way to find files on the filesystem either created before/after a certain point, or at a certain time.
Typically I rely on updatedb & locate for most of my file searching needs. Issues with those tools, though, are that it only has directory and file names, and it only creates a database of local directories, not anything mounted via CIFS|NFS or via -o loop (eg, .iso images).
So if I need to find files created after yesterday across the entire system (local and remote filesystems), I am currently needing to use find.
What other tools, if any, would accomplish this in a similar fashion?
I have tried ls and grep, but that requires (in my attempts so far) multiple searches:
ls -lR | grep Aug | grep 10 ls -lR | grep Aug | grep 11
My collection contains some MP3s which have some glitches like
displaying the wrong duration on loading minor jumps suddenly ending despite the duration claims another minute remaining noise
I'm looking for a tool that can detect as many of these glitches as possible and fix those that can be fixed (obviously e.g. noise can not simply be eliminated in most cases).
I have a website that has a massive list of royalty free stock photos and I want to download all of them. I have bought a membership for [URL] so I am able to download as much as I want from them for the next month.
Instead of going page by page and manually downloading each set of stock photos manually, I would like to automate this process. Here's my idea:
1. Download the website with the links to hotfile [URL]
2. Use grep to retrieve all the links to [URL]
3. Feed the links I recieve from grep into wget and download the works of them.
The problem I'm getting is when I use grep, It retrieves the entire line of html code where "hotfile.com" is shown. So here is an example of one link I receive in the output:
how to find a file with a ctime older than. let's say, 5 seconds? What I would like to do is to move all "new" files in a specific folder but not files that has not yet finished uploading.
I've got a quick grep question. I'm trying to work out a command I can use to locate all of the files in a directory that have sql database connection details. I want to do it by looking for the strings "localhost" and the name of the database.find . -type f -exec grep -l -E '^(localhost|DATABASE_NAME)' {} ;
If I were to lease a linux server on from a hosting company could someone among you give me a list of possible current uses/applications that would earn money these days?
I have been using Linux close to 2 years now. One thing that always bewildered me is audio support in Linux. These days I login to windows only for listening to music. After reading various blogs, i decided to give it a try in Linux with Amarok. There again I am facing a problem.
while i try to scan for music files it is not finding any files.
I have tried with mp3 .wma format files. While I try to add these files individually, Amarok is able to play those.
I am using Amarok 2.3.2 with KDE 4.4.5. Fedora - 12 is my flavor.
Given below is the log obtained with amarok -debug option
Quote:
TagLib: MPEG::Header:arse() -- First byte did not match MPEG synch. TagLib: MPEG::Header:arse() -- Invalid sample rate. TagLib: A frame of unsupported or unknown type 'TSC' has been discarded TagLib: A frame of unsupported or unknown type 'TSC' has been discarded TagLib: A frame of unsupported or unknown type 'TSC' has been discarded
I copied a back up of my windows 'my documents' fold and all of its' sub folders into my linux (Mint Debian) Documents directory. I found that many of my files can be found in more that one directory so, what I want to do is to find all the dups and deal with them. Is there a good linux application to resolve this 'duplicates' problem. (I don't want to touch the linux system files.)
I have a new system I built this week. I have the nvidia drivers installed, but desktop effects, googleearth and Quake4 all try to load the /usr/lib/libGL.so.1 from mesa-libGL and fail to start.I have two other machines, both with nvidia video, and don't have the problem with them. I've looked for differences on the systems, but haven't located anything so far.
I'm a Linux newbie and for my new job I have to do a lot of instant messaging to clients. They are on a variety of services, aim, icq, msn, myspace, facebook, yahoo, Google Talk, jabber, and .mac/ichat Is there a good instant messaging client that covers most of these on one application? I am using Mandriva with KDE 4.2.
I am looking for a free software that performs image comparison. I would prefer a non-GUI application as I need this to be automated. I am using Redhat Enterprise 3 and 4
What I want to do is force my DVD burner to use full writing power on the entire surface of a DVD. This should be able to invisibly damage the contained data, and should work on CDs, DVDs and BRDs too (in theory). I don't want to write data over it, so it's not a matter of having a RW disc or a R one. The result I want to obtain is simply an unreadable disc that does not look damaged. Do you know if there's a program able to ignore a disc being already full and perform this task?
Ive been using linux for a while but I am just getting into shell scripting, im currently trying to get a simple script for finding and copying files powered by the command:
Code:
This works fine from the command line but when put in a script such as:
Code:
Code:
with the keyboard inputs for $fc1 and $fc2 being *.doc and ~/test respectivly. The only problem i can see is the xargs -ivar "var" part possibly needing $var to be defined?
is there a recursive shell or Perl script to delete files with the same name as the parent folder? i wish to include the starting folder name as argument to the script.
Is there a program for Ubuntu that converts AVI files to VOB files? This is the only type of file that will burn to a DVD and successfully play on my DVD player. I can use DVD Flick on Windows but I'm sure this can't be the only way. There has to be a way to get VOB files on an Ubuntu operating system.
If i have files called, "hello123", "hello?","hello1" and "hellohoware" in the system and I want to list only the file "hello?" with the "find" command, how do I do it ?
find / -name hello? doesn't work as expected and also lists "hello1". I don't want "hello1" to be listed ...how do I list only "hello?" ?