I want to copy location of every .avi , .jpg file present in a folder or in subfolder present in a direcotry and save in a textfile how to doex : /home/username/Desktop/bookofeli/video/book.aviit should give full locaiton of path how to do
What i am trying is to check the file duplication in a folder and remove a file if it is a duplicate of another file ie the contents are duplicate; but names may be same.
Basically i am using md5sum to calculate the md5sum values of each file and redirecting to a file. And i am thinking of comparing the md5sum values.But i am finding it hard to decide how to complete the code after redirecting the output of calculation of md5sum to a file.
I noticed something a little odd I'm hoping someone can enlighten me on. I noticed in a couple of cases that a package has the proper version, but differs in two regards.
1. The package ends up with a .el4 on the end of the version for Red Hat 4.
2. The actual MD5Sum of the files the package provides differ.
An example below:
Code:
[root@RH4ES32-MCE bin]# for i in `rpm -ql GConf2`;do md5sum $i;done; md5sum: /etc/gconf/2: Is a directory 9f90335546f7c57ae6fb552cc2b919c5 /etc/gconf/2/path md5sum: /etc/gconf/gconf.xml.defaults: Is a directory
[code].....
So my package changed slightly to now show .el4 versus just 2-2.8.1-1 I've indicated in the first output above that the first couple of lines differ. I stopped my comparison at that point as they truly are different.
I am using Xfce as the desktop enviroment and Mozilla Firefox as the webrowser. Within the webrowser window, I do File>Save Page As. I save it, and the result is almost always foo.html and directory foo_files. But I think under KDE I could choose the format, one of them being something like "Single page" (only one file; the colecction of .png, etc is embedded into that file). And this is the format I want Xfce (or Firefox) to use when downloading to hdd.
I want to move all files and directories that are 1 month old out to back up into a separate folder. There will be a lot of files and I want to make sure it copies properly. The problem I'm having is integrating a MD5SUM into it to check integrity. MD5SUM is not recursive, so I figured it would work in a loop when it copies each individual file, I'll do a md5sum on each file and delete that md5 once its verified it copied ok.
[Code]...
I also need some sort of error handling to output all md5's that didnt pass the hash check.
I am trying to store the results of my code to a separate text file.But the problem is, as my results comes from a loop, my text file shows only the last result, not all of them.Like if the loop runs 5 time the text file shows the result for the 5th step.But i need to store all of them (1 to 5).Can I use awk to print the output field and store to another file and creat a new line so that the next output field goes to a new line?(just an idea, dont know).
#!/bin/basth for (( i=1; i<=5; i++)) do ./file.exe > output.txt done
My Ubuntu Remex Notebook had been install in windows7 by wubi, I want to reinstall it to solve some problems, but how to save my personal files in Downlads folder?
I've recently lost my window options, had to somehow manipulate my way to Xchat and ask some people how do I get it back (it was metacity --replace, and after I decided to stop the command and run it in background the X was completely useless so I had to do killall -u user). And that was after the internet connection stopped working for some reason (might've been the ISP).
The thing is, after using linux a long time, I still get the feeling that on dire situations, I don't know the good tricks (stuff like metacity --replace). I feel like a really need like a "rescue" cheatsheet for things like: how to save the X no matter what without pressing reset how to reset the system to "normal state" how to connect to the internet through the command line how to monitor what the X is doing (using ubuntu linux 10.04 btw)
This question may be silly and super easy for linux connaisseurs, but I was just wondering, for instance, I want to use the >find command to search for a file and send the results to a text file
I am looking for a terminal app that will allow me to save my password and maybe set up some sessions since I often log into multiple machines each day. Also is there something out there that will allow me to save common commands to a hot key?
I need to do some text file manipulation which I think should be done with standard commands in BASH. I'm looking at comma seperated text files (stock market data). It comes in the form of date, stock code, open, high, low, close, volume. What I need to do first is move all data with same stock code sequentially into individual files.
While doing this since the stock code will now be the file name I need to remove the stock code. Next I need to filter out overlapping data from different files with the same date. ie. where two files contain the same date on the one line only one line will be added to the combined file. I think there must be a tutorial out there for basic text manipulation like this, I just haven't found it yet.
I have a script almost working except for 1 thing. What I'm trying to do is read a file that has the files that need to be FTP'd using a bash script. I have everything working except the reading of the file. It works outside of the ftp script I've wrote but once I put it in the FTP script it doesn't.
Here's the Script:
#Here's where the problem is that I know of
I've been playing w/ the exclamation points to see if that could be the problem, but so far no luck.
I have a hidden folder with a lot of text files in it. I would like to search in this folder for all files containing a given text.The File Browser's" FIND searches only in the file names, not in their contents.The FIND function of Ubuntu does not allow me to search ONLY in the given hidden folder. So, how can I find my files within the hidden folder with the given text within them?
I have ubuntu 10.04.1 and a H.P.psc 1311 all-in-one printer scanner.Printer works ok but when trying to scan,with xsane,it goes through the scanning process and an image of the document comes up on the screen.
When I try to save this image to desktop or file/folder whatever I get the message; "Child Process Error. Failed to execute OCR command:GOCR:no such file or directory."
How to list the contents of a folder to a text file. I'm trying to list all my music, including all subfolders, etc. to a text file, but I can't remember the command.
I just switched from Windows to Linux. I've been using EditPlus for many years mainly because of it's ability to save locally, and then send the file via FTP, with easy keyboard shortcuts (ctrl-s to save, ctrl-alt-s to ftp).
I also need syntax highlighting and basic code editing features. Is there anything for Linux that can do this? I don't want to run EditPlus via Wine
I need to save locally and remotely simultaneously, or at least with a few keystrokes. I already know of plenty of ways to edit remote files.
may be an advanced question but I need to know how to do this. Here at work I am in charge of recruiting and we have about 1,000 resumes in already. All of the resumes are in a .pdf format. I need to rename every .pdf in the following format:{firstnameLastname}.pdfThe only way I know how to do this is to convert all the .pdf files to text, extract the name out of the first few lines of text, import into excel, and then use VBA to rename the files in mass:Here is my logic so far:~Deskop/a = houses all the .pdfresumesOpen terminal: Code: cd ~/Desktop/afor f in *.pdf; do pdftotext -raw $f; done That will convert all of the preceding resumes into text filesNow I would like to append the name of the text file into the last line of the text file. So, for example, for Resume1.txt, I want to append "Resume1.txt" to the last line within Resume1.txt. So after I run the command I open Resume1.txt and on the last line within I want to see "Resume1.txt" on the last line, at the end of the resume.How can I do this? I would like to use a loop and have the terminal append the filename to the body of the text file until all of the have been appended.
I'm looking for a fast way to verify a copy of a folder with 150Gigs of data, in 33 files. Some of the files are a few kb, while a few are 20-30Gigs. I've done a file count, which is quick, but doesn't verify that all the files are intact. I tried running md5sum on them, which works, but will probably take as long as copying the files in the first place. Diff works too, but is slow too.