General :: Archive All Files To One Single File In Every 30 Days?
Apr 12, 2010
I have a directory there are many files are writing to it , I would like to write a script to do that , can please provide the advise .
archive all files to one single file in every 30 days , and then remove these old files .
View 9 Replies
ADVERTISEMENT
Jul 22, 2010
AKA "zipping on the fly .. the slow-as-molasses way." The list includes full pathnames to each file, and they're all in subfolders of the same parent folder (which, unfortunately, is not the root folder of the drive or system on which the files reside). A cleaned-up and radio-ready portion of the list looks like
Quote:
.../taiga/ahqr-va-choyvp/bv0884-073.jpg
.../taiga/ahqr-va-choyvp/bv2635-073.jpg
.../taiga/ahqr-va-choyvp/bv3067-175.jpg
[code]....
What I'd like to be able to do is zip all the files in the list into a single archive, to avoid the step of having to copy them to the same location (presumably another folder on the HD) and then zip that folder. I'm more inclined to make provisions about extracting to a single folder at some other time. Is this possible in BASH, or would I have to consider a faster, more robust scripting language such as python or perl?
View 8 Replies
View Related
Jan 3, 2011
i have 10 vi files . these files contain some system related information. i need to combine the output of all these files into a single file. the final file should contain contents of all these 10 files and the output should be in a tabular format.
is there any command in vi that i can use to create a table ?
View 9 Replies
View Related
Apr 13, 2010
i have started using linux for less than 6 months. now i have come across a problem with pdf files in linux. i want to join different pages from different pdf files into single pdf file.i have come across softwares that do this but they perform this using page numbers from pdf files.but i need to do this based on keywords in different pages .for eg there 3 pdf files
india.pdf
contents:languages
.........
........
places
......
......
achievments
......
[Code]....
now i have to create a pdf file langunage.pdf ,combining the topic languanges from three pdf files america.pdf,india.pdf,china.pdf how can i do it?? whether there is any open source software for doing this?.
View 3 Replies
View Related
Nov 28, 2010
I am using Xfce as the desktop enviroment and Mozilla Firefox as the webrowser. Within the webrowser window, I do File>Save Page As. I save it, and the result is almost always foo.html and directory foo_files. But I think under KDE I could choose the format, one of them being something like "Single page" (only one file; the colecction of .png, etc is embedded into that file). And this is the format I want Xfce (or Firefox) to use when downloading to hdd.
View 1 Replies
View Related
Dec 6, 2010
I'd like to ask about archive mounter feature, can I mount zip file with read write mode? can gvfsd-archive do that?, or I must use fuse-zip to mount it? If I must use fuse-zip, how I wrap it so I can use it via nautilus or via gvfs-fuse-daemon
View 4 Replies
View Related
Feb 19, 2010
When trying to create a new compressed/archive file in Gnome Commander (GM) the file is created but the selected files are not added. I can open the new (empty) archive file and then add files to be compressed. I have tried using several different formats (zip, tar.bz and others) with the same results. The "file roller" is shown as a plugin but has no configuration other than the compressed file type.
View 2 Replies
View Related
May 6, 2011
I have searched for a way to copy file less than X days old and I found this:http://www.howtogeek.com/howto/ubunt...days-on-linux/ The syntax for deleting files less than 7 days old would be like this:find /path/to/files* -mtime -7 -exec rm {} ;I would like to copy the files to mntas, and I'm not sure what the syntax should be.ould this work?find /path/to/files* -mtime -7 -exec cp {} mnt
as ;
View 3 Replies
View Related
Apr 19, 2010
i am a newbie in linux ,i am writing a bash script to identify the files which are exactly 7 days ( a week old) i tried this command find /var/backup -mtime +7 -exec ls -d {} ;but this gives me even the files which are older than 7 days
[root@proxy access]# find . -mtime +7 -exec ls -d {} ;
./access.log.1.gz
./access.log.2.gz
[code]...
View 3 Replies
View Related
Jan 6, 2011
I am trying to put together ksh shell and I am new with writing scripts.How do you write a command to delete any files if it's 30 days old and also it's not currently being locked?
View 10 Replies
View Related
Sep 18, 2009
find -type f /path/* -mtime +7 -exec rm {} ;Is this the best way to delete only files (not directories) within /path that are older than 7 days? or is there a better way?
View 3 Replies
View Related
May 14, 2010
Command to find all of the files which have been accessed within the last 30 days?
View 1 Replies
View Related
Jan 22, 2010
I`m totally new to linux, in fact I`m a windows adminscenario:I need to run a script that will automatically move 30 days old files from particular folder to a particular folder.
View 3 Replies
View Related
Jul 19, 2010
I have a folder containing 5000+ zip files and I need to add a single file to each zip. How to do it quick way?
View 1 Replies
View Related
Jul 6, 2010
I'm trying to truncate a postfix Maildata directory for one of our users. I want to be able to move any files older than <n> days to a new location, but also copying the relevant directory structure. This should be doable in the one comman. I've used find to locate the files, and mv to move them, but I can't figure out how to build the directoryt structure on the fly in the new location.
This is what I have so far:
Code:
find /Maildata/editor -mtime +42 -type f -exec mv -v {} /EditorEmailOld/ ; > ~/editor.txt
View 4 Replies
View Related
Feb 28, 2010
I have an audiobook in 64 small mp3 files and I need them combined into a single file of any format in the proper order or order that I add them in. What program can I use?
View 9 Replies
View Related
Mar 4, 2011
I'm trying to zip or rar >100000 files into a single file so that I can upload it to my server much faster than ftp downloaded it. Total they're all only 4gb, but because of the number of files Nautilus freezes just opening the folder they're in. They're all .jpgs and all in the same folder and I've tried a few commands but I keep getting error messages.
Anyone have a command that will archive all the files from a folder into a single zip (or rar, tar etc)? I can't just archive the folder because then I would have to move all the files out of that folder and just opening the folder to move them would crash it, and I don't have ssh into that server.
View 3 Replies
View Related
Jul 22, 2010
I'm trying to create backup/archive my Ubuntu 10.04 system files (so I can restore it in case my system get corrupted). More specifically, I'm trying to zip the important files in my root directory not including my home directory (which includes my documents which I backup separately/more frequently) to an external hard drive attached via USB (called 'My Book').
Since File Roller didn't give me quite the level of control I was looking for, I created a script that I could execute to backup and archive regularly. Here's a snippet:
cd /media/"My Book"/"Linux Backups"
NOW=$(date +"%b-%d-%y")
LOGFILE=Backup_Root_FileSystem-$NOW.log
sudo zip -r -T -v Backup_Root_FileSystem-$NOW / -x /media/'My Book'* /media* /proc* /sys* /mnt* /dev* /cdrom* /home* /'lost+found'* | tee -a $LOGFILE
[Code]...
View 5 Replies
View Related
Dec 19, 2010
why the file manager, nautilus is it called? can't cope with folders with extremely large numbers of files in it, is there a fix? if not could be a future project for future distributions, i have some folders with 20, 000+ files in it, for rendering fractal animations, i am trying to free up some room, the trash can can't empty it from trash because there are so many files, and when i try to open the folder, it crashes and restarts completely
View 1 Replies
View Related
Mar 8, 2009
When trying to delete :
I must say that the tar file contain only soft links...not real files...
View 5 Replies
View Related
May 19, 2010
is there a way to view the Unix permissions for a file under Windows?
View 1 Replies
View Related
Jun 14, 2011
I want to run rsync on server A to copy all files from Server B when they are newer than 7 days.(find . -mtime -7) I don't want to delete the files on Server B.
View 2 Replies
View Related
Jul 20, 2010
i am downloaded some e-books in the format .rar. when i am extracting them i am getting error as There is no command installed for RAR archive files. Do you want to search for a command to open this file?
View 3 Replies
View Related
Jul 20, 2010
I have very expensive data on my laptop and many user using my laptop so i want to lock my single drive that other user not access my files and my single drive. please tell me right answer ASAP
View 3 Replies
View Related
Jul 13, 2011
I am new to the world of linux and when attempting to verify a tar archive I am displayed the following error. When running the command tar cvfW archivename.tar filename directoryname does not yield any errors.
View 1 Replies
View Related
Mar 23, 2010
Recently i am trying to check on the rsync speed for single file(2.4GB iso) directory ( 900MB directory with files inside ) When i run the rsync for single file: the speed i get is average 50MBps However, for a directory: average speed is 10MBps Is there any reason behind this ? i tried to google but unable to get the concept.
View 2 Replies
View Related
Jun 27, 2011
I just want to know if we can search for a particular string in all files of a particular folder in linux. if a folder has 100files only that hundred files needs to be checked for a particular string.
View 5 Replies
View Related
May 5, 2011
Is there any Linux utility to combine two or more binary files into a single binary file ?
View 7 Replies
View Related
Sep 24, 2010
I have a script that moves files after 60 days from the FTP folder over to a trash folder which is not accessible by the FTP. From there I can delete the files whenever the drive get's full. That works fine so far.
The only thing that bothers me is that files moved to ".Trash" are not in any folder structure anymore (one big directory with the deleted files in it).
When I do a "ls -la" on the array {} I see the file names including the folders. I'm not sure if the find command or the mv command forgets about the directories.
This is the script: bash -c 'date;find /Volumes/data1/test/ -mtime +60 -type f -exec mv {} /Volumes/data1/.Trash/ ;;date' >> ~/mylog
View 2 Replies
View Related
Jan 3, 2011
I am new to unix, I am looking for a script to delete files older than 7 days but i also want to exclude certain directories (like arch,log .....) and also some files with extensions ike .ksh, .ch, ..............) in directories and sub directories
View 4 Replies
View Related