General :: Script For Taking Backup Of Desktop Files
Apr 21, 2011
I have around 100 users. I want to take backup of files which are on desktop for every user. My user directory path is -: /home/dr/<user_name>/Desktop
1) Script has to run on a particular time everyday
2) Script has to take backup of all files present in "Desktop" directory
3) Make a tar with name "yyyy-mm-dd-desk-files"
4) Make directory outside "Desktop" with name "Desktop-Backup", if already exist then don't make this folder.
5) The tar have to moved in this folder.
6) Remove the files from "Desktop" directory. (i.e. Desktop should be empty)
7) Mail the status that "Backup Successful"
View 2 Replies
ADVERTISEMENT
Oct 15, 2010
I do use a virtualization environment using VirtualBox,Qemu. Many a times I have to format my laptop then in such situations installing again the previous programs and making same changes to the configurations is very painful. Is there some way I can make some sort of backup which I just need to install on my existing system and get all previous things immediately installed without going to install and setting of a lot of other applications and settings.
View 1 Replies
View Related
Mar 13, 2011
I used photorec to recover lost files and it brought up 70gb worth of files, when I was done looking through them I deleted these files. but these files still seem to be taking up my disk space. When I try to access my trash bin with root I get a message that reads...."The folder contents could not be displayed. sorry, could not display all the contents of "trash": operation not supported." if I open my trash bin when I'm not in root, the bin is empty.
How do I free up my disk space?
View 8 Replies
View Related
Aug 7, 2010
I am working in ubuntu 9.x (linux karmic kernal) .I have restored the content from CD to hard disk. In the mid way of this process, it was failed. I would like to know this below thiongs,
1) which position it got failed ?
2)Any offset option is there in linux to point the particular CD position ?
View 1 Replies
View Related
Jul 19, 2011
software required for taking remote desktop in mint 10
View 3 Replies
View Related
May 21, 2011
I'm just setting up a partition on a seperate HDD in my system. I plan to use the partition to backup the important files on my main HDD (to guard against HD crash).
The question I have is about where would be the typical location to auto mount this partition? Which would it be normal to go for:
1. /backup/
2. /media/backup/
3. /mnt/backup/
4. /home/chris/backup/
View 7 Replies
View Related
Feb 7, 2010
I have an old pc currently running ubuntu server 9.10. It was configured during install to connect to the home wifi router by a PCI ethernet card, which worked all well and good. However, at the moment I cannot connect to the router (I have moved the machine too far from it). I want to connect this machine (desktop) to the server so I can SSH into the box and backup some files. I need help creating a simple wired network connection between the two, as I have no clue as to where to start.
View 2 Replies
View Related
Jan 20, 2011
I am using Ubuntu 10.04 x86_64. I log in to the machine using nfs. For a problem with mounting my home directory, I had to copy all the contents of my home directory (including all configuration files) from a recent snapshot on to itself. That is, I did something like,
Code:
cp -r /home/user/user /home/user
All of my recent data and program configurations were in /home/user/user. So after the copy operation, I logged out and logged back in again to see that all my configuration and data was restored to what I wanted. But the problem is that now on my desktop I see hundreds of mounted volumes. These are coming from an hourly/weekly snapshot program. The tech support guys for my lab have suggested copying all relevant data to a backup and then deleting the home directory altogether. But I don't want to configure all programs all over again. I think I should be able to get rid of the problem by editing/deleting one or more desktop configuration files. I just don't know which ones. I tried looking around the gconf-editor but was overwhelmed at the amount of information on there.
View 3 Replies
View Related
May 6, 2010
I am not able to detect my USB hard drive in Linux rescue mode after passing "fdisk -l" as i want to take the data backup in my USB hard disk.
View 1 Replies
View Related
Jun 8, 2011
I have some confusion about one of my partition and the space it is taking. df -h output is given below;
# df -h
Filesystem Size Used Avail Use% Mounted on
/dev/mapper/ddf1_ADVDTARTINGp1
494G 18G 452G 4% /
[code]....
above information is showing that /var/lib/mysql partition total size is 379 GB and it is 68% used. However when I execute command du -sh /var/lib/mysql it shows following output.
# du -sh /var/lib/mysql
45G /var/lib/mysql
Now I want to know what files are taking space to make the partition 68% used. I want to list down all files in that partition with size.
View 8 Replies
View Related
Jul 27, 2010
I had this come up in logwatch: /dev/sda5 4.8G 3.6G 960M 80%/
Any ideas how I can find which files are suddenly taking up the space?
View 17 Replies
View Related
Jul 8, 2010
i'm looking for a reliable program to do backups to my desktop.(from the remote linux server....to my desktop )Preferably an FTP program that I can login to with Root access.... but most importantly, a program that will only copy over changed / new files after the first full backup. Thats the biggest problem I have currently.. when i do it with WinSCP, I usually backup the WHOLE entire /vhosts/ directory each and everytime. And its 2 GB's in size....takes forever when copying to desktop. I would only like changed / new files to be copied over after the first full backup.
View 5 Replies
View Related
May 10, 2011
I install and tested Restore EE Backup server on a test PC with basic configuration and its working fine.
[URL]
The issue i have is where is the location these backup snapshots or files are saving? I want to add a separate Storage to save the backup?
View 1 Replies
View Related
Feb 13, 2009
I am trying to make a demo of Sauerbraten for a friend of mine. I would like a short video clip of playing Sauerbraten, however I don't know how I can do this. Is there a program you download?
View 9 Replies
View Related
Mar 12, 2011
I have a situation where a directory has about 1.5 million files in it. On an hourly basis, I want to be able to find any files that have changed in the last hour, compress them, encrypt them and then copy them to both a local backup machine and an off site backup.
Is there any kind of utility or kernel module that creates some type of log of modified files? I know I can use find, but the search for -mtime in this directory takes quite a while and will not suffice for an hourly backup.
View 3 Replies
View Related
May 18, 2011
if there an application which does the following:
1. I have a laptop with an internal 200GB HD.
2. I run the application & it creates a list of all files (size & time-stamp) without actually storing them. Let's call this the "snapshot list".
3. I update some of the files on the laptop.
4. Now I run the application & it only copies the files which have changed on the laptop, that have different size/time-stamp from the snapshot list, onto some external media, such as a memory card. Of course, the files should be copied onto their proper location in the directory tree & not just pile up in one place.
Why is this useful? although the laptop has a 200GB HD I typically only update a small number of files, whose total size is maybe 10MB or so. If I could only backup those which have changed, I could do this with a tiny SD card instead of lugging around an external usb HD.
View 8 Replies
View Related
Jul 20, 2011
a server at work has been accessed through the desktop environment as root. I know this is a risk and since I have never done it before I was wondering if there are any files created by the desktop that could compromise the system and how could I clean it up.
View 2 Replies
View Related
Feb 2, 2010
I'm using gnome. I created some text files. After I change something in myfile.txt there's myfile.txt~ automatically created as a backup. Ok... now, on KDE I see these ~ files and I can delete them if I want but they're hidden in gnome... Gnome seems to treat them as special hidden files (just like .mydir directories) How do I make ~ files visible in gnome so I could edit or delete them?
View 1 Replies
View Related
Sep 13, 2010
I am using MySQL as the database system for my application on a Linux system. Every week I update the system and take backups (mysqldump) of the databases changed (2 databases). I then .tar.gz them and ftp the resulting file to a remote server, after which I remove the original backups and tar.gz files from the Linux server. Being a complete novice when it comes to Unix systems, I would like to know if it is possible to write a script which would do all this automatically, i.e. perform the following steps.
1) Backup database A to A.sql (mysqldump)
2) Backup database B to B.sql (mysqldump)
3) tar -cvzf dest.tar.gz A.sql B.sql
4) ftp dest.tar.gz to ftp@remoteserver.com
5) Delete A.sql, B.sql, dest.tar from local server
View 2 Replies
View Related
Oct 26, 2010
I backed up my Laptop with a script, as follows:
Code:
#! /bin/bash
sudo
growisofs -Z /dev/dvd -dvd-compat -r -v /home
I then installed a new version of Ubuntu 10.04 from disk and copied the files in /home from the cd to the hard. I am able to open, view etc. all the files in most directories except those in /home/documents. There are text files created by gedit, OOWP and several PDF files. I cannot open or view these files, depending: gedit and pdf files gets a Err.Msg. "Don't recognize file type" (it is clearly marked PDF) . The OO files look like rows of 'high bits' and a dialogue box opens giving me the options to change Char. Set, Font, Language, Paragraph break.
View 6 Replies
View Related
Apr 13, 2011
I have Kubuntu 10.10 64-bit, which was running fine until I did a bunch of apt-get software installs to permit me to compile YaWP weather plasmoid. In the process I am now at KDE 4.6.2, and my system is now non-functional, not able to start KDE desktop.
When KDE starts I get continuous krunner segfaults, krunner is taking 100% of the CPU, the desktop is there but the system is unusable. /var/log/messages is filled with the krunner segfaults, each line further indicating "error 4 in libkio.so.5.6.0". I tried reinstalling kubuntu-desktop, problem remains.
I expect I have created some software incompatibility with my installs, but I do not know how to diagnose/resolve.
View 5 Replies
View Related
Jul 10, 2011
Is there a way to have drag and drop access between Linux and a Windows remote desktop connection?
View 1 Replies
View Related
Mar 7, 2010
I have Ubuntu 9.10 installed using wupi inside windowsI just found about these files now, tried every possible way to delete the, including terminal commands mentioned here with no luck at allTake a look at these files in the screen shot here:http://img717.imageshack.us/img717/4169/badfiles.pngthey are totally un renamable or deletable, I tried even to login to windows and use unloaker and other utilities with no success at all
View 9 Replies
View Related
Apr 25, 2010
Ubuntu has been working fine since I dropped Windows six months ago , but recently I am experiencing problems when saving files that I have scanned in . They are taking an age to save and they are only small files less than 200kb ??? Saving anything takes an age now 5 mins +
View 2 Replies
View Related
Apr 29, 2011
I keep a backup of a bunch of files on a flash drive, so that whenever I change distributions I can just restore all my Android stuff (saves on re-downloading everything). One of these is the Android SDK.
In my ~/.bashrc I add the paths to some executables in the SDK, only if the directory exists, and only if the path is not already in $PATH. For the Android NDK this works fine, but for the SDK I get this:
Code:
snfo@snfo:~$ adb devices
bash: /home/snfo/Android/sdk/platform-tools/adb: No such file or directory
snfo@snfo:~$ ls -F /home/snfo/Android/sdk/platform-tools/adb
/home/snfo/Android/sdk/platform-tools/adb*
Everything else is fine though, just that one path is causing trouble.
Now, I've saw something similar to this before whenever you move an executable from one place to another. If you don't re-source your bash config it will continue to keep looking wherever it used to be located. But I've never moved these files.
View 3 Replies
View Related
Dec 8, 2010
I am backing up parts of my computer with DD, and i was wondering if there was a quick way to split the files created into 4.4GB sized files that will fit onto a DVD. Anyone have any idea of how to do this?
View 6 Replies
View Related
Jul 24, 2010
Today I tried to compress some folders containing backup files from last year. I right-clicked on the folders and selected compress as tar.gz. I let it work, and found that hours later, the folders were still compressing. How long is it supposed to take, anyway? I was trying to compress the two sets of backups simultaneously; together they're around 1.5 GB. They have many subdirectories.
View 4 Replies
View Related
Jul 22, 2010
I'm trying to create backup/archive my Ubuntu 10.04 system files (so I can restore it in case my system get corrupted). More specifically, I'm trying to zip the important files in my root directory not including my home directory (which includes my documents which I backup separately/more frequently) to an external hard drive attached via USB (called 'My Book').
Since File Roller didn't give me quite the level of control I was looking for, I created a script that I could execute to backup and archive regularly. Here's a snippet:
cd /media/"My Book"/"Linux Backups"
NOW=$(date +"%b-%d-%y")
LOGFILE=Backup_Root_FileSystem-$NOW.log
sudo zip -r -T -v Backup_Root_FileSystem-$NOW / -x /media/'My Book'* /media* /proc* /sys* /mnt* /dev* /cdrom* /home* /'lost+found'* | tee -a $LOGFILE
[Code]...
View 5 Replies
View Related
Dec 6, 2010
I've got a backup.list file that looks like this:
+ /ext/installs
+ /ext/media
- /ext
- /backup
- /dev
- /mnt
- /proc
/ext is a symlink to /mnt/vg1/vol1, under which are installs and media directories.
I'm running:
rdiff-backup --backup-mode --include-globbing-filelist backup.list / /backup
rdiff-backup keeps recreating a symlink: /backup/ext -> /mnt/vg1/vol1, which is kind of missing the point, because then it doesn't actually back up the files in /mnt/vg1/vol1/installs, etc.
View 1 Replies
View Related
Jul 18, 2010
I have two NASes. I work off of one, and the other is used as a backup. As I have it set up now, it's slow. Running a backup takes a week. Even for 7 TB, with 1,979,407 files, this seems a bit outlandish,particularly as both systems are RAID-5 and the network is all gigabit. I've been digging about in the rsync man pages, and I really don't understand what differentiates the various topologies.Right now, all the processing is being done on the backup NAS, which has the main volume from the main NAS mounted locally over SMB. I suspect that the SMB overhead is killing me, particularly when dealing with lots of files.
I think what I need is to set up rsync on the main nas as a daemon, and then run a local rsync client to connect to it, which would hopefully allow me to completely avoid the whole SMB-in-the-middle affair, but aside from mentioning that it's there, I can find very little information on why one would want to use the daemon mode for rsync.
Here's my current rsync command line: rsync -r -progress --delete /cifs/Thecus/ /mnt/Storage/input? Is there a better way/tool to do this? Edit:Ok, to address the additional questions: The "Main" NAS is a Thecus N7700. I have additional modules installed that give me SSH, and it has rsync, but it's not in the $PATH, and I havn't figured out how to edit the local $PATH in a way that persists between reboots. The "Backup" NAS is a DIY affair, built around a 1.6Ghz Via Mobo with a Adaptec Hardware RAID card. It's running CentOS 5 with a full desktop environment. It's the hardware I'm running rsync from. (Gigabit is through a additional PCI card).
Further Edit: Ok, got rsync over SSH working (thanks, lajuette!).I had to do a bit of tweaking on my command line, I'm running rsync with the args:rsync -rum --inplace --progress --delete --rsync-path=/opt/bin/rsync sys@10.1.1.10:/raid/data/Storage /mnt/Storage (Note: I'm specifically not using -a, because I want to change the ownership to the local account, to not freak-out SELinux)
View 5 Replies
View Related