I'm having some trouble to find a file system that allows me to backup my data to an external HD, and then access this data in other Linux (and sometimes windows and mac) machines.
The file system that I'm looking for must:
- have no user permissions: anyone can do anything with the data;
- have support to large files: I've used FAT for a while but it just sucks;
- maybe support access in windows and or mac with additional drivers;
- have journaling (or something of the sorts) to reduce the risk of data loss.
What would be the best way to have automated system backups? I'm trying to get it so my Xubuntu box automatically backs up the entire system including user settings on regular intervals, what would be the best way to do it? I have 2 hard drives with one that I do not use that I'd like to backup to.
I've been a DOS/Windows guy for 20 years, and recently became a SW test lab helper. My company uses CentOS for a lot, so I've become familiar with it, but obviously not as comfortable as I am with Windows.
Here's what I have planned:
machine: Core 2 Duo E8400, 8GB DDR2, 60GB SSD OS drive, ATI 4650 video card, other storage is flexible (I have 3 1TB drives and 4 750GB drives around that can be used in this machine.)
uses: HTPC, Network Storage, VMWare server host: SMTP, FTP server, and Web server virtual machines
I've figured out how to do much of this, but I haven't figured out how to do backups in Linux. I've been spoiled with Windows, with the built in backup system so simple to use. I find myself overwhelmed with the array of backup software, and unable to determine which to use. none of them seem to do everything I need them to do, but some come close, I think. I'm hoping someone here can help me out in figuring out which program to use and how to use it.
Here is what I need the backup software to do: 1. scheduled unattended backups, with alerts if the backups fail 2. a weekly full backup with incremental every 12 hours 3. removing the old backups when the new full backup runs, I would prefer to keep 2 weeks of backups, but that's not necessary 4. a GUI would be preferable, since my arthritic fingers don't always do as I want them to do. I typo things a lot, and the label worn off my backspace can attest to that.
How do you get Rsync to do incremental backups rather than full backups? At the moment I have a script that will create a backup folder (if it doesnt already exist) then copy the source files into the backup directory with the command
Target is where the files will be backed up to Sources is the dir(s) to be backed up Exclude files is the list of files not to backup log file is where the output will be saved to. At the moment it only does full backups, but I would only like to do incremental, how would this be achieved? Am I missing out an option in the Rsync that is required.
As you know everywhere we create a new file ,immediately a backup file with suffix ~ will be created with it if we open the file , i deleted a file but always i have problem with its backup how can delete them with deleting the file?
I am in the process of writing an rsync script to run unattended backups of my entire file system to another system located on my local network using ssh and password-less rsa keys.
I will absolutely will not use password-less keys with the root account and this is the limitation preventing me from accomplishing my goal because root is required by rsync to access the / tree and copy it to another location. I decided that if I compiled the script into a binary that I didn't have a problem with the password being contained within the binary itself but from what I've read there is no way to elevate to root and then back down to user level from within the script/binary.
I can create the script as the user and use chroot to make it owned by root but retain execution permission for the user but it will still cause the ssh login to be under root and therefore require either that I am there to enter my password or the use of password-less keys under the root account which I reiterate I will NOT do. Currently the script is executed by the user on the machine containing the files to be backed up.
I'm currently backupping our home data (pictures, videos, our CDs ripped to FLAC which I spent a lot of time to tag accurately ), totalling almost 300 Gb, on 2 external USB drives, one of which is meant to stay at a friend's. I left the factory msdos filesystem as it was, thinking it could be useful to be able to connect the drives to a windows machine with no problems. It's certainly useful to have «normal» data that I can take with me e.g. when visiting my family.
I'm simply using rsync manually, checking for suspicious changed or deleted file before commiting the change. I do that every 2 weeks or so.
Now I want to add a file integrity management to my backupping scheme: I want to be able to check that new data I'll be committing has not been tampered with (integrity check before updating tags on my main drive), and I want to be able to check that backupped data is still sane on my USB drives, especially if I need to recover from data corruption on my main drive.
Since I'm essentially mirroring the data, I thought run of the mill integrity software would let me just rsync the integrity database, and I'm done.
But after browsing through the docs of tripware, afick and the like, I fear they work only with absolute paths, so the database for my main drive wouldn't work for my USB drive, that's mounted elsewhere when I plug it in, obviously.
So, I feel I'm missing something. It looks to me I'm trying to solve a very common problem, how do people do it?
Did I miss a file integrity software that works with backups?
Is there a trick like using a symbolic link pointing to whatever file hierarchy I want to check, and have tripware/afick/... monitor that link?
Should I run a more elaborate backupping system than plain rsync? Which one? (Storebackup for instance looks promising since it involves md5 sums, but it's targetting a completely different problem, and I'm not sure I can use it at all for what I need.)
**Edit: path for mount was incorrect Distro Server: CentOS 5.5
Clients: Fedora(latest) OSX(latest)
Backround I am attempting to setup a server in my house mostly(for the first time) for backups and file sharing. It is important to me that file permissions are preserved. So its my understanding that I must use idmapd in order for this to work. As of now I'm only working with the linux distros while osx will be dealt with once these two work together. portmapper is up and running, along with lockd on both machines. Firewalls are also down on both machines for now. The server side was all setup using the GUI interface with no extra options selected. Problem When attempting to "mount -t nfs4 10.0.0.2/$sharedfolder /mnt" I get an error operation not permitted with no error printing in /var/log/message. If I use "mount -t -o nolock nfs4 10.0.0.2/$sharedfolder /mnt" it mounts just fine. Ive checked both machines multiple times to make sure that lockd is up and running. In the idmapd.conf file I the domain as "localdomain" for both machines but I doubt that is right; like I stated above this is my first attempt at a server. I'm assuming the problem is a whole missing step that involves some kind of id mapping server I need to setup.
I've been trying to create file sharing between my MacBook Pro and my Lucid Lynx box for time machine backups and media server purposes. I followed this guide:[URL]..Everything seems to work with these exceptions: I can see my LucidLynx box in my finder app in my Mac but only when I run these commands from Ubuntu:
If I restart my LucidLynx box then I can't see anything in finder. I can't log into my LucidLynx box from finder. I don't get a bad username or password error it just tells me the connection failed. *Note if I do enter an incorrect username or password it WILL tell me it's incorrect. I've looked at this link below since some people have used it in theses forums but it's a bit dated[URL]..
A while back I installed Dreamlinux 3.5 Gnome edition using ext2. When I attempted to use the email address books I imported from the Dreamlinux3.5 XFCE edition, which had been ext3, I discovered that none of the email addresses could be mailed to. I had to manually type in the addresses.
When I reinstalled Dreamlinux 3.5 Gnome using ext3, the same backup files that did not work in ext2 now work just fine. The question is, was this a "broken data" problem caused by the switch to ext2 file system or something else? Has anyone else experienced this?
I'm trying to find a secure way to backup files on my Prod Server to Backup Server. It must be automated, so I will need to run a command with cron which will login to Prod Server from Backup Server and backup data. 1. Do you think it would be secure enough to do this by creating an passwordless RSA private key on Backup Server and adding it's public key to authorized_hosts file on Prod Server? I can't think of a way to Automate this without having to enter any passwords without passwordless RSA key. Is there another. more secure way? 2. Should I create a special user for backup, which will only have read access to all files in the directory that I am backing up? If so, How can I run a check that this new backup user indeed has read access to ALL files in the folder that I intent to back up? How can I ensure the backup process will not skip files due to some permission problem? 3. I'm thinking of using rsnapshot tool, which uses rsync.
I bought a new SD card which I intend to put some MP3s on - except that I can't write to it because it tells me the destination is Read Only. No-probs thinks I: I'll just reformat it.
"Error creating file system: helper exited with exit code 1: cannot open /dev/mmcblk0p1: Read-only file system"
Various chmod commands all result in Read-only file system. I tried umount then mount commands, but it couldn't find it to mount once I'd unmounted it using the same /media/ file path (I assume it's the only one).
My Redhat Enterprise Linux 4 with 6x partitions (/, /boot,/home, /usr, /var, /tmp) of 6.0 GB IDE Hardisk was working quite fine. I decided to create LVM on /home and /var partitions but due to some errors occured and I delete the /home partitions. That's why partition table altered. I then delete 4,5,and 6th partitions (/home, /var, /tmp) partitions and now try to create one by one but following error is coming:-
[Code]....
The Super block could not be read or do not describe a clear ext2 file system. E2fsck b 8193 <device> I have tried following commands,but could not successful:- e2fsck -p /dev/hda7 (where hda7 was created but afterthat it was deleted) e2fsck -a /dev/hda7
I'm looking for ... or will develop if I cannot find any ... software that can keep archives of a tree of files. The tree of files is on a computer dedicated to backups. Various servers and desktop users periodically run rsync to update their slice of that tree of files. But that means old files are replaced. What I want to do is keep the old files around, with some limitations (such as a finite number of versions, finite age, etc). This would be similar to a repository like Subversion, except that data only comes in (there are no checkouts, though obviously a restore mechanism is needed). It would be like rolling backups, except the "rollover" would be on a per-file basis.
What I currently do, but want to get away from, is making a hardlink replica tree every few days with a command like "cp -al synctree/. archtree-for-today". But that eats up inodes very rapidly and is much harder to find things in. Making a file listing is very much slower this way, too.
Have just assembled a new computer and thought I would install the 64 bit version of openSUSE 11.2 in a "Windows free zone". After a hiccup or two I have managed to get a system of sorts running but on trying to copy files from my old computer(via a memory stick) it tells me that Vfat is an unknown file system.On my old computer I am running 32 bit openSUSE 11.2 as a dual boot system with Windows XP and have no problems moving files between the two different file systems.Is it possible to get a 64 bit file system to read 32 bit file system drives and if so how do I do it?
I'm a little bit confused with partitioning the filesystem in Linux. the difference between creating the file system with fdisk and mkfs (when formatting the disk). I can't clearly tell my problem, so please look at this picture:
i have generated .exe file from C file (ie filename.c ) after compiling in linux machine with -O option. I wish to know about how to run that .exe file when linux system starts up ?
In System Monitor, on the File Systems tab, the "Total", "Available" and "Used" columns don't seem to add up, and the "Used" percentage doesn't seem correct either.
My config: /dev/sda1 = 80 GB SSD drive, / partition. /dev/sdb1 = 50 GB FAT32 partition of an external 500 GB USB hard disk. /dev/sdb2 = remainder of the 500 GB USB hard disk encrypted using luks.
Screenshot: The /dev/sda1 figures don't really add up well, but they're close at least (how you get "50% Used" from any of those figures I don't know!).
However, for /dev/sdb2, they're miles off: "Free" = 146.2 GiB "Total" = 409.7 GiB "Available" = 125.4 GiB "Used" = 263.5 GiB
I am getting occasional errors during the boot process.One at the beginning and one or two when I switch to single user mode.I 'd like to run the system file checker to fix any possible errors.But when I run fsck in the terminal I get the message:Code:
mansour@ubuntu-notebook:~$ fsck fsck from util-linux-ng 2.17.2 e2fsck 1.41.12 (17-May-2010)
I backed my photographs with K3B in Suse, prior to installing Ubuntu. Now I can't access them. Well I can but not with Ubuntu or Windows 7. Suse (KDE) on the family computer reads them with no problem?? I'm confused. Is there something that KDE has that Debian doesn't. there's nothing wrong with my system exept for this. Reads cd's and dvd's.
So 2 days ago everything was all fine on my machine. Has been for about a month, but all of a sudden as of yesterday I have no sound, I am seeing IRQ interupts on boot, During boot I am seeing file system is not clean, , and swap space is being used for the first time while doing normal task, etc. These are 2 new hard drives in RAID 1 with ReiserFS. I should have used a newer FS but thats a whole other argument.
Anyways here we go. The system is Debian Lenny amd64 Physical RAM 4GB + 6GB swap /var/log/messages
Code: Feb 21 07:35:09 Sarah kernel: imklog 3.18.6, log source = /proc/kmsg started. Feb 21 07:35:09 Sarah rsyslogd: [origin software="rsyslogd" swVersion="3.18.6" x-pid="3994" x-info="http://www.rsyslog.com"] restart code....
I plugged in my USb drive into my computer yesterday and tried to delete a folder. I was unable to do so and got the following message
Cannot move file to trash, do you want to delete immediately? The file "my file" cannot be moved to the trash. Show Details Unable to create trashing info file: Read-only file system
So when I click on delete I get another error message:
Error while deleting. There was an error deleting Case Study Database. Show Details Error removing file: Read-only file system
At this point I can only click on Skip, Skip All, or Cancel.
I have not changed anything on the stick recently so I dont know what is causing the problem.
reducing the size of LVM. I did it by using the commands, lvreduce,fsck,resize2fs.After reducing the lvm size, my system is not booting...it is saying file system corrupted error.
I have asked this over on Launchpad and have found bugs filed on the eclipse bug tracker but it seems to be going nowhere so I guess ill ask here.
I am using Eclipse for some development work and having saved a file, defocus the eclipse window (to test changes in a browser) when I refocus the eclipse window I get the following error...
The file 'path omyfile' has been changed on the file system.Do you want to replace the editor contents with these changes?
I have searched and searched and this seems to be a CIFS/SAMBA problem. I even found one solution suggesting that changing his mounts from cifs to smbfs fixed his problem however that would appear not to be an option for me since using smbfs in the fstab causes cifs mounts. This problem would seem to occur in Bluefish as well as some other IDE's so it is not an Eclipse issue.
---------- LINKS TO RELATED ---------- Launchpad Question Eclipse Bug --------------------------------------
I'm only just starting out with the Linux ubunto 10.04 OS after yeas of wasted time on Microsoft os's,I hope I'm posting this request for help in the right forum thread, if not please accept my apologies, I have tried searching everywhere for help in installing a firmware file into the File System / lib / firmware directory and each time I get an access denied result. The file is for a DVB board and I have managed to track down the right Linux fw file for this particular piece of equipment, Could some kind helpful person either explain how to get this firmware file into the Root System directory or even send a link to another site that deals with this sort of problemI've downloaded all the programs via the Ubuntu Software Center that should be able to perform this task however all to no avail.The reason why I posted this thread in this forum board is that it (in my own personal opinion which may be wrong) seems to me to be a security problem
I'm planning to add 1tb sata disk to my lovely file-server under ubuntu 10.10,what i want is use this disk as additional storage for network user,indows and ubuntu?I mean when my ubuntu server down (worse case) I can easily take out the disk from ubuntu machine and plug in on windows machine
I recently upgraded to Ubuntu 11 and a few days later my ecryptfs filesystem began misbehaving in a weird way. In my home directory, many subdirectory names are duplicated verbatim. Here's an ls -F excerpt:
I can no longer access files in those directories (if I ls the directory, it appears empty; I can cd to it, but there's nothing inside). Not all of the directories are duplicated/damaged like this, but most are. A few non-directory files are also duplicated in this fashion, so for example:
Can windows read files from a home file server with an ext4 file system? or do I have to partition the drive with the server (ext4) and an ntfs partition with the files on?