i have a cheap-arse employer who doesn't provide any data backup for us. so i've setup my own very simple server with ubuntu 10.04 (no raid). from my linux desktop i want to backup to my file server. from my file server i configured samba to allow me to access my data from the many Windows PCs in the place. I also have a USB drive for when i go home or to another building.
most of my work happens on my linux desktop. so i use rsync to 'backup' from there to the file server. can i use rsync to keep my usb drive also sync'ed? in the same command?
when i access files on my file server from a Windows PC via samba, and change something, will rsync pick up that change on the server, and migrate the changes back to my linux desktop?
Thought I'd post it here because it's more server related than desktop... I have a script that does:
[Code]....
This is used to sync my local development snapshot with the live web server. There has to be a more compact way of doing this? Can I combine some of the rsyncs? Can I make the rsync set or keep the user and group affiliations? Can I exclude .* yet include .htaccess?
When I run rsync --recursive --times --perms --links --delete --exclude-from='Documents/exclude.txt' ./ /media/myusb/
where Documents/exclude.txt is
- /Downloads/ - /Desktop/books/
the files in those directories are still copied onto my USB.
And...
I used fetchmail to download all my gmail emails. When I run rsync -ar --exclude-from='/home/xtheunknown0/Documents/exclude.txt' ./ /media/myusb/ I get the first image at url.
I have a tiny shell script to rsync files between two servers and remove the source files.
This script works fine, when it has been initiated manually or even when the rsync command is executed on the command line.
But the same script doesn't work, when I try to automate it through crontab.
I am using 'abc' user to execute this rsync, instead of root, as root login to servers are restricted in all of our servers, by us.
As I mentioned earlier, manual execution works like charm!
When this rsync.sh is initiated through crontab, it runs the first command(chown abc.abc ...) perfectly without any issues. But the second line is not at all executed, and there is no log entry i can find at /mnt/xyz/folder/rsync.log.
I'm using Ubuntu 10.04 LTS server and Postgresql 8.4. I have a .sh script that is run by cron every other hour. That works fine. The .sh script includes an rsync command that copies a postgresql dump .tar file to a remote archive location via ssh. That fails when run by cron; I think because it is (quietly) asking for the remote user's password (and not getting it). I set up the public/private ssh key arrangement. The script succeeds when run manually as the same user that the cron job uses, and does not ask for the password. I am able to ssh to the remote server from the source server (using the same username) and not get the password prompt (both directions), so why doesn't rsync work? I even put a .pgpass file in the root of that user's directory with that user's password, and the user/password are identical on both servers.
I think the problem is rsync is not able to use the ssh key correctly. I tried adding this to my script but it didn't help.
Code:
Here is the rsync command embedding in the .sh script.
I'm considering setting up a virtual machine running Windows, with Ubuntu 10.10 as the host OS, for those cases where I have a Windows-only program.I understand that using a VM will lose some performance, but are there other limitations to what the OS in a virtual machine can do compared to "running on bare metal"?
For example:
Can a VM play games, like Dragon Age Origins or Civilization V? (Possibly with poorer framerates and/or lower resolution, but does it play at all?) Can a VM rip DVD/Blue-ray using AnyDVD or similar Windows program? Can a VM handle new hardware that requires dedicated drivers, but the drivers are only available for the OS running inside the VM? (Ex. graphics card, digital camera, card reader for smart card authentication.) Is it possible to say anything about "general limitations" of VMs, or is this wholly dependent on the specific VM?
* A static IP from my ISP * A router/modem in my home * Two machines: MachineA and MachineB connected to the router
Currently, the router is setup so that ports 80, 21, and 22 all forward to ports 80, 21, and 22 on Machine A.
However, I would like to be able to SSH into Machine B as well. What is the best way to accomplish this? I figured I would setup Machine B's SSH server to listed on a random port like 2995. So then I would configure my router so that incoming requests from port 2995 are forwarded to port 2995 on Machine B.
I'd like to know if it's possible to set the DISPLAY environment variable in order to display on two machines, the local and the remote, at the same time.
I am Working On Citrix Xen Server.I have Installed two Virtual Machines(Centos 5.3).Now Apache is Configured and its running on the First VM.Can I Set up a Apache Clustering On those VM?.My Aim is "If Apache On the First VM Down,then Apache on Second VM Should Automatically Start".Is there Any Tutorial to Setup Apache Clustering On Virtual machines.
I've installed a virtual machine on my computer. and I want to boot from its hard disk(which is on my /home/xen/domains/test1/disk.img). what I need is create a grub entry for this.
Two machines connected via LAN Older machine is running RedHat 3 Kernel 2.4.21-50.ELsmp on i686 Newer machine is at a current level of SUSE Linux. Newer machine has a DVD drive. Here's the question. Is it possible to, remotely, mount the DVD device to the older machine? If so, can it be used to install software packages?
suddenly ,I get a project to shutdown all xp's by a Linux server(RHEL5) using a single command.by net command i did that.but in this case i have to specify all the ip addresses manually .Is there any way that Linux server obtain the ip addresses automatically on a network and then shut down all the pc's
I usually monitor machines in the network via ruptime. But one machine (lambda) acts very odd. It never shows up in the output of ruptime, unless ruptime is run from a particular client (delta). For example, say I am logged in lambda. This is what the output of ruptime and few other commands are:
[Code]...
Linux delta.grp.vpncc.org 2.6.18-53.1.21.el5 #1 SMP Tue May 20 09:35:07 EDT 2008 x86_64 x86_64 x86_64 GNU/Linux Only difference between delta and other machines is the fact that delta is the file server that hosts the nfs directories shared in all the machines. All the machines are in the same subnet. All of them run either RHEL4, RHEL5 or CentOS5. lambda runs RHEL 5.4, delta runs CentOS 5.2. I tried restrting rwhod daemon on lambda, but no change. why lambda is playing this hide and seek with ruptime?
I just narrowly passed another exam on UNIX Academy. Their DVDs are incredible. While I made some mistakes here and there.I was really stumbled upon a question that was formulated something like "SNMP for managing Linux machines remotely".I can't remember the exact definition. I'm not sure what it was about.I was under impression that SNMP it is a carrier protocol for commands like "ping" and "traceroute". I was trying to read about it online but it is way too broad and technical, they're loosing me on very first page. I suspect there's some kind of product or technology that uses SNMP in some way to handle Linux, I'm not sure.
I have a server that runs a minimal instalation of debian 5.0. The server has only 128mb of ram. I am trying to install virtualmin but I keep getting the message: spinner: fork: Cannot allocate memory
I have read that this is because the system is out of memory. I am wondering if there still is a way to get this installed om my machine. After the installation is complete I have methods of reducing the amount of memory the application usses.
I have a network of 20 machines, all running Ubuntu 10.04.
Each machine has about 200[GB] of data that I'd like to share with all other 19 machines for READ ONLY PURPOSES. The reading should be done at the FASTEST POSSIBLE WAY.
A friend told me to look into setting up HTTP / FTP. Is it indeed the optimal way to share data between the machines (better than NFS)? if so, how do I go about it?
UPDATE: Just to clarify, all I want is to be able (from within machine X) to access one of machine Ys files and LOAD IT INTO MEMORY. all of the files are of uniform size (500 [KB]). Which method is fastest (SAMBA / NFS / HTTP / FTP)?
Im new to the forum and fairly new to Linux as well. my apologies if this is posted in the wrong section. My problem is How can you search for a file in multiple machines (like a server farm) ? For example i log onto machine num1 and want to search for a file named "xxx.yyy" which may be in one of 4 machnies. What i do right now is manually log into each machine and run the find command. However i have heard that it is possible to do it via a couple of simple commands. I have looked into pssh and cssh as well as ssh tunneling (along with public key authentication to stop the machines requesting a password every time i log in) and unfortunately i was unable to find an answer.
There are some encfs folders with private data on the server and all data is exported via nfs to all other omputers in the house.I can mount the encfs folders on another computer (using encfs command) to work with the data, but I never dared to mount it on more than one computer simultaniously, because I fear the encrypted data might get corrupted if more than one computer mount and access it at the same time.
So I want to ask about your experience: Is it safe to mount an encfs folder on several computers at the same time? All computers use "hard" and "sync" as nfs mount options to minimize risks of data loss. But can I access the folders simultaniously, or do I risk corrupting the encfs encryption and lose everything?
I was wondering if in RHEL5_3 if you can add netgroups to your kickstart file like when you are jumpstarting Solaris boxes? Trying to see if I can have a group of machines use a certain DHCP or DNS server during my kickstart.
I am wondering, whether this is possible. Can I change processor architecture type in virtualization software for different virtual machines? Like I want to do some compilation on Intel itanium and I have physically intel xeon CPU.
I have developed a website. Now i would like to monitor the IP address of machines who are accessing the server. The simplest possible solution to this that comes first into my mind is to use jsp inbuilt tags.
I have a few servers that are exposed to the internet. When someone tried to brute force hack in to the ssh, ossec adds their IP to the hosts.deny. Then the hacker (read: script kiddie) moves to the next IP up the line and hits my next server, etc, etc.
I end up getting 20 emails for all the servers that they hit.
My question, is there anyway to sync the hosts.deny file across multiple servers so that if they are locked out of one, they are locked out of all?
I am trying to run a rsync backup Script to backup my data from my HD to my NAS Drive (currently mounted as Samba share, but can change to NFS if that would improve things, if I can work out how!), but when I run the Script I get the following error on every file it tries to copy
Code:
As I mentioned I am mounting my Share from my NAS as Samba and below is the fstab entry
Code:
The rsync command with options is as follows
Code:
The error seems to be with the -a option, but I have no idea why.