Server :: Backup Nightly Some Items And Using A Command Line?
Jan 30, 2011
I have 2 'how-to's' with rsync neither are working. I need to backup nightly some items and using a command line, I am prompted with a password, enter and it works. I now tried these 2 items in the shell script, both of them still when run prompt for a password. I have tried;
#!/bin/bash
export RSYNC_PASSWORD=passwordhere
- and -
also created a blank file called pw (to test) and then tried
#!/bin/bash
rsync --password-file=/pw xxxx
Both of those options when run still prompt for a password. These severs are running old RH9 servers.
I have Mandriva 2006 on a computer that I regularly backup important directories every night to a flash drive. I use Cron for my backup with scheduled backups nightly. For my script file to do the backup, there is a line that says to copy these directories to /media/sda/backup, so I figure it was copying to the sda named usb device. Well it wasn't. It is copying to the /media/sda/backup directory on my harddrive. But I would like to copy it to a usb device. How do I find out the name of my usb device that I have plugged in?
I have a really quick question-I would like to set up some type of scheduled event to back up my entire /home folder to a USB drive.I know about all of the various programs such as simple backup, etc. and have used them before. As far as I know, these programs cant do what Im trying to do.Does anyone know what I could use to back up my files at a specific time to a specific USB device?Preferably, I would like to just have a simple
Code: sudo cp /home /media/Cruzer run every night at, say, 2am.
I have been looking around online and I am seeing that there are several solutions for doing a nightly automated backup on Linux. I was wondering what people here actually use for doing such and why they use one particular backup method over another.
What I am looking to do is every night (at say 3am) I want my system to backup my 200gig Documents folder to my external hard drive. Does Ubuntu have a tool built in by default to do this or do I need to add something from the repos/online?
Sadly I am unable to login to a particular user via the GUI, but that is another thread [URL].
I know that you can create a single backup file of all your emails and settings when you have access to the GUI in Evolution. Since I am unable to login to gnome (or KDE) but I am able to login to the command line by su to the user I would like to create a that evolution backup file. I will be reinstalling Ubuntu.
iam trying to sync file server data into backup server machine by command- rsync -avu path/of/data ipaddress-of-backup-server:/path/where/to/save after running it ask for root password and manually it is successful.but i want to make it automatic.for that i also tried cronjob and also generated authentication key but iam not successful in login automatically..anybody know how to authenticate root to login for storing data in backup server.
i want to make sure my mail server is behaving as expected. the past two nights, i've received logwatch notices below. the section in question is the 20 or so deferrals from some .ru e-mail address. i've since removed the user that registered on my wordpress blog (wlvp@yandex.ru) and added the three IP address to iptables DROP, but these "e-mails" still look like they're in some queue unless i'm reading wrong. i've removed known usernames in case you find the numbers useful. why are all these deferred messages here? if they're stuck in a queue, can i remove them?
Is there a way/command to back up all data from a Red hat Linux 4 serve[Including user rpofiles, data, group info, encrypts] either to a Red hat Linux 5.4 machine or as an Image file or manageable resource?
i've gotten my fedora 12 to the point where i can run python3 scripts from command line and can call up python 2.6.2 idle with the command 'idle' from command line. what command will call up python3 (3.1.2 to be exact) idle?
I was having so much trouble with ubuntu 8.04 that I deleted off my computer and did a fresh install of 9.10. I downloaded the 64 bit iso from the internet, burned it to a disc and installed it. When it asked if it should be the server version I thought I said no. Now when I boot it only boots to the server version and all I get is command line. Can I get out of this and get my regular ubuntu screen back or did I install the wrong version?
Whilst I have used GUI-based Linux distros for the last few years, I am now struggling somewhat when it comes to setting up a server from the command line. All the howto's I've read so far tell you how to set up certain things, but I don't know what I actually need to set up.Rather than asking the same questions all over again on a forum like this, and generally being a bit of a noob, I wonder if there's a resource somewhere that someone could direct me to in order to know what I need to do to set up a fileserver.
The hardware is all done, and I will probably set up a router distro like IpCop to manage the network, but I'm rather lost as to what packages I need, and how to know what the server is/isn't doing.I've used plenty of command lines in the past (ms dos, amiga dos, BASIC programming etc), just not a linux command line. All I've ever done is mount a few drives, use nmap and started x
I was using Red Hat 7.3 forever and decided it was time for a change. I went to Fedora 10 but it was really buggy. CentOS 5.2 is VERY stable.Here is my problem.The server is command line only -- I tend to hate GUI"S.I setup Samba no problem disabling the ports needed through the firewall and that was straightforward.CUPS is a nightmare for me since CentOS locks down the cupsd.conf and then the firewall does its thing. I allowed port 631 through the firewall but then got lost on the cupsd.conf. It's been too long and the old redhat one won't work with CentOS (not surprising since it's a VERY old system) straight-forward CUPS tutorial for a command line interface. I just need it to be:
when i use rsync command to backup my image file , it shows the following error message.
bash: line 1: /usr/bin/rsync: Argument list too long rsync: connection unexpectedly closed (0 bytes received so far) [receiver] rsync error: remote command could not be run (code 126) at io.c(463) [receiver=2.6.8]
The command which i used is rsync -avrl -e ssh cms@server:/data/cms/data/images/* /mnt/Backup/Intranet_cms_backup/images
Whenever I try and use php from the command line, I get this message:PHP Warning: PHP Startup: Unable to load dynamic library '/usr/lib64/php/modules/imagick.so' - /usr/lib64/php/modules/imagick.so: cannot open shared object file: No such file or directory in Unknown on line 0The problem isn't the missing module, it is the warming message itself -- it is breaking scripts that rely on the output.I tried shutting off errors, changed my error reporting to:
I just got done installing Ubuntu Server 10.04 on a home server. Now what should I do? I was hoping to have 3 computers connected to it so that we can all share files wirelessly if that's possible.Does Ubuntu Server have a GUI or command line only?
I am accessing a linux server remotely from my putty. I started the server and now I want to close the command line. when I do cntr+c or cntr+z it kills my server aswell. how will I close my terminal without closing my server? I tried cntr +d but it is not doing anything
Running: Red Hat Enterprise Linux Server 5.2 (Tikanga) I need to be able to automate transferring a few files over from one server to another using scp or the sftp protocol. I have received a text file which looks like a key file along with username and passphrase information for the target server in question.
Instructions were given to me to import the provided text file in puttyGen then save the imported key as a private key to be used by scp or sftp. My assumption is this is for windows utilities, which I am not using. My frustration comes in trying to automate logging into this server via sftp or scp to automate some file transfers. I am asked for a password every time because the public and private key methods failed to find my keys. How can I call scp or the sftp utilities and use the provided key file (the one I generated using puttyGen or the original one provided to me) to login to this server? I've tried taking the generated ppk file from puttygen and adding it with the ssh-add command but that still did not work.
I install software called Sunflow v0.07.2 in my Ubuntu. When I try to run it,it says "API error : JVM available memory is below 800 MB (found 728 MB only). Please make sure you launched the program with the -Xmx command line options". What is that xmx command line.What should i do to run Sunflow.
I have just got myself up and running on a ubuntu 32 bit server with postfix using this guide here [URL]...at-ispconfig-3 and everything is working great ( i didnt install ispconfig )
I want to get a command line mail client running but I am having some permission issue.
sudo aptitude install heirloom-mailx then when i type mail this happens mail /var/mail/jj: Permission denied
i also tried mutt etc but always get permission denied.
Total, total noob here, but I would like to know if this is possible, and maybe some tips on how to do it if it is.I just got myself a dedicated server, and I guess I assumed I could just remote control it via my own screen, much like using vnc or remote desktop, but that wasn't the case.The server has Debian 5.0, a "minimal install" is what they called it, I assume it's just installed and not configured at all.1. Connect to the server, and I guess the only way is via telnet? Since I have no VNC or remote desktop (yeah, microsoft, I know, but I don't know what it's called for Linux/GNU).2. Can I install VNC or somehow "activate" or "open" the GUI via telnet, much like a remote desktop kind of solution? I have used Debian's GUI on my laptop a couple of times, and I'm a total Windows-user, but I can navigate the graphical interface. I'm just COMPLETELY nulled when it comes to command line.So yeah, I guess I'm wondering how the heck I can remote control the server via a GUI, preferably VNC, and if I can install it via telnet or however I'm supposed to connect to the server?
Would anyone be so kind as to point me to a good reference for how use the command line to configure a directory to be shared in the directory properties?I know how to do it using graphical tools such as Nautilus, Konqueror, and Gnome's shares-admin, but I'd like to know how to do with via command line, because, once you know the commands, the command line is always quicker.I guess you could say I'm looking for a command line method for sharing directories in much the same way that chmod sets permissions from the command line.
Having an odd problem running a mysqldump via crontab. I have the script running on other servers and they work fine, so not sure how to actually troubleshoot, but the script looks like the following;
If I run it as a cronjob as root, it finishes in a second and a 20k file is there. If I run it from the command line as root it does the backup (takes a few minutes) but does complete the backup and can be unzipped and read successfully.
On our server we have a certain directory, say /storage, that contains many large files. They are all compressed (gzip). Many of our users are not computer-savvy, and so when one of these files is needed, they will copy it to their own directory. Consequently, we have multiple terabytes of duplicate data. I'd like to enforce an alias whereby if someone tries to use cp on a file from /storage, they will instead create a symbolic link. My idea was something like:
alias cp='cp.storage' File cp.storage:#!/bin/sh truePath=$(readlink -f "$1")
[code]....
The conditional checks whether the file being copied begins with "/storage". The problem with this is that if someone wanted to use cp with any options on a file not in /storage, those options would be obliterated. Can someone guide me as to a good way to accomplish this? Either a way to get the options from cp into cp.storage, or another approach not using alias this way. Everyone will be using bash.