Server :: How To Restore Dump Command
Sep 7, 2010I am still new in linux (Redhat) i used dump command to backup the root of the linux server: #dump -0u -f /dev/st0 / the command is achieved. how to restore this dump.
View 2 RepliesI am still new in linux (Redhat) i used dump command to backup the root of the linux server: #dump -0u -f /dev/st0 / the command is achieved. how to restore this dump.
View 2 RepliesNeed confirmation if the following scenario works for making my client and server as identical?
My local(source) Linux server @192.168.0.2
My remote Linux client @192.168.0.70
On the local system :
#df -m
Filesystem Mounted on
/dev/hda3 /
/dev/hda1 /boot
tmpfs /dev/shm
On the local system , issue the followings to make client and server as identical :
#dump -0uvf - /dev/hda3 | ssh root@192.168.0.70 -c "restore -rf - /"
#dump -0uvf - /dev/hda1 | ssh root@192.168.0.70 -c "restore -rf - /boot"
#dump -0uvf - /dev/shm | ssh root@192.168.0.70 -c "restore -rf - /tmpfs"
I have used Dump Command to dump the application files. For Full backup the level 0 is working fine. For incremental backup I used the level 1 or 2 it is getting the error as
DUMP: Only level 0 dumps are allowed on a subdirectory
DUMP: The ENTIRE dump is aborted.
The code I used
===============================
#!/bin/bash
#Full Day Backup Script
#application folders backup
#test is the username
now=$(date +"%d-%m-%Y")
[Code]...
I got a backup file which first 1000 bytes are as follows:
Code:
00000000 54 41 50 45 00 00 03 00 8c 00 0e 01 00 00 00 00 |TAPE............|
00000010 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 |................|
*
00000030 02 00 87 05 51 64 50 48 01 00 00 00 01 00 00 00 |....QdPH........|
[code]....
I tried to restore it:
Code:
$ restore -r -f zzz_labxxxx
Checksum error 32615101403, inode 0 file (null)
restore: Tape is not a dump tape
What application could I use or even try? The man who made the backup can't be reached anymore and we had been left with the backup.
I'm running into some problems using pkill in my scripts. I'm using a program that needs to dump.
View 4 Replies View RelatedDoes the dump command back up entire file-systems or is it capable of backing up subsets of a file-system? And is tar capable of taking device names (for file systems) as input to be archived?
View 1 Replies View RelatedI want to generate core dump files from my program when it crashes. Its a pretty big process and has about 10-11 threads in it.I have followed the documentation to enable core dump by setting ulimit to unlimited etc. I quickly tried "A demo program creating a core dump" from the following webpage, which succeeds in Segfault and dumping a core file in the directory that I configured.However, I tried running my original program and caused it to crash. I did this by making calls to kill(), raise() or the same null pointer access as shown in the webpage above. In each case, my program crashed but did not generate a core dump file. Am I missing something?My program is in C++ and my environment is Redhat 9.0 (kernel 2.4.20)
Going through the "Why do I NOT get a core dump?" section on the same webpage as above, I can see two potential problems. One - there are issues with the suid/sgid (bullet # 6). I am not able to change any settings with suid because my system does not contain either /proc/sys/fs/suid_dumpable or /proc/sys/kernel/suid_dumpableTwo, my program has threads in it and the bullet # 8 is the problem.
what could this bei saved my correct iptables file @ Code: /etc/iptables.up.ruleswhere webmin is looking for it.webmin config is to automaticly boot this file and addes a line at.
Code:
/etc/network/interfaces
file
[code]...
I need to discover more detail error kernel panic in my linux box, some body know any way do dump in kernel Linux?
View 1 Replies View RelatedMy application team is asking me to generate the kernel-dump.
Here are details about my server.
OS: RHEL 4.7/32 Bit
Kernel Version: 2.6.9-89.0.23.ELsmp
Processor: Intel(R) Xeon(R) CPU E5520@ 2.27GHz
Hardware: HP Proliant 380G6 series server.
I am using Electric Cloud applications. Sometimes it creates some kernel panic and immediately got rebooted. Kernel-debuginfo rpm is not installed. In some thread, I read the kernel-debuginfo rpm's version should match with the kernel version. In my case I couldn't even find the exact version of kernel-debuginfo version.
I'm quite new to linux, but I've managed to grasp some basics. Now my intention here to create a virtual directory, which I resorted to creating an Image File so that I can mount it and have my folder have a dedicated storage. I will mount this image as a loop device. Well it's not much of a problem, but I would like to know whether this is suitable. Say I want to create a 25GB Image.
Code:
dd if=/dev/zero of=/home/disk-img/25GB.ext3 bs=1G count=25
Is this recommended? I'm using block size as 1G which is really huge, so I was wondering, if this is actually recommended. From what I read, some said that it's only advisable to use 4096k or lower, but what I found was that these suggestions are very dated (year 2003), and it is now 2010, so I would like to know if it makes any big differences.
I am using RHEL 4.7 (32bit) on HP Proliant 380G6 series server. We are using Electric Cloud Agents on these servers. Nowadays we are facing some memory issues and its creates some kernel panic and then restarts the server. When i reported the issue to my application team, they asked me to come with the core dump. I googled it enough, then i set ulimit value as unlimited. (previously it was 0, then i made a entry in /etc/profile file as follows
ulimit -c unlimited) But still whenever my server restarts due to that kernel panic, it couldnt generate the core dump. My application was installed on /opt
The attached document has the kernel panic logs
Using CentOS 5.5. I have a handful of users that I need to have connect to my server via sftp and start in the same directory. for example, user1, user2, user3, etc.. will connect via sftp and upon connection will all be in the /some/dir/path/ftp-root directory.I know one way is to create these users all with the same 'home' directory, since by default a user starts in their home directory when connecting via sftp, but before just doing that, I wanted to find out if that is really the appropriate method to use? alternatives? Is there some setting on the sftp server end that could direct all users to one starting directory so that these users don't have to have the same 'home' dir? I'm using the sshd daemon that comes with CentOS 5.5 (with all current updates/patches)
View 4 Replies View RelatedHaving done a short DoD wipe of hard drive (Dareks Boot & Nuke),I installed Windows XP on the first half of the drive and again zeroed the other half of drive for installing Debian.Please see attachment of screen shot for command lines input and output.After doing a grep for non zero characters on the second half of the drive(sda2)I was puzzled to find a grep command line search for non zero characters actually turned some up. I have no idea why they are there or what if anything they mean
View 4 Replies View RelatedIs there any command available inorder to read the server crash dump files?
View 4 Replies View RelatedOne of my clients needs a backup of his svn repository. I see that this is possible using svadmin dump command. I see where the location of the source repository is, but I don't see anything in documentation as to where the actual dump file is located. I need to know where the dump file is so I can scp or rsync the file to another server for backup.
View 5 Replies View RelatedAnyone know if the iptables-restore command operates atomically? I want to make sure that the entirety of in-kernel iptables are switched over to those specified in the input to iptables-restore, all at once, with no intervening emptiness, incompleteness, or mixing with prior table. The man page doesn't say about this.
View 3 Replies View Relatedwe have a customer that ran a sudo chmod +x -R * command on his / filesystem by mistake and now the machine cannot be accessed on the network Has anyone any idea what chmod command to run to restore the system to its original state ?
View 1 Replies View RelatedI do computer forensics here in Afghanistan and I am trying to keep a clean image of a dual bootable hard drive. Here is what I try to do...
1. Boot into UbuntuLiveCD
2. I run "sudo dd if=/dev/zero of=/dev/sda conv=sync,notrunc bs=64K" to wipe the drive with all zeros.
3. I then install Windows by creating a new partician about 50GB.
4. I then install Linux by creating a partician in ext4 mounting it at '/' in addition I create a swap partician.
5. Next configure everything just the way I want it. I install all the drivers and software I need for my windows partician and build out the remaining part of the disc as a "data drive."
6. Then I use "dd" again to try to image my "clean slate" of a system. Remember I am dual booting. I dd the /dev/sda and gzip it.
7. When I go to restore it, I boot from the live CD again and unzip ig and "dd" it back onto /dev/sda.
8. I run fdisk -l and I get:/dev/sda1 * 1 6375 5120000000 7 HPFS/NTFS/dev/sda2 6376 11724 42965842+ 83 Linux/dev/sda3 11725 12453 ...... 82 Linux swap / Solaris.This means to me that it can "understand the file system"
9. But then when I take out the Live boot CD and try to get my "clean slate" machine back, the system goes into Grub Rescue mode with a grub command line "grub rescue>"
10. I tried using the tutorial on Grub2, but...
a. It would not understand the command "linux"
b. When I try to do insmod, it says it doesn't recognize the file system.
I'm setting up a server with Jaunty Jackalope version. I'm trying to test setting up a basic iptables rules... No matter which command I put in, it is failing on the first command when I run iptables-restore < file location (the first rule always fails). I'm doing this on the root user and first typing in the iptables rules in a test file. I've tried the first command starting with % sudo, iptables and -A. All have the same result. I've also tried letting the HTTP rule be first with the same result.
[Code]...
are there any sweeper / cleaner apps which can take back my installation as to when it was new. I know home directory could have config files etc..but that doesnt matter, i want all packages and applications reset to how it was when i had a clean install. I am using Natty
View 2 Replies View Relatedhere is the thing. I've deployed BackupPC in a server at work, and everything is working fine. Now, what I need to do this thing. I have a remote server with a website and a postgres db running. I've been able to set up everything to be backed up using rsync. But I would like to make the same process to restore the backup immediately in a local server trough rsync, that has to be ready just in case of failure of the remote server. What I've tried to do is to run the DumpPostUserCmd, so whenever a dump is performed in the remote server, I can have my local server updated. What I've find out is that when you perform a restore through the web interface, the command that performs it is (using ps ax in the backuppc server):
/usr/bin/ssh -q -x -l root myserver.wheretodotherestore.mydomain /usr/bin/rsync --server --numeric-ids --perms --owner --group -D --links --hard-links --times --block-size=2048 --relative --ignore-times --recursive . /path/to/restore
So, what I could find out is: /usr/bin/rsync is the command that the process runs in the server where I want to put the restore. Then the options, and finally, source and destination of the restoration. But, as you can see, the source is '.' but I cannot guess where to point that!
I've been working with out servers for the past several years. We have a mix of OS. Fedora, Ubuntu, RHEL, etc. I've accidentally trashed the libc.so.6 link and now nothing but shell commands work.
I tried to echo "/lib/libc-2.10.2.so into the /etc/sestatus.conf file, but that didn't have any affect. Does anyone know where the search path is maintained besides PATH so I can try to point the apps to the /lib/libc-2.10.2.so file? I have a bunch of SVN repos on this server and right now, no one has access.
I am having a problem restoring a database. Server is Ubuntu 8.04.4 LTS. It appears to wig out at a certain point, when this happens the output on the screen shows;Quote:
0,4),('2011-02-21 03:30:32','Queue Wait Outside (100)','Queue Wait Outside (100)',11,0,4),('2011-02-21 03:30:49','Queue Wait Sleep (30)','Queue Wait Sleep (30)',5,0,4),('2011-02-21 03:30:53','Queue Wait Sleep (30)','Queue Wait Sleep (30)',13,0,4)
[code]...
how to take tar backup and restore in mysql,Iam new to mysql,i searched in google but i did not get the exact one.
View 4 Replies View RelatedI have a production server to which you need to be approached with caution. I have to update the 107 packets between other kernel php and several libraries.
This is the first production server on such a scale, and he did not want me to get something went wrong. And if you can make a restore point after installation so that if it does not start I will be able to restore support the package versions?
I installed openSUSE on one of the spare computers in work to test it out before I install it on my own laptop. I tried connecting Kmail to MS exchange and did so successfully using POP 3 but never checked the box saying leave messages on the server and it removed the emails from the server. Now I have nothing in my Outlook Inbox on my XP machine. how I can restore the emails from Kmail to the server.
View 6 Replies View RelatedI'm trying to restore all my files to my new ubuntu server (a netbook actually) and can't! after copying for a few moments the samba shares stop working. in synaptic it says samba 2:3.4.7 as im typing this i realized I was able to copy a few large files when i was starting out so there may be some software added since thats the issue. In my scenario i found that the server does recover after a minute or two. but when copying large files it causes ALL networking to stop. funny that if i go to the server screen it is fine. Ive experienced this connecting with a windows 7 pc and a mac. the server will start responding again in a couple of minutes. when it locks up it locks up ALL networking: cant connect to the webserver, or ssh either when this happens. Just starts working again after 2 minutes. I am unable to sync this with my backup drive unfortunately.
I did some experimenting today and found that I was able to stay connected via ssh for well over an hour with no problems. as soon as I start copying files is when it happens. I use this as a server in a home office so I need to resolve this. for now going to share a drive off my router i guess. have been poring through the logs and reading everything i can on this the last few days.. I switched from network manager to WICD after reading about it online somewhere. that actually did help to apparently reduce the frequency of the issue.
does anyone have any tips? this is a new machine/install of Ubuntu 10.4 LTS desktop version (its just easier than the server version for me even though using as a server). its a new MSI X Slim X320. I bought it as it has the X530 Atom processor which is super efficient and I run this as a solar powered web/file/torrent/app server. In that regard im very pleased with it.
i want to know whether there is any module to restore my linux machine before 5 days as like in windows
View 1 Replies View Relatedwritting shell scripts,i searched in google but could not get the exact,pls can anyone help how to take mysql backup & restore using shell scripts.
View 3 Replies View Related