I am looking at getting a DLT drive for my network; however, I have never used the tar command with a tape drive. What happens if the data is larger then 1 tape? Does the tar application automatically span tapes or do I need to use switches so it spans multipule tapes? Right now my Full backup will take 2 or 3 tapes.
I have accidentally ended up in deleting my root directory while I blindly fired command while watching movie.
I fired following command #rm -rf ~/<SPACE>*.out instead of this command #rm -rf ~/*.out
Things already done: 1) Created /root directory relogged to get some of basic settings of gnome and Desktop. 2) Things went well now when I login my desktop ,gnome environment and other things looks to be working well only prompt on my terminal has changed. I can fix it any ways.
Things I want to ask: 1) I haven't studied much about contents of /root directory to best of my knowledge is it like other user's home directory with some basic configuration files for mostly required applications. SO my question is have I lost any thing important system file or something? 2) If I have lost any important configuration or system data how can I recover it without reinstalling whole system? (My opinion about this is, It is quite possible but to do so, as far as I know capabilities of linux. But I still want comments from experts before I try any things on it because I don't want to backup my whole HDD and reinstall the whole stuff again for me and also my sister's stuff in MS.)
I have a Red Hat Enterprise (AS) 4.8 system and I need to know how to totally rebuild the system from dump tape. I have been making some full level 0 dumps of the system to the attached DAT72 tape drive... In the case the boot disk goes south, I need to reload from tape, onto a new disk drive. I know how to do this in Solaris. I assume you boot from CD to like a mini-root, then configure and mount the drive on temp mount points, restore the sys data, then load the "boot blocks" (like installboot on solaris).
At some point in time my VMs exceeded the 5.8GB limit and overflowed into swap (all dev VMs). I have since shut down the VMs and restarted Firefox (memory hog). Right now I have 0.73GB in RAM and 0.29GB in swap. I want to move all of the data in swap into RAM where it will be faster. So I will end up with 1.02GB in RAM and 0GB in swap. Is there a quick command to move data contained within swap to RAM? I know of the swapon/swapoff command. I was wondering if there was another method which clears swap.
Does anyone know of any existing software to aid in this task? We already offer backed-up data storage to our users, but no means of say arching original data sets for long term storage(NIH, etc rules). Something that could be user initiated, where they login into the archive server, tell it I want such and such files archives, and on its own it takes those files and gives them to a tape library(drive+robot).
Currently I have bacula backing up the contents of my server on a daily basis (Mon-Fri) on to a tape. There are in total 9 backup tapes so far. One for each day of the week (Mon-Thurs) and one for every Friday in the month (so there are 5 friday tapes). The tapes are all from the same pool and once the tapes are full they should then get recycled. As I understand and from what I have seen, the data from my server backups up to the tape and is then marked with the append status. Every time that tape is used, data is then appended onto the tape until the tape is full. At this point it is marked as full but wont be recycled until all the tapes from the pool are marked as full (so until all 9 tapes are full).
My question is, is there a way of overwriting the data on the next backup on the tape, rather than appending or requiring that all tapes in a pool are full before a tape can be recycled. As the mon-thurs tapes currently cant be recycled until all the Friday tapes have been used. The reason why I dont have seperate pools for the Friday backups and the other days of the week or even seperate pools for each day of the week, is that if a tape is accidently not changed or someone puts in the incorrect tape by accident a backup will still occur. Which is better than no backup!
I'm trying to use wget to retrieve some data from our tape backup utility (HP Command View 1/8 G2 Autoloader). The URL requires two parameters for the info I want to retrieve. I have searched for a few hours and have tried numerous combinations to get the data but the parameters aren't being executed. I have escaped the URL as well.
I am running Ubuntu 10.04, and i recently purchased an tandberg LTO-4 SAS tape drive. I want to access it and backup data on it. Do I simply just connect plug it into the server,and I should be able to backup/transfer data to the tape drive? Or are there intermediate steps before I can do that. Here are some results from commands that I have typed:
I have been using an LTO-5 Ultrium-3000 tape drive connected to an ATTO HBA without problem. I can control the tape drive using "mt -f /dev/nst0" and have been able to make successful backups using cpio, tar, and dump/restore. I followed some instructions on the web about how to install the HPE Library and Tape Tools application (version 4.21) which relies on conversion of a rpm to a deb file. The software seems to have been installed correctly and runs. However the hardware scan function does not recognize my tape drive. The following is suggested in the user manual if the tape device is not recognized by the software under Linux:
1. Login as root. 2. Edit the following file: vi /etc/modules.conf 3. Add the following line as appropriate: add options scsi_mod max_scsi_luns=128 4. Reboot the computer.
The problem is I don't have an /etc/modules.conf and am not sure exactly which file would be equivalent? If this is even the correct solution.
My tape drive is controllable and functions well using "mt -f /dev/nst0 status" so it seems to be a matter of LT&T software to detect the tape drive.
At the risk of providing too much info here some, possibly relevant, output from lshw
*-pci:3 description: PCI bridge product: 6 Series/C200 Series Chipset Family PCI Express Root Port 1 vendor: Intel Corporation physical id: 1c bus info: pci@0000:00:1c.0 version: b5
I have a situation where I am trying to move some data from a Linux computer to a Windows computer. In all there is 700GB of data to move in about 1.5 million files, so I don't want to do this over the network.My first thought was to use an external USB hard drive and create an NTFS partition and copy the files from the linux computer to mount on the Windows computer. After 4 days of copying without completion I abandoned that idea. I thought the NTFS might be slowing it down, so I created an EXT3 partition. 4 Days later it was still copying. I did some calculations and there was no way the USB 2.0 connection was that slow. I then used ddrescue and cloned the drive to be copied overnight and it took about 12 hours. i was able to mount the USB drive under Linux and access the files appropriately. The only problem is that I can not access that USB drive on my Windows 7 computer. I have tried Explore2fs, DiskInternals Linux Reader, and Ext2 Installable File System For Windows, but none of them is recognizing the external drive.
My problem is that I need to move a moderately complex system from an old system to a new system. The old system is a core 2 duo running on an asus p5k-se (p35 chipset) M/B, Nvidia 8500 gt, 3 x sata II hard disks, 1 x sata dvd, 1 x ide hd, 4GB ram. It runs opensuse 11.1 kde 4 as a desktop system + samba server, apache server, database server + other non-opensuse software. In addition some of the opensuse software is not the default 11.1 versions as later versions were required. The nvidia driver is from the nvidia repo. There are several file systems, some under LVM.
The new system will be a core5 760, asus p7p55d-e M/b (p55 chipset), nvidia 240, 3 x sata II HD, 1 x sata dvd, 4GB ram and possibly 1 x ide HD. This M/B also includes USB 3 & sata III. I have no USB 3 devices but this may eventually change. I have no plans for sata III and believe that it may be better to attach any sata III SSD to the sata II bus.
What I would like to do is to move the hard disks from old system to the new system. What I would like to know is, if the system is left at 11.1, is the system likely to work with the new hardware. If the answer is no, if the system was upgraded to 11.3, would the transfer of the hard disks then work? In order to upgrade from 11.1 to 11.3 using the DVD i believe I should get rid of all non-default repos. I assume that I should also get rid of anything that was installed from them, e.g. the nvidia driver and any software versions installed manually outside RPM. Also is there any default 11.1 software that should be removed either before or after the upgrade to 11.3
My co-worker install CentOS with /var (ext3) is 1TB:
I do the following tasks to move sdb1 to VolGroup00:
After that I moved /var back to / folder and reboot but the system hangs at "Starting system logger". Uninstall and re-install sysklogd package don't make any change. What I don't understand is I can start syslogd from command line:
I'm trying to download all the data under this directory, using wget: [URL] I would like to achieve this using wget, and from what I've read it should be possible using the --recursive flag. Unfortunately, I've had no luck so far. The only files that get downloaded are robots.txt and index.html (which doesn't actually exist on the server), but wget does not follow any of the links on the directory list. The code I've been using is: Code: wget -r *ttp://gd2.mlb.***/components/game/mlb/year_2010/
I was trying to set up a partition on my netbook's hard drive, and foolishly forgot to backup my home folder. Now, Ubuntu (10.04, btw) won't boot. It allows me to get into the manual recovery shell, though. Now, I'm perfectly willing to reformat, but first I'm hoping I can recover the files from my home folder without having to take my netbook to the ridiculously overpriced computer repair centers in my area.
I have an external hard drive, but when it's plugged in during the recovery shell, it won't register as being a valid directory. I know the commands to copy or move a directory, but without my external reading as a valid directory to move my home folder to, I'm kinda stuck. Is there any other way to recover the contents of my home folder without having to go to a computer repair place? Neither of the ones in my area know much about linux at all, and I really don't want to have to pay $80+ to have someone else recover my home folder.
I have some data files that should be distributed with my program. Using dist_pkgdata_DATA in Makefile.am, I get these files installed to /usr/local/data/share/package-name. The problem is that data is read-only, and my program needs to modify it. Playing with dist_sharedstate_DATA, dist_localstate_DATA, dist-data_DATA varibles, I got different installation directories, like /usr/local/com, usr/local/var, but data is always read-only.
How can I distribute modifiable data files with my package? I need some common directory for all users, or maybe local data in a user directory.
I have a drive with ext3 that was my old /home drive.When I moved to 9.04 I went with ext4 and just linked to my old home drive.My old ext3 drive is becoming unstable so I want to back it up.I have a 500gb ata drive (unfortunately not a sata drive) thatto wipe and put 10.04 on. I want to move my data from this backup driveIm going to back up to, to the new 10.04 until I get another hd as abackup drive.My question is, I know there is some issue between ext3 and 4 and I dontwant to get into a bind. What would be the best format for moving
I would like to move a samba fileserver from ClearOS5 to Ubuntu. It has raid1 (software) and I would like to save one partition from it, md3. The other partitions could be used for Ubuntu. How would I go about this so that I don't erase data from md3 and is it even possible?
What is the best and simplest way to compare two directory structures without actually comparing the data in files. This works fine: diff -qr dir1 dir2 But it's really slow because it's comparing files too. Is there a switch for diff or another simple cli tool to do this?
I am running an openldap server on fedora core 10 and now running into a need of get all users data from Active Directory. Actually I have a php based application which will be using that data from OpenLDAP and it will need to be updated on weekly bases. how can I do it and any script.
I am using a touch command to setup a specific range of names, identifiers, etc. Now I am trying to move the files from a certain range to another directory I created. Sorry to post just can't remember how to do this but I know it is fairly simple for you guys. I have the output set as fred*_[!g].* and I am trying to get these files to my created directory.
I am trying to move my home directory from my install partition to a new partition. I cloned my installation from a previous ~78 gb HD using g4l to a new 250 GB drive. Now that I am using the new drive i created a new partition to used for files called "files". New partition is sda3 and the boot partition is sd1. I am trying to follow this guide [URL] but I am having no success.
The output of: Code: find . -depth -print0 | cpio --null --sparse -pvd /media/sda3 is
Code: pio: /dev/sda3//./.jungledisk/cache/jd2-a114b643324c576f1c36e3f17a9043f4-us/Files/cf-1381.tmp: Cannot open: Not a directory cpio: `/dev/sda3' exists but is not a directory cpio: /dev/sda3//./.jungledisk/cache/jd2-a114b643324c576f1c36e3f17a9043f4-us/Files/cf-1336.tmp: Cannot open: Not a directory cpio: `/dev/sda3' exists but is not a directory cpio: /dev/sda3//./.jungledisk/cache/jd2-a114b643324c576f1c36e3f17a9043f4-us/Files/cf-1387.tmp: Cannot open: Not a directory cpio: `/dev/sda3' exists but is not a directory .....