I have a Dell PE2900 SCSI'd out to a Quantum SL3. I'm attempting to read/display the contents of the tapes or maybe dump the displayed contents or table of contents to a file.
I can see and access the library: ]# mtx status Storage Changer /dev/changer:1 Drives, 16 Slots ( 0 Import/Export ) Data Transfer Element 0:Full (Storage Element 1 Loaded):VolumeTag = NGL930L3 Storage Element 1:Empty Storage Element 2:Full :VolumeTag=NGL985L3
[Code]....
But when I try running tar -tzf /dev/changer or tar -tvf /dev/changer I just get a flashing cursor and the drive sits at IDLE.
I have a Red Hat Enterprise (AS) 4.8 system and I need to know how to totally rebuild the system from dump tape. I have been making some full level 0 dumps of the system to the attached DAT72 tape drive... In the case the boot disk goes south, I need to reload from tape, onto a new disk drive. I know how to do this in Solaris. I assume you boot from CD to like a mini-root, then configure and mount the drive on temp mount points, restore the sys data, then load the "boot blocks" (like installboot on solaris).
on a Linux pc I want to restore the contents of the DLT tape to /mnt directory,how to do itIf I just want to see the contents of the files, is it possible to read the contents irectly from tape without restoring to hard disk?
I am trying to use "hexdump" on the files in the /dev/ directory. This is ubuntu 9.10.There is some more in this forum: [URL]Hexdump returned small blocks of bytes or couldn't see the files even as root. I expected to see the actual data streams from the various devices. How come this doesn't work?
how easy it would be to read the contents of a physical disk that was part of a larger logical volume. The disk contains a "Linux LVM" partition that spans its entire size. My problem is that one of my disks died, and I have to send it back for a warranty replacement. However, the disk is dead, and I can't zero it out. I'm just trying to assess how difficult it would be (or at least how likely it would be) for a tech that's checking out the disk to get at the data.
I have been using an LTO-5 Ultrium-3000 tape drive connected to an ATTO HBA without problem. I can control the tape drive using "mt -f /dev/nst0" and have been able to make successful backups using cpio, tar, and dump/restore. I followed some instructions on the web about how to install the HPE Library and Tape Tools application (version 4.21) which relies on conversion of a rpm to a deb file. The software seems to have been installed correctly and runs. However the hardware scan function does not recognize my tape drive. The following is suggested in the user manual if the tape device is not recognized by the software under Linux:
1. Login as root. 2. Edit the following file: vi /etc/modules.conf 3. Add the following line as appropriate: add options scsi_mod max_scsi_luns=128 4. Reboot the computer.
The problem is I don't have an /etc/modules.conf and am not sure exactly which file would be equivalent? If this is even the correct solution.
My tape drive is controllable and functions well using "mt -f /dev/nst0 status" so it seems to be a matter of LT&T software to detect the tape drive.
At the risk of providing too much info here some, possibly relevant, output from lshw
*-pci:3 description: PCI bridge product: 6 Series/C200 Series Chipset Family PCI Express Root Port 1 vendor: Intel Corporation physical id: 1c bus info: pci@0000:00:1c.0 version: b5
I am looking at getting a DLT drive for my network; however, I have never used the tar command with a tape drive. What happens if the data is larger then 1 tape? Does the tar application automatically span tapes or do I need to use switches so it spans multipule tapes? Right now my Full backup will take 2 or 3 tapes.
I have dell poweredge 830 server with tape drive and RHEL 4 running on it....the issue i am facing is,i am unable to insert the tape as i had ejected the tape forcefully from it....
i tried to do a listing of the contents backedup on tape and it got struck in middle throwing below error,
/dev/st0:device input/output error. after which i was unable to eject the tape using mt -f /dev/st0 rewoffl
i removed the tape by holding the eject button and now when i try to insert another tape, it's unable to take the tape in to tape drive...
I've tried doing all except reeboting the server, can any one help me out in this issue, hope the blow information may help in debugging the issue... code....
I am trying to read the contents of a file into something else. I have a file.txt that I am working with, I want to read the file and take the data and run some commands with the data that it read. So if it read www.yahoo.com I want to be able to nslookup. Does that make sense? I have been trying to use the read command but that does not seem to work. I even was trying to read filename | > filename to see if I could even read any of the data at all. Nothing is working.
How can read the file /var/etc/allInOne.cfg and distribute its contents on multiple cfg files using C language. /var/etc/allInOne.cfg contain the data and the path of each text file.
The source file "/var/etc/allInOne.cfg "look like this: line1 line2 ... line10 filePath:/var/etc/file1.cfg line12 line13 ... line14 filePath:/var/etc/file2.cfg linen .. filePath:/var/etc/filen.cfg the result will be :
The expected result is: /var/etc/file1.cfg will contain line1 to line10 /var/etc/file2.cfg will contain line 12 to line13 /var/etc/filen.cfg will contain linen to linen-1
So I finally managed to get my Samba file server working () but now I have a question. On the server I have 4 folders, each being used for a certain topic. How would I set the folders as read only but not the contents, so that way remote users can read/write to the folder but can't delete or rename the folder itself? Also, if I restart the Linux box it will cease to show on the Windows Network unless I delete the Samba share and remake.
I have used Dump Command to dump the application files. For Full backup the level 0 is working fine. For incremental backup I used the level 1 or 2 it is getting the error as
DUMP: Only level 0 dumps are allowed on a subdirectory DUMP: The ENTIRE dump is aborted.
The code I used =============================== #!/bin/bash #Full Day Backup Script #application folders backup #test is the username now=$(date +"%d-%m-%Y") [Code]...
I want to generate core dump files from my program when it crashes. Its a pretty big process and has about 10-11 threads in it.I have followed the documentation to enable core dump by setting ulimit to unlimited etc. I quickly tried "A demo program creating a core dump" from the following webpage, which succeeds in Segfault and dumping a core file in the directory that I configured.However, I tried running my original program and caused it to crash. I did this by making calls to kill(), raise() or the same null pointer access as shown in the webpage above. In each case, my program crashed but did not generate a core dump file. Am I missing something?My program is in C++ and my environment is Redhat 9.0 (kernel 2.4.20)
Going through the "Why do I NOT get a core dump?" section on the same webpage as above, I can see two potential problems. One - there are issues with the suid/sgid (bullet # 6). I am not able to change any settings with suid because my system does not contain either /proc/sys/fs/suid_dumpable or /proc/sys/kernel/suid_dumpableTwo, my program has threads in it and the bullet # 8 is the problem.
Ive managed to install samba, I've shared a folder. I can access from a Windows 7 machine via \ubuntupublic. I can put files in the folder form the ubuntu machine and edit them on the windows box. I can put files in the folder/share from the Windows box but then I cannot edit them on the Ubuntu machine (they are read only and have a "Lock" over them). I can fix this by going to the properties of the file/folder in Windows and manually assigning "Everybody" full control (then the lock disappears and all is well.) I want read/write access to all the folders contents from both machines all the time (security is NOT a concern I WANT the permissions wide open) what am I doing wrong?
I'm trying to get my new Wireless USB dongle to work with Xubuntu 9.10 Karmic Koala. I've extracted the .rar file that was on the drivers CD that came with the dongle. According to the ReadMe....
The Model Name is: RT2870 Wireless Lan Linux Driver Driver Name: rt2870.o/rt2870.ko Hardware: Ralink 802.1n Wireless LAN Card Description: This is a linux device driver for Ralink RT2870 USB ABGN WLAN Card. My issues:
1. I do not see the driver file anywhere in the folder contents... I suppose that I have to "build" it. 2. In the makefile the chipset is default set to "CHIPSET = 3070" & the folder that holds all the extracted files is named 2008_1128_RT3070_Linux_STA_v2.0.1.0 so I'm confused as to whether this dongle has a 2870 or 3070 chipset inside. Does it matter? 3. Sorry, I'm a bit of a n00bster when it comes to all this, but I'm having some problems with the build instructions. It reads:
Recently I tried to get BGP table dumps from public route servers. I telnetted into one of those public route servers and ran "show ip bgp" command. My question is: how to save the command output to my local machine? I cannot run "show ip bgp > tmp.txt" on the remote route server.
You would think it would be easy to do this (and essential to maximizing and predicting the usage of tapes), but apparently this isn't so. The program MT(1) actually had some commands that give the block positition (if supported by the drive) from which the remaining space could be deduced (even if you couldn't predict exactly how much space the next archive would take up). However I'm using MT(1L) which no longer has such commands.
In my program, I fork() to get a child process. Because of some problem, child process terminates by a segmentation fault. Parent process is still running. I have compiled my code with -g option. I have done: ulimit -c unlimited. I am not getting core dump of the child process. How can I get the core dump of child process?
I would like to have dump backup just my home directory but am having problems the command I am using wants to back every thing and takes hours upon hours it has been running for about 10 hr and only 21% is done. This is the command dump -0u -f dp_hd /media/CENTON USB/ /how can I get this to back up only my home directory
I need help writing a script that will copy everything from tape to system directory. I have a Linux box with 3 TB of Hardware space. I am using the following commands
1) mt /dev/st0 rewind
2) tar xvf /dev/st0
3) tar xvf /dev/st0 fsf 1 (Using this to move to the next segment of the tape) and then
I have set of DVD disks. I'd like to rip them to .mkv files, but with all information - i.e. all subtitles, and all audio tracks. Is there any tool on Linux that I could use to do it? I found some Gentoo howto about ripping, but it requires writing shell scripts, and I'd rather use something with clickable interface.