Need help on how the scsi and multipathing works in Linux. From the docs i have read, i understand that by the use of multipathing we can assign multiple paths to a SAN partition. If there is a problem then one path will failover to other path.However, i am not clear on how linux recognizes the SAN partitions using the multipath drivers. For e.g. I have a HP Proliant server on which we have the following mounts:
Filesystem Size Used Avail Use% Mounted on
/dev/cciss/c0d0p3 59G 11G 46G 20% /
/dev/cciss/c0d0p1 494M 27M 443M 6% /boot
I'm doing a few tests with fedora 15. I'm surprised because they changed the naming way on network devices. eth0 is now called em1.At every fedora I have found the configuration files in /etc/sysconfig/network-scripts, but today I can't find them.
i m facing same error in most of the HCL servers. the problem is that it throws error while booting and sometimes not throws error. the error is :-
Feb 13 13:17:25 fe13s kernel: Adapter 0: Bus A: The SCSI controller was reset due to SCSI BUS noise or an invalid signal. Check cables, termination, termpower, LVDS operation, etc.
Feb 13 13:17:30 fe13s kernel: Adapter 0: Bus B: The SCSI controller successfully recovered from a SCSI BUS issue. The issue may still be present on the BUS. Check cables, termination, termpower, LVDS operation, etc
Feb 13 13:29:15 fe13s kernel: Adapter 0: Bus B: The SCSI controller successfully recovered from a SCSI BUS issue. The issue may still be present on the BUS. Check cables, termination, termpower, LVDS operation, etc code....
i was fooling around with ATI graphics settings and after i went to reboot my Desktop it kept comming up with odd symbols instead of a log in screen, then i started to try and do things with a LiveCD and screwed things up even more so now any time i attempt to boot, even into recovery, it stops at the line
[2.573600] sd 6:0:0:3: [sde] Attached SCSI removable disk
i have tried to do multiple fixes with no avail and i am not very savvy with linux as i have only been using it under a year, so it would be awesome if anyone could help this is 10.10 Maverick and the 32bit version,
I am testing PCLinux 2009.1 prior to installing. My desktop currently runs on ME and has been out of use for some time. It still works however and I can still get online. I'm currently using a laptop running XP. I would like to make use of my old desktop and tried to test PCLinux 2009.1 today. I followed the instructions obtained online and booted from the Linux cd. The pc booted up ok and displayed the first screen allowing me to select the "Live CD" option. After enter my system hung up at "Probing SCSI Devices:Segmentation fault".
Suppose during a script execution I am attaching one or two new disk having different vendor id to host machine, how do I know which disk corresponds to which file in /dev/ directory? i just want to perform some operation on those device from some script, how will i know which file in /dev directory correspond to which disk(having same size but different vendor id).
I have a Windows 2003 server with fiber attached volumes (NTFS) that I would like to mount readonly on a linux system to back it up to tape. The fiber device will allow me to present the volume R/W to one host and R/O to another, however, the R/O system doesn't see any of the changes made by the R/W server. In other words, how can I make a readonly volume refresh, scan for changes, or update without un/re-mounting it?
Is the "mount -o --bind" option what I want? From the MAN is doesn't seem right... the option "sync" seems slightly more promising but I think I'm just grasping at straws here. The best I have come up with is a cron job to unmount then mount the volume periodically.
In my understanding, the way /proc/scsi/scsi gets populated, /proc/paritions also gets populated in the same fashion. i.e. the description for first entry of /proc/scsi/scsi can be seen in the first entry of /proc/partitions and same for rest.
So, With this assumption, in my project, I used to relate first entry of /proc/scsi/scsi with first entry of /proc/partitions to get its total size and same for all entries.
But, I observed some differences in following scenario, where
1) The first 4 entries in /proc/scsi/scsi are SAN luns attached to my system and for which the actual device names in /dev/ are sda,sdb,sdc and sdd.
2) The last 4 entries are the internal HDDs on same system. In /dev/, their respective device names are sde,sdf,sdg & sdh.
(Output attached at end of the thread)
But in /proc/partitions, the device order is different.
You can see their respective sizes in /proc/partition output as well.
So, my question is, in this particular scenario, I can't relate the first entry of /proc/scsi/scsi with first entry of /proc/partition. i.e. scsi0:00:00:00 is not /dev/sde, because it is actually /dev/sda.
It seems that my assumption is wrong in this scenario.
Is there any way or mechanism to figure out actual device name for an entry in /proc/scsi/scsi in /dev/ directory?
How can my application should relate /proc/scsi/scsi entries with their respective device names and sizes?
i have a centos 5.5 server running tomcat,oracle xe,apache etc..i would like to be able my windows machine to access this server by name, and not ip, for example have http://backserv:7080 for oracle xe web admin and so on..i know one way is DNS, but if i install oracle on a windows machine in peer-to-peer network, i can allready access with by it's name. o i'm guessing i should do something with samba, wins or something like that.never mind i got now for the REAL complex stuff of insalling web sphere 7 without user interface (need to look for a response file).
I am using a QLA2462 and have got FC storage attached through a brocade switch. As soon when I add more storage for this server on this switch my device naming shifts and this new storage comes in between e.g. as sdb, and the other 2 go to sdc and sdd. I looked around but I can't find a good solution for SuSe 11.2.
I have a problem defining persistent device naming on a Debian Lenny server.I have:RAID1 controller on the server machine with two SCSI disks.external storage with RAID5. I have / mount on the first partition on the server SCSI disk and /storage mount on the external storage. I'm experiencing a problem: The system recognizes the system disk (RAID 1), as sda or sdb randomly.I want: To control the recognition, and tell the system that sda (sda1) will always be the system disk.The motivation: GRUB is configured to work with sda, and when the system disk doesn't, boot process fails, and I end up in the initramfs shell-like interface.
Booting the kernel . . . mount:mounting /dev on /root/dev failed: No such file or directory mount:mounting /sys on /root/sys failed: No such file or directory
I'm writing a script for asterisk to monitor trunk failure, i do a loop for every trunk it got nad would like to name variable like server1=, server2= naming the server upgoing as the trunk is. here is the scripts:
what i would like to do is name the variable server, username and status with the count variable, like this server$COUNT to have server1 when on trunk one, bu as soon as i add the $COUNT after the server, it seems to try to make it a command, it says that:
Code: ./test.sh: line 45: server1=18.104.22.168: command not found
I need to mount a drive over fiber channel. It's a 4 terabyte drive. We're using a paragon driver to access other volumes on the SAN in a similar way. Works fine. I mount the other volumes as mount -t ufsd /dev/sdd1 /mnt/VolumeName Works fine. This is also an ntfs drive. The error I get when I mount: mount -t ufsd /dev/sdg1 /mnt/VolumeName01 is NTFS-fs error (device sdg1): ntfs_fill_super(): Device has unsupported hardsect_size
I have installed 2 fc cards .I have attached FT-SAn to these cards now when i see miltipath -ll output , it shows some path for the multipath can you bfief from where these paths are taking the value . significance of each digit
Setting up a new Centos 5.5 server with an LSI Logic 7402XP fiber channel HBA card connected by fiber to an enclosure with 28 SATA drives.This is just a bunch of disks.I want to mount them /disk1, /disk2,disk28. No RAID or LVM. After connecting the dsisks and rebooting there are 28 new devices /dev/sdc, /dev/sdd, ...
/dev/sdad. Going to /sys/class/scsi_disk will list all of them. Pick any one and go into device and you can cat the vendor and model and state is "Running." But try parted /dev/sdc and it just quitts. fdisk and mkfs produce errors. Attepting to read or write the device directly with dd fails. The mpt fusion kernel mods are present. I'm pretty sure this hardware is on the suuported list. And I've tried swapping out the card and cables in case of bad hardware. What am I missing to be able to partition and mkfs these disks?
getinfo is at http://pastebin.centos.org/36873 [root@localhost dev]# cd /dev ; ls sd* sda sda3 sda6 sdac sdb1 sdb4 sdc sdf sdi sdl sdo sdr sdu sdx sda1 sda4 sdaa sdad sdb2 sdb5 sdd sdg sdj sdm sdp sds sdv sdy
I created a new disk on our scsi san. I then ran the following command: echo "- - -" > /sys/class/scsi_host/host1/scan. did that command for each host. and in dmesg, it shows it found a device /dev/sdg. but when i do a fdisk -l. It never lists /dev/sdg. I just did this other day on another server and it worked fine like that. This is RH 4.8.
I currently have a massive problem with my main webserver: one of the SCSI drives where I keep my htdocs seems to have failed (sd 0:0:0:0: [sda] Result: hostbyte=DID_NO_CONNECT driverbyte=DRIVER_OK,SUGGEST_OK end_request: I/O error, dev sda, sector 7797601).
Since I have no backup (dumb!dumb!dumb!) of it, I tried to copy what I could on the fly, but that failed miserably. The next idea was to umount the volume and try remounting it R-O, but I have a bunch of apache2 processes that are apparently reading/writing on it and I can't seem to find any way of terminating them. I tried via init script, killall and kill -9 but they just don't respond. Until I can terminate those processes, I won't be able to unmount the drive and try saving what's left of the data.Oh and the cherry on top of the cake: the server is 1100 miles away so I can't even access it physically..Is there any way I didn't think of to get rid of those apache processes?
I am setting up a new server and am in the midst of testing RAID. This is an Ubuntu 9.10 server. RAID1 (/dev/md1) is spread across 12 one-terabyte SCSI disks (/dev/sdi through /dev/sdt). It has four spares configured, each of which are also one-terabyte SCSI drives (/dev/sdu through /dev/sdx). I have been following the instructions on the Linux RAID Wiki ([URL]....
I have already tested the RAID successfully by using mdadm to set a drive faulty. Automatic failover to spare and reconstruction worked like a champ. I am now testing "Force fail by hardware". Specifically, I am following the advice, "Take the system down, unplug the disk, and boot it up again." Well, I did that, and the RAID fails to start. It outright refuses to start. It doesn't seem to notice that a drive is missing. Notably, all the drive letters shift up to fill in the space left by removing a drive. The test I did was to:
Is removing a disk from the bus a reasonable test in the first place? Meaning, is this likely to happen in a production environment by other means than a human coming by and yanking out the drive? Meaning, is there a hardware failure that would replicate this event? Because, if so, then I don't know how to recover from it.
I am running fedora 8 on a Dell 2950 with a qlogic fibre channel card.is attached to a Dell/clarion AX100.I had a drive assigned to this server which had been working for approximately a year now.Recently during a power failure things turned ugly. Rebooting didn't fix things.I unassigned/reassigned the drive, rebooted, still no good.Funny thing when I unassigned the drive and reboot the system.still see /dev/sdb but no partitions and fdisk -l does not display /dev/sdb.I even did "echo 1 > /sys/block/sdb/device/delete" and the device no longer showed in /dev, but after a reboot it is back.I pulled and reseated the fibre card, same
I am trying to install Ubuntu ver.10.10 server on HP DL360 with 2 internal 36GB untra3 scsi drives. The installations seems to have gone fine but upon booting, the OS is stoping at the BusyBox menu with initramfs prompt. None of my boot or root filesystems is mounted here so I can not go further. I've searched all over the net for answers but nothing seems to work.
I'm no expert on DHCP. My problem is that i have a CentOS 5.5 server on which i want to install a DHCP server. I have two NICs where eth0 has access to the internet and where eth1 should act as an DHCP server.
I have installed dhcpd and this is how my dhcpd.conf file looks like.
ifconfig looks like this
When i start dhcpd on eth1 i get no error messages but when i connect any devices to eth1 they don't get any IP. I cant find anything in any logs about devices trying to get an IP address. I dont have any firewall rules in iptables.
Since May 12,2009. Our system lifekeeper has the error log "lifekeeper error: DEVICE FAILURE on SCSI device '/dev/add'", but it ran normally. Until last week, it failover to the standby server. The disk still running, the error still come out.