CentOS 5 Server :: Way To Allow Two Servers To Use Same File System On SAN Using ISCSI?
Mar 9, 2010What is the correct way to allow two servers to use the same file system on a SAN using iSCSI?Is it GFS on the two servers?
View 3 RepliesWhat is the correct way to allow two servers to use the same file system on a SAN using iSCSI?Is it GFS on the two servers?
View 3 RepliesI built a CentOS 5 Linux cluster with GFS storage using local RAID volume and share it with gnbd_export/import on two web-servers. Now I need to expand that storage to another servers local volume. I saw the picture in the manual, but I don`t know, how can I create that scheme.
I can use gnbd_export on the second server and gnbd_import on the first. In that case I will have two volumes on the first storage and I can expand volume group, logical volume, etc on it.
I have a fairly simple iSCSI setup using two devices, but they have swapped names on different machines. running CentOs 5.3 ia64, and using iscsi-initiator-utils-6.2.0.868-0.18.el5
vm1:
[root@vm1 ~]# fdisk -l
Disk /dev/xvda: 4194 MB, 4194304000 bytes
255 heads, 63 sectors/track, 509 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
[Code].....
Any way to get iSCSI to mount the devices as consistent device names ?
We have a Centos 5.6 server mounting two iSCSI volumes from an HP P2000 storage array. Multipathd is also running, and this has been working well for the few months we have been using it. The two volumes presented to the server were created in LVM, and worked without problem.We had a requirement to reboot the server, and now the iSCSI volumes will no longer mount. From what I can tell, the iSCSI connection is working ok, as I can see the correct sessions, and if I run 'fdisk -l' I can see the iSCSI block devices, but the O/S isn't seeing the filesystems. Any LVM command does not show the volumes at all, and 'vgchange -a y' only lists the local boot volume, not the iSCSI volumes. My concern is that, the output of 'fdisk -l' says 'Disk /dev/xxx doesn't contain a valid partition table' for all the iSCSI devices. Research shows that performing the vgchange -a y command should automatically mount any VG's that aren't showing, but it doesn't work.
There's a lot of data on these iSCSI volumes, and I'm no LVM expert. I've read that some have had problems where LVM starts before iSCSI and things get a bit messed up, but I don't know if this is the case here (I can't tell), but if there's a way of switching this round that might help, I'm prepared to give it a go.There was absolutely no indication there were any problems with these volumes, so corruption is highly unlikely.
I'm planning to add 1tb sata disk to my lovely file-server under ubuntu 10.10,what i want is use this disk as additional storage for network user,indows and ubuntu?I mean when my ubuntu server down (worse case) I can easily take out the disk from ubuntu machine and plug in on windows machine
View 2 Replies View RelatediSCSI Server, can I change the default port 3260 to other port ?
View 1 Replies View RelatedI have a weird performance issue with a centos 5 running a nfs server and a rh8 client. I think the fact that it is rh8 client should be downplayed. It is just that with rh8 client the performance degradation seems more clear. See test details below OS in server is Centos 5 x86_64 kernel 2.6.18-92.1.22.el5
1Gb connection between machines File to test over NFS is a 1GB file. First of all I wanted to measure how the network alone performs while using NFS. So in the server side I run a "cat" command on the 1GB file to /dev/null. Please note that the disk read speed is about 98MBs. At this point the file system has the 1GB file cached in memory. In the client side a "cat" on the same file gives me a speed of about 113MBs. It seems then that the bottleneck in this instance is the network and it is very close to nominal speed. So the network performance is really good. (BTW I know that the server got that file from cache because a vmstat or iostat shows no disk activity.)
The second test is reading from disk with no caching involve. In the server I flushed the 1GB file from the memory. For instance by reading another 5GB file and I repeat the same thing as above in the client (a cat on the 1GB file). Now, the server has to go to disk.(vmstat or iostat shows the disk activity). However the performance, now, is about 20MBs, I was expecting something closer so 90MBs. (since the reading speed in the server in the first test showed 98MBs).
This second test was repeated for ext2, ext3, xfs with no significant differences. A similar test using a RH8 NFS server and client gets me close to 60MBs for a 1GB file not cache by the file system in the serverSince network speeds and disk read speeds are not the bottlenecks ... what or where is the limiting factor then?
I need to enable GFS for a shared storage file system, I need to share it to different ip's so that every one get syn data on that file system
ex : /data on 192.168.0.10
I need /data to be in GFS and need to share to 192.168.0.20 & 192.168.0.21
so that this two machines can have sync file system
/data is in shared storage
I have configured a "Syslog" server on /var directory as a separate ext3 partition - to receive the logs and events from the clients & the firewall as well. The directory needs to grow dynamically as the logs are populated. Is there a way i can make the filesystem grow dynamically as and when the directory is full.
View 6 Replies View RelatedThe first server I installed installed fine. The second server, installed with the same config, went to "kernel panic not syncing no init found try passing init= option in kernel" error. I tried reinstalling but it keeps going to that error after install reboot. The storage is ISCSI connected via Intel Server Adapter, which allows it to boot from ISCSI. Not sure if that's the cause for the problem, but the first server is connected to the same ISCSI and installed just fine.
Is there a way that I can make sure ISCSI module installs during installation? Although I think it is installed since it's able to copy the files and setup /dev/sda. I just wana make sure that it installs during setup.
Centos 5.3 includes Ext4 and improved support for encrypted file systems but it appears to be aimed at laptop/desktop systems, in that a password must be entered at boot time.
Is it possible to have a server with an encrypted root file system boot up without entering a password?
Mandos will do it...
http://wiki.fukt.bsnet.se/wiki/Mandos
...by serving up the password from another server...
http://packages.debian.org/squeeze/mandos
...to a client loaded into the initial RAM disk environment...
http://packages.debian.org/squeeze/mandos-client
...but it's not available on CentOS, and is only in Debian unstable.
Is there a similar (or any) solution for CentOS?
In particular, I'm envisaging encrypted virtual machines being served passwords from their virtual host.
Alternatively, the data that *really* needs to be protected could be encrypted while the system core remains unencrypted. But then the keys to decrypt the file system must be stored in the unencrypted portion, so this is not an effective method.
I have never done this before but stood up a new Debian (testing) x64 system. It only has 146GB available for RAID1 so I created a 500GB iSCSI LUN on my NAS device on the network and am really confused how I can attach my Debian to the iSCSI LUN I created. Right now the OS is installed all locally on the machine but I would like the iSCSI LUN to be the /home directory for mail storage. Is this possible or do I need to mount the LUN to a newly created folder / mount point that is locally attached?
View 1 Replies View RelatedI have an iSCSI target on my Ubuntu 10.04 LTS server that is giving me access to 3 x 2TB volumes. I am using LVM to create 1 big volume and using it as a place for my backup software to write to. Upon initial setup I tested to see if the connection would still be there after a reboot, and it was. After going through the trouble of configuring my backup software (ZManda) and doing a few very large backups I had to reboot my SAN (OpenFiler) for hardware reasons. I took my server down while performing the maintenance and brought it back up only after the SAN work was done. Now, the LVM volume is listed as "not found" when booting.
Using iscsiadm I can see the target but LVM doesn't know of any volumes that exist using those targets. I need to get it back up and running and then troubleshoot the reboot issue.
I seem to get logged into the NetApp filer fine - it shows the session, but I don't get a device. And don't panic - the system is CentOS 5.4, but the system is named OpenVMS (don't ask . . . ) iscsiadm shows me logged into the NetApp and the session is up.
[Code]....
I have used open-iscsi for a couple of years on fedora and oracle linux, but since I use ubuntu for my desktop, web servers and mysql servers I thought I would be nice to do so on our new MySQL-server which should be connected to our Qlogic SAN. The installation went well after figuring out how to put two network cards on the same network, which seems to be the Qlogic way to do multipath. But I cannot make open-iscsi to log on automatically after reboot. Is there something different with open-iscsi on ubuntu?
Everything works if I logon and do:
iscsiadm -m node -l
multipath -v2
And then mount the volumes but if I try to put them in my fstab like:
/dev/MySQLdata/data/var/lib/mysql/ext4_netdev1 2
/dev/MySQLbackup/backup/var/mysqlbackupext4_netdev1 2
I've been trying to figure out how to set up a partition on my CentOS serve r as an iscsi target, so I can access it from another CentOS client.I've been reading the manuals and various pages on the web, and nothing is very clear. I just want to be able to create a partition on my server, define it as a target, and then have my client initiator mount it.
View 2 Replies View Relatedi have installed centOS5. i want to configure iscsitarget. i know that iscsi target is built-in in centOS5. how can i use iscsiadm utility to configure iscsi target. please describe in detailed.
1.) i want to make 4 drives (4Drives OR 4LUNs)in iscsi target each with 5GB space.
2.) please describe and highlight where we mention disk space(5GB)
I have an iSCSI device that is large, ~17TB usable, and I've created an XFS file system at the device level. I mount the device no problem and am able to touch and remove files without issue. I start my application and within about 5-10 minutes I start seeing the following entries appear in /var/log/messages
[Code]...
I have a CentOS 5.3 server, running in a VM. I get the iSCSI session to my Buffalo, but the disk doesn't appear.
Sep 29 09:06:29 qaservice iscsid: iscsid shutting down.
Sep 29 09:06:29 qaservice kernel: Loading iSCSI transport class v2.0-871.
Sep 29 09:06:29 qaservice kernel: iscsi: registered transport (tcp)
Sep 29 09:06:29 qaservice kernel: iscsi: registered transport (iser)
[Code].....
Linux 2.6.18-238.9.1.el5
Apache 2.2.3
MySql 5.0.77
PHP 5.3.3
Samba 3.5.4-0.70.el5_6.1
Problem: my server eats up all the memory in the box after a few days. My guess is it has something to do with these processes:
[Code]...
I'm trying to jump Ubuntu Server 10.04.1 x86_64 on an HP DL360 G5. After loading the modules for install, I'm greeted with an iSCSI prompt that cannot be bypassed. Hitting Finish brings me right back to the same iSCSI prompt.The issue seems to be identical to y kickseed file is as follows:
Code:
#platform=x86, AMD64, or Intel EM64T
# System authorization information
[code]....
I'm wanting to run 10.10 server from a root disk located on an iSCSI server. My server is FreeBSD 8 running isc-dhcpd and is sharing out a 20GB iSCSI disk. I've run the server install CD (currently testing in VMWare), and it finds the iSCSI share without a problem. I'm able to install the OS just fine, but the problems start when I try to boot the installation. I'm booting off a gPXE iso until I can get PXE chain-loading figured out. My DHCP config looks as such:
Code:
host mm-iscsi
{
hardware ethernet 00:50:56:3f:3f:bf;
fixed-address 10.50.3.245;
[Code]....
I'm trying to install Fedora 11 x86_64 on system with Adaptec U160 and 2 Ultra160 HD. It is not part of a network and iSCSI is not wanted. But the installer keeps selecting iSCSI automatically and won't let me change it. I try to do a "replace linux system" with my own partition scheme-which preserves the partitions already on this drive. I eventually have to exit the installer.
View 1 Replies View RelatedI have a server wich is connected to an iSCSI storage and gets harddisks from this storage. Sometimes I have to add new disks to this server. Everytime I add a disk and make an /etc/init.d/iscsi restart on the server the disks don't have the same device name as before the iscsi restart.
It should be possible to gave the disks persistent names using udev rules. Now I tried to create different rules in "/etc/udev/rules.d/99-static-iscsi-names.rules" e.g.
# /dev/sdc
KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -s %p",RESULT=="360a98000503355344c4a576864467877" NAME=="sdc%n"
In "/etc/rc.local" I added "/sbin/start_udev" and in "/etc/scsi_id.config" I added the line "vendor="NETAPP",model="LUN",options=-g"
[Code]...
I am trying to install iscsi-target on centos 5.4 i386. When I go into the iscsi-target directory and type make I get the following error:
[root@cluster-storage iscsitarget-1.4.19]# make
cc: /lib/modules/2.6.18-164.el5/build/include/linux/version.h: Datei oder Verzeichnis nicht gefunden
cc: keine Eingabedateien
/bin/sh: line 0: [: -le: unary operator expected
/bin/sh: line 0: [: -le: unary operator expected
/bin/sh: line 0: [: -le: unary operator expected
/bin/sh: line 0: [: -le: unary operator expected
/bin/sh: line 0: [: -le: unary operator expected
/bin/sh: line 0: [: -le: unary operator expected
/bin/sh: line 0: [: -le: unary operator expected
/bin/sh: line 0: [: -le: unary operator expected
/bin/sh: line 0: [: -lt: unary operator expected
make -C usr
make[1]: Entering directory `/opt/iscsitarget-1.4.19/usr' .....
make: *** [kernel] Fehler 2
I'm trying to configure an ISCSI/DRBD high-availability cluster and I'd like to know what is the best option between OpenAIS and Heartbeat. I've seen they both are included in Centos Repos, yet OpenAIS requires installing 2 addition repos to install Pacemaker (EPEL and Clusterlab repos).
View 1 Replies View RelatedI'm trying to increase the size of an iSCSI device. On the LUN side, the provision size has been expanded already. I expanded the size of this device from 15GB to 30GB.
I then rescanned the devices:
[root@ps-dev-nfs ~]# iscsiadm -m node --targetname iqn.2001-05.com.equallogic:0-8a0906-db7938c02-c6a994bda8a4a773-ps-dev-nfs -R
Rescanning session [sid: 1, target: iqn.2001-05.com.equallogic:0-8a0906-db7938c02-c6a994bda8a4a773-ps-dev-nfs, portal: 10.3.2.10,3260]
[Code].....
I am interested in turning my home server into something that I can store backups on. I do photography and therefore have a lot of photos. I use Mac OS X for my photo editing, so it must be accessible from my Macbook. I am new when it comes network storage servers, so what would be the best solution for me to be able to backup my photos seamlessly? I would like it easy enough so others can backup files without any terminal commands and such. What would you suggest? CIFS? RAID? iSCSI?
View 3 Replies View RelatedI am running RHEL5.5 its a fresh install and we are testing Xen Virtualization. We are wanting to use our iSCSI SAN for the VMs. I have created the initiator iqn, and discovered the target address. We are connected to the target, but there is no new block device in /dev.
[root@xxxxxxxx etc]# netstat -a | grep iscsi
tcp 0 0 *:iscsi-target *:* LISTEN
tcp 0 0 xxxxx-netapp.okc-:44840 xxxxxx-iscs:iscsi-target ESTABLISHED
tcp 0 0 *:iscsi-target *:* LISTEN
[Code].....
I have two servers, both running CentOS, but one running Cpanel as the control panel and the other running Webmin.
I would like to sync the web folders using rsync. However, when I sync the files from the Cpanel host the permissions look as follows:
Code:
But that doesn't work on the Webmin server which needs the following:
Code:
It seems the Cpanel install has the group as 'nobody' whilst webmin has it the same as the file owner. I can fix this by running:
Code:
But as the file transfers will be frequent, I don't want the files to be inaccessible until chown is run.
So, what I really want to know is - how can I change things on the 'webmin' server so that files with ownership owner:nobody will run?