I'm setting up a raid 5 on several hard disks with a layer of lvm on top for good measure.I know the recent kernels support growing software raid, but since centos runs 2.6.18, I wanted to make sure it'll work. Does the centos kernel support growing raid devices?
I have installed CentOS 5.3 with Xen on a PowerEdge 2650 machine with 6GB of RAM. As usual, PAE was already enabled so I did not have any problem with utilising all of the memory. However after upgrading to the latest release of the kernel (2.6.18-128.7.1.el5xen), memory available decreased to 4GBs. Then I switched back to the old kernel and 6GB was there. Then finally have switched back to the new kernel and 6GB is again there. So now PAE seems to be enabled but what concerns me is the inconsistent behaviour. (Also I am not sure the reason was the upgraded kernel.)
I'm trying to find out when QME2572 (Qlogic) card became support by the kernel. We have a RHEL 5.1 system that is moving to new hardware, however the kernel at this release doesn't support the new hardware, due to the Qlogic card change. I tired the Redhat KB and Bugzilla. Is there a Kernel change list etc I can search. Never really played around with the kernel too much so I'm just after some pointers for looking up this information. Offically its not supported until Redhat 5.3, I'm trying trying to research kernel info so I can tell the customer they have to upgrade.
i have recently bought a system which has intel atom and my Lan card hasn't been installed on the system, i am using centos 5.2, i have read in some articles that Centos 5.4 directly support this LAN Card,so will upgrading the kernel for the Operating system help, or i have to download the whole iso for 5.4 and install it over the existing if upgrading the kernel will help which file i need to download to upgrade the kernel and where can i download that file.
I try to set up a 5 nodes cluster and a shared Coraid Storage with conga but it fails with "Shared Storage Support" checked.The message is:'A problem occurred when installing packages: Packages of set "Clustered Storage" are not present in available repository' and it is shown under every node on the next sceen after I submit.The pc where conga runs is on the same subnet (192.168.xxx.xxx) and it has the same /etc/hosts of the other nodes.In that pc runs a proxy too and the nodes go out through it (that pc has 2 NICs)Every node (2.6.18-128.1.14.el5-xen-x86_64) is patched whith the last yum update (this morning) the same is for the pc (2.6.18-128.1.14.el5).Every node has 4 NICs , 2 NICc towards the storage the others in bonding towards the WAN.Every node is exactly alike, they have been installed with the ks.cfg generated from the first node and they all have the support for Clustering, Virtualization, and Clustered Storage.
I have been using SLES 10 SP1 so far with about 6TB Raid system without problems.I have upgraded the OS to CentOS 5.3 i386 and I have noticed the kernel can not recognize raid system larger than 2TB.Is there any parameters that I have to set ? or the i386 distribution simply does not support larger raid, so I have to use x86_64 version, instead?
I am running a CentOS 5.5 system which does NOT support VT Technology, so I am unable to use Xen HV. I am very happy with OpenVZ other than the fact that it doesn't support Windows, only Linux.I'd like to have VMWare Server 2.x installed, but it appears that OpenVZ changes the Kernel to, in my case, 2.6.18-194.3.1.el5.028stab069.6. The default kernel on my system is 2.6.18-194.3.1.el5.i386.If I change grub to load 2.6.18-194.3.1.el5.i386, VMWare will work but OpenVZ will not. If I change grub to load 2.6.18-194.3.1.el5.028stab069.6, OpenVZ will work, but VMWare will not.If I try to run /usr/bin/vmware-config.pl while in the 2.6.18-194.3.1.el5.028stab069.6 kernel, I get an error of:
None of the pre-built vmmon modules for VMware Server is suitable for your running kernel. Do you want this program to try to build the vmmon module for your system (you need to have a C compiler installed on your system)?Using compiler "/usr/bin/gcc". Use environment variable CC to override.
What is the location of the directory of C header files that match your running kernel?[/usr/ src/ linux/ include]I'm so lost as what to do now..is there anyway I can compile both of the kernels together somehow?
I just want to know what is the default Fedora 32 bit kernel - does it support more than 4G memory. I would like to save myself from compiling custom kernel and all following fiddling (nvidia drivers, virtualisation software etc) and 64bit OS is not for me - too many apps that I need are missing, of poor quality or need more manual intervention. Actually if bigmem / whatever it is called is configured in the default kernel, this may be the reason to switch. I intend to use 4 to 8 GB memory and it's a desktop, so server builds (like Ubuntu) are not for my taste.
A while back I compiled a custom kernel, 2.6.35. I forgot to add UDF support when configuring and compiling. Will I need to re-compile the kernel to get UDF support or is there some other way I can add it?
I use Slackware 13.37 32-bit version on ThinkPad X60s. I compiled two Window Maker dockable applications monitoring the processor temperature, fan speed, and battery status: wmpower (see: [URL]) and wmlenovo (see: [URL]). Both these applications refuse to work.
wmpower displays the messages: Code: CPU frequency scaling available No power management subsystem detected No power management support... wmlenovo displays the message: Code: No ACPI support in kernel
Did I missed something obvious during the configuration of the system and ACPI doesn't work for me or there's some more serious problem with my system?
I am a long time (1.something) slackware user and maintainer of a mirror site.I'm suddenly having problems with my favorite distro. (1) I've been mainly running Slackware64. When I was experiencing recent problems on a Slackware (32) 13.1 system, I discovered the huge-smp kernel does not support more than 4G of RAM. This is an obvious bug, and I am shocked that there is no fix out yet. Surely I can't be the only Slackware 32 user with >4G of RAM. I've verified it on a 12G i7 system and an 8G Athlon64-X2 system.
Will Redhat support CentOS under any circumstances? I have heard at my employer that if you setup you system with CentOS and then on down the line you find you need RHEL support that you can purchase a license from Redhat and they will support your CentOS installation.Is that an official policy at Redhat? Unofficial? Or completely untrue?
I am installing a new large 4 Tb RAID (hardware) under 9.10. Does GPT support come by default in the new 9.10 kernel? How can I check? I gather I need to use PARTED to make a GPT partition table on the raid. Anyone know details of how to do this? [URL]...
I'm experimenting with a load balancing setup and have found some pretty good guides, but they all assume that ipvs support is built into the kernel already. e.g. http://www.howtoforge.com/high_avail...apache_cluster
So of course "modprobe ip_vs_dh" simply outputs "FATAL: Module ip_vs_dh not found."
I have been unable to boot my CoreExpress Atom Z530 based board from flash, because my 126.96.36.199 kernel does not seem to have support for the (US15W chipset) IDE controller configured. I have read many web links that point to CONFIG_PATA_SCH as the key, but that does not work for me. I _can_ boot RedHat 2.6.18-164.e15 which does not have CONFIG_PATA_SCH configured, but I believe they are using initrd.
I compiled my own kernel with tun/tap and bridge support. Both modules load fine at boot time (I could read that in the dmesg output). Now I want to use it, and the /dev/net/tun node is not there, so my application gives that error. I'm trying to bridge openvpn connections. Is it possible that udevd is not doing his work?
PS: I'm on a WD MyBook World Ed NAS device. It's ARM, so I cross compiled the kernel from my debian linux machine. I also installed debian on the NAS.)
I was wondering if there's a way for me to get real-time kernel support on regular Ubuntu 9.10 through some sort of update. I know Ubuntu Studio has a real-time kernel, but I'd prefer not to use it. I'm a fan of the original Ubuntu.
I just installed the latest version of Ubuntu 10.04 LTS, and am wondering if support for my dual AMD processors is built into the generic kernel, or do I need to compile a custom kernel? I have an HP tx2-1025dx touchsmart.
I have a PCI SCSI controller with this chipset that was working fine in 10.04 to drive a Polaroid Sprintscan 35 Plus scanner. Having just run the upgrade to 10.10, the scanner is no longer being seen. If I go into the controller BIOS the scanner is seen there, and running lspci shows the controller as
Does somebody know what about native support by kernel for Broadcom's wi-fi chipset - bcm 43225? In kernel sources 188.8.131.52 in section "staging drivers" I saw something like "Broadcom 80211n". I selected this module for compiling, but in process I got error something like wrong type of variable or something else when module was compiling.
I want to add Sandy bridge support to linux 184.108.40.206.I have linux kernel 220.127.116.11 ,and Sandy bridge changes were added from kernel 18.104.22.168.Please can anyone suggest how to backport the kernel support for Sandy bridge.
I have just started to work with BeagleBoard and built the Linux kernel (version 22.214.171.124) and download to the board and it start. I have also connected to the board over serial to Putty terminal.
Now I want to connect to the Beagle Board using USB instead of serial and I am looking for some info on how to do this.
My goal would be to connect to board over USB and then be able to: * Download my own binary files/programs from PC over USB to BeagleBoard * Communicate over USB instead of UART/Putty to get log and execute commands
Question: How do I build the Linux kernel with USB support, meaning what settings should be ticked in menuconfig? Or is it already included?
How do I enable USB OTG in menu config?
I assume I would need some driver on my PC with Windows XP to connect to my BeagleBoard over USB. Were can I find this USB driver?
I'm done following this guide about Kernel Building. [URL] .... and I got done successfully, the .config file was the very default obtained from "make menuconfig" the only extra thing added was:
Code: Select allCONFIG_USB_EHCI_HCD_ORION=y
All dependencies had to be installed about 2GBs for kernel building to function properly, mainly due to the needing of kernel headers. It loads successfully on my desktop whether or not it needs that module (if it was added correctly) but my goal was to add xhci support so it will load from my laptop, still won't load from the laptop giving me a initramfs prompt. Adding that code line to the .config file didn't seem to do what I wanted.
Although since I did a "make-kpkg clean" command as requested by the manual I dont know if it used that line or not to be honest. since I think it deleted the .config file in its efforts to clean old configs so I don't know I'm new to kernel building.
Is there a way to programmatically determine whether a particular driver support is built-in rather than a loadable module? I'd be grateful if somebody would share how. I have written a shell script using the RTC (real time clock). I can check whether the rtc-cmos kernel module exists and load the module accordingly as needed, but I don't know how to determine when the driver is built-in. Of course, if the driver is built-in then the module does not exist.
I seem to recall there is a method to query the kernel config file (/proc/config.gz), through which I probably could grep for the CONFIG_RTC_DRV_CMOS parameter. I also can check whether /proc/driver/rtc exists. If that file exists then either the driver is built-in or the module was loaded.I realize the rtc-cmos driver is built-in with the standard Debian kernel build, but I still would like a way to query where the driver is supported.Is there a straightforward or direct method to query the kernel whether a particular driver is built-in?