Server :: Using Either FreeNAS Or OpenFiler As The OS?
May 24, 2010
what is the best way to go for setting up a NAS. It will be used for 2-3 weeks worth of HD Video storage and needs to have a redundant power supply, swappable SATA drives. I'm thinking of using either FreeNAS or OpenFiler as the OS.
We are limited to a small amount of available server hardware in South America. To buy an 8TB HP NAS server costs $11000, quite pricey for us. In Canada I'd just get a couple of Buffalo Terrastations, but that's not an option now.
So I'm thinking of going with an INTEL SR2400 ($550), maybe even two of those for more redundancy. Anyway, I was thinking of using 1TB or 1.5TB drives, but according to other web sources, that's a bad idea with RAID, as it's more prone to hardware errors.
I've been tinkering with both recently to see what can suit my needs as a simple raid1 mirrored backup server. I used FreeNAS for all but minutes before I had a raid1 array running and shares set up through CIFS. Using Ubuntu on my laptop I was able to see them as well. I had some more figuring out to do, as I wanted each share to be blocked off from the other. aka - I didn't want "fred" to be able to access "bob's" share. So then I move on to Openfiler to check it out. I hear it's simpler. It's web interface, while no doubt much slower than FreeNAS, was easy to mingle around. At first glance, I thought I'd like it more.
I began to set up my raid1 array. After realizing the final release I was dealing with had a bug, I found some commands to run in a root shell to fix. Okay, so now we're moving along... raid1 array created. Then I had to create a volume group. Then a volume. Then shares. And I still don't have it running over cifs. I'm curious if FreeNAS by nature is truly simpler to set up raid array's with, or if maybe I'm just totally misunderstanding Openfiler and that be the reason why I'm thinking this.
So what i would like to do is possibly get an atom 330 board with a pci-e slot and a 12port raid card and 2 or more GB of ram.Wondering if this would be a better setup then what i have now.Now there are some fall outs to what is setup now which are that Openfiler can not detect if a drive has failed so it requires me to monitor this server on a regular basis. I considered doing a software raid 6 but wasn't too keen on that since i read it takes a fairly good machine to compute.
I'm trying to share a file via Samba on a Ubuntu server that is actually stored on a FreeNAS box. The FreeNAS drives are mounted via NFS and the Samba share contains a symlink to file on the FreeNAS drive.Browsing the Samba share I can see the file and size, but any attempt to read the file fails. It complains about authentication but all credentials across all machines are the same.So, is it possible to share a file this way or is there another way to do this?I know I could create all the profiles on the FreeNAS box but for convenience and ease of maintenance I was hoping to do this via the Ubuntu server
This should be a quick one. I'm trying to backup a single directory and it's subdirectories on my Lucid Server to a freenas box across my network. This is what I'm using to do that..
rsync -r -a -v -z * --delete freenas:dSIBackups It almost works perfectly except for one problem. When a file is deleted at the source, this command doesn't seem to delete it on the receiving end. I assumed that the --delete would do that but aparently not. Can anyone think of a reason that this would happen?
I think I can eliminate Media Companion as the problem since all other samba servers work with it.
I want to trouble shoot this but don't even know where to start. How do I figure out what makes OPENWRT samba server different from the other 100% working FREENAS and PC-01 Servers?
I'm a photographer and I have a requirement find a better method of storing my photos other than multiple USB2 drives via USB hubs. Currently I use a Macbook Pro and 6 external drives connected via USB2 or FW800. 3 are a copy of the first three, kept up to day manually by running an rsync backup. I'd like to run a FreeNAS or OpenFiler NAS box using 2TB drives mirrored via software RAID. But - I would like to have the flexibility of also plugging into the drive physically for the faster throughput when necessary. My question is, is there a file system that both *nix and Mac OSX will play nice with?
I downloaded the lastest x86 stable release of openfiler and installed it on VBOX durring the install I selected to manually do the partitions and added a
/boot swap (x2 ram) /
Partition scheme. I selected DHCP client on the network configuration. Now from what I understand once I have done that I should be able to point my browser to https:<IP>:446 and login as root and configure everything within the GUI. But when I log in there is nothing to configure. My host system is Ubuntu x86_64 4GB Ram amd dual core processor.
I would like to make a small NFS server for a small LAN. Normally, I would build a dedicated cheap and cheerful linux box to do this. However, I was wondering if all of this could be done more easily using a commodity standalone device like e.g. "NetGear ReadyNAS Duo NAS". I presume devices like this run their own proprietary OSes, and I would prefer instead an opensource OS based device. I do like the look of these devices as they seem simple and small.
So my real question: What would linuxers advise for me given that I want a minimalistic NFS fileserver? I can make my own dedicated linux desktop machine. However, is a standalone device similar to the above, but running something like FreeNAS, also an option?
I have set up freenas with 3 1tb hard drives. I have set up the SMB shares for the drives and can view each shared drive from each of the machines on my network. I can copy files from the hard drives, on the freenas but when I try to copy a file to the Freenas hard drives I get a message that I need permission to do this. I have all my shares set as anonymous how do I change the permissions so that I can save files to the drives.
For some strange reason, I can't seem to be able to mount the nfs share from my FreeNAS system on SL6. I'm able to do it just fine from Ubuntu 10.04, Linux mint 9, Fedora 14, CentOS 5.5, and OS X Snow Leopard, so it has to be something specific to SL6. The below command does not work: mount freenas:/mnt/share /test.
I get a mount.nfs error message that says "requested NFS version or transport protocol is not supported." I also tried this command which yielded the same result: mount -t nfs FreeNAS:/mnt/share /test Am I doing something wrong or is this just a bug with SL6?
I've found that simply sharing a folder on my Ubuntu PC with other Linux PC's on my network proves unwieldy from an 'ownership' standpoint. For example, if I create a file from my Ubuntu Laptop and save it onto my Ubuntu Desktop 'share', I find that if I try to access it from the Desktop that the group/ownership prevents my access so I end up having to chown it. I'm finding sharing in a mixed environment to be a pain.
So, I recently configured an old desktop box with FreeNAS and I want to move all of my files from my main Ubuntu PC over to it, and have those files accessible from various clients (both Windoze and Linux) from thoughout my network. Now I've got the FreeNAS all set up and shared, and I'm ready to transfer files to it. What's the best way to mount Mr. FreeNAS on my various desktops and laptops such that sharing is not a problem from a security (group/owner) standpoint? Should I be using CIFS/SAMBA or is NFS mounting better? Will there be sharing problems or ownership issues when accessing from Windows?
At work, I recently installed FreeNAS 0.7.4919 on a computer and set it up to be a samba server. Using a Windows XP Pro SP3 computer on the same subnet as the FreeNAS server, I can ping the FreeNAS server's hostname and it works just fine. However, using an Ubuntu 10.04 computer on the same subnet as the FreeNAS server, I am unable to ping the FreeNAS server's hostname. When I try, it says "ping: unknown host [the FreeNAS server's hostname]". I can ping its IP address just fine, though. Why is it that Windows XP Pro can ping the FreeNAS server's hostname but Ubuntu 10.04 can't?Here's the output of the ifconfig command on the Ubuntu 10.04 computer:
Code: eth0 Link encap:Ethernet HWaddr 00:14:c2:cd:a6:39 inet addr:10.37.74.141 Bcast:10.37.75.255 Mask:255.255.252.0
I am very new to Ubuntu and have been having trouble mounting my FreeNAS drive. I installed Ubuntu 10.10 on partition sda2. I wanted to keep FreeNAS completely separate from here, so I used Virtualbox to host FreeNAS as a guest o/s on a second hdd, sdb1 mounted at /media/NAS-Data. I can access the NAS from all computers except my Ubuntu box. I have CIFS/SMB and NFS (among other) services enabled on FreeNAS.
I would like to run a program that needs access my music. I followed many of the "How To's" on the forum, but am not sure if they didn't work or if my setup is different and can't work the way that has been described. My last effort was to mount the file system using NFS, but I get a timed out error.
When I run showmount -e 192.168.0.44, result is /mnt/cNb-NAS-data 192.168.0.0. I've tried many variations to mount, but none have worked. For all I know, again I'm very new to Ubuntu, the file system is already considered mounted (/media/NAS-Data), and I just need to find the correct path to access my data. This is probably obvious, but when I navigate to NAS-Data, it has the Virtualbox NAS.vdi file.
Was hoping someone might be able to either help me get the correct path name or mounting instructions in order to view these files from Ubuntu.
I just set up a FreeNAS server and have a shared drive set up that I can currently access from Linux, Windows and OSX. I'm having a problem getting the trash folder to work for files deleted from my Linux machines though.
I know this may be more of a FreeNAS forums question, but I've tried asking there and haven't gotten a response. And the recycle bin is working when a file is deleted from a Windows machine on my network so it is an issue specific to Linux.
So my question is, what services, protocol, etc... has to be used (on the server and/or on my Linux machine) to have a functional recycle bin on the FreeNAS shared drive? If I delete a file on the shared drive from my Linux machine, I would expect it to go to a trash folder and still be on the shared drive. It's working from Windows but not Linux.
So I've been researching this on my new NAS. I just tried to register the IP on the new system to have an issue pop up that says: "msk0: watchdog timeout (missed Tx interrupts) -- recovering" The error repeats then ends with no set IP. I tried the loader.conf fix but I cannot write to the system its installed on (embedded on a harddrive + data + swap) but it gives me a issue with write privileges which I thought would have been fixed by chmod +w /cf.
I had a samsung 1TB HDD that I used for storing data, on an xp machine, so it was formated as NTFS.I moved this HDD to another machine and installed Freenas on it, and the installation worked fine (fyi, I used the tutorial posted here :[URL]..During the installtion, Freenas installed it's system files to a new small UFS partion. After finishing the setup, I realised that I had changed the file system of the other partion (980gb, previously NFTS) to UFS and now I don'T know how to go back. I had about 400gb of data on it and I'm pretty sure it's still there, but don't know how to get it back.
I tried messing around with recovery software such as R-Studio, and I was able to see some of my files so I know they're still there. After quite a bit of googling around, the only solution I seem to find is using gparted which is a tool to modify partions file system without loosing data, but I'm afraid to use it.
So is there a way to browse NTFS data on a UFS partition and convert it so FreeNas can see my files ? Or is there a way to put the partition back to NTFS so I can back up my data to another drive before I lose something valuable ?
I have installed a linux server in my office to run 16 machines. Its main use will be a internal mail server but will be also running websites.
I have installed Ubuntu 9.10 server x64 and have got apache running.
I am looking for the simplest more robust solution for smtp, pop3 and imap. I have only ever used qmail before and found it a pain to configure and its getting old so I though I should probably try something new. I have not much experience with running pop3 or imap on linux so would love a suggestion on that.
have to create a webhost on an running fedora server which runs multiple webpages + a coldfusion serveri have to add an coldfusion virtual host to these.what i would do:*crate a new user & group*enter vhosts.conf and copy an existing host and modify it for the new one.*create an new folder and copy the main files (phpstarter and webroot) *chown the files for the right useri think an apache graceful would be needet
Right now i have a HP DL 180 Server with 130 Gb Hard Disk & 8 Gb ram after Raiding0+1. i want to configure Domain Controller Server for my office for 200 to 300 Users. what should the partition size must be mentioned in my 130 Gb Hard Disk, is that going to be Sufficient for ME ?
i am bit confused about /Usr /Var /Boot partitions, as i need to manage perfectly in 130 GB
if i go with 4 Gb swap and remaining for " / " is that will be fine ? should i need to specify partition sizes separately for / tmp /var / usr ..
I'm running a linux cloud server with the following config 1.2ghz Processor allocation 752MB Ram
The site loads slow and clicking a link almost freezes the page for a second. Also, the page loads could be much faster. We've been running mysqltuner and have pretty much optimized all slow queries. Is there anything we can do to fine tune the server for faster and more responsive?
iam trying to sync file server data into backup server machine by command- rsync -avu path/of/data ipaddress-of-backup-server:/path/where/to/save after running it ask for root password and manually it is successful.but i want to make it automatic.for that i also tried cronjob and also generated authentication key but iam not successful in login automatically..anybody know how to authenticate root to login for storing data in backup server.
I will be relocating to a permanent residence sometime in the next year or two. I've recently begun thinking about the best way to implement a home-based network. It occurred to me that the most elegant solution might be the use of VM technology to eliminate as much hardware and wiring as possible.My thinking is this: Install a multi-core system and configure it to run several VMs, one each for a firewall, a caching proxy server, a mail server, a web server. Additionally, I would like to run 2-4 VMs as remote (RDP)workstations, using diskless workstations to boot the VMs over powerline ethernet.The latest powerline technology (available later this year) will allow multiple devices on a residential circuit operating at near gigabit speed, just like legacy wired networks.
In theory, the above would allow me to consolidate everything but the disklessworkstations on a single server and eliminate all wired (and wireless) connections except the broadband connection to the Internet and the cabling to the nearest power outlets. It appears technically possible, but I'm not sure about the various virtual connections among VMs. In theory, each VM should be able to communicate with the other as if it was on the same network via the server data bus, but what about setting up firewall zones? Any internal I/O bandwidth bottlenecks? Any other potential "gotchas", caveats, issues? (Other than the obvious requirement of having enough CPU and RAM).Any thoughts or observations welcome, especially if they are from real world experience in a VM environment. BTW--in case you're wondering why I'm posting here, it's because I run Debian on all my workstations/servers (running VirtualBox as a VM for Windows XP on one workstation).
I have got a running mail server on CentOS5.4 Final using Zimbra free edition. . Assume that domain is organization.com.local And it is running fine with users created and bugzilla running on the same machine can send mails using this server. But it can obviously send messages to the internal users. There is this option to configure it to use external mail server to send mails outside the network but my external smtp server needs an authentication.Is it possible to configure it this way to use my external server to send mails from internal domain to say gmail?
i recently installed apache2 mysql php exc. to run a realtime stat site for my servers.the only thing is that the buttons/ images are not showing up.i checked that my GD was up to date and installed perl5 GD. dont know what else to check.my site with problems go to:http://24.20.177.228/stats_public
I am working on linux server with below specifications.Linux EDT 2008 i686 i686 i386 GNU/LinuxWhile checking the status of the server using the command 'opmnctl status' and when server is down the output is not getting redirected to file.I m using the command as,opmnctl status > abc.txt.
i have been trying to complete the following project1) Configure a FTP server where we can upload and download files.........2) server must run at 9 pm & stop at 9 am automatically ............although the first task was easy ,i have no idea how to accomplish the 2nd task(not to mention I'm a new user)
I've two internet based server ( xx.xx.xx.xx and yy.yy.yy.yy ) The Y server is running VNC server and is responsible for answering to VNC sessions. But I need to hide the IP of Y server so I want X server to be as VNC Proxy and redirect all VNC sessions to Y server.
I guess the best way is to use iptables but actually I can't get it working so
I'm trying to setup RAID 1 on a CentOS 5 server for a zimbra email server.I get a partion schema error. Can I do this?The server is a HP Proliant ML150 G3 server with two 80GB HDD.