CentOS 5 Server :: Automation Of Ftp Server \ Configure A FTP Server Where Upload And Download Files?
Jan 8, 2011
i have been trying to complete the following project1) Configure a FTP server where we can upload and download files.........2) server must run at 9 pm & stop at 9 am automatically ............although the first task was easy ,i have no idea how to accomplish the 2nd task(not to mention I'm a new user)
I've setup a Lamp Server for Testing, The Lamp Server is Up & Running on CentOs 5.5
I am now trying to setup a VSFTP server where local users can upload files to there home directory so that Apache can serve web pages straight from the directories of system user home/accounts giving users the ability to run their own web sites which are hosted off the main server [tutorial here: [url]
So far i have been able to serve/display index.html files from the users home directory [url] but so far i cant upload files to any user home directory, every time i try to upload a file with filezilla i get this error message: 553 Could not create file. Critical file transfer error
I have searched online for similar problems like mine and so far i've tried alot of the solution but none seem to work. I'm confused, dont know where i went wrong, i put the users in a group called ftpusers and here are the permissions on the users (test, ftpuser & testftp) home directory. have a look an tell me where i went wrong :(
Also the root directory where the web pages are served from is called public_html here are the permissions
Here is my vsftp.conf file can someone check it to see if i made any errors in there:
I just want to place upload/download limit to each ip address ( say 20MB per day for each ip or specific ip) using squid proxy i tried through delay pool but it control only download/upload speed not place any download/upload quota restriction. following are my codes I have 1Mbps line
delay_pools 2 delay_class 1 2 # here 700kbps(87Kbytes) Net-Total usage limit, with 50Kbytes per user
I also used
"reply_body_max_size" parameter to control upload limit.
"request_body_max_size" only restrict download based on size.
how to setup quota ( download quota per day per ip)
i need to configure postfix on centos to relay email from the internet to the Exchange Server and i also need that emails sent from the exchange within the same domain be sent to postfix then resent to exchange because i have spamassassin and clamav installed on centos to filter all incoming and outgoing mails ...
I tried to install php from source file on centos server and i got the error as # ./configure --with-apxs2=/usr/local/packages/apache/bin/apxs --with-mysql --prefix=/usr/local/packages/apache/php --enable-force-cgi-redirect --disable-cgi --with-zlib --with-gettext --with-gdbm
checking for specified location of the MySQL UNIX socket... no checking for MySQL UNIX socket location... /var/lib/mysql/mysql.sock checking for mysql_close in -lmysqlclient... no checking for mysql_error in -lmysqlclient... no configure: error: mysql configure failed. Please check config.log for more information.
I have one server at home, connected through a router that gives me a static IP:192.168.1.2 - My PC192.168.1.3 - CentOS serverI enabled the port 80 into router forwarding system and I can connect properly to CentOS server, using the IP my internet provider gives me. My router automatically forwards a changed IP to my dyndns.com account so I can also access the server through a domain name (I pay to have it linked to one of my domains).My next step is to configure the mail server (dovecot) , so I can send emails from my local server(192.168.1.3).I was wondering if anyone has any tutorial links or tips to have the mail server configured through a home internet provider + home router.
I will be relocating to a permanent residence sometime in the next year or two. I've recently begun thinking about the best way to implement a home-based network. It occurred to me that the most elegant solution might be the use of VM technology to eliminate as much hardware and wiring as possible.My thinking is this: Install a multi-core system and configure it to run several VMs, one each for a firewall, a caching proxy server, a mail server, a web server. Additionally, I would like to run 2-4 VMs as remote (RDP)workstations, using diskless workstations to boot the VMs over powerline ethernet.The latest powerline technology (available later this year) will allow multiple devices on a residential circuit operating at near gigabit speed, just like legacy wired networks.
In theory, the above would allow me to consolidate everything but the disklessworkstations on a single server and eliminate all wired (and wireless) connections except the broadband connection to the Internet and the cabling to the nearest power outlets. It appears technically possible, but I'm not sure about the various virtual connections among VMs. In theory, each VM should be able to communicate with the other as if it was on the same network via the server data bus, but what about setting up firewall zones? Any internal I/O bandwidth bottlenecks? Any other potential "gotchas", caveats, issues? (Other than the obvious requirement of having enough CPU and RAM).Any thoughts or observations welcome, especially if they are from real world experience in a VM environment. BTW--in case you're wondering why I'm posting here, it's because I run Debian on all my workstations/servers (running VirtualBox as a VM for Windows XP on one workstation).
I have got a running mail server on CentOS5.4 Final using Zimbra free edition. . Assume that domain is organization.com.local And it is running fine with users created and bugzilla running on the same machine can send mails using this server. But it can obviously send messages to the internal users. There is this option to configure it to use external mail server to send mails outside the network but my external smtp server needs an authentication.Is it possible to configure it this way to use my external server to send mails from internal domain to say gmail?
I want to upload files from my computer to an FTP site and I don't want to upload files that are already on the server. So I need a tool that finds out which local files that are different from the ones on the server, or that don't exists on the server.
I'm using a cheap provider that does not support rsync or ssh, so I can only use FTP. I generate the files before uploading them, so comparing timestamps is meaningless. I've tried lftp with the mirror command. It's slow (I think it uploads all the files). I upload the files from different computers, so I can't use sitecopy, which uses a local database to keep track of which files are on the server. I'd like to be able to upload all changed files with one command. Preferably no GUI application. And it needs to run in Ubuntu.
I was thinking about creating a tool similar to sitecopy, but which stores checksums of all the files on the FTP server on the server itself. But then I thought that there may already be such a tool.
I have tftp-server running on Centos 5. Clients which are on the same subnet as the server are able to get and put without problems. I have a client that is across the internet that is having trouble getting files from my tftp server. A tcpdump reveals that the client is requesting the same file over and over again. In /var/log/messages, I am see the following error repeated over and over until the client finally gives up.
localhost in.tftpd: tftpd: read: No route to host
I have a weird performance issue with a centos 5 running a nfs server and a rh8 client. I think the fact that it is rh8 client should be downplayed. It is just that with rh8 client the performance degradation seems more clear. See test details below OS in server is Centos 5 x86_64 kernel 2.6.18-92.1.22.el5
1Gb connection between machines File to test over NFS is a 1GB file. First of all I wanted to measure how the network alone performs while using NFS. So in the server side I run a "cat" command on the 1GB file to /dev/null. Please note that the disk read speed is about 98MBs. At this point the file system has the 1GB file cached in memory. In the client side a "cat" on the same file gives me a speed of about 113MBs. It seems then that the bottleneck in this instance is the network and it is very close to nominal speed. So the network performance is really good. (BTW I know that the server got that file from cache because a vmstat or iostat shows no disk activity.)
The second test is reading from disk with no caching involve. In the server I flushed the 1GB file from the memory. For instance by reading another 5GB file and I repeat the same thing as above in the client (a cat on the 1GB file). Now, the server has to go to disk.(vmstat or iostat shows the disk activity). However the performance, now, is about 20MBs, I was expecting something closer so 90MBs. (since the reading speed in the server in the first test showed 98MBs).
This second test was repeated for ext2, ext3, xfs with no significant differences. A similar test using a RH8 NFS server and client gets me close to 60MBs for a 1GB file not cache by the file system in the serverSince network speeds and disk read speeds are not the bottlenecks ... what or where is the limiting factor then?
I'm configuring a new public mail server running Centos 5.5 and Sendmail 8.13.8. I would like that the sendmail configuration is correct. The server will handle many domains using Virtualmin. Everyone can send mail only if authenticated, which I already tested.
How can I upload configuration file? I receive an error if try to upload a zip file.
I use a PHP IDE that has no built-in ability to upload a project to a site. So, I'm looking for common easy to use tool for Linux that would able to upload modified documents to the server instead of uploading of the whole site. I also accept shell scripts that would be able to do this.
I have installed the Apache, PHP and MYSQL in the rackspace cloud server environment. Can anyone please guide me How can I configure email server in that with postfix or some other with multiple domain.
my task is to store the data to a file at server sent via HTTP POST, i written suitable cgi script in C, but the PROBLEM is that i can't complete the task, I guess i require server permission to do so, how to configure server or how to get access right from server to store data to a file.
i'm teacher and have been assigned to take care a small computer lab in my school.in my computer lab there is a centos box with 2 network cards, switches and computers.at this moment all the computer can access the internet directly without connected to the router or server..(bypassing the centos box).so now i would like to make my life more " difficult" where i want all the computers access internet through my centos box so that i could have a better internet usage management.
I've been reading some forum and even centos document regarding establish networking but it seem difficult for me to understand all the terms and jargon overnight.I also have been try to configure the 2 nic using the "trial and errors" approach but no luck..i'm really hoping that someone could guide me so that i could configure the 2 nic correctly hence my centosbox could be more beneficial..below is the scheme how my network will be..internet------>modem--------->centosbox--------->switches------->computers(eth0)(eth1
I have been able to find enough information that I need to enable 'time-udp' in /etc/xinetd.d/ But there isn't an entry for time-udp. How do I enable time-udp (Time of Day server) on a Red Hat system? It's RHEL 5.6 64bit.