what the procedure is to file bugs against slackware so I will post here. The rc.lighttpd I have works but the function to check if lighttpd is running has an exception when there's no lighttpd.
I have a remote server which has been working perfectly for about 1 year. It is in a secure location and used for serving a local web-app, and we remote administer it through an OpenVPN tunnel, and previously a Hamachi tunnel.It has in the last few days become partially unresponsive. What I mean is that I can ping it fine through the VPN, and I can SSH into the box, but when I try to run some commands, it seems to freeze partway through, and I get the output interrupted and a blinking cursor. I have also tried to access the web-app through the VPN and while the page header in the browser changes to the correct page name, the actual page does not load.A reboot does not seem to have cleared the problem, although I cannot be certain at this moment that the reboot command actually completed, or froze partway through
Lighttpd anti-hotlinking for images i just want these domain to link my imgages (test1.com,newtest2.cn,800keke.net,800org.com.cn),the other site will be redirect to [url].
lighttp configuration :
Code:
This configure onle effect to test1.com. no effect to (newtest3.cn,800keke.net,800org.com.cn. i still dont know why..
I'm trying to configure lighttpd to send SCGI requests to different ports, depending on what file(s) are accessed. Is this possible? This is what I've tried, and it hasn't worked.
I have been searching in the forum and google but still not lucky enough to figure out yetI have a lighttpd server runningbecause apache consume so much CPU and memory) andqmailtoaster (just setup).Here is the configuration in cgi modules:
I have followed all the steps as mentioned on documentation.except automake (I am not clear as where to make automake)
I logged in as Root on a Ubuntu system and then
Code: cd /opt svn checkout svn://svn.lighttpd.net/lighttpd/branches/lighttpd-1.4.x/ cd lighttpd-1.4.x ./autogen.sh ./configure make make install
After this what do I need to do to be able to start the lighttpd ?After this I did not found any script /etc/init.d/lighttpd. so what more has to be done ?
I ran across the above article, which described a DoS attack in which requests are sent very slowly to the Web server. I'm running lighttpd 1.4.28 on a Gentoo Linux server, and I'm wondering if there is anything I could do in preparation to defend against such an attack.
A bug report [url] seems to indicate that there was a patch in place already against this sort of attack, but I wanted to be sure that was the same thing and if there was anything else I needed to do.
This is what I did till now and all was installed with success: yum update wget [URL] yum install lighttpd chkconfig --levels 235 lighttpd on /etc/init.d/lighttpd start Must I configure something else too? if yes... what?
I've got this problem for a few weeks and I cannot figure out. I'm pulling my hair out. I have a server installed PHP, lighttpd and redis. Sometimes, I got the following messages in the error log of lighty: Code: 2010-09-24 13:57:33: (mod_fastcgi.c.3011) backend is overloaded; we'll disable it for 1 seconds and send the request to anoth er backend instead: reconnects: 0 load: 567 2010-09-24 13:57:33: (mod_fastcgi.c.3011) backend is overloaded; we'll disable it for 1 seconds and send the request to anoth er backend instead: reconnects: 0 load: 626 and:
I know that you can access and run any script of the web by wget:
Code:
wget mydomain.com/page.php
But this is literally accessing it externally through the web, i think that it is safer and faster to access the script internally. I am using lighttpd to host my php pages, and is there a way to do that? I have had some hosting experiences, the cronjobs on the hosts let u input:
I installed Apache server with Debian 5.0.2 Lenny. I am trying to write a script which would analysis web log files. I found the log files on /var/log/apache2. There is an access log file, `access.log`. My question is what configuration file determines the location and the name of the access log file. How can I change them? I used CustomLog in /etc/apache2/apache2.conf like below.LogFormat ": %h %l %u %t "%r" %>s %b" common CustomLog /home/test/my_log_file common Apache2 generated /home/test/my_log_file. But no logs were written in the file even after I run `/etc/init.d/apache2 restart`. Ichanged the log file location. It still didn't work. However, Apache2 still wrote logs in the file `/var/log/apache2/access.log`
I want to access files on my ubuntu server wireless. Is there a way I can do that? I'm sorry if this is a stupid question, but I'm kind of new at this whole server thing.
My Redhat Enterprise Linux 4 with 6x partitions (/, /boot,/home, /usr, /var, /tmp) of 6.0 GB IDE Hardisk was working quite fine. I decided to create LVM on /home and /var partitions but due to some errors occured and I delete the /home partitions. That's why partition table altered. I then delete 4,5,and 6th partitions (/home, /var, /tmp) partitions and now try to create one by one but following error is coming:-
[Code]....
The Super block could not be read or do not describe a clear ext2 file system. E2fsck b 8193 <device> I have tried following commands,but could not successful:- e2fsck -p /dev/hda7 (where hda7 was created but afterthat it was deleted) e2fsck -a /dev/hda7
i am using natty 11.04. can you please guide me how to clear recent history in dash under search files folders?also tell me how to customize dash application menu etc...
What they don't make clear on most manpages is that you can't simply usebzip2 thisfile.xxx that gives bzip2 free rein to include the original extension in the filename. You must use the -c tag (and better than that, the -c and -k tags, the latter of which leaves the original file untouched post-compression), then -if you wish- the level of compression to apply to a file (in the range of 1-9, with 1 being next-to-no compression and 9 the highest level of compression), followed by the filename/s to compress, then after a single space, a right-hand carat (">"), then another single space and the name you want to give the compressed bzip2 file, remembering to add the extension ".bzip2" (no quotes) to your filename.
Sounds pretty obvious, doesn't it? Well, I thought so too once I read a particular manual page and reasoned it out. Then I remembered the 15 or so minutes it took me to Google myself to that manpage, and I decided it was good Net citizenship to put up a little step-by-step like this one on an appropriate forum on which I was already a registered member.
It isn't just the C++ heads or the Python jockeys who need things broken down to their simplest bits for them from time to time. Newbies and near-newbies do, too. I took it as read that was the reason for this subforum, so here I went.
Running SunGard Banner software on RHEL 4.2 x86-32 bit Linux server Oracle Application 10.1.2.3 samba enabled. Users run processes/reports that are logged in a daily log file. In our daily job submission log files the user password shows up as clear text.The password shows up as $PSWD (sample from the logfile):
i've had my dedicated server for about 16 months now and my log files are huge and I host a few small sites but one of them is quite large and can get a log file (access.log and error.log) of 1-2gb a month so over a year can take up almost 30gb just for that site so I clean them out regularly but I'm wanting a more automated way to do thisi've been thinking of using the find command using something like this but it isn't workingCode:find /var/www -name "*.log" -type f -size +10M -exec echo > {} /;
My var partition is full. The main culprit is NOW the mail file. What is the command to empty the mail files?The mqueue and clientmqueue files are now low on usage but the mail file is responsible for filling up the var partition.
I have a dedicated server and I am having email issues etc (seemingly) because the /var directory is 97% full
I would like to know if it is safe to clear it and how to clear it (assuming it will not disrupt/kill server services to do so).
I have a 'Matrix' control panel so i can view the storage etc but it does not have an way of clearing the /Var directory.
I have Putty Access to root but do not know much about command line access.
I found a few threads but the information is not clear to me as there seems to be an assumption of (basic?) knowledge I don't yet have.
My linux support guru that usually does this kind of thing for me is away and not contactable and my server is grinding to a halt and unable to store/send email.
I have only a very basic understanding of command line but really need to get this sorted ASAP.
Is it possible and SAFE to delete files via FTP from the /var/cache/apt/archives?
i was install ubuntu server with desktop and when i try to open my ip adress in browse it show me this Forbidden
You don't have permission to access this file on this server. Cheyenne/2.2.8 Server at localhost Port 80.how can i fix it.My IP address is 88.163.109.209, you can see the directory forbidden error if you type in my IP address.
I have a file 'my_file.txt' stored on 'myserver1.col.edu' Now, I am using a different server 'myserver2.col.edu' to do some work and I want to access 'my_file.txt' on 'myserver1.col.edu' to read (possibly edit) WITHOUT physically copying the entire file across. Is there a way to do this - perhaps through ssh?
I am new to linux and know some basics, no proper knowledge of servers. I want to know is that how can i access my companies "fileserver" from linux (gui as well as from cmd line). This file server i can access easily from windows machine but i don't know how to open/access file server from linux machine..If can tell what tools are needed and also but is the process from command line.File server is windows server and my machines is Redhat linux.
I am trying to access log file which located in /etc/log/apache2. I could get into the directory using `su`. I was able to run ls command under the directory and everything was file. I could run a command,
ls -d /var/log/apache2/*
However after I switched to my account, I got an error. sudo ls -d /var/log/apache2/* ls: cannot access /var/log/apache2/*: No such file or directory
I want to use this command in a bash script to get a list of log files. Should I write the script as root and run it as root?
I'm running a Debian/Samba PDC on a Windows network. We desire to monitor a few network shares, so that whenever a change is made to a file in those, we have a log of who did it and when. Some capability of seeing what the change was, or a way of reverting it, would be nice luxuries. But username and timestamp are most important, if possible.
On the debian forums, someone advised using Tripwire for this purpose. didn't give much other advice about it so I kind of struck out on my own researching tripwire. got it installed, played wth it, and found it problematic, to say the least. It seems a bit much on the complexity and security side, and it seems more oriented towards maintaining system integrity, than monitoring documents. So far I've not managed to get a policy update working on account of a cascade of errors about a few hundred files under /proc disappearing, despite no changes to the system.so I'm wondering if anyone here has advice.
1. Does debian have this sort of functionality built in? is there a system log I can parse to get this information? 2. Is tripwire the right application for our purpose? 3. Is there anything better suited, more user-friendly or more parsimonious. I don't need something to monitor all system files, guard against intrusion, and make me cups of tea, just to monitor a few folders that I specify