I have a web server in my kitchen with apache running on it. Since the upload speed is quite low due to my isp I would like to execute a bash script that uploads a file to another server through a website (which is htaccess protected) The idea in general: Someone with access to my website browses through a folder, copies a file path to an input form and presses "upload". Rather than executing a bash script directly I could have a cron job running in background that finds the path and then uploads the file to the other server I have userspace on and is accessible via sftp/ssh. The file would be than erased later after a couple of days or so. That person would be able to access the file with higher speed some time later without logging in via ssh and doing all that manually.
When I try execute a variety of basic commands (including ps, ls, ifconfig, locate), I receive the error 'no such file or directory'.
Here are some suggestions that I've found online, that I have tried without success: I did a 'whereis ps' and found the file in the /bin/, and have checked that '/bin' appears when I do 'echo $PATH' I did a filesystem check which showed my hard drive as being clean I tried doing a 'sudo chmod 777 ps' but was told that I dont have permission. I don't think permissions for these files would have changed though (and I can't check as I can't run the 'ls' command).
I am trying to configure Apache on my server to work with ssl, but everytime I visit my site, I get the following message in my browser: SSL connection error. Unable to make a secure connection to the server. This may be a problem with the server, or it may be requiring a client authentication certificate that you don't have.
Error 107 (net::ERR_SSL_PROTOCOL_ERROR): SSL protocol error. Just some background on the situation: I am using Ubuntu 10.04 desktop edition. I installed apache by installing zend server (it installed apache automatically). I then installed openssl. Non-https pages work fine on the site. I tried getting trial certificates from multiple certificate sites but nothing is working (same error). I was previously hosting my site on another server on which ssl worked just fine. I also tried using the key and crt file from that server, but I got the same error. The domain name and IP are still the same though. My SSLCertificateFile and SSLCertificateKeyFile are pointing to the correct directory and files.
I'm using ubuntu 9.04, I want to design webpages in PHP for that I installed apache 2 on my machine. I'm using my mobile to connect to the internet when ever the mobile is connected and the wired connection established the server running properly. But when the wired connection disconnected my browser page didn't show the localhost home. How can I configure the apache on my machine.
I installed apache on my local machine to test my sites. I have one problem:I created a new site configuration in /etc/apache2/sites-available and enabled it. The problem is that cgi is not working, it just prints the content of the files. (The default location for cgi, /usr/lib/cgi-bin works without problem.)Here is the content of the site configuration file:
I have Apache Server working online under Fedora Core 6. But before I installed and configured everything, I've been testing in Fedora 12. The problem surges here, when I start the httpd service, every supuse 404 action on a web browser, takes me to localhost. I mean, if I enter google.com, no error, just goes to localhost, http://asdasd, no error, gives back localhost. I used to ignore the problem 'cause I thought it was a problem on my apache, but when I installed the Server on the Fedora Core 6 machine, I found that I have the same problem there. Of course, it only occurs when I am browsing through the same machine that has httpd started. Does anyone know how to change that??
I am very new to linux. The first time i ran a linux machine was one Saturday. Anyway I am trying to set up an apache web server, all I want to do is play around with html and post it on my unbuntu server so it is available to the internet. I am also trying to install samba but I am having trouble with that so for now I am using winscp. Ok, so I made a folder in my home directory for webstuff, and set up apache to look in that folder.
mkdir /home/username/webstuff I put my index.html file into the /webstuff folder. But when I go to my website, it says 403 error unable to access "/"
Did I make the folder in the right place? I do not want people being able to access my / folder so maby I made the web stuff folder in the wrong place? Also I thought I would just put my inded.html file inside of the default one apache gives you, but when I tried to transfer index.html it said permission denied. How to set up permissions so I can use apache and transfer my html files from my remote desktop to my server would be great!
how can I run CGI scripts that are without extensions under the document root?i can not use ScriptAlias on the document root, i can run .pl scripts but i cant run any script without extension...ofcourse all files are executable on the fs.
Do I have to do a basic configuration / initialization of mysql, like described in the following link, before I can use it? [URL] The background is, that I want to use mythtv and therefore I need a working mysql. When I tried to create the needed database, as described in the file README.SLACKWARE (included in mythtv.Slackbuild package from slackbuilds.org) with the following command, I got an error: # mysql -u root -p < /usr/share/mythtv/database/mc.sql
The error message was: ERROR 1146 (42S02) at line 3: Table 'mysql.servers' doesn't exist -- integrale
I don't remember that I ever did a basic configuration but mysql was already used by joomla, and it worked right out of the box.
I know this is a bit out in left field but my gf's dad has created a fairly large application in Excel 2003 and now we need to run it online. I setup mono and xsp2 on my Ubuntu server in order to test but have no idea how to convert a monster .xls file in VB to ASP or maybe even PHP (one can hope, right??)
Most of my googling odysseies lead me to articles which talk about compiling apache & php from the source as this is touted to be more flexible. But the problem we have is this...we already have an apache 2 web server running. Now how do we serve php pages of it.
I recently took my Compaq Presario 2175US notebook completely apart and resoldered the power jack so that it could be used again (after 2 years of dormancy). After successfully putting it back together (with a handful of leftover screws ) , I was able to successfully install Ubuntu 10.04 (Windows XP seemed silly at this point and I really like using Open Office). Anyway, after a couple of days looking for the wireless driver for the internal wireless, I found the answer and got that working.
Unfortunately, the video does not seem to be working other than a basic VESA configuration. The notebook has the ATI Radeon IGP 320M video chip installed. I have searched for many days (about a week) on the internet for a way to get this working, but have been unsuccessful. Although I have found many posts all over the net (here included) that list folks with similar problems, and having tried many of the "fixes", I am still unsuccessful at getting the video card to work in any more than a minimal manner (no games, no 3D of any kind...etc.)
I'm using Ubuntu server 9.10 with 2 NICS (Internet-router-eth0, eth1-LAN). I use iptables to generate rules for 20 computers, but when I execute the script, ALL TRAFFIC DROPS, including the server. What am I doing wrong?
Code: #!/bin/sh #eth0 192.168.0.50 - connected to Internet #eth1 192.168.1.51 - connected to LAN #192.168.1.52 - workstation1 #set default policies iptables -P INPUT DROP
iptables -A FORWARD -i eth1 -o eth0 -p tcp --dport 80 -s 192.168.1.52 -j ACCEPT. The reason I'm doing this is, I just want to open necessary ports in the server and restrict LAN usage.
I am trying to set up a LAN with some basic file and printer sharing, as well as either a VPN or FTP server. I am setting up a server for an office with about 4 computers and one remote computer at a home office. I would like the server to back up a selected group of folders on all of the computers as well as share some printers between them all as well. If all of the other computers are windows computers will Ubuntu work for something like this? I would also like to host a website and e-mail server as well if the server is not loaded down too much. I know I can get windows server to work, but I would like to use something else if possible.
I've been running Samba for a couple of years now. Bumped up from 9.04 to 10.10 server (64 bit), and nothing wants to work. I have a really simple share that I want to be accessible by all Windows clients without a login (guest access). Here is my config:
I have a nagios server monitoring around a 1000 hosts in our office. To make things easier, i've integrated apache into Active Directory for the website authentication. This works perfectly, however the Nagios website uses alot of CGI scripts to display information I get the message below when trying to show CGI scripts with an AD user. Local users work fine. 403 Forbidden You don't have permission to access /nagios/cgi-bin/status.cgi on this server. I tried adding the AD users to the cgi.conf file within Nagios (which works for local accounts) but it doesn't seem to work even with this change.
I have finished reconfiguring my 1st PC Centos 5.5 with Squid and Dansguarian. Works fine the configuration is basic.lets name it to 1st PCeth0 - modem ip (public)eth1 - 172.16.1.1 (LAN)Then I've decided to change and add another Centos 5.5 for Firewall Service that is directly connected to the modem and then shares internet connection and some ports to 1st PC.
lets name it to 2nd PC, it also has two NICSeth0 - modem ip (public)eth1 - 10.0.0.1 (ip to connect to the 1st pc with squid services)so I changed the ip of the 1st PC with theseeth0 - 10.0.0.2eth1 - 172.16.1.1 (LAN)and connects the eth0 (10.0.0.2) of the 1st PC to the eth2 (10.0.0.1) of the 2nd PCthen I've tried browsing on the network but it does not work...is there a simple iptables command for the 1st PC and 2nd PC?ere are the iptables script and lists of my 1st and 2nd PC 2nd PC (firewall to be)
I'm trying to get a VB program to make a client connection to my PostgreSQL server running on an Ubuntu 10.10 server. Here's what I've done: Client side - installed and registered the OLE DB .dll's from the PostgreSQL OLE DB Provider project. server side - the configuration described here: [URL]... Have added this line to postgresql.conf:
listen_addresses = '*, 220.127.116.11, localhost' I've also tried this as simply: listen_addresses = '18.104.22.168, localhost' also, port = 5432 I have added the following to pg_hba.conf: # IPv4 local connections:
My question is I trying to install php5 and add mcrypt,, but it is unable to run "aclocal" command. It show this message.
I have check with the path of /etc/profile, I'm sure that I have added the path as below. Beside that, before "aclocal" command I able to run another command which is "phpize". It is able to run successfully, after I added the path as below...
I have Webmin installed on an Ubuntu server. I currently have a successful apache server running on port 80, however I want to create a virtual host on port 81. When I try I go to servers->Apache Webserver-> Create Virtual Host I change the port to 81 and the document root to /var/port81www then I click create. How ever when I goto 192.168.1.5:81 (local ip, I know I have to port forward but its not even working local) it does not work.
I have a Dell PowerEdge 2850 running Ubuntu 10.04 Server with SSH and Samba. My problem is that I am unable to execute my adduser.sh script which reads from a text file and adds users to the box and Samba. I have ran chmod a+x to make it executable and placed it in /usr/local/bin.
When I run sudo adduser.sh I get "sudo: unable to execute /usr/local/bin/adduser.sh: No such file or directory".
When I run adduser.sh I get "-bash: /usr/local/bin/adduser.sh: /bin/bash^M: bad interpreter: No such file or directory " I have been using Kubuntu as my home workstation for some time now but I have managed Windows servers but since I was given the freedom to setup this server I chose Linux.
how to bind a script to a F key (F12) that will run as root even when not logged in. I have a headless server on client premises where it'd be easier for them to press F12 to run this script that will be rarely needed than to give them SSH instructions etc. I know this must be do-able, but I can't get my Google-fu on for this question. The only way that I can possibly think of doing it is to touch a file whenever that key is pressed and have the script idly checking for that file every few seconds in a loop.
I've recently been trying to restore a Debian installation back to it's previous state after a serious operation system crash. Efforts are largely going well however I've run into problems since reinstalling Wine. Previously I could execute a Windows executable by simply entering ./executable.exe in the bash shell but this no longer seems to work as now I have to include the "wine" command in front of the executable. how to configure Debian to automatically use Wine to execute using the ./executable.exe format?
I have been having trouble setting up a daily backup script with cron. It would basically never worked. Searched the net for answers but didn't find anything. I finally figured it out !! When root crontab is edited the execute flag is removed from #/var/spool/cron/crontabs/root. I change it with #chmod a+x /var/spool/cron/crontabs/root and all is good.