I have executed the command
Code:
sudo wget -r -Nc -mk [URL]
(referring the site : [URL]) for downloading entire website using wget for offline viewing on Linux,
At the middle of download I have shutdown my laptop (downloading was not over) and when I started my laptop again. I have executed the same command to continue downloading but I have got the error :
Code:
test@test-laptop:/data/Applications/sites/googlejam$ sudo wget -r -Nc -mk [URL]
--2010-07-25 19:41:32-- [URL]
Resolving code.google.com... 209.85.227.100, 209.85.227.101, 209.85.227.102, ...
Connecting to code.google.com|209.85.227.100|:80... connected.
HTTP request sent, awaiting response... 405 Method Not Allowed
2010-07-25 19:41:33 ERROR 405: Method Not Allowed.
Converted 0 files in 0 seconds.
test@test-laptop:/data/Applications/sites/googlejam$
I installed and configured a squid transparent proxy on my linux os at work. Also it is veryslow but every thing is ok while I do not try to use port 443,so when I try to use sites like mail.yahoo.com or other which are using https(443) port and the method used is CONNECT I see some errors in access.log like:
I cant update php in my RHEL Server 5.1 using yum. I am using RHEL Server 5.1 and I cant use yum to upgrade my php. Other site told me that use wget to solve this problem. How to use wget to upgrade php? This is my first time to handle linux server..
While reading some papers on securing apache with selinux, I have tried to bind httpd to port 3000 expecting to be blocked by the selinux, since port tcp 3000 isn't on the http_port_t list. However I was able to start the service...
I'm preety sure selinux is enforcing. Also, if I bind httpd to tcp 81 selinux denies the start of the service, as expected!Did I miss something? Why is httpd allowed to start binded to a port that's not explicitly allowed?
I have a link to a pdf file, and I want to use wget (or python) to download the file. If I type the address into Firefox, a dialog box pops up asking if I want to open or save the pdf file. If I give the same address to wget, I receive a 404 error. The wget result is below. Can anyone suggest how to use wget to save this file?
If a wget download is interrupted (like if I have to shutdown prematurely), I get a wget.log with the partial download. How can I later resume the download using the data in wget.log? I have searched high and low (including wget manual) and cannot find how to do this. Is it so obvious that I did not see it? The wget -c option with the wget.log as the argument to the -c option does not work. What I do do is open the wget.log and copy the URL and then paste it into the command line and do another wget. This works but the download is started from the beginning, which means nothing in the wget.log is used.
I have recently decided to venture into online blogging and other things, and my friend told me wordpress is a great tool for this. I have set up my mysql database, and apache and php are all working fine, and when i copy my wordpress folder into my /var/www/, and edit the config file for my database, i point my browser to[URL] to get it up and running. But i only get the error: "Fatal error: Allowed memory size of 16777216 bytes exhausted (tried to allocate 7680 bytes) in /var/www/wordpress/wp-admin/includes/plugin.php on line 302". Some googleing seems to suggest that this may be a problem with a php.ini file, but im assuming this is a windows based solution, and im not sure where/what a linux alternative is.
Just an FYI for anyone who may be having this particular problem. A short while back, I was trying to attach a picture to a Twitter post, and dropped my network connection. No big deal...connection came back, and things went on. Next time I launched Choqok, it popped up a message saying "Server Error: This method requires authentication". It was puzzling, and didn't appear to impact my use...until I went to send a direct message, and it would give that error and crash. After quitting Choqok, the file (/tmp/ksocket-user/klauncherXXXXXX.slave-socket file) was still present. Deleting that file manually cleared the error up. I've seen this mentioned in a couple of other forums, but none with a solution posted.
If anyone else has that error, and this method resolves it, please let us know. I'm using openSUSE 11.4, but it should apply to any version/distro of Linux using Choqok.
I am installed cUrl from http://packages.ubuntu.com/lucid/amd64/curl/download for my ubuntu 9.0 with php 5.2 ,Apache 2.0.It was installed by gkebi package installer. but when i display phpinfo() curl not found since i was attempt to use phptwitter class. what is the proble i have? When i am trying to install again package installer shows same version of curl already found, but phptwitterclass shows following error 'Method "curlQuery" of class "Twitter" reported an error: CURL library not installed.'
I recently had to move to a new machine, everything went well except for one thing. I did fresh installation of LAMP server all with default configs. Every time I'm using PHP script to that invokes include, require or require_once I get the following error:
Code: Fatal error: Allowed memory size of 20971520 bytes exhausted (tried to allocate 7680 bytes) in /var/www/index.php on line 2 index.php file: PHP Code: <?php include "index.php";?> icukapi.php file: PHP Code: <?php echo "test";?>
My memory_limit in php.ini is set to 20M. I tried to increase that however it didn't quite work. PHP seems to allocate all possible space and return that message every time i try. If somebody has an idea of how to fix it I would be more than grateful. I spend quite a long time searching for an answer however the all things i found suggested increating memory_limit which in this case doesn't work.
Apache is run as www as is all the files/folders. People are uploading via FTP, scp, so the problem is if I chmod so everyone can read, then rsync as a user it works until new files are added which then my ; if rsync fails with a permission denied. Now I can add a chmod in the script so everyone can read, but since www can already read, I figured I would just change my script to use www. I added the ssh key to his authorized_keys file, but when I try to just ssh in I see this in the secure file;
server sshd[29539]: User www not allowed because account is locked sshd[29539]: Failed none for invalid user www from ip port 54983 ssh2
Now I read a few places already saying I need to add a password to the account, etc. but before I jump and try all I read, 1st major one, will this now break apache? Will this affect any startup things, etc. and .... will that unlock that user for ssh in or is there another preferred method?
I have a problem when I want to use su I get this error:Code:su: pam_start: error 26I have googled it so I found this topic (http://www.linuxquestions.org/questi...r-26-a-615024/) but it didn't really help me. There was a reply on that topic and his question was what the output of this was:
There is a symlink from /var/www to a personal directory. FollowSymlink and chmod 755 are all set. It works perfectly until each morning I will get a "Symbolic link not allowed or link target not accessible" error. When I do a "sudo service apache2 restart", the problem will go away.
I have used Dump Command to dump the application files. For Full backup the level 0 is working fine. For incremental backup I used the level 1 or 2 it is getting the error as
DUMP: Only level 0 dumps are allowed on a subdirectory DUMP: The ENTIRE dump is aborted.
The code I used =============================== #!/bin/bash #Full Day Backup Script #application folders backup #test is the username now=$(date +"%d-%m-%Y") [Code]...
when i wget aro2220.com it displays --2011-03-02 16:35:58-- url... 127.0.1.1 Connecting to aro2220.com|127.0.1.1|:80... connected. HTTP request sent, awaiting response... 200 OK Length: 177 [text/html] Saving to: 'index.html' 100% 2011-03-02 16:35:58 (12.4 MB/s) - 'index.html' saved [177/177]
However, when I look into the file it is actually blank saying something like "It works! This is the default web page for this server" which can't be correct since that is not what Aro2220.com actually displays.
Furthermore, when I try to wget files I've put on the server for myself it returns a 404, file not found.
trying to update a 11.1 installation. As I start the installation I get a message "openSuse 11.0 partition /dev/sda5 Change the mount-by to any other method for all partitions." Also the message said go back and reboot and make the change. What is this all about, how do I do this.
for some reason i am not allowed to change the permissions on my hhd i have a 3G partitioned off of it but i cant acess the other 290G what i can do to get permission
Ubuntu does not allow me to save in /home. It says: You don't have the rights to save the file. Check if you entered the location correctly and try again./home is an 850 GB partition which I mounted there at the installation of Ubuntu. It was meant to save personal data on, but now I can't save anything on it.I can only open what's on it, but I can't save anything
I'm using a local proxy server VPN'd to another network.
How do I setup either Firestarter or Gufw/ufw to ONLY allow in/out from ONE port? (The one port the proxy uses)
Ex: Firefox is proxied to 127.0.0.1, all ports, and then the proxy picks it up, and sends out on port xxxx, and recieves on port xxxx, then sends back thru 127.0.0.1, back to Firefox.
Any setting/rules I've treid on either Firestarter or Gufw kills the proxy>VPN (Proxy won't connect to remote network)
Addendum: If I start the proxy FIRST, then the firewall, all is good. I'm thinking the proxy uses a port to connect with remote network first, then switches to my configured xxxx port...hmmm
I am just wondering why this is happening. I just tried to use a tutorial to customize my GUI and give myself a custom launcher for Libre Office, and when I tried to copy the edited .desktop file back into my "/adam/usr/shared/applications" folder, it told me I didn't have permission to do this. I tried to edit the folder permissions and it told me that I was not able to because I am not the owner. Is there any way to fix this or do I have to re-install Ubuntu?
I want to use MYSQL administrator remotely, if I try I just get Host 192.168.1.216 is not allowed to connect to this mysql server. All the fixes I see on the Internet are for specific databases, but I want to access everything?