Fedora :: Scan Url And Downloading Files
Feb 1, 2011I understand wget is used to download files. Is there a way I can search a url for what files are available for me to download. I need to install a plug-in from an adobe website.
View 2 RepliesI understand wget is used to download files. Is there a way I can search a url for what files are available for me to download. I need to install a plug-in from an adobe website.
View 2 RepliesI accidentally deleted configurations including GPodder's. Now I want GPodder to scan for thousands of media before redownloading them. I cannot find that. It shows them as undownloaded. Is there a way to tell it do that?
View 1 Replies View RelatedHow to determine what type of files clamav can scan? For example, if there is no unrar installed it can't scan files in it. So is there any way to find out all types of files that clamav can't scan?
View 2 Replies View RelatedI keep on downloading tar.gz files into my downloads folder and i cant do anything with them. what i need to do to install the file so i can use it? An example, i am trying to install Frets on Fire, and am failing bad.
View 9 Replies View RelatedI am vijaya, glad to meet you all via this forum and my question is I set a crontab for automatic downloading of files from internet by using wget but when I kept it for execution several process are running for the same at the back ground. My concern is to get only one copy, not many copies of the same file and not abled to find out where it's actually downloading.
View 1 Replies View RelatedI know everyone may discard this post because you do not need AV on linux BUT i am trying to use my linux PC as a download horse and would like to scan the files for viruses before I move them to my Windows machine. I downloaded and installed the latest AVG, how do I scan specific files, is it the avgscan command? I check /opt/avg/avg8/log for the scan results of a folder scanned and nothing. I tried avgscan -o -p ~/downloads where -o shows even clean files but after i run the command, there is nothing shown in the prompt and nothing in the log file.
View 4 Replies View Relatedscan my network and check for files within a LAN, in a mixed Linux and Windows environment.Something like LAN spy for Windows. Is there a package or program like that for Linux?
View 2 Replies View RelatedI had a hack on my oscommerce website recently. I have put in the relevant security patches but I need to check whether the hacker left any code changes in my files. What is a good file comparison software for linux? I need it to scan though the current files and folders and compare it the original default oscommerce installation so I can check the code.
View 14 Replies View RelatedI installed squeeze on my eeepc 1015ped and downloaded the correct firmware-brcm80211 drivers but every time I scan for my network using iwconfig wlan0 scan or wicd, my computer completely freezes. I previously had a solid install running xmonad, and wicd was working like a charm (using the same broadcom driver) but i tinkered too much with it and decided to do a fresh install. I haven't quite run into a problem like this before.
View 2 Replies View RelatedI know that there is little need for me to install an anti-virus etc - but - I was thinking, it is a good idea to scan folders and files that I send to colleagues that run windows.Whats the best way and programme to do this? I guess I simply install an AV programme and thats it!
View 5 Replies View RelatedI have a file which contains the data i retrieved through prstat and an array that contains all the unique process ID's of that particular file. i want to compare each and every line in the file with each and every element of the array so that i can create multiple files for the multiple value in the array.
View 6 Replies View RelatedI've been using PAN for quite some time - recently installed 0.133 from the Ubuntu software centre. Worked fine for a while, no issue. Then, a week or so, it started downloading .msg files in company with the binary files I was getting. Sometimes one .msg file for one binary, sometimes quite a few. It seems to have some correlation to the size of the binary, that is the larger the binary, the greater the number of .msg files were downloaded.This morning, it would ONLY download .msg files. I could see the decoded binary in the PAN viewer pane, but it isn't present on my system. I have made NO changes to any configuration files, other than installing the recommended updates.
How do I correct this?Can anyone tell me what these .msg files are, and how to stop them from downloading?Are there as-good or better newsreaders out there that I can/should try?
How to set up a script to log on to a server through SSH, copy file from the server to the local machine, and then run a script on the downloaded file. More specifically, I've got a minecraft sever that is run on ubuntu.
I know that I can do
# ssh username@ipaddress
to log on to the server through the terminal.
After this it asks me for a password. Once I type in the password I have access to the directories on the server. How can I set up a script to log on to the server and enter password? Once I do this, how do I automate it to copy a file from that server to ~/Desktop? If I can do this, I have a script that will run from there.
I've learned that I can do
scp -r remoteuser@remotebox:/remote/directory /local/directory to copy files from a server to my local machine, but it still asks me for the server password. how do I make it so that the password is automatically entered?
I want users of my website to be able to download files from it.
When they request a download, I would like them to see a window with the following attributes: ability to choose where downloaded file goes; progress of the download, and estimate of remaining time; a button to cancel the download.
There are hundreds of sites out there that do this, so obviously it's possible, but can it be done from a php script on the server? If not, does anyone know of a script in java(/script)?
I would also like to monitor when a download failed and succeeded, and if the failure was not because the cancel button was pressed, pickup whatever system information is available about the failure cause.
I ran a chkrootkit scan and found this: The following suspicious files and directories were found: /usr/lib/pymodules/python2.6/.path /usr/lib/xulrunner-1.9.2.8/.autoreg /usr/lib/firefox 3.6.8/.autoreg /usr/lib/jvm/.java-6-openjdk.jinfo
How do I get rid of this suspicious file?
I have some scripts that need to use a newer version of PHP, Im running Debian 6 which has PHP 5.3.3 support, I found I could install php 5.4 using [URL]. This worked, it updated my PHP to a newer version, the only issue is that when the install completed apache now downloads the PHP file instead of rendering it.
Im guessing this has something to do with the Apache configs, but I don't know what to do.
Code:
Select alltom@vps:~$ dpkg --list |grep -E '(apache)|(php5-)'
ii apache2 2.2.16-6+squeeze12 Apache
HTTP Server metapackage
ii apache2-doc 2.2.16-6+squeeze12 Apache
[Code] ...
I was thinking of migrating my apt-mirror repository to the recommended ftp scrips: [URL] .....
I pre populated my pool with already downloaded files, and setup the scripts.
However, if I run the bin/ftpsync, and monitor rsync with lsof -p, I can see that it is still downloading files from oldstable (wheezy) despite exclude options.
I'm guessing it's a configuration error, but I can't seem to figure it out. Any thoughts? My etc/ftpsync.conf is as follows:
Code: Select allMIRRORNAME=`hostname -f`
TO="/server_storage/srv/mirrors/debian"
RSYNC_PATH="debian"
RSYNC_HOST=ftp.us.debian.org
LOGDIR="${BASEDIR}/log"
[Code] ....
Actually, I don't think it works like I thought it did. A few guides I found listed the exclude options, but the sample config file has this:
Code: Select all## If you do want to exclude files from the mirror run, put --exclude statements here.
## See rsync(1) for the exact syntax, these are passed to rsync as written here.
## DO NOT TRY TO EXCLUDE ARCHITECTURES OR SUITES WITH THIS, IT WILL NOT WORK!
#EXCLUDE=""
So it looks like it doesn't exclude the suites at all.
I need to download some very large files (circa 75 GB) from a remote server via SFTP. I've been using SFTP via the command line on my Linux netbook. Around halfway through, the transfer stops and says "stalled." Can anybody recommend a reliable way to download these files?
View 1 Replies View RelatedWhen I try to load a php file my browser downloads the file instead of displaying the page. Here is my apache conf file.
#
# Based upon the NCSA server configuration files originally by Rob McCool.
# This is the main Apache server configuration file. It contains the configuration directives that give the server its instructions.
# See [URL] for detailed information about the directives.
# Do NOT simply read the instructions in here without understanding what they do. They're here only as hints or reminders. If you are unsure consult the online docs. You have been warned.
# The configuration directives are grouped into three basic sections:
# 1. Directives that control the operation of the Apache server process as a whole (the 'global environment').
# 2. Directives that define the parameters of the 'main' or 'default' server, which responds to requests that aren't handled by a virtual host.
# These directives also provide default values for the settings of all virtual hosts.
# 3. Settings for virtual hosts, which allow Web requests to be sent to different IP addresses or hostnames and have them handled by the same Apache server process.
# Configuration and logfile names: If the filenames you specify for many of the server's control files begin with "/" (or "drive:/" for Win32), the server will use that explicit path. If the filenames do *not* begin with "/", the value of ServerRoot is prepended -- so "/var/log/apache2/foo.log" with ServerRoot set to "" will be interpreted by the
# server as "//var/log/apache2/foo.log".
#
How do I download all the files form here: [URL]. I am on freeBSD 7.0 and I tried wget with the -r switch and it gives me URL's only. Maybe this is simply not an ftp site I don't know. How I can download all those files with the same directory structure.
View 5 Replies View RelatedIs there an easy way to get all of the multilib files from [URL]? I've tried:
1.) wget - won't use wildcards on an http server, and trying to get the whole folder just gets me index.html
2.) ftp - it's not an ftp server and I can't login anonymously via gftp. I wasn't expecting this to work, it was just something else to try.
3.) rsync - If this could work, I may not be using the right syntax. I tried (-n for a dry run first):
Code:
rsync -avn [URL]
And it just sat there doing nothing until I hit ctl-c.
Obviously I could download each file from my web browser but I figured there had to be a more elegant Unix-y way without all the clicky-clicky. Are the files hosted on an ftp server anywhere?
I'm trying to download Star Wars the Clone Wars torrents from two different feeds. The scantime below is temporary while I'm trying to get it to work. Will change it to 15 min later on.Below is the verbose output.
Quote:
INFO --- RSSDler 0.4.2
DEBUG writing daemonInfo
INFO [Waking up] Wed Oct 27 18:12:11 2010
DEBUG checking working dir, maybe changing dir
[code]....
I currently mirror the updates repository to my computer using rsync. I was wondering if I could save space and bandwidth by only rsyncing the .delta.rpm files? Are there any disadvantages to this or does zypper/YAST handle updates just fine with it?
View 3 Replies View RelatedI have an Ubuntu 9.10 server set up at my house. I have Apache2 and PHP5 installed on it. Every time I go to the server on a web page and try to load the PHP index page it downloads instead of displaying.
I have virtual servers set up and have the files stored at /home/cusinndzl. If anyone needs to take a look I can let them into the webmin panel.
Ubuntu Lucid 32-bit desktop machine. I run "aptitude update" and "aptitude safe-upgrade" manually from the command-line regularly to upgrade my machine. I usually don't pay that much attention to the downloading, but I noticed yesterday that aptitude appeared to be downloading four files simultaneously, rather than the one at a time that I've always seen. Would this be correct? How can I change that? I can't say I've ever noticed it before.
View 3 Replies View RelatedI want to limit bandwidth for downloading files with squid. I want to reserve bandwidth for other traffic (esp. web browsing). I know about delay pools but I don't understand well. Some users use download managers to download large movie files. I don't want to block downloading but I want to give them limited bandwidth. May be 5KBps or 6KBps because I have only 512Kbps (64KBps) connection.
View 1 Replies View RelatedI am looking for a file sharing program to install on my dedicated server that will allow me to upload large MP3 files and allow my clients to download them. these files are recordings of counseling sessions for families who are seeking help for their children.
What I am looking for is similar to the system this company uses [URL].
New apache install, php files are downloading rather than displaying, does anyone know whats causing this. When I compiled php with apxs it automatically added this line:
Code:
LoadModule php5_module modules/libphp5.so
I restarted apache, and also try adding this line:
Code:
AddType application/x-httpd-php5 .php5 .php4 .php
and then restarting.
For my project, I'm interested in the Scatterometer Products of Oceansat 2 from an Indian page [URL]
It's no problem for students to get a password to access and download their data for free. Nevertheless it's quite complicated to download the files by hand, since you have to mark every file by hand and click on a download button at the end of the page. When I tried it with my Script (which is below), and an internal server error 500 occured. I hope you're not too busy and could have a look on the script where the cookie and IP are entered manually for trial purposes.
The construction of the page is:
The adress, where you have to login:
http://218.248.0.134:8080/OCMWebSCAT...controller.jsp
The adress, when your're logged in:
http://218.248.0.134:8080/OCMWebSCAT...controller.jsp
[Code]....
I put a few Itunes songs into my music folder and tried to play them. It didn't. Then, I got to reading in the forums and noticed that there was a "one click" configuration for Gnome users. After doing some more reading, I saw a lot of references to restricted codecs and things of that nature. After beginning the one-click process and then aborting midway through installation, three questions come to mind:
1) Why all the warnings about legalities of downloading the codecs/files? If I bought the music and I use the codecs to listen to it, where does legalities come into play?
2) While installing some of the different files, etc in the one-click process, I received a few warnings that a particular file was not from a trusted source (I don't remember the info verbatim) and then it gave me an email address, presumably from the developer, and asked if I wanted to install it anyway.
3) If I do go through the one-click process, will I be able to listen to I-tunes or am I pretty much screwed on I-tunes on linux?