I am using endian firewall 2.4 and squid 2.6, everything is working fine for me. In squid report i am getting download list by user or IP based. Now my question is, is this possible to get user based uploaded file using HTTP POST. Is there any configuration have to be done is squid.conf, if answer is yes then please give me the what is configuration i have to give. Example: if suppose user uploading a image file into facebook.com, i want to track the file name and which IP address and then where its uploaded.
I've got a bachelors in Computer Information Systems from Metro State in MN, but sadly, we didn't spend much time familiarizing ourselves with UNIX based OSs. I'm sure I'll have plenty more questions to pose later as well. I'll pay it back with help for anyone who's working on an evil Nortel router.
I believe I've got VSFTP set up and configured (mostly) the way I want. Here is the config file:
I am trying to read a file uploaded by a simple <input type="file"> form and directly write it to a mysql blob - without saving it to the filesystem. I tryed something like:
Code:
It writes some few Bytes to the DB and there is no error, but it's not the actual file that is being written.
I'm having a very strange problem. I use lampp on centOS linux. My application has a upload Script (in PHP) which uploads file to file system. after upload and moving file to correct location, uploaded files are getting deleted. I check file upload and moving of file by putting all the status in a text log file. files are getting uploaded properly and after upload I'm able to move the files to it's correct location.
Alright running a ubuntu based webserver. The app will be accepting user uploaded files from my client's clients. My client will then need to download an access the files. I'm looking for a solution to scan for windows malware at the time of upload so I never expose her machine directly to her client's uploads.
I have an rsync server and am now setting up cwrsync on my windows machine. I want to be able to run cwrsync over ssh with public private keys. I followed a tutorial over here to set this up. It, however, still prompts me for server password and works only if I provide the password. For some reason the public/private key process is not working.
- I generated a key using the command: ssh-keygen -t rsa -N '' (I verified the key gets generated on my Windows machine)
- I uploaded the generated file id_rsa.pub to server /root/.ssh/authorized_keys
I am also prompted for a password if from command prompt I run this command to log into server: ssh -i c:docume~1user.sshid_rsa root@<server_ip_address> On server I have changed the configuration file (/etc/ssh/ssh_config) to say:
RSAAuthentication yes PubkeyAuthentication yes
I then restarted the sshd service, however, to no avail.
im using a lease line on real ip for internet connection in my office.i have shared my internet conetion with a squid 2.6 stable6. i m having almost 200 pc in my office.since last few days my squid is creating a problem.as soon as i restart it ,it work fine but after a few minute it becomes extremely slow ..it almost dies.when i go to its cache log i found the error that your cache is running out of file descriptor...once i increased the no of file descrptor from 1024 to 4028 , the problem temporaily sorted out but after a few days same problem is repetaed ....and still the problem exist
I am using squid to controlling access to the internet all is working fine expect one of the user who is using outside organization portal to connect internet. But whenever he tries to enter in the portal by typing (EXAMPLE)url. Permission denied error from squid occur.
How can i allow this portal in squid. So squid will allow this to access.
I have a vsftp server running, and we have problems sporadically where people (not the person who uploaded the file) cannot delete files uploaded from OSX machines. We never have this problem with files uploaded by PCs. Anyone ever heard of this or know what I should check? The users uploading the files all belong to the same group. Also, I think if you log in to the FTP server with another OSX machine, it will let you delete them. I'm looking at the folders now, they have drwxrwxr-x. When i try to delete them from a PC, it gives me a 550 Remove directory operation failed.
I have generated a list of directories that I would like to use ls and grep on, but it is not working. I am using the commandCode:cat directories.dat | xargs lsand I get a whole lot of these errors:Code:ls: cannot access ./foo/bar/baz/grault/*: No such file or directorybut when I try the directories manually one at a time I find that they all exist and all have files in them. Same thing if I try to grep anything. What is going wrong?
I have a bunch of tar files in a folder with the name in 'file_name.MMDDYYYY.HH.mm.SS.tar' format. I need to write a shell script that picks latest and greatest of the above tar files and extracts it to a particular folder
I've published a samba share on a Ubuntu 8.10 server, and mounted the share on a windows box. For some strange reason the files are not sorted by file name (nor by any other sorting type) if I list them from the Windows command line (dir). I could always use the "dir /on" parameter, but I hope there's a way to make samba sort the files for me
I am trying to write a script to pick the directory name from a list of file. Here is a detailed picture.Have a file name LIST which contains the follwing for example/apps/oracle/product/test1/apps/oracle/product/test2/apps/oracle/product/test3I need a script that reads these line from LIST and creates foldersin /apps/oracle/product/test1/backup/date/test1 after reading the first line /backup/date/test2 after readin the second line/backup/date/test3 and so on.
I have a system that will not boot as /usr has been destroyed and I would like to get a list of installed packages before re-installing. I know that it's possible to get this using dpkg or apt, but I cannot run those.
Where in the filesystem is this information stored and what's the best way to get a list of installed apps from the files?
With file command we can get the file type. File returns a unique answer for each file type. I want to get the total list that file can return for linux & solaris. For example, I have a partial list of file's return output. Unfortunately I can't see the total list from manual page.
ASCII C program text ASCII C++ program text ASCII C program text, with CRLF ASCII C program text, with very ASCII English text ASCII English text, with CRLF line
I have a text file with a long list of RPM's. I need to check if each RPM is installed. I'm sure I can cat out this file and run "rpm -qa" against it, but I'm having trouble with the syntax right now...
It's always a good to backup a configuration file like sources.list before you edit it. To do so, issue the following command: sudo cp /etc/apt/sources.list /etc/apt/sources.list.backup Where does it backup to and how do I access it? I want to put the backup on removable disk or upload it
i downloaded extracted obsidian.pl to .xchat2. loaded the plugin in xchat. now how do i build the file list? are the files genarated from root_dir? is it the build_file_list? varibles from the obsidian.pl file ive been trying to find some kind of setup file or webpage for obsidian fserver but to my avail there isnt much info on this plugin.
[Code]...
ok so i managed to get a list of files generated now it seems i cant acess my fserv when i type /ctcp <nick> <trigger> all i get is this "Received a CTCP BLOODYSCRIPT! from hraesvelg" nothing else. i run the the ctcp command from another computer in the house. it fails to respond with !list
How can I generate a list of files in a directory [e.g.: "/mnt/hdd/PUB/"] ordered by the files modification time? [in descending order, the oldest modified file is at the lists end] ls -A -lRt would be great:[URL] But if a file is changed in a directory it lists the full directory...so the pastebined link isn't good [i don't want a list ordered by "directories", I need a "per file" ordered list] OS: openwrt..[no perl -> not enough space for it :( + no "stat", or "file" command]
I received the following output from an rsync (3.0.0) command that was executed: sending incremental file list sent 77214 bytes received 484 bytes 155396.00 bytes/sec total size is 254531170 speedup is 3275.90 What does "sending incremental file list" mean?
I'm sure I've done this before and am having a brain fart.
I have run into this a couple of time this last month:
I have a list of torrent files (blahblahhexblah.torrent) saved in a text file. I would like to read the text file and populate the client directory.
I can echo the file using:
But I cannot remember how to pipe the output into the directory, nor can I find any resources that describe the process.
Interestingly enough, I've had a couple of opportunities to do similar tasks with passwords and privileges and even across the LAN in the last few days. Solving this problem will help me solve the others for next time.