General :: Where Is Time On Squid Log File
Apr 5, 2011I viewed a squid log file and I am confused on what the following 13 digits indicate..../var/log/dansguardian/access.log:"1301956620.901".
View 2 RepliesI viewed a squid log file and I am confused on what the following 13 digits indicate..../var/log/dansguardian/access.log:"1301956620.901".
View 2 RepliesI want to egrep or a grep in the squid log file .. from this time 1287483952.440 to '1287483452.440, how can I do it.
View 3 Replies View RelatedI am using endian firewall 2.4 and squid 2.6, everything is working fine for me. In squid report i am getting download list by user or IP based. Now my question is, is this possible to get user based uploaded file using HTTP POST. Is there any configuration have to be done is squid.conf, if answer is yes then please give me the what is configuration i have to give. Example: if suppose user uploading a image file into facebook.com, i want to track the file name and which IP address and then where its uploaded.
View 1 Replies View RelatedI am running squid 2.6 In the
Code:
#tail -f /var/log/squid/access.log
i can see
TCP_HIT/200
TCP_REFRESH_HIT/200
TCP_MISS/200
Kindly guide me about it what mean by TCP_HIT,TCP_REFRESH_HIT and TCP/MISS ?
im using a lease line on real ip for internet connection in my office.i have shared my internet conetion with a squid 2.6 stable6. i m having almost 200 pc in my office.since last few days my squid is creating a problem.as soon as i restart it ,it work fine but after a few minute it becomes extremely slow ..it almost dies.when i go to its cache log i found the error that your cache is running out of file descriptor...once i increased the no of file descrptor from 1024 to 4028 , the problem temporaily sorted out but after a few days same problem is repetaed ....and still the problem exist
View 3 Replies View RelatedIs there any web-tool that provides real time squid users utilization and websites access.
View 2 Replies View RelatedI am using squid to controlling access to the internet all is working fine expect one of the user who is using outside organization portal to connect internet. But whenever he tries to enter in the portal by typing (EXAMPLE)url. Permission denied error from squid occur.
How can i allow this portal in squid. So squid will allow this to access.
I need to get the modified date on a file in linux to use in a script.I tried using 'ls -l' on the file, but this caused problems when the date turned from a single digit into a double. The reason for the problem was because I was parsing the result string on spaces.How can I get the date of the last time a file was modified so I can use it in a script? For example, if a file was modified on 1/11/2010, I need the 11.
View 2 Replies View RelatedWe have implemented a Ubuntu with Squid (2.7) & Dansguardian, Webmin and the webmin modules for these services. Server was running fine for about 6 months then suddenly there are hundreds of TIME_WAIT's when using:
Code:
netstat -antup
This causes the clients to get this error on their browser:
Code:
commBind: Cannot bind socket FD 10 to *:0 (98) Address already in use
Even 1 single client working will fill 3 pages of TIME_WAIT
Clients are Mac's bound to OD & AD (magic triangle). We reverted the vm image to a snapshot back when was working fine but problem remains. We created a server from scratch and same issue. We use ldap_auth to the active directory for authentication:
Server is 2008 R2
I installed webmin on Debian apt-get install webmin. After installation following message came.
Code:
Setting up webmin (1.510-2) ...
Webmin install complete. You can now login to https://IP:10000/
as root with your root password, or as any user who can use sudo to run commands as root. So I infer that the installation was successful.
But when I do
Code:
http://IP:10000
I get following error
Code:
Error 111 (net::ERR_TUNNEL_CONNECTION_FAILED): Unknown error.
I need to goto a folder, open a script file (same file every time). The file has different function calls, I need to comment all other function calls and just un-comment the one I want to use.
#call_1
call_2
#call_3
[code]....
everytime if more that 10-15 people using internet at same time squid proxy slow to responde everyweb just keep loading and here my squid proxy settings
1 nic
squid proxy 2.7 stable 9
ubuntu 64bit
intel duo core
[Code]...
my squid show like this when i try create swap directory
[Code]....
The tail of my log is the following:
BUT:
So, how is it possible that some process writes to the file but file time modification still remains untouched?
I want to extract the file creation time. How to get it. I am using fedora core 4
View 3 Replies View RelatedI have a file called foo with a date string in it - and nothing else in the file
2010.01.13-22:28:28
I want to send this file content to date -s command so I can modify using a cron job.
I tried several ways to pipe the file contents to date command but did not work.
I have a log file that I would like to examine during some changes under process that writes to this log.
Is there some way to open this file and read in real time changes written to it ?
I want to search in my apache log, for events which have occurred say between 11:00 AM to 2:00 PM. I have got few scripts/commands but they are not conclusive, some of then are trying to do an exact match(awk) and for some i am just getting the pattern wrong (eGrep)
I am running RHEL.
Apache log eg:
================================================================
10.92.120.129 - - [24/Jun/2011:01:29:07 -0500] "GET /favicon.ico HTTP/1.1" 302
================================================================
now if i want to find out the logs from 1 PM to 2 PM
some example i had was:
cat <apache.log> | grep -E '07/Apr/20110[1-2]:[0-5][0-9]|02:00)'
small script in which i can pass the start time and end time range, and it should give an output of the rows which fall under that time range.
Im looking for an app pr line of code that could let me observe a process, save the info in a number of variable and then put the gathered info on a file.
Ive been trying with variations of top but no luck. I am running several CentOS virtual servers, VM is 2gb ram 2 processor.
Maybe a script that works over a specified amount of time while writing lines with the info on a text file so at the end i can have a sort of table with the data.
The thing is Im going to stress test the server and I would like to have the data to make some statistics.
How I can tell tar to assign a new creation date to extracted files? Let's say I have an archive filled with old files of varying ages. Upon extraction I want all files to have the same time stamp (that of the time of the extraction).
View 1 Replies View RelatedI have two machines, both are centos 4.4 (one is a virtual machine, the other is not) they are treating file timestamps differently. Below is an example from each of the machines to demonstrate.
Code:
[behaving as expected]
-bash-3.00$ ls -1t --full-time
[code]...
I am tying to read a file in with nawk whilst trying to take input from a pipe. I've come across the getline option and no matter how hard I try, I can't figure out the correct syntax. What I want to do is to take some input from the pipe and make a comparison with all of the values in a file and print a match.
Code:
suppose i have two file with same name fstab one file is located in /etc and the other is located in /root/ If i make a change in /etc/fstab file the changes has to reflect in /root/fstab . Is there any command to do this?
View 6 Replies View RelatedHow do I list only access time and file name? For example if I do ls -l --sort=time then
[Code]...
I know this command exists I just can't seem to find it. I want to see the last few lines of a file as more are added in real time. Can someone point me in the right direction?
View 2 Replies View RelatedI was wondering if there is a command to show a real-time creation of files. I basically executed a command that will created thousands of files and takes a long time. I want to check if it is still creating additional files or if ti got frozen.
View 3 Replies View RelatedI have a few servers that are exposed to the internet. When someone tried to brute force hack in to the ssh, ossec adds their IP to the hosts.deny. Then the hacker (read: script kiddie) moves to the next IP up the line and hits my next server, etc, etc.
I end up getting 20 emails for all the servers that they hit.
My question, is there anyway to sync the hosts.deny file across multiple servers so that if they are locked out of one, they are locked out of all?
How can we write a file and display in terminal at the same time. Like for example, when I do..
php -f file.php > testfile
That should save right.. but I want to display it in terminal otherwise.
I need some help recovering from a "slight" screwup. We just moved 3 TB of data from one RAID Array to another. Low lever archive files. This was done with a regular cp (for some reason) and now we have lost all the timestamps on the files, and we urgently need to get the timestamps back on these files.
We are running Ubuntu 9.10 Server and we have mounted the following
1. /mnt/old-raid ##Old raid from the old server
2. /mnt/new-raid ##New raid on the server
I know we can read out the timestamp on the old server using the command stat -c '%Y' <<filname>>
I know we can change the timestamp of the file, using the command touch -d '<<date>>'
To get from the stat -c date to the input date in touch we need to use date -d @<<timestamp>> +'%d %b %Y %R'
So my question is, how can I create a loop that will list all files in a folder, get their timestamp and update the old timestamp with the new?
I am using:
user@unknown:~$ sudo command -option > log
to save the results of "command" to the file "log", but I'd like to also get the result on the terminal, is this possible?
I am using ubuntu 10.04 lts.