I recently set up a server in my house using Apache 2 running off Ubuntu 10.10 Netbook edition (no, the computer is a desktop). I tried running the website on a few computers in house and all the images and scripts seem to load fine but it is unable to read from the data.xml file through javascript. This leaves all of the web pages empty because all of the content is parsed from the xml file through javascript. I tried debugging with google chrome and after running this code to load the xml file:
if (window.XMLHttpRequest) {
// For IE7+, Firefox, Chrome, Opera, Safari
xmlhttp = new XMLHttpRequest();
}
[Code].....
the xmlhttp.responseXML is null and the script crashes. Could this be a permission issue? I tried setting the data.xml file to read-write permission for any user but that didn't work.
I have 2 Ubuntu machines up & running on a small network. Both are running SAMBA servers. One several years old & has a rather old version of Ubuntu. The other, quite new & has a version about 9 months old.The newest has a problem. When working with large & complex Word documents, it will periodically generate an an error on the work station.
"Word failed reading from this fille (xxx specifies file we are working on). Please restore the network connection or replace the floppy disk drive". The it crashes Word & unsaved work is lost.The error never occurs when using the old server, only the new.
I don't understand the results of a simple performance test I ran using two basic scripts (running on a high end server):
perfVar.zsh :
#!/bin/zsh -f MYVAR=`cat $1` for i in {1..10}
[code]...
Performance test result:
> time ./perfVar.zsh BigTextFile > /dev/null ./perfVar.zsh FE > /dev/null 6.86s user 0.32s system 100% cpu 7.177 total > time ./perfCat.zsh BigTextFile > /dev/null ./perfCat.zsh FE > /dev/null 0.01s user 0.10s system 91% cpu 0.118 total
I would have thought that accessing a VARIABLE was way faster than reading a FILE on the file system... Why this result ?Is there a way to optimize the perfCat.zsh script by reducing the number of accesses to the file system ?
I have a 7.2 GB file (VMWare virtual machine file) that I am trying to copy from its original location to the another folder OR to external hard drive...each time I try to do this, I always get the following error after the copying process reach 'exactly' 1.4 GB
Error reading from file input/output error
And I have to either Cancel or Skip
I've tried to split the files to smaller pieces but the idea didn't work as I still get the same error whenever I try to compress/ split or do any operation with this file. how I can copy this file?
I just built a server, which displays 8 cores in the i7 processor. However, because of the 2 threads per core, it shows up as an 8 core system. When viewing htop, it shows system loads at 0.29,0.18, and 0.13 for respective averaging times. How would I calculate actual system load for this processor? Would I divide the average by 4 or 8? Or, is there another method?
Running 10.04 64bit. Installed the latest Gnucash from repositories. Using Quicken 2003 and exported an account to a qif file. When trying to import to Gnucash the error "A bug was detected while reading the QIF file." displays. This, at least is more than I was able to do previously (prior to 10.04 and several earlier versions of Gnucash running in 8.04 through 9.10). Before it would cause Gnucash to crash and shutdown. I select a specific account with categories from Quicken to export. I can import to KMymoney and use it but that program is not compatible with the wife and does not do reconciliation reports by default.
i installed kernel 2.6.34 to fix my lid closing issue, and that went great. but now when i go to reinstall my broadcom i get this error. Code: Reading package lists... Done Building dependency tree Reading state information... Done bcmwl-kernel-source is already the newest version. 0 upgraded, 0 newly installed, 0 to remove and 1 not upgraded. 1 not fully installed or removed. After this operation, 0B of additional disk space will be used. Setting up bcmwl-kernel-source (5.60.48.36+bdcom-0ubuntu3) ... Removing old bcmwl-5.60.48.36+bdcom DKMS files...
With F11 installed Apache is having permissions issues reading files out of the html directory. Only wants to work with permissions set to read for other. [Thu Jun 11 23:25:28 2009] [error] [client 127.0.0.1] (13)Permission denied: file permissions deny server access: /var/www/html/index.html Tracked down the permissions issue. Is there a good reason not to change the group to apache and remove world read?
I am trying to create a script that reads a list of url's from a text file and then executes the same script for each line. This is what I am looking for:
elinks -dump $(url's in file, one by one) | grep "reply" > jobs.txt
I am using a backup system with cron + tar. Since the server is very busy, I get often the cron-email: "The file XYZ has changed while reading". This message is a bit annoying and I see it as critical point in my backup system. I believe that this file is then not in my backup. (Is that correct?) Let's imagine the hard disk dies and I have to recover the system and my personal data, and in the night the mysql-table XYZ was not in the backup, because it was in read-usage. I would then have lost this table forever. Is there any way to tell TAR, that it should force the file to be included (if in read-usage, then wait 2 seconds and try again)?
When using cssh, I define some cluster tags in my local .csshrc file. Other parameters seem to be picked up by cssh, but the tags are not.I have no /etc/clusters or /etc/csshrc files. I create a fresh, default .csshrc file and append a line like:myhost = frodoWhen executing "cssh myhost", I get: Can't call method "name" on an undefined value at /usr/bin/cssh line 988This is the same message I get if I make up some destination, like: "cssh doesnotexist". So it looks like rather than picking up the tag "myhost" from .csshrc, and translating the command to "cssh frodo", it does a dns lookup on myhost.I've tried making a /etc/csshrc file like this (but without the equal sign), but that doesn't work either. This is all on Ubuntu 9.10 with cssh version 273.
I want to install this file for reading some windows .dat files[URL]...I looked at some code and it looks like its trying to connect to an outside server. This program is supposed to give non-Microsoft e-mail programs the ability to see attachments included in an e-mail message by a Microsoft e-mail composer.
I have a .bkf backup file, created by the Backup utility that Microsoft provides with Windows XP. Is there a way to read the contents of the file using a non-Microsoft OS, preferably Mac OS X or Linux?
My requirement is that I wanted a shell script which should read the alert.log(oracle) and should send the alert to my email if any error encountered on daily basis. I have a script which does the same job but at the same time it creates a new alert.log whenever any error occur. But I don't want the new alert log I wanted to read the same alert log daily and if any new error come should alert as email.
I have a binary file, which I need to process using my C++ application. Only thing I know is first chunk of the file is long, second chunk is int, third chunk is char etc... The binary file actually contains something like below. (which is represented in hex base).
D7 07 00 00 00 00 00 00 37 18 00 00 DE 07 ............ so on.....
I need to procees the file in the following way.
* I know the first data segment in my file is long. So it takes 4 bytes. * so I need to read the first four bytes. That is D7 07 00 00. * Then I need to reverse this as 00 00 07 D7. * Finally I need to get the decimal value of above hexa decimal line. ( 00 00 07 D7) * i.e. 00 00 07 D7 (in hex) = 7D7 (in hex- after removing leading 0 s) = 2007 (in decimal)
I have two files of data with different numbers of columns and rows. I want to read this two files in two arrays and then compare for example the second column of first data file with the third column of the other text file and if the difference between of two numbers is less than a threshold then the program print the information in the rows which fulfill this condition in the third text file. I have written below program but the problem is that it does not go through all rows of second file.
declare -a a declare -a b r=` awk '{n++} END {print n}' second.txt ` echo $r awk ' {
[Code]....
Actually I have two data files one of them contains 44406 and the other one has 12066 rows and I want to check whether the difference between the components of two specific columns is less than a threshod but I have simplified it here. I had written this code and then I have realized that this code just goes through the number of rows as the same as the first.txt file and ignores the rest. I could not find the problem yet.
I made a string key-value mapping struct in C, and functions to add and remove entries. I would also like to write a function to read in this file format:
I need to Read a path of a file witch is written in Text file i used this
Code:
FILENAME=$1 while read line do echo $line done < $FILENAME
it worked and showed me the Line witch was written in my file but now my problem is how am gonna use that line as a path i mean for example if am gonna execute a linux command on that file like dpkg -i /path/to/the/file how am gonna export it from The $Line variable and use it after the command.
I am trying to read certain lines within a file and give the output of the certain lines that dont equal my value, I think showing you would be easier. There is multiples of these inside one file...
Code:
LV Name /dev/vg00/lvol1 LV Status available/syncd LV Size (Mbytes) 300lable/syncd
[code]....
I want to read everything in the file, if the status is not available then it should display the name (directly above status). If they are all availbale then do nothing. I think I know how to do it which includes putting the info in string form and placing in hash but it is proving to be out of my skill range.
To save on the writing of WAY to many files with very little in them, I want to put it all in one file and read a specific few lines. There will be six variables to be read at a time. Format is as such:
//Set 1 string name 5 12
[code]....
From name to 5th number is a set. The name will be of different lengths for each set. This will be a big file of probably 40+ sets. My problem lies in reading one and only one set be it set 5 or set 34. It needs to be done in C++.
Installed Fedora 12 in vmware and during the installation of a professional program versioned 2003, following error encountered.
tail: cannot open `+124' for reading: No such file or directory gzip: tmptarfile.tar.Z: not in gzip format tar: This does not look like a tar archive tar: Error exit delayed from previous errors extract error, installation cannot proceed.
Checked the google and tried ncompress and export_posix2_version=199209, but the problem persists.Is fedora compatible to this program, MEDICI 2003 ver?