General :: 64 Bit - Different Commands Shows Different Results?
Jul 10, 2010
The system of my VPS is centos 5. I want to know if it is 32bit or 64bit. > uname -a Thu May 13 13:49:53 EDT 2010 x86_64 x86_64 x86_64 GNU/Linux
looks like 64 bit.
But
> getconf WORD_BIT
32
Looks like 32 bit.
Which is correct?
I believe the linux of my VPS is 32 bit. Because I downloaded a 64bit mongodb, found it can't start. And when I downloaded a 32bit instead, it works well.
This question may be silly and super easy for linux connaisseurs, but I was just wondering, for instance, I want to use the >find command to search for a file and send the results to a text file
I'm timing how long it takes to run a command foo. I'm looking to append the results from the time command to a file, and discard the results from the foo command. I tried the following, but it didn't do what I want:
$ time ./foo > /dev/null >> output_from_time_command.txt
if you do the command conky in terminal, it starts conky ofcourse, but it also shows output to that terminal so you can't do any other commands to that terminal, Is their an option like you can do with the '&' sign in other cases? If you do the '&' sign with conky it still gives output, also the conky -d command gives output...
I am trying to determine CPU usage on a radio controller running Linux. We have a script that another member of my work wrote using vmstat which I could not get any consistant results from. So, in an effort to understand what was going on I wrote a script using ps. Sadley after running several tests using both scripts, they do no match up at all, vmstat always gives me a much higher value. However, if I simply run the commands at any given time they do match up, therefore I believe that there is an error in one of the scripts. I am very new to linux and so Here is the vmstat script:
When using the grep plugin to VIM, I can search the current directory for all occurrences of a string within a set of files, like this::grep Ryan *.txtThis outputs something like this:
file1.txt:3:Ryan was here file2.txt:10:Ryan likes VIM file3.txt:5:superuser.com is a fav of Ryan
I (for curiosity's sake) am running du -a inside /usr/lib/git-core and it does not show results for all the files in that directory. Why does it leave out an arbitrary set of files?this is what cd /usr/lib/git-core; du -a returns:
I'm encountering a strange problem. I need to open and forward all UDP and TCP ports related to VoIPtelephony (5000:32000) in the Suse 11.1 server that's acting as router/firewall in our setup. The ports must redirect to a Asterisk server in the local network. (This server has the IP adress 192.168.0.3)I've opened ports in Yast (Firewall>Ports>Advanced) and putted in some masquerading rulesirewall>Masquerading):0/0,192.168.0.3,tcp,5000:31000,5000:310000/0,192.168.0.3,udp,5000:31000,5000:31000when I do a nmap localhost I get:Starting Nmap 4.75 at 2010-01-08 16:52 CETInteresting ports on localhost (127.0.0.1):
Not shown: 991 closed ports PORT STATE SERVICE 21/tcp open ftp
I've setup a git-svn repo with cron to fetch from the svn repo daily. I have a script to do the fetching, and this is what is invoked by cron. Everything is fine with the repo, and the script works fine when executed manually. However, when it runs under cron, empty files get dropped into the .git directory. The files have names that look like they are some base64 output, e.g. juTrvjP6m8 and kcKf3hu3b4. Two of these files show up for every cron run. I thought these might be commit hashes, but they're not, git-show says it's an unknown revision. I set-up the repo as follows:
[code]...
The last line pushes the repo to a separate (bare) public repo from which others can clone. I'm piping the output from the cron job to a file, which looks like this:
[code]...
The line "fatal: unable to run 'git-svn'" is alarming, but the fetch seems to go ahead anyway. Where are these empty garbage files coming from, and how to stop them? Am I in for bigger problems in the future?
In my .bashrc I have the following lines to turn on colors for grep and ls alias ls='ls --color=auto'export GREP_OPTIONS='--color=auto'.I've tried changing the alias to export LS_OPTIONS='--color=auto' but that doesn't work.Is there anyway to use an export instead of alias. And are there actually any benefits to one way over the other?
i'm in need of my hard drive & usb flash drive numbers, but when i do fdisk -l i get no results - it just gives me the prompt as if i've pressed the enter key on my k/board.
I sorted a large mailing list and the result is in almost sorted order, but not quite.The sort was invoked in the default manner, just "sort".See (below) a segment of the output file.(Note: the actual files have a street address following each name.I have omitted the addresses to avoid complaints about disclosing personal information on-line.) I expect all people named KATZ to be grouped together, but they aren't. This phenomenon occurs elsewhere in the output file, so it has nothing to do with surnames beginning with the character string KATZ.
I am at the moment using Ubuntu 10.10 with the default color scheme. If I open a bash terminal and type ls -l / I get the results with most information in white on the standard purple background, most directory names in blue on the normal background, tmp in blue on a green background, a file name in white on the normal background and links in teal on the normal background.
So in this situation I am wanting to figure out what the green background behind tmp signifies. I have searched for information about bash color codes and I find hundreds of links regarding how the CHANGE the colors. I have yet to find one which explains what the colors mean.
Below is an example output of what I see when I run the 'ls' command on some directories in linux (this is from a tomcat/common/lib directory). However I'm not clear on why some of the filenames are appearing inside [square brackets]
so I was wondering how I could do a simple find which would order the results by most recently modified. Here is the current fine I am using. (I am doing a shell escape in php, so that is the reasoning for the variables. find '$dir' -name '$str'* -print | head -10
How could I have this order the search by most recently modified. (Note I do not want it to sort 'after' the search, but rather find the results based on what was most recently modified)
I am looking to edit a great script made and posted by wje_1q Here I would like to be able to save results to a txt file is this possible
#!/usr/bin/clisp (defparameter *character-set* "0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZ") ;(defparameter *character-set* "ABC") ; < --- this line is for testing (defparameter *word-length* 10)
I come from more of a programming background but have been giving the task of sending backup results from a linux box in an email to external email addresses, I have spent days looking trough google for info but cant seem to find anything simple, I have looked at postfix but because I don't understand networks, hosts, nameservers etc. What I need to do to allow mail get sent from the linux box to external mail address. I have outlook connected to an exchange server on windows and I can ping the linux box from my windows command so surely it cant be to much involved.
I've just installed Debian 6 on an old pc so i can use it as a file server.As far as i know installation went fine but it didn't detect my Edimax wireless network card.The problem I'm having is using FDISK. I tried to explore my CD drive and i got an error message "Unable to mount location". I wanted to see if there was an installer to install my network card becuase the manufacturer says it's supported by Linux. When i use the su - command and try using fdisk -l it just displays the command prompt and no hdd info. I tried it the other day becuase I want to mount all the physical drives and one logical volume. This is only the first step for me configuring this computer as a server and i'm suck already and ready to give up and try Windows server (whcih i don't really want to do).Has anyone got any ideas how i can mount my cdrom and why FDISK isn't working?