I want to traverse a directory and get a list of files that contain a set of patterns. I assumed I could use grep for this, but I having trouble getting grep to only return files that match ALL patterns. Here's what I've come up with so far:
However, this gives me a list of files that match ANY of the patterns in the searchpatterns.txt file. I want to match ALL of the patterns. I've looked through the man page, but can't find anything that allows me to change the "OR" to "AND" for multiple patterns.
I have a number of files:FooBlahhFooI only want to be able to grep for names in a file that contain Foo and not BlahhFoo. However I am not able to pull only those files away. How can this bee done. My grep/zgrep knowledge only goes this far at this point. I'm still learning but I'm stuck on how to make my arguments more precise zgrep 'Foo' SomeFileIMade.gz > /home/user/FOOFILE
i have registered two domain names that i want to use to connect to my ubuntu server. I was wondering how to do this i was looking at bind9 but that didn't work that great. The server is behind a router with firewall i can connect to it using the external IP address but i like to use the two domain names if that is possible.
Here is my scenario and what I want to try and achieve.
Server A = 192.168.1.5 (serverA.home.com) Server B = 192.168.1.6 (serverB.home.com)
Currently I have serverA.home.com domain name set up using a free dns service online. When I am anywhere in the world I just type the domain name it hits my static ISP IP via the free dns servers online, hits my home router, gets port forwarded to 192.168.1.5 Server A and I am up and running.
I have bought a second domain for serverB and want it set up like serverA but I am totally baffled on how I can resolve 2 domain names on a single network? I need (somehow) to try and tell the dns servers that when I enter in serverB.home.com it will hit my static ISP IP as above and then hit my home router but then get directed/split to Server B and not to Server A.
Basically all I want to do is name the machines and get traffic resolved to them.
I'm trying to find exact matches of some users in the /etc/passwd file using "grep -w", but it doesn't always work. For example, I have the following users:[URl].. So, let's say, I want to search for the user "stewart" (which doesn't exist)
Using a list of names (over 4000 of them) painstakingly gleaned from the source file years ago for a database file, I want to match the names against the source file so that they can be updated with the tags <forename></forename> in the original source file.
I placed the list of names in @forenames (only posted a few of them here).
Perl script is:
I am able to get the name bracketed by the tags to appear on the console screen but don't know how to apply the output to the source file. Perhaps I need to do a match on the words then some kind of edit to surround the matching words with the xml tags? I'm a rank novice doing this as a labour of love for a friend.
Is it possible to have one static IP address with a NAT network forwarding each domain name to certain internal/DMZ IP addresses? I know you can do it by port but if both websites are on port 80 can you forward to the corresponding server on the dmz.I ask this because I noticed the website braemere.com.au had to be typed into a web browser and entering the IP which is 220.127.116.11 did not take me to the website.
So I'm looking to host potentialy three different...sites...I want to call them from the same machine. One is a radio station and the other 2 are just straight websites. So I wanted to know if I had to use different machines to accomplish this or can I alter maybe the http conf file or what. I did find this in the forums but wasn't quite sure if this applied to me. [URL="http://www.linuxquestions.org/questions/linux-software-2/serving-multiple-domain-websites-on-one-server-329914/"http://www.linuxquestions.org/questions/linux-software-2/serving-multiple-domain-websites-on-one-server-329914/[/URL]
I'm new here, so please be gentle! I recently installed F12 64 KDE on my Acer Aspire 1810TZ, could connect to the WLAN router but neither the native Konquerer nor Firefox could open any pages...
resolve.conf has the DNS according to DHCP offer, from command line dig properly resolves the name, ping (e.g. google) works, telnet to domain-name on port 80 works, GET too. When using the browser with the domain-name I tells me "Server not found", tcpdump doesn't display any activity. But when I manually change the domain-name with the IP (received by command line dig) then the browser has no problem how do browsers interact with the network resources of the OS?
For clairification, I can ping. I have tried several IP addresses and 100% success rate. When I noticed the problem I was trying to run sudo apt-get update && apt-get upgrade After some time I noticed these error messages to start with
I tried to ping the adddress security.ubuntu.com from my Windows machine to verify that I could connect and was surprised when I could. I then pinged the address 18.104.22.168 which is what my windows machine resolved the name as and it went though.
My thoughts on this are that when my Ubuntu Router came up, for some reason it did not incorporate the ISP's DHCP servers into the ip address it obtained. Sadly I know to view ALL IP infomation in windows via ipconfig /all command but I do not know what this is in the *nix world. need commands that I can use to check and troubleshoot this apparently DHCP issue so I Can start to update my server and expand on its services?
I recently installed bind9 on mandriva 2008.1, after having done the necessary configuration.. I still can't find my domain I configured cant ping on other machines on the LAN but can actually ping on any other website on the internet even though name server is configured to point to local machine..... I dont understand what I'm doing wrong.
I want to use single label host names on my local LAN, without using any domain (at least no registered one). My machines should be named in a fashion similar to 'myserver', 'mydesktop' etc. so that i from a browser on any LAN machine may write 'http://myserver', and get the webserver on 'myserver' (so NO domain part, e.g. myserver.domain).
Do i need to create a DNS zone for each host or is there a way to put all host in a single zone, if so; which? Would such a name be considered a root domain?
I have been logging into a server remotely and trying to set up a mailing list on it. The server is the newest version of ubuntu server: uname -a: Linux Themis 2.6.28-11-server #42-Ubuntu SMP Fri Apr 17 02:48:10 UTC 2009 i686 GNU/Linux.I noticed I could not download packages with apt-get or ping domain names, and I can't even ping 127.0.0.1.And do you think there is something wrong with the network card?
I have a centOS 5 server running apache 2 with it's web root at /var/www/html. By setting up a dns entry to point to the computer's IP, I was able to go to testing.myservername.com and reach the contents of that directory.My question is, how can I go about mapping multiple future domain names to folders that are under this webroot?For example how would i bind www.temporarydomain.com to the /var/www/html/temporarydomain/ folder and also be able to map www.anotherdomain.com to /var/www/html/anotherdomain/?
here are the steps that i have taken already to resolve and troubleshoot the issue. i can ping Google. i can browse the Internet i can browser the repo Index of / through my browser. i have insured that ipv6 is disabled through network configuration through yast.