Got site hosted, up and running now trying to work out the website particulars.
Right here it goes, not sure if it is at all possible but any input would be much appreciated.
Here's the deal. Developing an easy straight forward online store integrated with EBSWorldpay . Not looking for any fancy stuff. What i would like to do is put a flash url link that would direct a user to a html file that contains a hidden form and it would submit it on load redirecting a user straight to EBSwordpay checkout. Simple HTML Redirect with submited form.
Check the link here it has the source code of the form that is required.: [url]
Trying to use Example 0.1 ( So instead of having a separate page with a "BUY" for the form to be submitted, I would like to get the form posted straight away which would in turn load the payment site. Would create a seperate html file for each item and have redirects on the main website linking to those onLoad submit forms.
I want to ask, what should my computer look like if I plan to compile and run the linux kernel a lot? I mean what kind of hardware would someone recommend I use. Also is there a way I can fix this and maybe speed up the compile process. I tried make again and memory was again exhausted.
i have installed squid 2.6 on my centos. i have writen a shell script to ping a network and write to a file. write '1' if network up and '0' if network down. After that,a perl script will read the file and do the redirection.Perl will redirect to a fix URL [URL] if the network down and do nothing when it up. i have put my perl script in squid.conf at url_rewrite_program /my_file_path.
below is my shell script for pinging:
Quote:
#!/bin/bash while [ 1 ] do HOST=143.148.137.134
[code]....
My problem is client browser are not redirect to www.google.com even the network is down. It should go to the fix URL when the user click any URL in network down situation.it just appear cannot resolve host.
I have a linux test machine which I would like to run a copy of a production webserver. This is a legacy application which does not use a property file for its server name. Throughout the application, the server name is hardcoded (example: open connection to myServer.myCompany.com).
Is there any linux trick which I can use to redirect all requests for a certain host back to localhost? I know in Windows that I can add an entry to the hosts file and have it redirect back to localhost. How do I do this in linux?
Recently I need to redirect domain name that is a mobile minisite called m.domain.com/abc to external hosting web server ip with like 123.123.123.123/abc in our apache web server. This website is hosting in external. Our apache is 2.0.53, because I am new in Linux, so I don't know how to work out this. I want to know that how can I achieve this? Is it just use the DNS to map it or need to config our apache to work out this? I really urgent to know what can I do for it?
I'm trying to set up a framework where people connected to same wifi connection can enter a local site for developing purposes.I want them to be thrown to a local copy (development copy of the site) when they type in www.development.loc in the browser.I don't want to connect the world, only people connected to my wifi.Anyone willing to help by stating what I need to edit to :1. Allow me to access local copy of site that's on computer (located : /var/www/developmentsitename/) using www.development.loc in browser./and2a. What do I edit to create the server accessible by other computers connected to same wifi connection. 2b. If another computer can connect to this site now, can we create a virtual desktop setting in which workers can work as if they have their own partition on the server to work on and upload work onto the development server.
Location and syntax of named.conf in Ubuntu's version of BIND 9. I already have an HTTP server (in the form of Apache) along with forum software set up on my Acer netbook. However, the only way I can access the forum software is if I type in my IP address. How do I set up my named.conf to get the IP address to redirect to a domain? I already have my resolv.conf configured to allocate a domain to the IP address, but testing the ping of the site returns a "No such domain" error.
got a VPS and for testing purposes a homeserver with apache running. what two tools would i need to redirect my static ip to my dynamic one?
the traffic shall all go through the dynamic one. i've got a domain name registered with the VPS also. nameservers are configured.
is there any setup required on my local (dynamic ip) apache settings for using the name? and if i stop the service, will it still work? do i have to setup a webserver on the VPS for it? none running yet.
I am a Linux Distro user from 2006. Last year till i am using openSUSE 11.3. Recently i get a problem, When i enter a link My Firefox redirect to iwsearch.net (Infoweb). I install it few days ago. I this its my Browser (Firefox) problem! But I install Opera, Chrome no change same problem. This is very annoying. Too much pain for me. How to remove iwsearch.net.
Also another request to openSUSE Network Manager developers, Please add a options on Wired Connection. The options is Original MAC and Cloned MAC like Ubuntu 10.10 (Gnome), But i need it on KDE, because now i fall in love to KDE. So, i need (i think most of users need) MAC cloned systems, My ISP conection is MAC systems. Pain for change from Command mood I don't like to change from Konsol. We have two Wimax provider- banglalionwimax.com and qubee.com.bd , We are whole Linux user in this Bangladesh can't using USB modem, only Support Gigaset Modem. Its Indoor modem. But USB modem is portable and flexible. Please its hard request do something for USB modems.
I am trying to grep multiple numbers from file, grep does have the -f option for that.
Code: grep -f <`seq 500 520` /etc/passwd I know this could be done with
Code: for i in `seq 500 520`; do grep "$i" /etc/passwd; done But my question is fare more behind this example. It is possible to redirect one command output which will be treat as a content of file for another command ?
Is there a way to configure Apache so that if a user tries to access a webpage that doesn't exist under mydomain.com, that it redirects that user to the index.php?
In the ordering of files I keep I need links to directories. Sometimes I even need to move directories to new locations. I have tried using symlinks, but they become dead when I move the directory they point to. I have tried hard links, but I haven't found any Linux file system that would support hard linked directories. How can I achieve that a complex structure of directories (currently with symlinks for directories and hard links for files) keep symlinks live when directories are moved?
- is there any utility that updates symlinks when a directory is moved?
- is there any Linux filesystem that supports hard linked directories?
- is there any good Linux interface to the new NTFS (the only file system I know to support automatically updating directory links, called directory junctions)?
I have a variable (call it $valAS) that gets calculated . Also, i have a files Bassu3. if $valAS= 45368 (2nd part of 2nd line in Bassu3) after calculation, then $result=45368, but if $valAS != 45368 then $result= $valAS
Bassu3:
Quote:
name.srv name 45368 (value always changing) name 99999 name 88888
what is the best command to use to parse strings?I have a variable $str and need to parse this string.Can you provide an example of the command used to get a substring of $str based on the index values of start and end
My php knowledge is very poor (only worked with strings so far), and I am faced which a task that is a real challenge for me: I have a variable, that contains data of different type, in this order: byte, byte, string, string, string, string, short, byte, byte, byte, byte, byte, byte, byte, string.
Strings are of variable length. How could this data be parsed into variables of the right type, and then all converted to strings? What are the functions to use? Strings are unicode ones, and they are delimited by "
I have to parse a file containing billions of records and populate them in the Data structure. I have used a lot of C++ class and creating objects of the class I am storing the information retrieved by parsing the file.
Now as the file become huge and number of objects become very large my code is getting bad_alloc error as it is not finding any space avalable in the heap for allocating new object.
To run a parallel chemistry program, I need to build the host list. The cluster already has SGE(grid engine) installed and it generates a host list file with the following contents:
The important bits are hostname(ex compute-1-1) and number of cpus to use(ex 2). And for this program, it wants them in this form, a shell variable: HOSTLIST=hostname:cpus=X hostname:cpus=X .... I've tried this script, but it doesn't work
I have a basic HTML file set up which allows my to input some data. I have a MySql database set up behind the scenes with a table for this information to go in. When I click the submit button in my HTML file this PHP file opens and it comes up with "Parse error: parse error in C:wampwwwcomp39xinsert_student.php on line 32"Here is my code I am using in the PHP file:-
<html> <head> <title>COMP 39x FInal Year Project - Student Added?</title>
I don't think this is a "perl one-liner" of find and replace. I'm trying to auto-fill some information in a listing of files. The simplest example is that in the files the following exists:
I would want the script to find this and populate it with something like -- Date : 20101004-1758
I have a few more similar fields to autofill, and I'd like to do this from within a larger perl script I'm developing to process these files. So, how I perform in-place file modification from within a perl script?
I'm trying to figure out how I can get a request count per CIDR/24 from an apache log in combined format - e.g.: Code: 24.147.44.122 - frank [10/Oct/2011:13:55:36 -0700] "GET /apache_pb.gif HTTP/1.0" 200 2326 "[URL] "Mozilla/4.08 [en] (Win98; I ;Nav)"
I'm stuck using BASH for this and I generally write everything in Python, or even Perl if necessary so BASH isn't the most comfortable for me. I've got enough to extract the IPs and get a count I just need a slick way to come up with a sum on a per CIDR basis.
There is a file with that format of each models' information.I don't think that's good format, but I cannot change that format. I needed to modify the model name, 'model = xxx' as 'model = abc'.And I don't know how to parse and modify 'model = iii' and 'model = ddd'.The only clue to parse 'model = ddd' is the second 'model = ' after the second 'system information'. But how to parse the second keyword?Is it possible with 'sed'?I sometimes have to modify the information of the file.using shell script if possible.Python is ok. (Shell script is better for me.)
I would like to extract Room number, Lastname,Firstname,invoice (205880080),arrival date, departure date, and total(229.46). Can you at least give me a hint on how to proceed? I have tried a lot but I am stumped from the beginning.
***History*** Room: 124 B Payment: Bell/TRAVELSCAPE.COM Lastname*FIT*,Firstname 4A, 0K, 0B Guest Bell *205880080 FT Bell *205880080 July 31, 2010