i m using gcc to compile certain programs.they get successfully compiled but fail to execute.Here is a sample code and the error [tommy@tommy cprog]$ vi test.c
#include<stdio.h>
int main()
{[code]....
a similar problem arises with codeblocks for the same program.the error is the same even when it is just main() and absence of return 0;
the problem exists also on c++.i m using Fedora 15
I went to an interview last week and there was this guy who asked a simple question that i have been trying to solve for a couple of days. I tried google but i just cant get the search keywords right. The result were just useless. Well, the question is : "How can we allocate a limited memory to a process before we start its execution" well, the question is related to an X11 system so may be some flags must be set to limit its memory.
In gui validation if user enters string in integer requiring field, a dialog box pops out. So on clicking the button i want to stop the program execution and return to the previous jframe(either the same one or newly created object of it).
How can i handle this in java. I tried calling main(); but it doesn't work. Also System.exit(0); doensn't work.
I have some questions.First, how can i obtain a process execution time ( after it terminated)??? Second, in a multi-core system like SMP how can i run different processes on different cores explicitly? i.e. how can i assign a process to a specific core?
In my program(C++), I have a ping function call which will ping the specified IP and returns the result. The problem is, my code does not print the o/p to the terminal window after calling the ping fucntion for the first time. but, the cout statements exist before the ping function call, works fine. When i debug it using strace i found that output filedescriptor is getting closed some where through close(1), before returning from ping call. What would be the issue, can u please some one explains me what the problem is.
I am outstation and having laptop with windows and internet connection. I want to check one small program on linux. kindly tell if some free linux server is available on internet where I can upload my program, compile and execute. The program is generic and there is no restrictions regarding linux version.
I'm trying to figure out how to continue execution of a Python program while a background task is going on.
In bash, it is adding "&" to the end as in:
Code: #! /bin/bash find / -iname "*.py" > /tmp/all_pys & echo "looking for all python scripts and putting them in /tmp/all_pys" (yes, I know I could just put the echo BEFORE the find but I just use it for example)
Say I do that in python (not necessarily running the shell, but perhaps using scriptutil.ffind or something similar), but want to allow the user to be doing something else while running that search in the background, how would I thread, fork, subprocess (not to be confused with the module and Popen) [or whatever I should call it] that?
what i need is to execute a script before reboot actualy the system reboot faster than the script execute how can we force the system to wait for the script complete execution
I have been trying to get my development box up and running and I decided to do some CGI work in Perl and I am getting a 500 error on Apache 2. I have checked all the perl settings and the script runs fine on the CLI. I am simply passing a textfield from a HTML form to Perl. I was using the standard CGI.pm calls but I modified the script just to use print statements to make the output page. The script has one textfield in it, as it was just supposed to test the CGI.
I have the UserDir directive active for my user and I did some modifications to allow CGI execution from my home directory /public_html/cgi-bin/. the script will not run from the main /var/www/cgi-bin directory either,that also returns a 500 error. the permissions are 755 on both copies of the script in the cgi-bin directories.
I am trying to set up a simple script to automatically invoke streamripper to record my favourite radio shows. The script works fine in a terminal window (Konsole), but when scheduling this as a cron job (KCron via SystemSettings) a Konsole window pops up catching the streamripper output.
Now stramripper can behave with the --quiet option, and I set output redirects "> /dev/null 2>&1". Inside the script or on the command line of KCron.
Quote:
sh myscript.txt >/dev/null 2>&1
...or just as the script without the "sh" command. Same result. Still the same annoying Konsole terminal window pops up, now with no other content than what Cron obviously generates at the end of the script execution:
Quote:
-------------------------------------------------------------------------------- End of script execution. Type Enter or Ctrl+C to exit. --------------------------------------------------------------------------------
Well, the whole point of this cron scheduling was to avoid manual intervention and create the MP3 file of the radio show automatically. how to avoid the terminal window completely?
I am trying to write a script to set the SGE job execution order. I named each job with 'job1', 'job2' and 'job3'. I want my script to do: When 'job1' execution is complete, 'job2' is executed; when both 'job1' and 'job2' are complete, 'job3' is executed.
First, I tried this following script and it worked well:
We are on our first Linux platform and I am trying to coordinate a distributed application backup across multiple machines. I am trying to write a script in which I would have used RSH to execute scripts on the other servers. We are no longer allowed to use rsh, and someone suggested ssh. I am using that instead of telnet, but I am not sure of the syntax"rsh server [-n] path/executable" is what I would have used, just not sure of the syntax for ssh
I'm not sure if this is a Debian-specific question or a KDE question.
FYI: uname -a Linux DESKTOP 3.16.0-4-686-pae #1 SMP Debian 3.16.7-ckt11-1+deb8u3 (2015-08-04) i686 GNU/Linux
(Running KDE).
Question: Upon logging in (typically, for the first time each day) I would like to see a console window/shell or a simple terminal-emulator/shell automatically open up.
The reason is: In order to get my .bashrc to execute I have to log-in, then bring up an xterm or some other terminal-emulator.
Background: My .bashrc has a once-per-day routine/section that will set up my desktop to look the way I want it on my initial log-in (primarily with my Java applications up/running and displayed and waiting for me).
Then, later on that day whenever I bring up new bash shells, my .bashrc is executed but that once-per-day routine/section is no longer executed.
I have seen command like $rm aa dd cc bb ee then something like $ >> zz // it removes zz also Where aa dd cc bb ee zz are the files in my directory. Do not know exactly.
I have an application running in my machine.I copy the source code to make it run in another machine.I copy the libraries that are needed and paste it to the other machine.I am able to build without any problem and I get the executable. As soon as I start executing,the application crashes with the following message.
0x00110416 in __kernel_vsyscall () #0 0x00110416 in __kernel_vsyscall () #1 0x004a41d3 in __waitpid_nocancel () from /lib/libc.so.6 #2 0x0044259b in do_system () from /lib/libc.so.6
And achieve something like this when I view the contents of test2.txt:
Hello World Your boss is Me
The problem is, it's not that easy to achieve. I don't want to do the "echo Hello $USER1 >> test2.txt" because I'm planning to implement this on a MB-sized text file.
I'm trying to save the output of the script I wanna run but for some reason I cant manage to do it with "script.sh > Output_4.txt"
The thing is that the script was already run and it extracted some directories so when I try to run it again (to save the log to a Output_4.txt) bash says (see screen-capture) and of course the log's empty!!!!
Now my script is able to start server, But I am still have some problem with my script. When the start server command is executed, the control does not pass the line and does not execute further of that line.what is the problem and how can I get smooth execution of the my script. My Script:
I have never set-up a chroot-jailed environment before and I am afraid I need some help to do it well.To explain shortly what this is all about: I have a webserver to which users send python scripts to process various files that are stored on the server (the system is for Research purpose).Everyday a cron job starts the execution of the uploaded scripts via a command of this kind: /usr/bin/python script_file.pyAll of this is really insecure and I would like to create a jail in which I would copy the necessary files (uploaded scripts, files to process, python binary and dependencies).
I already looked at various utilities to create jails but none of them seemed up-to-date or were lacking solid documentation (ie. the links proposed in How can I run an untrusted python script)Could anyone guide me to a viable solution to my problem? like a working example of a script that creates a jail, put some files in it and executes a python script?
Problem started after installing a new firewall in my network. originally after installing the firewall there would be a 10 second delay after entering a command before it was processed, the command would be executed however there would be an error, i forget what it said, something about not being able to resolve my domain name (and it would give my domain name). i realized this was a problem with my NAT, so i configured DNS forwarding on my firewall so that any attempt to access my server using my domain name from inside the network would point back at my server. after setting up the DNS forwarding i no longer receive the error about my domain name being unreachable, however there is approximately a 30s delay between entering a command and its execution.
the problem is hardly critical i do not see any more serious problems, it is just a pain to do any work when every time i try to do something i have to wait for it to execute.
How do I create a command to launch a program and then have the terminal wait for a specified time and then move on to the next command?I'm wanting to create a startup script, and I need program B to wait until program A has finished loading up.
I am using CentOS release 5.4 ( 2.6.18-164.9.1.el5xen ) and created an HPC cluster by using NIS ( for user authentication ) NFS ( as file system ) and mpich1 as parallel compilers and utilities and TORQUE as job scheduler. I want to make sure all users should use scheduler for job submission and should not submit the job directly ( qsub job.sh ).
I want to prevent all users from executing executable files created by self , from its home directory .
Suppose if a user create an executable a.out and if he tries to execute by ./a.out it should display an error.We should also allow users to execute normal user level linux commands .How can I implement such a set up in my environment
i need to take the backup some folders encryption formate with password i used gpg encription in the script but it is not taking the password it throos errorthe syntax is#!/bin/bashtar -cvzf test.tar.gz target ; echo "test" | gpg -c test.tar.gz --stdin
Get the list of virtual addresses being accessed by any application during its execution time. I don't want the complete memory map but just the pages which were accessed during an application's execution?
I am calling a service using http post through wget, the command is successfully executing but for each execution its creating a file and saving variable names n data n it. I want to execute this command without creation of a file. Would anyone suggest me what needs to be done in this regard.
My command: wget --post-data 'var1=99&var2=200' http://xyz.example.com:5555/invoke/Samples:httpInvoke For every execution, its creating the files with names: Samples:httpInvoke1 Samples:httpInvoke2 Samples:httpInvoke3