I would like to do something which I guess is pretty common: convert my collection of CDs to good quality mp3, named according to artist-song and properly tagged. Having to rip hundreds of CDs (i am not rich, i just live in China! ) I would like this to be as automatic as possible, like: insert cd wait 4 minutes (sw gets tracks' name and tags from the internet, and rips) CD is ejected automaticallyinsert new cd ...and so on If I was on windows, i would probably be doing this with itunes. Juicer does not support MP3. I read great things about rubyripper, but i dont understand how to make it start automatically.
I currently have a little box with Mediatomb installed which I use to stream to my Linn DS. Is there an application that I can use that will automatically rip a CD when it is inserted into the machine and then eject when it is complete?
In linux how do I to get the ascii decimal/hex/oct code from a char or a ascii char from a decimal/hex/oct code. I see from the gnu sed user manual that I can use the d# o# x# to specify a character but I am not sure how to use that. If some one has a better way to get the the ` and ' chars through I want to sed with the backtick '`' character and ''' single quote character.
I am having some weird problems with calling commands stored in a variable (I need to do this to assemble a command with a bunch of parameters automatically and then execute it).Example code that will replicate the weirdness:
What happens to the quotes? I have tried various combinations of single quotes, escaped quotes, etc, but it seems like quotes in a variable are not evaluated as quotes when that variable is executed.
There is probably a very simple fix, but its wrecking my buzzI call the following on the command lineR CMD BATCH '--args Y_filename="one.txt" out="two.txt"' brew.RI want to call this from a bash script but replace one.txt with $1, and two.txt with $2 (ie the first two arguments from the command line)I have tried in vane"R CMD BATCH '--args Y_filename="$1" out="$2"' brew.R"
$ execute_some_long_command <command is executing> <Accidently press middle button that inserts bunch of garbage (including, for example, `rm -Rf ~/*`) into console>
How to let execute_some_long_command finish, but not execute inserted things?
I want to watch a number of processes in "D" status repeatedly with following command:Code:# watch -n 1 'top -b -n 1 | awk '{if ($8 == "D") {print; count++} } END {print count}''but it didn't work. I also tried with double quote. Can I use 'watch' command is this case?PS: I know 'while' and 'sleep' can do the same but it is dirty workaround.
I need to find each line in a file which does NOT begin with a double quote (") and append that line to the previous line. I have been successful doing this using the following command: cat filname.csv | sed -e :a -e '$!Ns/ [^"]//;ta -e 'P;D' > newfilename.csv
My issue is the substitution. As you would expect after the line is appended to the previous line the first character is removed. I need it to not be removed. I tried: cat filname.csv | sed -e :a -e '$!Ns/ [^"]/&/;ta -e 'P;D' > newfilename.csv but it just hangs.
Goal: Input: "line 1" line 2 Output with existing sed command is: line 1ine2 I need it to be line1line2.
In my Windows environment, I use email client such as Microsoft Outlook to connect to our email server to send email with the following configuration:
Incoming server (POP3): 995 - (requires with SSL) Outgoing server (SMTP): 465 - (use encrypted connection SSL)
[code]....
And the mail server requires user ID login and password.how do I setup a text command based email client in my Linux (Centos 5.1) to send out email through the existing email server above, which is in another machine? The email client has to be text command based because I need to use command line to send notification email from anothar application installed in my Linux (Centos 5.1) Since the email client will only be used to send email notification, I don't require setting up of an email server in my linux.
Using CentOS. I have a cron setup to run this command: Code: /var/test.sh | mail -s "Test Cron" mr182@somewhere.com The email is sent but the output of the script is not in the email body, it's just blank. I know there is some output because there are some echo statements in the script.I don't want to get an email for all cronjobs, just this one.
I create a bash script that writes another bash file. But in the generated bash file I want to write a bash command in the file and not executing it.Here's my bash file:
Code: #!/bin/bash cat > ~/generateGridmix2data.sh << END
make Dovecot automatically archive the servers copy of the email if/when a client connects to it using POP? I am trying to achieve the the same thing as Gmail's feature.
So I've tried finding a solution to this on my own for the past few days (believing it is a simple solution), but for the life of me I can't figure it out. I'm running 11.2 and I have Evolution as a start-up program and I have finally got my wireless working, so it connects automatically at start-up also. I figured out how to have network manager automatically unlock the default keyring, but...
Each time evolution starts up I am prompted to unlock the default keyring. How can I give Evolution permission / access to the default keyring automatically so it can check me email upon start-up without me having to type in my password?
I'm trying to send an email using mailx, in a bash script, but I can't get it to work. In the terminal I can, and this is how I do it:
Code: $ mailx johndoe@gmail.com Message^D
Within seconds of doing this, I get sent an email. The problem is with the bash script I'm trying to make. Among other things, I tried this:
Code: #!/bin/bash mailx johndoe@gmail.com < "Message" I honestly don't know what I'm supposed to do, and I've Googled a bunch of things too, and didn't have too much luck. Is there anyone who could help me out?
Edit: Figured it out. This is what I did, and it works for me:
Code: #!/bin/bash echo "Message" | mail -s "Subject" "johndoe@gmail.com"
I need to find a way to download the attachment from a daily report e-mail to me. The kicker is it will need to be down with a cron tabbed bash script.For example, which linux based CLI client is best suited to be scripted?
and i have to type bash everytime to get it running. i found a bandaid fix by having it type bash at the "custom command" profile preference... but i still consider the problem to be unsolved, because i still have to manually type 'bash' whenever i log remotely (putty). it seems like something happened, and now bash is not loaded by the terminal by default like it should.
PS: i already tried 'cp /etc/skel/.bashrc ~./bashrc', and it didn't seem to do anything.
I am using Ubuntu 9.10, fairly fresh install. Had one hard reset since I installed it, which could be when the problem started. Also, I was messing a bit with profiles, switching their ID's and adding/deleting them (not sure if that would be the problem).
Just recently I started using a cluster in my school to run some heavy tasks, which might require long periods to get done. The thing is that I can only send one task after the previous one has been completed (cluster use policies) and since I don't know in advance how much time each task will take, I'd like to have a way of knowing when one given task is already done (other than keep checking it).I am allowed to submit the jobs in bash scripts, so that's what I have at my disposal.
here is a fragment of a bash that i have to send an email when a back Up is made. basically a usr runs the sh - it asks for the job num and I read that from standard in . I do a tar of the drv to tape and I wish to send an email that job such and such has happened. My problem is : the subject of the email. I can set it - but if i try to use a var ( as in the $jobNum ) that Does work in the mail body - the subject comes out blank. I have also tried to buld a text file and shove that in the SUBJECT. my Q is how can I use a var in the subject.
# email subject ### so I have tried bulding and reading a tmp file -no works EMAILSUBJECT="/tmp/emailsubject.txt" echo "backup job" $jobNum >$EMAILSUBJECT
I just recently learned about the wonderful little lpr command- and using man -t (bash command) to beautifully print man pages for reference- but is there a way to print both sides of the paper using a printer so equipped?
I have a big bash script ,its goal is to download movie one by one . But I often get into a problem: if this script is executed in cron,it often does not completely download the movie.I often find the movies it downloaded are several KB while the movie is actually 20MB.So I think it is because it did not wait for finishing one task ,and jump to download another.So I want to know ,is there a way to force the bash script to wait until one movie downloaded completely and then start to download another movie ?
Bash's command history is great, especially it is useful when adding the history -a command to the COMMAND_PROMPT.However, I'm wondering if there is a way to log the commands to a file as soon as the Return key is pressed, e.g. before starting the command and not on completion of the command (using the COMMAND_PROMPT option would save the command once the prompt is there again).
I read about auditing programs like snoopy and session recorder like script but I thought they're already too complex for the simple question I have. I guess that deactivating that script logs all the output of the command would lead already in the right direction but isn't there a quicker way to solve that probelm?
I'm trying to write a script that will continuously ping a server and then send out an email when the server is down, and then when it is back up. Then, continuing with monitoring. I would like to not run this in cron, because I don't want to script to run with multiple instances.
For example, Ping a server every minute. -If successful, do nothing. -If failed to ping, then send out email stating that server is down. -Once ping is successful, then send out email stating that server is up.
I only want it to send an email once after a failure, so the end user isn't get an email every minute that it fails. Once it is successful, then send the email (one time), stating that the server is up. Then, continue to ping and if fails again, repeat the process.
I'm working in Debian, and I've noticed that when I resize my PuTTY window, the console inside isn't resizing to the new dimentions, so things like nano are running at 80xwhatever instead of the more useful 130xsomething that I've offered it.
set up BASH scripts on the server to automatically download and process data, and then upload it to my website. Is it even possible? Do servers allow website owners to place BASH scripts that can run automatically, or keep running indefinitely?