I want to record an internet radio station starting at 2:00am tomorrow morning. The specific program on the radio station lasts until 6:00am. The command I need to run to record the station is: Code:mplayer http://wjcu.jcu.edu:8001/listen.pls -ao pcm:file=indie_heat_of_the_night.wav -vc dummy -vo nullI'd use cron, but 1. I'm not sure how to and 2. it seems unnecessarily complicated for something that I only want to run once. If cron is the only/easiest solution, I guess I'll just have to resort to that, but I'd rather not.
I am running to a weird issue: for some reason when I try to run a crontab script which has these two operations: #!/bin/sh DATE=`date +%F`
but when I reduce this script to the first sql backup only and I use another cron job to put this dump file to the folder it goes through...weird isn't it?so 1+2 in 1 cronjob don't work but 1 in 1 cronjob and 2 in another cronjob it works perfect. any thoughts?
I am using Linux 64 bit Redhat Linux. I am trying to setup simple crontab as follow...1. Edited crontab file using crontab -e2. Listed the file once to verify it using crontab -l. This will display as.. 18 5 * * 2-3 ksh $HOME/testScript.sh > $HOME/testscript.out3. Logged in a root and restarted cron deamon using "/etc/init.d/crond restart"As per my understanding now my testScript should start running at 5:18 am Thuesday
I'm trying to back up my database daily at 2:30am. is this the right format? 30 2 * * * mysqldump -u root -pPassword database > backup_$(date +%y%m%d).sql
I need to actively make sure some files, in a specific directory, are chmod 750 and owned by transmission:media-daemons. Other users will save to this directory, with other permissions and UID/GID but I must make sureto reinforce this default.
So I have this on my /etc/crontab:
Code:
# /etc/crontab: system-wide crontab # Unlike any other crontab you don't have to run the `crontab' # command to install the new version when you edit this file # and files in /etc/cron.d. These files also have username fields, # that none of the other crontabs do.
is it possible disabling a crontab job without deleting the crontab description entry (by crontab -e)?I could also accept to change the entry itself. Now it's:0 0 * * 0-6 /home/me/cron/script.csh
I used crontab to set a file playing at a certain time, this works fine; however I want this to run even if no-one is logged in (but the computer is on). I can't get it to do this
Line is: 30 06 * * * env DISPLAY=:0.0 /usr/bin/totem /home/adunaic/Playlist.pls
Okay I think the problem was with needing a GUI.
Using this instead works, but i only hear the sound when I log in. I think something needs to be doen to start the audio perhaps?
So at work I usually forget to shut down my computer at the end of the day 5:30 and they switch the breaker off at 6pm which cant be good way to shut down a computer. So i was doing some googling and this is what i come up with.
this is the example i found online Code: #crontab -e -u root Code: 0 20 * * * /sbin/shutdown -h now since the above turns it down at 8pm i need to edit it to 5 30pm
Code: 0 17:30 * * * /sbin/shutdown -h now would this be correct?
Can anyone shed some light in this? Using Fedora 14-64, new install, 185 Opteron x 2 gig ram, sata hard drives formatted Ext4.However, in my home directory I have a folder for all my digital photos of which I have more than 20,000, and in another folder I have images and clipart of which I have almost 8,000. That is a lot of read only access to a significant number of files in my home directory.
How can I tell Fedora to not update the LAST ACCESS TIME of those files (specifically images) that will never actually be changed other than just being read. I want to leave that feature enabled for the rest of my home directory. I am trying t; improve my disk performance in Nautilus because whenever I access the folders with my images the system literally slows to a crawl and sometimes even the mouse stops working for several minutes until Nautilus has finished having its heart attack.
I am using openSUSE 11.2 and had the same issue once before in openSUSE 11.1 After some time, two domains I use very often can't be accessed any more, these are: Google and [URL]... (am doing some OO development recently). At the time where I write this post, only these two are "blocked".
When I try to access these two domains using Firefox, w3m or telnet, it tries to open the connection forever. Ping works fine. I tried opening the IP directly in the browser, but same problem, the connection never reaches. I know it's not a router issue, as the other computers can access these domains. Still, I tried to restart the router, but the issue persists. I tried restarting the network service (/etc/init.d/network stop + start), also nscd, without success. The only way I found to make it work again is to reboot. This problem reappears from time to time, after using the computer for a long time, and accessing these domains a lot of times. Note that I also use suspend to disk, so the "blocked" state is kept after resume. Only reboot "cures" it.
How do you find a file modified March 17, 2010, between 3:30 pm and 4:05 pm? I know that I must be missing something somewhere.How do you search for info like this? I goggled "search files time Linux" and got about 38,300,000 results. I looked through the first four pages and did not see what I was looking for.Do I need to calculate how many minutes ago that is and give that to find.I really want to do this in the GUI so that I can operate on the files found without typing in so much stuff.
How to refresh a page automatically?Say for example i need to refresh page in ubuntuforums to get new questions.I feel lazy to refresh the page often.Is it possible to refresh the page automatically in a specific time interval?I have tried ReloadEvery Firefox Add-on.But it refresh all the tabs.What i want is i want to refresh a page in a particular TAB.
I'm running Ubuntu 10.10 in a computer with a 7200rpm HD, 8Gb Ram, and a Core i7, and the time it takes to load the desktop is insane. There are threads that mention the problem of gnome looking for nonexistent floppy drives and to solve the problem by disabling that in the BIOS (option I don't have). Anyway, besides that problem, by running at startup a gnome-terminal with the iotop command, I noticed that two processes have a huge i/o load on the system: google desktop and ubuntu one. I would like those programs to run as part of the startup process, but be launched after several seconds (to allow the rest of the programs to load). Is there any way I can achieve this? I think there should be a way modifying the commands under startup application, but I cannot find anything that works.
I'm looking to find a way to schedule my computer to wake up at say 7:00am. Every night before I go to bed, I put my computer into suspend so the fan doesn't wake me (old computer). I can't seem to find a task scheduler that allows me to be able to wake the system.
How do I get vlc to start at a specific time in a media file. I can do it easily in mplayer2, mpv but not vlc. Let's say the media file is named abc.avi (file format,container not important for this example.) Also the running time is around 2+ hrs. In mplayer2 and mpv I can simply do :-
I have looked high and low for this, but to no avail. I am looking for a script to shutdown my computer when it reaches a specific time (say 10:00). I know about the shutdown command but if I use that I cannot shutdown my computer manually. I would like to be able to shutdown manually AND automatically.
I've been quite puzzled by the behavior of unattended updates - it seems it runs the updates at or before 7am (on all the server installs I've got). They are only set to do security updates unattended, but what I am wondering is if there is a way to change the time that they install? I want the updates to install early in the morning, at like 3am or such, so I can reboot the machine when I get up if needs a reboot.I found a few mentions about it, but nothing specifically talking about the time.[URL]..
I've just put together a 10.04 server using a GA-D525TUD atom board. I've installed gnome desktop too. I'm trying to get 4 sata drives to go into standby, but hdparm seems to be ignoring the settings. If i issue hdparm -y /dev/sdx the drive duly goes into standby and stays in standby until accessed. I've put entries in rc.local along with the hdparm.conf settings, but still no joy.
I am trying to extract just a few files/folders from a Time Machine backup, but can't seem to find them. The drive is automatically mounted in Ubuntu, and am able to access after enabling view hidden files, the HFS+ Private Data Directory. But that is a jumbled array of thousands of numbered folders, with each taking a fair amount of time to open on this aging Dell running 10.10.
I've tried running the standard Places - Search for Documents, with 'Show hidden and backup files' enabled, but that won't pull up any of the search times I'm going for (and seemingly won't find anything at all on the drive). So, is there any way to decipher the directory tree so as to be able to access this data from Ubuntu? Or perhaps a file embedded somewhere in there that lists out the original structure, so that I can use it as an index to see what number correlates to what originally named folder?
When I leave my computer and later return, depending on how long I've been away, I find the screen-saver running, the screen turned off, or the computer suspended, i get the error as specified below on a black screen:
Quote:
Ubuntu is running in low-graphics mode.
Your screen, graphics card, and input device settings could not be detected correctly. You will need to configure these yourself.
The message gives several options. Despite some experimentals, The only solution which I have found is to reboot and everything becomes fine as before.
Is there a way to set Linux to automatically log in to a specific user account and at the same time lock the screen? I want to save time and trigger various software that always should start up on boot, while leaving the computer unattended during startup (extra important and practical for remote control boots), by enforcing a 'screen lock' so that no-one can see what happens behind the login screen without entering the login credentials.
I have a question about using crontab with /etc/crontab...
I had a cron job that I needed to run as root. At the time I thought that sticking it in /etc/crontab would be a good idea. However, I used the crontab command to edit /etc/crontab, which I guess is not standard procedure? Specifically, I configured /etc/crontab as my local user's crontab (i.e. sudo crontab /etc/crontab) then added my cron job as I would a local user crontab (i.e. sudo crontab -e).
Originally, my cron job looked like this:
30 * * * * root /my/batch/script &> /dev/null
After adding the new cron job I started seeing errors. Something to the effect of "can't find command root" or something similar. So I removed the 'root' user definition from the cron job and the job started running fine. However, because this is /etc/crontab, there are other system related cron jobs that have been defined to run under the root account (e.g. "17 * * * * root cd / && run-parts --report /etc/cron.hourly" runs as root, etc.). So these pre-existing system cron jobs, which up until now have been running smoothly, are now generating "can't find command root" errors. But I think that the system cron jobs _are_ successfully being run someplace because logrotate seems to be working.
So what I _think_ is happening is that /etc/crontab is being run twice: once as the system crontab, and once as my sudoed local user's crontab. When I run crontab -l I see nothing, but when I run sudo crontab -l I can see the contents of /etc/crontab. I am reluctant to delete my sudoed local user's crontab, because then in the process I would be deleting the system crontab, and I do not know how I should restore the system crontab's contents. (I am still not sure as to the most appropriate way to edit the system crontab).
How can I get out of this mess? I want /etc/crontab to go back to the way it was before--running _once_ as the system crontab. As for my new cron job, I'm willing to reconfigure it anywhere so long as I am still able to run it as root. Any ideas? (I am using Ubuntu 8.04 Server LTE)
I have installed an application manager(monitoring application) on my linux server. Now, i need to have backup schedule for my application. The application itself has executive file to backup database.But when i put this file in my crontab to schedule the backup program it wont run!50 09 * * * root /opt/ME/AppManager9/bin/BackupMysqlDB.sh