If a user want secure coping of files through scp from one system to another system then how this can be done. This user have only been granted securing coping of files through another system but without ssh.
A directory /test/test1 is created & under /test1 there are some files & subdirectories with data in it. I had copied the files (text & script files) with command as,
cp -irv /test/test1/.* /test as per the requirement but what i see in destinarion i.e, /test that no files or directory has been copied in /test & the files/directories is also removed from source i.e, /test/test1
so my query is how can i get the files/directories back with data?
Q. - "Next we will copy all files, directories and sub-directories from your current /home folder into the new partition: sudo rsync -axS --exclude='/*/.gvfs' /home/. /media/home/."
I would like to know what the dots in /home/. and /media/home/. mean? Or to put it other way; will the command bellow copy my /home to /media/home (on different hard disk)?
I often find myself working in environments that force MS Outlook down my throat. I know there must be others here who have dealt with this frustration, so I'm hoping to get a convenient workaround.My strategy so far has been to make an Outlook rule that immediately copies every email to a PST file (for export). This is the first rule that runs. The pst file is kept directly on network filespace, so the linux and windows sides can both get at it. On the linux side, a pst to mbox converter runs, and then emacs can access the mbox mail.
It all works, but it's cumbersome. The script must periodically move the PST file and immediately copy an empty PST file in its place. I would have appreciated it if Outlook would only handle the file when adding mail to it. The problem is, after I replace the file with an empty PST file, Outlook pukes, and even deactivates the rule for copying emails. So I must manually close Outlook, empty the PST file, then relaunch Outlook just to run the rule to copy email.Is there a way to use cygwin or something to launch outlook in the background, run outlook rules, and exit, every 10 minutes?
Well, I am facing problem when doing lab questions.
I must use DLXLinux bundled in Bochs (bochs.sourceforge.net).
I am required to use the /usr/local directory.
In /usr directory, there is no directory named 'local' but there is one thing called 'local@'. So, when I try to use mkdir command to create 'local' directory in /usr , there are error "cannot make directory.....".
I want to run a cronjob every 15 minutes that checks a directory for files. If the directory contains more than ten files I want it to send an email to me.
All I have is this...
*/15 * * * * ls -l | wc -l | [filename] | mail -s "This is just a test" [email address]
I would rather not write a bash script. Is there an easier way to do this? I was looking into some commands like find and grep.
I'm quite new to linux but I have configured a simple ftp server and it's working great. I have a FTP-Shared folder with upload and download subfolders. Under upload's and download's I have identical category subfolders like mp3's, movies, software etc. in both. As the guy's upload, I would like to create a line crontab where I can move all the content under /FTP-Shared/upload/mp3/* older than 14 day's to FTP-Shared/downloads/mp3/ recursively (Like in cp command), but the timestamp must be searched on the first directory and not sub files example: /mp3/Club Dance/CD1/Hallo world.mp3This is how far I got:[root@clients ~]# /usr/bin/find /FTP_Shared/upload/Mp3s/ -depth -mindepth 1 -mtime +14 -type d -exec mv -f {} /FTP_Shared/download/Mp3s/ ;This command moves the directory and files, but it is not recursively
I've just installed Ubuntu 10.04, trying out linux for the first time. I'm wondering if it's possible to link my documents folder in Ubuntu to my documents folder in Windows, so that when I open the folder in Ubuntu, I am linked to the folder in Windows.
I'd like to mount the directory /var/www/mysite to the directory /home/daniel/mysite, but also have the user of the mounted files mapped from the original user (www-data) to my own user (daniel). So that the file /var/www/mysite/index.php who's user is www-data will appear in the mounted directory as /home/daniel/mysite/index.php and be owned by daniel - and alternatively, if I create a file /home/daniel/mysite/test.php with my own user, it will be created in the original directory under the user www-data Is it possible? If not, what alternatives do I have so I can use an IDE and still make sure all the files belong to the HTTP server's user?
I am in my current directory. I want to copy a directory somewhere else into this current directory. Lets say I want to take it from direc1/direc2 and the directory I want to take is called demo.
Code:
That is what it shows in the man pages, but when I do that, it says cp: no match
If I have a directory /foo with a few files in it, how do I symlink each entry in /foo into /bar/? For instance, if /foo has the files a, b and c, I want to create three symlinks:
How can i auto mount more than directory in the same directory ? i want to automount 2 home directories in the /home and still be able to enter the other home directories !
The problem that i've another account on the system with home directory joe when the user1 home directory auto mounted i become unable to enter joe home directory !
I'm using Mac OS X's Terminal.app shell to compile and run Fortran programs. One such program resides outside of my home directory (it is in the Applications folder, which resides on my hard drive but seems to be outside of my home folder). How can I navigate into this directory using Terminal.app to run the programs that reside there?
I have a directory /sanvol/ifr/ which contains 90,000 files and 2 subdirectories containing 60,000 files. I need to copy the contents of /sanvol/ifr to /sanvol/edm/. The cp command will not work because there are too many files.
I tried the following: find /sanvol/ifr/ type f -name '*' -exec cp -p {} /sanvol/edm/. ;
but this copies all files including whats in the subdirectories to /sanvol/edm/
I need to replicate exactly the contents of /sanvol/ifr/ to /sanvol/edm/ and preserve file attributes and dates.
I have an interdependent collection of scripts in my ~/bin directory as well as a developed ~/.vim directory and some other libraries and such in other subdirectories. I've been versioning all of this using git, and have realized that it would be potentially very easy and useful to do development and testing of new and existing scripts, vim plugins, etc. using a cloned repo, and then pull the working code into my actual home directory with a merge.
The easiest way to do this would seem to be to just change & export $HOME, eg
cd ~/testing; git clone ~ home export HOME=~/testing/home cd ~ screen -S testing-home # start vim, write/revise plugins, edit scripts, etc. # test revisions
However since I've never tried this before I'm concerned that some programs, environment variables, etc., may end up using my actual home directory instead of the exported one. Is this a viable strategy? Are there just a few outliers that I should be careful about?
I am using my media server as my podcast collector. I am in the process of learning the ins and outs of NFS so i can mount a NFS directory and transfer my podcasts from server to player. For now i am using scp to transfer podcasts from server to desktop then to player. The problem is the path to the directory of one of the podcasts is /home/user/gpodder-downloads/The BILL&TIMMY Show Podcast.
whenever i try and run my scp command it fails because it thinks that TIMMY is a script i want to run in the background. I have tried to back-slash escape the character, i've tried single quoting and double quoting the character and i still get the same problem. as it sits now i have to move all podcasts to another directory and then transfer them to my desktop...but i would like to transfer the podcasts without un-necessary steps.
I have a server with Fedora 13 with which I would like to get NFS working. I have looked up multiple howto's and tutorials, but I'm having a problem not addressed by any of them.Official how-to, another how-to, and another how-to.I have verified that nfs-utils, nfs-utils-lib, portmap, and system-config-nfs are installed and running. I have verified that I have, in fact, shared the directory that I want to share, and that the proper permissions are set.
I had to go through some gyrations to get the Belkin wireless N router to allow my server to have a static IP. However, I can ping the server from the nfs client (a toshiba satellite running mint 8), and vice versa. I have (for now) disabled firewalls on both computers. I think I have disabled SELinux on Fedora 13 (for now).When I attempt to connect to the server from the client, the output looks like this:Quote:
aragorn ~ # mount -v 192.168.2.101:/test /home/kelev/test/ mount: no type was given - I'll assume nfs because of the colon mount.nfs: timeout set for Sat Dec 18 12:21:09 2010
I've done a low level format on them so they're completely empty. When I use them with my windows machines, they're absolutely fine. When I plug them into my Ubuntu machine, there is a hidden directory created called 'RECYCLER' which I'm assuming is for deleted files?However, it also creates a .exe file in this directory called 0x2D9FA278 which has an Icon with an H in it and a comment of 'Facebook Photo' This has the effect of making all the directories on the stick into shortcuts! I googled the file name and it seems to be some sort of Trojan, but I don't understand how it's go into my Ubuntu machine, I've scanned with ClamAV and it finds nothing.
I downloaded a mouse theme form gnome look and installed it in the themes. But it has not appeared in the pointer themes section in custimation even though it said that it is installed correctly.When I drag the file to install it again it says something along the lines of it cannot copy a directory over a directory.Where can I find where the mouse/pointer theme is located and delete it. I have searched filesystem, google and these forums and not had any luck yet.
Recently I mounted a larger partition into my home directory since I was running out of space, Everything went smoothly, but it caused me to wonder about something I cant figure out. While playing with the mount unmount commands when I was copying everything over... before editing my fstab.
Is there a way to access the files that existed in a directory before you mount a partition to that directory? after mount the original files are gone.unmount and they are back, Where do they go?
I want to update all the machines in the network from a central repository which is on my master server and whose archive directory is shared through samba.I searched in the man page of sources.list and found that there is an option for this but can't able to implement this. Can anybody kindly tell me the way to do the same.