General :: Copying Files From Multiple Locations To One Destination
Feb 21, 2011
Suppose I have a tree structure like this:
/home/mahmood/sim/a/b/file1.cpp
/home/mahmood/sim/a/b/file2.h
/home/mahmood/sim/a/c/file3.txt
/home/mahmood/sim/d/file4.txt
How can I copy all of them to /home/mahmood/sim. So that when I run "ls" in /home/mahmood/sim, I see all files:
file1.cpp
file2.h
file3.txt
file4.txt
Can 'cp' search for all file and copy them in another folder?
I have 60+ directory's each containing multiple .doc files. I need to move them to a single directory and keep their file name intact. I don't think cp will do that with out listing all the file names. I was thinking of something like: cp -r /dir/*.doc /newdir . Or should I use a combo like find -type *.doc|cp /newdir?
Description: I am a newly appointed system engineer taking care of linux servers. We have a new set of data coming in which need below configuration: How to do a script with function?:
for files with ".txt" in sm copy each of the files to folder : sm1 and sm2 (log every copy) if succesful: remove original log into the log file if not successful: (not successful copying 1 particular file to all the folders) retain and retry log into the log file mail out the admin with that particular file name
I have already do try a bit: cd /export/home/ for dir in sm1 sm2; do cp -p sm/*.txt $dir/ done Is my starting right? How to do the rest parts?
i want to copy one or more files or directory from one drive to multiple drive simultaneously. It like a cloning a disk. But i dont like clone entire disk. i want to copy/clone only certine files or folder. if any one can know how to copy one source to multiple destination simultaneously.
When you run the following cp command in the BASH terminal, how does Linux know which files are the source and which are the destination when copying multiple files from one location to another?How does Linux know that the services, motd, fstab, and hosts files are the source and the /home/fred/my_dir is the destination?This question came up in a Linux class and I was not sure of the answer. I was thinking it is based on the source path entered ending with a file path and the destination being a directory, but was not sure.
I have hundreds of directories in various subdirs that I need to remove. I want to remove all of these dirs, but can only find solutions on how to do remove files (or how to remove subdirs from within the current dir).
I think I need something like
find -iname 'testfile*' | xargs rm -i
where I want to remove every directory that contains the word 'testfile' within the directory name. I know xargs wont work for dirs,
I am working with DM355 target board. Here we record. The video coming from IP cameras. Now I have to write c program for copying. The recorded avi files with date and time to NAS server using scp. I wrote a script to copy single file to NAS server.
My company has 2 locations. I have a server running BIND, Apache, and MySQL. I'm setting up a second server just in case the primary goes down. I'm sure it's bad form to it the way I'm doing it, but how might I go about configuing my backup? Should I do it as a Secondary DNS server?
I have recently purchased an external hard drive in order to backup my home partition. In my PC I have a "1.5T" drive with several partitions on it, containing OSes and the home partition. The home partition is 1.3T according to df, the external drive contains one partition that spans the entire disk,df reports it as 1.4T in size. Both partitions are ext3. When I use rsync to copy files from the home partition to the external partition, the external disk becomes full, despite the destination - supposedly - being larger than the source. I don't understand why copying files from one partition to a slightly bigger partition should need more space than on the source partition. Does anyone know what is happening ?
Details : I created the partition on the external drive with gparted; gparted reported it the already have several gigabytes in used space immediately after the partitions creation - I thought at the time that this must be normal. The home partition contains many files of all sorts, including lots of big audio and video files. If you are wondering, for all my important files this external disk is only secondary backup, as they are also backed up to the "internet".
These are the mount points :
/mnt/tmp/ : home partition, /dev/sdb6 /mnt/external/ : external partition, /dev/sdc1
I used rsync to copy the files, I know there are more efficient ways to do this, but I wanted to use the same command that I will subsequently run to sync the backup.
Next I tried adding the --sparse switch, as I was wondering if the problem may come form sparse files. I don't know however if rsync would go back and shrink the sparse file by just adding the switch and executing the command. I also added --one-file-system, for good measure. Here is what I ran next :
rsync: writefd_unbuffered failed to write 4 bytes to socket [sender]: Broken pipe (32) rsync: write failed on "abcd.avi": No space left on device (28) rsync error: error in file IO (code 11) at receiver.c(302) [receiver=3.0.6]
[code]....
Looking at the destination after a partial copy seems to indicate that the problem is not symbolic links being "expanded". I have not checked the source filesystem for sparse files, nor the destination to see if these files could be larger there, as this does not seem trivial.
Here is some additional info :
$ df /mnt/tmp/ Filesystem 1K-blocks Used Available Use% Mounted on /dev/sdb6 1415342836 1414173740 369096 100% /mnt/tmp
We have an rsync cron job set up to mirror all the files in a "..dashtdocsdocs" folder to the same folder on another server. It copies all the files over correctly and deletes any files in the "docs" directory that aren't in the sending directory, but it also deletes any files we put in the target directory's parent folder (..dashtdocs or other subfolders like ..dashtdocsimages) even though they've been excluded in the .rsync-filter file.
So for example server A has ..dashtdocsdocs and ..dashtdocsimages. Server B has ..dashtdocsdocs but if I manually copy the images folder over to ..dashtdocsimages, the images folder gets deleted from the target directory every time rsync runs.
I'd like to keep just the docs directory synched and update other folders manually, but they keep getting deleted. It looks to me like it's running a delete-excluded option, but that option wasn't used.
I have two storage drives that I will be sharing by FTP. One is internal 1TB ext4 HDD and another one is an external USB 1TB NTFS HDD. Both drives get mounted to /media and I am trying to set an additional mount point for each. For internal HDD everything works perfectly. I simply went to /etc/fstab and copied the line related to it. Now I have:
Code: /dev/sdb5 /home/eugene/.MOUNT/sdb5 ext4 defaults 0 0 /dev/sdb5 /media/sdb5 ext4 defaults 0 0 which does exactly what I need.
I tried doing the same for the USB drive which produces unexpected results. The lines are Code: /dev/sdc1 /home/eugene/.MOUNT/sdc1 ntfs defaults 0 0 /dev/sdc1 /media/sdc1 ntfs defaults 0 0
This has the following results: - in /media/Y (Y is label of this HDD) I have this HDD and can access all its contents which is good - in /home/eugene/.MOUNT/sdc1 I don't have anything and this is bad - in /media/sdc1 I have only one folder from this HDD and this folder is empty (on the HDD this folder is not empty) and this is somehow weird.
So I've got a home server hosting a website for my restaurant, but I'd like to get another server up to get some redundancy going.
I have another machine I'd like to set up at another location to take over retrieving requests sent for the website whenever my home server goes down. I've got my domain through [URL], but the domain is hosted through [URL] for their dynamic dns service (because im not using a static ip).
So I'm guessing having another server set up is just a matter of setting up dns records, however I don't know where to begin with setting that up. Any words of wisdom out there?
Today i am trying to learn how to use sed. I set up a testing folder with the following files:
AAb.lol AAc.lol AAx.lol test.sh
My goal is to create a script (test.sh) which renames all the files to their original name without AA. I want to end up with this:
b.lol c.lol x.lol test.sh
sed seemed to be the perfect tool so i went ahead and created a script which i think should clear the job.
[Code].........
mv: missing destination file operand after `$i' From that 2nd line i can tell that $NewName is just empty. I also read something about sed needing the -e option for scripting purposes but i just don't understand it.
Is there a way to sync Tomboy notes to multiple locations? I would like to be able to sync them to my UbuntuOne account and at the same time to my local NFS server, but from the looks of it Tomboy only lets you choose one location for syncing. Maybe there's a workaround for this or something?
I am looking for a solution where the data to be shared across multiple locations through login. In this regard, the following options are required.
1) The files/folders created can be viewed by the group and the owner has an option to edit and delete. However, if required the owner/admin can grant permission temporarily.
2) log of files/folders being accessed
3) The access of files/folders is based on username and password.
One thing striking me is FTP Server to enable the above solution. If so, i just want a step by step guide. Otherwise, let me know any other better options.
Missing ifcfg-eth[2-5] fileset for ZNYX 345Q Quad Port 10/100 cards. I have showing in the gui network device that my ports for my ZNYX ZX345Q Quad Port card my ports are Auto eth2, Auto eth3 etc. My Motherboard and Intel cards show as System eth0 and System eth1.
There ARE corresponding entries for those in my /etc/sysconfig/network-settings/ directory, but there are not ifcfg-eth[2-5] files to correspond to these adapters. Can I just write my own files and that will do it?
How does Fedora 12/13 load these drivers into the kernel without having these ifcfg files?
I'd love to know if there is another way Fedora controls NICs / other system resources.
I have a new network attached storage unit that I'm trying to transfer my data to. On this NAS, it has a very basic linux installed with SSH enabled. Browsing through the programs installed on it, i found smbclient. Am I able to copy files directly from my old NAS to my new one using smbclient?It would sure beat transfering 950GB from my old NAS through a computer then onto my new NAS.
I am running openSUSE 11.2 KDE 4.3.4 on my ThinkPad R51 laptop.I would like to copy 'File A, File B, File C' (for example) that has photographs of a holiday , to a blank CD inserted into my drive and cannot find a way to 'copy and paste' into the volume.
I just read the Linux scp command issue question and it reminded me that I regularily forget to specify the colon in the host part of a scp command, and thus copying a file locally instead of copying to a remote host, e.g. I do scp foo host instead of scp foo host: But I never use scp to copy a file locally. So I wonder if there is a way to make scp fail if both (the source and destination) arguments refer to local files.
i am accessing linux through putty and i wrote somany programs in unix using putty and gedit but now i need to copy all files into windows. how to copy directory(linux) to folder(windows) without installing any softwares?If it is necessary to install software to copy files then tell me the process of using that software.
I'm trying to copy files from my current server to a new server. Both servers have SSH installed. These are the commands I'm using. However I'm getting connection refused. I did a google search and found out that maybe the reason of this error could be due to the fact that my current server doesn't have SSH. However I use SSH often on my current server so I can say that it has SSH for sure.
OpenSSH_4.3p2, OpenSSL 0.9.8e-fips-rhel5 01 Jul 2008 debug1: Reading configuration data /etc/ssh/ssh_config debug1: Applying options for * debug1: Connecting to IP_OF_CURSERVER [IP_OF_CURSERVER] port 22. debug1: connect to address IP_OF_CURSERVER port 22: Connection refused ssh: connect to host IP_OF_CURSERVER port 22: Connection refused
I need to copy a large number of files, it comes to 1 lakh from one server to another. When I tried various commands using scp , ftp etc. It is saying "Arg list too long". In which way can we copy all the files. The Two servers are under Linux.
I've a directory containing around 2.8 lacs of files. I want to move them to another directory.If I use cp or mv then I get an error 'argument list too long'. If I write a script like
for file in ls *; do cp {source} to {destination} done
then because of ls command , its performance degrades.How can I do this?
I am facing problem in copying a large number of file 18 lakh (18,000,000) files from my personal hardisk to another hardisk each file is very small and size of folder is around 3.95 GB copying files using copy given by Windows is frustrating and I am not even able to compress file its giving me error that its not readable.And problem is I am not able to open this drive in Linux it showing me error there saying do diskchk in Windows and Windows disk check is also not able to repair this drive and goes into some mode unsolvable.Is there any way to open disk with error to open in Windows and if not any way I can copy data faster?ERROR: Disk labled EDU is corrupt go to windows and chkdsk /f there and reboot into window 2 times.