General :: Scp With Range Of Files To Be Copied To - From - Or Between Different Hosts - Error - No Such File Or Directory
Jun 27, 2011
Working fine: ==> scp my_log-bin.01393[0-9] root@192.168.103.66:/backup/ error - No such file or directory: ==> scp my_log-bin.0139[30-99] root@192.168.103.66:/backup/
How to set the default file permissions on ALL newly created files in linux - but differs in important ways:
I want all files created in (or copied to or moved to) a certain directory to inherit a set of default permissions that is different from the system default.
Rationale: The directory in question is the "intake hopper" for an application. Users in a group place files in the directory, and the app (running under another user id in the same group) takes them and processes them. The problem is that the owner of each file placed in the directory is the user that placed it there, and the permissions are defaulting to "rw-r--r--"; I want to change that to "rw-rw----". The app doing the intake can't do that explicitly, because the user id the app is running under doesn't own the file in question, and the default permissions don't allow the app to chmod on the file! Obviously, the user could do a chmod after putting the file there - but I want to keep the "drop" by the user as simple as possible. (These folks are not linux-literate, they just drag and drop the files from their windows desktop to a (Samba) network share - i.e. they don't even know they are interacting with a linux system.)
umask seems too powerful: I don't want to set default permissions for every file created anywhere by these users - just those created in (or placed in) this directory.
I'm sure that the issue I'm having is easily solvable once I gain some understanding about copying files - and file permissions in Ubuntu. Here's my situation:
I have an external HDD where I like to back up some files (I mess around with distros on my main machine and feel less stressed knowing the important stuff is backed up). I have an ext4 partition on the external drive where I have copied files, both through the terminal (cp 'filename' /dev/sdc3) and by drag and drop (gnome-terminal).
The problem is, once the files are copied, most are inaccessible. I can view them, but some directories and individual files say I do not have permission to open them. Others are accessible. This is from the same user profile that copied them.
How do I see what's going on? More importantly, how do I make files on external drives available to any user or OS (that can handle ext4)? I want to make sure that if my whole system gets effed that I could still do a reinstall of my OS and then access those backup files.
How would one block an IP range access to a Debian-based Linux system for say 47.1.1.1. - 48.255.255.255? Would it be with the hosts.deny file? If so, how would it be written in the file? Also, would the system require being restarted for the changes to take effect, after writing to the file?
If I execute the following command: cp -R /myfiles /mydestination
If myfiles contains several sub-directories and files, in what order will they be copied? For example, directories might be named 0123a, 9993c, myfolder, xfolder.
They are not copied in alphabetical order OR in date order OR in the order they appear when using a standard ls command as far as I can tell, so what actually does determine the order?
Edit: I am trying to determine the order that the cp command uses in order to determine how far along my copy command made it before it stopped. For example, I was hoping to be able to determine it copied 3 of the 4 directories successfully.
I have an embedded linux system (Debian 'Lenny') which booting from a microSD flash. If I make a copy of a file on the flash file system (cp test test1) and then power off (disconnect power spontanious). Connects power again and the system come up, but the file test1 is gone. How can I secure that test1 is NOT disappear if the power get lost?If I copy file and then restart system with reboot command, the file test1 does not disappear.
I recently had data recovered and it was sent back to me on what I think is an NTFS drive. I copied all the files over to a file share I have on a Linux box, that's ext4. Now I have that share mounted on my OSX machine, and I can't move or rename most of the files. However, in a couple cases I was able to rename a folder after the third try. Another time I was able to rename a folder once, but not again. All the permissions are showing up the same on the command-line -- I can't see any differences between the permissions on any of the files/folders. Note that I can create new folders and add files no problem, and then rename and move those all I want.
I have ran into a permissions problem. It seems that any file I have created with Fedora I can browse to in Firefox. I cannot browse to files that I have downloaded from the internet or copied locally across our network. I changed file permissions to 755 and 777 for these files. I even thought that maybe it was the php file content that was causing the issue. So, I copied the file content from file 2 into file 1. File 1 was the file I could always browse to since I created it in Fedora. I could still browse to file 1 after it had file 2's content.
I look at the permissions and they appear the same for both files. I am not sure why I cannot open the other files. I downloaded phpMyAdmin. I cannot hit any of these files in a browser. I also copied some files from a a backup location we have internally. I cannot browse to any of these files either. I used chmod 777 filename. Even after doing that I could not hit any of the files. I moved the files into my document root directory ( /var/www/html ) and I still cannot browse to these files.
I have two systems running on linux. system one is running with RHEL 5.4/X86_64 hardware, system two is running with RHEL 5.3/i686 hardware. One filesystem is shared from system two and mounted as NFS on system one. Now when i do a copy from local filesystem to the NFS share from system one,it shows as follows
Quote:
-rw-r--r-- 1 xkinved rbak1 30 Mar 3 2011 king
But if i do copy with -p option then it shows right time stamp. Both machines are running with slight(minutes) different in time. Does this could be cause for this problem? The problem is happening while i do FTP from some other machines too.
I have iomega appliance, which is based on Debian distribution. There is an NFS share that I have created which is without password. Since it is without password, there are some viruses copied. I want to find out which IP address is the source of these files. In other words, I want to know which PC is copying these infected files on the NFS share
I'm currently stuck at 6.13 GMP-5.0.0 of the LFS installation. After running make, I receive the following:
error while loading shared libraries: libbfd-2.20.so: cannot open shared object file: No such file or directory make[2]: *** [libmpn.la] Error 127 make[2]: Leaving directory `/sources/gmp-5.0.0/mpn'
I'm trying to use ssh-keyscan to get some known_host file population going on, but I have a ton of hosts I want to scan, all with multiple aliases in /etc/hosts. Is there a way to use my current /etc/hosts file to do an ssh-keyscan instead of making a special list of hosts that (from what I've read) ssh-keyscan needs?
When I try to list files in directory. I am getting i/o error #ls -l /test I am getting i/o error. Why I am getting this error and what are these i/o errors.
If I need to append a set (or sets) of data to a file(or files) on remote hosts what is the best mechanism by which to do that? My first thought was ssh but the command syntax to append to a remote file isn't clear to me. Can anyone point me in the right direction here?
I copied g95 executable under /usr/bin which is clearly in my PATH. When I type g95, it says "bash: /home/username/bin/g95: No such file or directory". Why doesn't it look for somewhere else that is in my PATH?
the system currently have a directory with all the invalid files. how bad is it to move a single file to a directory containing 3 million files already?
I would like to create a cronjob that will delete all files within a directory 1 hours after it is created to the folderI found this cron find /path/to/file/* -ctime +1 -exec rm {} ; but it's deleted all files.I want to make an exception, all file should be deleted except one file (letsay file a.zip)
Probably an easy (which means stoopid) question...I am trying to reroute a website using my hosts file so that it matches my servers certificate file for testing without effect dns and the live site.When I went to edit my /etc/hosts file it is non-existent. I have, I am assuming in it's place, hosts.allow and hosts.deny. Can anyone explain why I do not have a hosts file?
I am trying to compile the code. I am getting this following error. Am I missing any package?
x86_screen.h:12:22: error: X11/Xlib.h: No such file or directory In file included from osal_event.c:12: x86_screen.h:25: error: expected specifier-qualifier-list before �Display� make[1]: *** [osal_event.o] Error 1
I am migrating an application from HP-UX to Redhat Linux and I have some files of scripts. In one of the script files I have the line /pylive/smartPAY/spserver/scripts/sp.con which is the correct path and file name for an existing file. However I get the message "No such File or Directory/pylive/smartPAY/spserver/scripts/sp.con" The script worked fine in HP-UX, why not in Redhat Linux?
I am trying to compile a memory map module cmem.ko from TI codec tools. This is for the DSPlink software development for the OMAP3530. I am compiling this on the host(Ubuntu). First error could not find cc1 and i re-installed gcc and added it to the path. That fixed that error but then I get to this:
cmem.c:38:19: error: stdio.h: No such file or directory cmem.c:39:20: error: unistd.h: No such file or directory cmem.c:40:22: error: sys/mman.h: No such file or directory
When testing my script for a question, I find that when I execute the expanded commands, I get no error message, but when I source the file with the command, I get 'No such file or directory'. Here is what I mean:
I am trying to compile a file named process_30sec.c(thats a C program) through the Ubuntu terminal and it is showing the following error in command terminal: process_30sec.c:15:23: error: asm/bitops.h: No such file or directory I have just started using Ubuntu&Linux so I am not familiar with lots of commands and things going around.I am using the following command to compile the process_30sec.c file:
kutub@ubuntu:~/Desktop$ gcc process_30sec.c
The above mentioned program file contains a header file <asm/bitops.h>, that I need to use in the program.It seems that this header file is not recognized by the system, and it needs to be included somewhere in the system so that the terminal would allow me to compile this program.
I'm trying to install a WiFi Wireless IEEE 802.11N/G/B WLAN 150Mbps Network Adapter USB2.0 on Slackware. I get the following message when I enter "make install." cannot stat rt3070sta.ko no such file or directory