Software :: Directory Listing Of Files That Don't Exist?
Oct 16, 2010
I think as a result of a script that started duplicating files in a loop, the allotted capacity on my VPS filled up with multiple nested copies of the same files... After a reboot, I could delete most of them, but got rm stalled in certain directories...
after isolating which ones, I found this: a directory listing that lists the files, and at the same time tells me they are not there!!!
ls: cannot access userkey.php: No such file or directory
ls: cannot access workshop.php: No such file or directory
ls: cannot access quiz.php: No such file or directory
ls: cannot access webservice_rest.php: No such file or directory
If I runls -R1I get a recursive listing of all files under the current directory.However, if I dols -R1 *.avi, ie I want to search only for files with the file descriptor .avi, I get an errorQuote:ls: cannot access *.avi:No such file or directorySo it seems I am using ls incorrectly. What's the correct way to use wild card pattern matching when using the -R switch? Or maybe that isn't possible?
how to check two arguments whether it is a file and the other is a directory, check if files exist in directory and copy it if it is a newer version. Seriously i need idea to do tis, jst some simple sample wil do
new to ubuntu and linux, and using Lucid 10.04 LTS ok, i'm trying to get XBMC going, and following a nice step by step instruction on wiki.xbmc.org, but now i'm stuck at this step... In Ubuntu the SVN Repositories are not automatically added. You must add them manually. First, download the SVN Repo Installer from: [URL] Extract it to the ~/.xbmc/plugins/programs directory. If this directory does not exist, run XBMC one time and then exit back to Ubuntu. The directory should now exist.
i ran xbmc once (after the video driver problem was solved, though it messed up my dual monitors, gotta figure that one out yet.) i got the zip file, it's in my downloads. now my newbness really shows... i can't find an 'extract' command for gnome, can't find the .xbmc directory using the file browser, can't figure out how to hunt for folders instead of files, and don't know how to look inside folders i don't have permission to, that is, i don't know if there's a 'sudo' like option for the file browser. i've been searching the forums, but without the correct search terms, i'm wading in an ocean. i really want to give ubuntu an honest try, but i feel like a foreigner. EDIT: btw, up to this point, the forums have been invaluable, you all are great.
I have RHEL5 with apache installed and running. I have a directory off of / called software that I would like to have show up as a directory listing so if I go to [URL] I get the directory listing. I have tried searching around and nothing seems to work for me. So out of the box how do I get this to work? My server properly resolves to [URL] so I am good there.
Within the mv() command is there any way to not need to type the entire directory location? In the example there are directories with varying depths... I want to move a file from depth 2 to a directory of depth 3 when the directory of depth 3 is within the directory of depth 2. Basically, can I cp() andor mv() a file to a subdirectory without having the enter the entire directory location?
[root@xxx dir]# ls dirD1 [root@xxx dir]# cd dirD1/dirD2 [root@xxx dirD2]# ls
This is the script I'm running tar tf some.tar somefolder_insidetar And output it's a list with all folders, files, and SUBDIRECTORY Files, the only thing I need it's just show the contents (folder and files) of the current directory choosed, not listing subdirectory files, or subdirectories inside subdirectories.
I have set up Ubuntu 9.1 with Samba etc. as a web server. The files are in /var/www and are shared through Samba as web. The root directory and sub directories have index.htm files in them but unfortunately except for the root directory users can see the directory listing rather than the directory automatically producing index.htm. This is obviously causing security problems. How to prevent directory browsing by users coming into the server over the internet by a redirect URL from another site. E.g. [URL] where bigwebsite is a domain registered with a change ip server because I have a service which dynamically allocates an IP Address.
how to pipe the current directory listing into sort so that the output is the date in descending order (primary sort key). If there are multiple entries with the same date, I'd like the times sorted in ascending order. It seems simple but for some reason this isn't working:
ls -l | sort -k 6r -k 7
For some reason it doesn't seem to ever get to the second sort key when using column 6 (last modified date).
Is there a command to return a recursive listing of sub-directories and the number of files in them? I have found plenty of ways to give me the total number of files in a directory structure, but none that gives a list of the sub-directories with the number of files in them. "du" gives me a listing of directories with their sizes, but I couldn't find an option (or any other way) to give me the number of files as well. Ideally, I'd like to get list with "Size" "Files" "Dir name" - And the order of the columns doesn't matter. Is there a "simple" command line solution or do I need a shell script for that?
I have, for example, a folder called "MyFolder" and it contains 3 files: MyFile1, MyFile2, MyFile3. The only file that I do NOT want a particular user/group to even see that it exists is, for exmple, MyFile2.So, when they do a directory listing on MyFolder, they should only see MyFile1 and MyFile3. How can this be done in Linux? The important thing is that it is not just preventing them from "executing" MyFile2, but to prevent them from even knowing that it exists by not including it in a directory listing.This is a simpified example using one file, but in reality, I have lots of files and some of those that I want to block are also subfolders.It is very important for me to hide the existence of certain files/folders when the user does a directory listing. It's also important that the files stay in their current folder (that is, I can't use a workaround which requires moving all the files into a separate folder and then securing that folder).
I have a user who just got a new laptop and when he connects to the FTP server, it verifies his username and password like it should, but when it goes to do the directory listing, it just kinda sits there. if i refresh the view, the folders show up. wonder if there's something in a log somewhere for vsftp i should be looking at. it could just be his ftp client or air card too, but i want to be thorough on the server too becuase it will help me learn.
I obviously a noob to Suse but after installing 11.4, and filezilla, i cant connect to my ftp server. Ive opened ports 21 22 and 990 explicitly in the firewall and all I get is a 425 error Cant Retrieve Directory Listing. It logs me in ok but doesnt allow data connections?
I am trying to make my Apache server show symbolic links in a directory listing, but have so far been unsuccessful. In my latest attempt, I have placed the following code in .htaccess, in the directory with the symlinks that I want listing:
Code: <Directory /> Options All </Directory> Im httpd-vhosts.conf, I have also placed the following code within the relative <VirtualHost></VirtualHost>:
How would I limit this to searching for the text 'SomeString' or 'SomeOtherString', but only if the file has extension .php, .inc or .js? Also - what piping to xargs does here? I don't understand how this command actually works.
My desktop just froze for some reason and I had to force shut down. When I logged back in I could not get into my /home/username directory, a message popped up that it does not appear to exist and if I would like to put it as root.
I want to copy a file from an old directory to a new directory and ideally I would like that the 'mv' command itself could create the directory for me in one step...That is to say that the directory would not exist until 'mv' has been called as in:Code:mv olddirectory/file.dat newdirectory/file.datso mv would create the directory 'newdirectory' in the above call....I am not sure if this is possible however, that is to say I think it may be necessary that the directory 'newdirectory' already exists before I make the above call to mv, as when I make try the above call when 'newdirectory' doesn't exist I get the following error:Quote:mv: cannot move `test1/4215/4215-001R.fit' to `test2/test.fit': No such file or directoryso to sum up, my question is:is there some paramater for 'mv' that I can change so that it will create the new directory as well as copy the file and if not, are there other commands that might do this?
On a KDE4 environment after downloading some music from rapidshare with JDownloader the archives self-extracted with the symbol in some of the file names. Those files couldn't be renamed or deleted, the file manager said that the files didn't exist - very weird. The files should have had some swedish characters in their file names. Now I'm stuck with those files on my machine. Anyone knows how to get rid of them?
I have a requirement to list files using find command My folder contains below list of files with out extention.I have a requirement to exclude only ABC.123.* type files and list others. Even though files having MNO contains this pattern i should not exclude. Even if file ends with .txt or .doc it should not be excluded. That is ABC.123.1234.txt should not be excluded.But I am not getting what is required. Can any one please let me know if I am doing wrong any where. As per my requirement I cannot use grep, -regex, or -regex attributes to find command.
its a very basic question but iam not getting it right nowi have to list all the pdf files on my desktop even the pdf files which are present in folders on the desktopls *.pdfonly list the files present on the desktop, but not the files in the folders on the desktop containing the pdf files.
I have a newly built Redhat 5.5 box. When connecting over VPN and executing either "ls -al" or "df -h", the command prompt hangs without returning any output. Have got same result with TeraTerm and Putty clients. However, other commands run fine. Also,local users are able to run "ls -al" or "df -h" successfully.Storage is local disk. Any idea what could be mis-configured.
I've got a Fedora 10 server with a simple read-only samba share.I'm able to mount and browse the share from a Fedora 12 client, but all directories appear as empty--and I can see on the server that they contain many files. This happens whether I browse using smbclient, or mount using mount.cifs.I've got smb/nmb ports enabled on both the client and server. File permissions on the server look right.The server smb.conf setup: