General :: Create Hard Links Within A File System To A Directory?
Aug 1, 2010
I am trying to create hard links within a file system to a directory, but unable to do that. is there any limitation to create hard links to directories within file system ?
I am going to add following line to the /etc/fstab file /dev/sda4 /MyApp ext3 noatime, errors=remount-ro 0 5 Does it also create directory "MyApp"? Or which kernel file is responsible for creating the directories? Also on linux, which kernel file is able to create some soft links and additional directories?
i just uppgraded to Centos 6. Linux 2.6.32-71.29.1.el6.i686 on i686. now im getting some troubble whit this file: wget [URL]... i create a downloads directory and download the source file and--
mkdir $HOME/downloads cd $HOME/downloads
But when i try to build courier-authlib whit this comand: #sudo rpmbuild -ta courier-authlib-0.63.0.tar.bz2. I get this:
Code: $ echo 2 * 3 > 5 is a valid inequality. This will create a file in the current directory named '5' with the number '2' in it, the names of all the files in the current directory, followed by the number '3' and 'is a valid inequality.'
What I do not understand is why 'is a valid inequality' gets written to this file. I thought it would write '2', all the file names in the current directory, then '3' into the file called '5'. Why does the 'is a valid inequality.' get written to the file also?
I would like to create a directory public_html/archive that is a mirror of public_html/images, except that deletions in the images directory don't cause corresponding deletions in archive.
I was going to manage all of this in code, but then I figured, maybe the system can handle it for me in a robust way.
Note that the website will reference archive in real time, so it has to be an instant mirror and not something done nightly, etc.
Code: mkdir: cannot create directory `/dev/cgroup/cpu/user/5900': No such file or directory bash: /dev/cgroup/cpu/user/5900/tasks: No such file or directory bash: /dev/cgroup/cpu/user/5900/notify_on_release: No such file or directory It seems like it's probably from this part of .bashrc:
[Code]...
What does this code do, why, and what's causing it to go wrong?
I'm working with a dual-boot laptop running Ubuntu 10.0/Windows 7 and a Debian 5 VPS while the OS's shouldn't have much impact on my question.
What I would like to do is create a html page that I can upload to my VPS which lists all of the files/folders on my local 2TB hard drive (Specifically media such as Movies, Music, TV Shows...). The media obviously will not reside on the server, but I would like to at least have a list which will allow me to select, for instance, a bands artist so that it redirects me to the albums in the directory below.
Ultimately, I'm looking for Open Directory Browsing without actually having the media on my server. I have been attempting to create something to this effect using lynx, however, I'm not sure if it can be done with this command or if it's even possible for that matter.
I want to make symbolic links for all them to my current directory /test2
I tried
But it failed. It seems like I can't make symbolic links for all the 5 files simultaneously.
Often times I need make symbolic links for multiple files with some common pattern (just like ".txt" here). I really hope to avoid making symbolic link for each of them one by one...
I have a big problem, I can`t enter to the system, said this:
mount: mounting /dev on /root/dev failed : no such file or directory mount: mounting /sys on /root/sys failed : no such file or directory mount: mounting /proc on /root/proc failed : no such file or directory
I'm trying to crawl a directory on a website and basically download everything in it. The structure is simple enough(but there are also multiple folders), but there is one thing that makes wget choke up.Both of the links work, but they are both the same thing. So wget will download the same file twice. How can I make wget ignore the first one? but this doesn't seem to actually do anything. It will still download the duplicate URLs
I have been running Ubuntu as my sole OS for about a month now, and aside from Computer Janitor, I have found no drive/fs cleaning tools, or defragging tools. I know this isn't Windoze, but surely there must be a way of maintaining your hard disk. I mean installs/uninstalls must leave detritus, and file moving must cause SOME fragmentation. Are there any apps, systems or methods in place that provide this functionality?
How can a directory mount the hard drive? How can a list of files sitting on a hard drive that is not in RAM run your computer and then mount the physical device that stores the code that it is running? Obviously the CPU can access the hard drive before this so how does this massive abstraction make physical sense?I'm thinking that if this is abstracted to this extent so is the "/.../.../" path nomenclature as well. Is this just to change directories and or an environmental variable? Is the file really farther the more slashes that are in it?
I installed ubuntu off of my laptop and put the Ubunto file system on a 1TB hard drive off of a drive enclosure. Now the only way I can access Windows is if the drive is still connected to the laptop. NO EXTERNAL HD No BOOT MENU when I boot up.
I'm looking for some advice about how to implement the following functionality in my pet project.
There are two users on my system, user1 and user2. When user1 logs in he can do what he wants etc.. when user2 logs in I want to somehow link the entire file system to another place. In more detail when user2 logs in and does something like ls ~, he should see contents of /home/user1/extra/home/user2/ instead of /home/user2/ and when user2 does ls /usr/bin he should see contents of /home/user1/extra/usr/bin/
Is there a simple way to do this kind of operation. I have looked up the ln command, but I am a little unsure about how to show a completely different file structure to a particular user.
I did play around with ln and found that you can make "soft links" to directories. The problem being how can I link the user2 home directory /home/user2 to something like /home/user1/buffer/home/user2 . can I use something like ln -s /home/user1/buffer/home/user2 /home/user2 .. I guess not. I didn't want to try it as I wasn't sure so as to not hose my system.
As I understand it creating an image of a Linux system makes an exact copy of the OS and any user files/configurations/programs etc. What i would love to do is create an image of my work PC and install it at home on my desktop. Can someone briefly explain the process of creating and installing images of Linux systems?
Home OS - windows Want - An image file that can be executed in a virtual machine(VMPlayer or VirtualBox) or booted directly on my home PC.
I am unable to create either a file or a directory in a specified path in Linux. I am getting the error "No space left on device". I have checked with df -k and df -i. Free disk space is 28 % and free inodes are 28 %. What else could be the reason?
and my current working directory is sub1link, is there a quick way to either: change directory to link source parent (i.e something similar to cd .. but take the user to /dir1/ change directory to link source (i.e switch from /dir2/sub1link/ straight to /dir1/sub1
Is there any program/script that can create a torrent file for each file in a directory? I have been looking all over but can't seem to find anything of the sort. I have 700+ files I REALLY don't want to make my self.
Does anyone know why files in my /tmp directory are not able to rm even using root login? not only that, I can't even chmod or do anything to files in /tmp directory... it always saying "read only file system" warning
CentOS 5.2 64bit 2.6.18-92.el5xen. Use rsync with --link-dest for nightly backups, works well. Was recently asked to start weekly backups to an external drive for off-site storage. The regular syncing works but hard linking seems to be ignored. So the backup is long with no space saving advantage. Here is an example of the command being run:
I had a drive that kept kernel panic'ing so my data center recommended using the spare hard drive to reinstall OS on, and import the data from the old drive. (they checked the hardware, it wasn't the hardware) The new install is done, and I need to mount the old drive and get backups off it since my data center does not provide management whatsoever.
It's the same OS on both (Cent OS 5.4 32-bit) I'm an advanced user on windows, but linux gets me. I can ssh in, do basic stuff like setup IP ranges and restart services. I normally navigate the box through SFTP so I have a gui. WHM shows me my drives as such
Found Disk: hda Found Disk: sdb
so I'm assuming SDB is my old drive and the drive I need to access. I attempted to follow instructions on