Using C++, I want to process sub-folders on my home folder sequentially each with a special naming format and containing some binary files in it:
Code: 1/ 2/ 3/ 4/ 5/ 6/ ...
Give above folders, I will process files in 1/ at first, 2/ at second, 3/ at third, and so on.
For some n/ folder, if I realize that n/ actually does not exist in local file system, I do not want to wait for it. Hence I will keep processing (n+1)/ folder, and so on.
However, when processing some (n+m)/ folder, previously not processed n/ folder may have been created on local file system. In this case, I do not want to miss processing it, but somehow detect its creation and process it. After processing n/ folder, I want to continue from (n+m+1)/.
i am facing a problem regarding permissions. how can i set 775 permission for all newly created files and folders. when i give chmod -R 775 /data permission is getting to all files and folders. but when i create a folder i wont get that permission. i want this 755 permission should be permanent for all old and newly create files
I just maked an ext4 partition by the help of gparted. Ubuntu is my only OS no dual boot. Using Ubuntu Maverick. The problem is partition must be open as root to do any work else it wont even allow me to open file,create folder,cut copy paste or anything.
I'm a little bit confused with partitioning the filesystem in Linux. the difference between creating the file system with fdisk and mkfs (when formatting the disk). I can't clearly tell my problem, so please look at this picture:
We are running Oracle 10.2.04 RAC on Red Hat Linux 5 and when ever our SAN storage admin created new LUNS we have to reboot the servers so that server can see newly created LUNS. This causing downtime to our application. How we can add LUNS dynamically without rebooting servers. We are using device mapper mulitpath from red hat version 0.4.7-30.el5 and we have QLOGIC HBA's.
i've written a bash script to add new users to our system. the script works so I won't bother you all about that. when a new user is created with it, they can immediately login to our domain from any terminal, which is good. However, the newly created user is unable to login to debian at all, and so cannot access the server. when attempting to do so, they get a message like "the system administrator has disabled your account". This is a good thing really as normal users have no need for debian login, but I do need to add a few admin users who will need direct access to the server machine.
This is the code I'm using to add the user. The rest of my script is just a wrapper and GUI. I figure the login shell may have something to do with it, so I tried changing the shell of a user to the default /bin/bash. This resulted in the user being able to login - sort of. Gnome doesn't load though, and there's a cascade of errors across the screen about things failing to save or load settings. mostly stuff like nautilus, X, and gnome. the desktop background is black and there's no interface. Logging in with a previously existing account works fine though. Clearly I have an issue somewhere.
I have an Intel setup with 64bit Ubuntu. I have an NVIDIA graphics card. When I used compiz, I found that my mouse would follow newly created windows. (e.g, I would bring up terminal and if I hit F1, the mouse would move to the upper right of the screen by itself. I have since turned off compiz which resolves the issue. where I can turn it off/on.
I just created a LUKS filesystem following these instructions. Everything seemed okay at first. It mounted with no problem and I moved some files there. I then unmounted it and remounted it to see if I would need to use a special command. It mounted right away and even allowed access to normal users. So, I rebooted to see if anything would change. Before I go on I should say that my partitioning scheme is weird. Not knowing any better I 'upgraded' to 11.04 when my update manager told me a new version was out. This didn't go well and I had to do a fresh install to put 10.10 back on my machine. After this the way it partitions the drive has been weird. What I had was /dev/sda1 which has my installation on it including /home. But, where it gets weird is /dev/sda2 would not manually mount. Looking at the disk in gparted it showed /dev/sda2 THEN under that, as if they were sub partitions or something, I had sda6 and sda7. I had been using 6 and 7 for various things and they mounted fine, so I decided to encrypt 7. After reboot I only have sda1. Everything else shows up as unallocated and ever way I try to mount I get device does not exist.
I only did the procedure for sda7 but 6 has been affected as well. There is no longer a sda2 the way there was before. This always bothered me anyway since I wanted sda2 for my /home but it wanted to call it sda6 and put it under sda2 like I said, I could never fix that, now this.
I just want to know how the default size of a newly created file or folder is 4.0 kb.Does this value is mentioned in any configuration file,if that is the case can we edit that file and can we change this default value.
I had a samsung 1TB HDD that I used for storing data, on an xp machine, so it was formated as NTFS.I moved this HDD to another machine and installed Freenas on it, and the installation worked fine (fyi, I used the tutorial posted here :[URL]..During the installtion, Freenas installed it's system files to a new small UFS partion. After finishing the setup, I realised that I had changed the file system of the other partion (980gb, previously NFTS) to UFS and now I don'T know how to go back. I had about 400gb of data on it and I'm pretty sure it's still there, but don't know how to get it back.
I tried messing around with recovery software such as R-Studio, and I was able to see some of my files so I know they're still there. After quite a bit of googling around, the only solution I seem to find is using gparted which is a tool to modify partions file system without loosing data, but I'm afraid to use it.
So is there a way to browse NTFS data on a UFS partition and convert it so FreeNas can see my files ? Or is there a way to put the partition back to NTFS so I can back up my data to another drive before I lose something valuable ?
Running NFS on Fedora 10. Exports fine. I tested it locally. I tested the NFS configuration by trying to access the exported directory from my local machine, before testing it from a remote machine. While logged in as root, I created a new directory "/mnt/nfstest".
Then I mounted the NFS share at the new directory I created: [root@eric root]# mount -t nfs localhost:/mnt/nfs /mnt/nfstest
When I tried to mount on the remote client: [root@frank root]# mount -t nfs eric:/mnt/nfs /mnt/nfstest
After a while I got: # mount eric:/mnt/nfs /mnt/nfstest mount.nfs: mount system call failed
I tried strace but wasn't sure what I was looking for, but I've attached the results as a .odt file.
The question is, as far as I know Ubuntu distro adds a user created with useradd to supplementary groups automatically. For instance, I want to enable sudo for all newly created users on my LiveCD and want them to be added to the group 'wheel' on creation. I'm sure it is possible to do it in Fedora, but how?
CentOS 5.4 install, likewise open standard install (For active directory authentication).I have a license service which requires a license.txt be in the users home directory.The group owner for license.txt must be the same as the license service. Whenever a new domain user logs in, it creates the all the appropriate files but the group owner for license.txt is the users domain group. My current workaround seems like more effort than it's worth, is there another way to get this process solved easier/more secure?
- copy the license.txt into /etc/skel
- created a script to check for the presence of license.txt, check it's permissions and change them if necessary
- gave the domain's group sudo [nopasswd] access to the script (the script is not writable)
I have several directories, each owned by root and a group of the same name,By setting the sgid bit, I made sure that newly created files and directories are owned by the correct group, and that directories have the sgid bit set too.On each newly created directory or file, the permissions are set to 755. This is because this is the default umask, and I cannot change a users umask. I actually only want files created below a particular directory to have group write access, inheriting this behaviour to newly created directories properly.I'm not on samba or NFS, I have to do this for SSH users.The filesystem is ext3.I started to fool around with ACLs, but couldn't find what I was looking for.
I have NFS fileserver that has served me well for more than year. But recently I noticed that it has started to reboot on its own very frequently, almost once a day! It is most likely not a power related issue as I tried changing UPS/power sources, but no help!So my question is:Is there any log file where I can check which is causing the reboot? There may not be a single logfile, but I need some point to start the investigation!
Just installed Linux (Ubuntu 10.10-net book-i386) on an acer 5310 laptop, everything is working great except for the internet. Google home page takes over a minute to load on Firefox, the actual search results takes only a second but then it takes 2 or 3 minutes to load any site i click on. I have broadband, only 1mb, but it worked perfect on the previous OS, also using Firefox.
I'm trying to come up with a startup init script that will check to see if the system was shut down gracefully, or if it is rebooting from a poweroff or something similar?Anyone know of a way to check for this condition with the least amount of room for false positives or vice versa?My intial thought is just a startup script that will will check for a file on startup, and on a proper reboot/shutdown just touch the file. But id like to avoid that type of script if possible,
I am just starting out in LINUX and I know the basic commands but I am a having a problem. I scoped the man pages but I can't get it. Maybe one of know... Can anyone tell me the cmd to figure out the system a file was created on? I just can't figure out this problem.
I'm in a bit of a unique situation here. I have a remote system running two instances of RHEL5.1 (dualbooting) and I need to install RHEL6 on the other partition from the active one. I understand that you can't upgrade 5.1 to 6.0 so I would need to do a fresh install. My basic question are, is there a way to install an OS from an existing OS to another partition? Can anaconda be run from an already installed system? Is it possible to manually install a system just from RPMs?
I've come across things like debootstrap for Debian and haven't found anything similar for Red Hat systems.
As I'm currently in a different part of the country from this machine I would like to achieve this without having to be physically at the box or reboot the machine (reducing risk of being completely cut off from it).