Fedora Servers :: How To Delete LDAP Data Base Bdb
Aug 29, 2011
Please let me know how to delete/flush complete LDAP database ? I tried deleting all files from /var/openl dap- data. but still top level data base attributes remains.initially I had set shadowExpire,shadowFlag etc, I want to remove all these.Kindly help me to build my new database.
Loaded plugins: fastestmirror Repository updates is listed more than once in the configuration Determining fastest mirrors YumRepo Error: All mirror URLs are not using ftp, http[s] or file. Eg. 13 is not a valid release or hasnt been released yet/ removing mirrorlist with no valid mirrors: /var/cache/yum/i386/13/base/mirrorlist.txt Error: Cannot find a valid baseurl for repo: base
I want to ensure I have done all I can to configure a system's disks for serious database use. The three areas I know of (any others?) to be concerned about are: I/O size: The database engine and disk's native size should either match, or the database's native I/O size should be a multiple of the disk's native I/O size. DMA: Disks that are capable of Direct Memory Access (eg. IDE) should be configured for it. Write-caching: When a disk says it has written data persistently, it must be so! No keeping it in cache and lying about it.
I have been looking for information on how to ensure these are so for CentOS and Ubuntu, but can't seem to find anything at all. I want to be able to check these things and change them if needed. The actual hardware involved is very modest. The point is to get the most out of what hardware we do have, even though it's "not very serious hardware" from a broader perspective.
I'm new LDAP kind of stuff.I want to configure LDAP server and Client on windowsXP. could you tell me which Open source LDAP Server and client are best fit for windowsXP. Is OpenLdap not suitable for windows?
I am using LDAP to manage the autofs and everything works fine. I have a situation here that I can't figure out how to do. I need to use 2 direct map in auto.master, but the LDAP service doesn't allow me to add 2 "cn: /-" entries in the directory.
The following is my auto.master map: dn: ou=auto.master,ou=autofs,ou=dolphin,dc=example,dc= com,dc=my objectClass: automountMap objectClass: top ou: auto.master dn: cn=/-,ou=auto.master,ou=autofs,ou=dolphin,dc=example,dc =com,dc=my objectClass: automount objectClass: top automountInformation: ldap://ldap.example.com.my/ou=auto.iso.indirect,ou=autofs,ou=dolphin,dc=examp le,dc=com,dc=my cn: /-
I have another ldap entry auto.data.indirect but I have no idea how to inject it into the directory. Any ideas how to implement it? I am using Fedora 13.
I am trying to configure and compile the pam_ldap and nss_ldap for MIPS architecture. for this they need ldap library. so I am trying to compile openldap code for MIPS first, but the configuration is failing.The command and options used for configuration is as follows.
I have a openldap server running on one machine (fedora10) and pam_ldap.so and nss_ldap.so running on the other machine.
I have added a new user to the LDAP server database, this user is not created on client machine.
1. Can i login to the client machine using this new user?
2. Now if i try logging with this new user I am getting error messages, the error messages are as follows at client side
Sep 2 10:34:36 localhost sshd: Invalid user kim from 10.254.194.148 Sep 2 10:34:36 localhost sshd: input_userauth_request: invalid user kim Sep 2 10:35:16 localhost sshd: pam_ldap: error trying to bind as user "cn=min soo,ou=people,dc=samsung,dc=com" (Invalid credentials)
I'm rather new to Fedora server, but I'm attempting to run a music FTP server, where anonymous users can submit songs into one particular folder (so i can personally tag them), while other user accounts have full read-write. Here we go: I 2 directories, /music and /untagged
I want anonymous users to be able to read both directories, but only be able to upload to /untagged, and not be able to delete anything. I want users that I select to have full read-write-create-delete privileges. how would I go about this with vsftpd?
I was transferring some files from my laptop (running FC6) to a server at my work (don't know what kind) with "scp -rpC" and it stalled, don't know why. Now when I try to delete the files from the server so I can start again I get the following error message
i'm trying to build a web server but I keep having a problem. Everything loads fine until I get to the "Install base system" part of the installation and it puts up a red screen warning me that the install has failed. My uncle gave me his ubuntu discs and even some different hard drives, but I keep getting the same problem at the same point of the installation. Here is the rundown of everything i did.
System #1 - ASUS P5G41-M m/b with 4GB memory Intel Core 2 quad q9300 cpu two western digital 640GB hard drives
I was testing it out to run a raid 1 just to learn how to make a software raid array. its my first try at making a raid array. The first time i used ubuntu server 10.04 32bit and the system failed at the install base system point. I then used boot n nuke to clean the drives and tried 10.04 64bit. it failed at the same exact point. my uncle gave me 2 western digital 1.0TB hard drives and I tried again. I got the same results. At that point i gave all the hardware back to my uncle and he gave me a server board to try out because he bought it and hasnt touched it. so I built a second system.
I have LDAP authentication working via SSSD using authconfig-tui and a few minor modifications to sssd.conf (ie: max_id etc). The problem I am having is it would appear /etc/ldap.conf is being ignored and/or setups that work perfectly on RHEL5, F11 and F12 no longer work on F13. Specifically Im referring to "pam_check_host_attr" and "nss_map_attribute". It refuses to honor either of these options and I can only assume a number of the other options in our ldap.conf. For instance, "nss_map_attribute" is defaulting to the standard "homeDirectory" rather than "homeDirectoryLinux". This is related to a bunch of OSX clients we have and its not optional to use another setup. The host restriction is also a major issue.
Where apart from ~/.kde4/share/apps/kmail does kmail keep account data, such as login data/names of pop/smtp servers? I wanted to install an entirely clean version of kmail, so I uninstalled/re-installed kmail via YAST. I manually deleted ~/.kde4/share/apps/kmail. After installing kmail I found that the account data of my Yahoo account were still listed in the kmail configuration panel. Shall I delete the Yahoo account within kmail prior to uninstalling the kmail package?
How to backup in such way that, data will be backed up in every 1 hour...but every time not full data;; only the data/file which have been modified within last 1 hour. Its on locally. And how to back up local data to remote system. Because I have tried with ssh-keygen also , but everytime its requiring password.
I am attempting to copy files from one server to an external USB drive on a second server. Both servers are running custom RedHat Linux, kernel 2.6. Both are setup as Check Point SecurePlatform (one is a log server and the other a management server). I am trying to archive files from one (HP DL380 G5) to the second (HP DL380 G6). I am not able to archive directly to the external usb drive when connected to the HP DL380 G5 (data gets corrupted and switches to read-only access). The external USB drive has no issues when connected to the HP DL380 G6 server, thus my reason for trying to copy the files across servers.
When I attempt to use scp to copy files between the server, I am prompted for the password. Once entered the debug shows the authentication is successful, but then no files are copied (see log below). I have tried searching for potential answers, but none have panned out. Unfortunately I can't yet post the actual scp debug log output as some of the text is viewed as a URL and not yet allowed for me. I have changed all the atsign symbols to '(a)' to avoid the URL inference.
Output log from scp attempt follows: [Expert(a)server1]# scp -p -r -v /var/fromhere/* copyfile(a)server2:/media/disk/tohere/ Executing: program /usr/bin/ssh host server2, user copyfile, command scp -v -r -p -d -t /media/disk/tohere/ OpenSSH_3.6.1p2, SSH protocols 1.5/2.0, OpenSSL 0x0090707f debug1: Reading configuration data /etc/ssh/ssh_config debug1: Applying options for * debug1: Rhosts Authentication disabled, originating port will not be trusted. debug1: Connecting to server2 [192.168.10.3] port 22. debug1: Connection established. debug1: identity file /root/.ssh/identity type -1 ..... debug1: Exit status 1 lost connection [Expert(a)server1]#
We purchased a new database system at work last October, ditching the old system because of a lack of support from the vendor. This is a retail Point of Sale and Backoffice database system. I am not sure what system the new one runs on, but the system we replaced was a Firebird data base. The reason I am posting is because we are now in need of the information contained in the old database which was not completely imported into the new system.
Basically the problem is this: The database in on a Windows XP system and I found a copy of SQL Manager Lite 2008 on the system, which after quite a bit of studying, I figured out how to extract the database into a removable file. I have this file (178MB) on a USB stick in a file called Backoffice.fbd.
My studying suggests to me that I can get into this database with MySQL. I have never used this and have no clue how to do this. All I want to be able to do get into the database and create tab deliminated spreadsheet files for each of the database sections (Customers, Repairs, Sales History, stock files, etc.) Is it possible to do this with Ubuntu and MySQL and if so, can expert suggest one or two things to get me started. While a guided tutorial would be nice because I am not an expert, I am willing to learn on my own if someone could point me in the right direction.
I tried to clone my laptop ubuntu partition (9.10) using ddrescue into my portable harddisk.I run command "ddrescue r3 /dev/sda6 /dev/sdb6" sdb6 is my portable harddisk partition with lots of data. In my head I was thinking that it will create image file (like iso or img)that can be restored to another disk.Suddenly I relize that it delete my existing data (in sdb6) then I interrupted ddrescue.Is there a way that I could get back my data in sdb6?Partition in sdb6 is FAT32m, now when I try to open sdb6 using Windows OS, it always aks me to format
I am working in a office where only one internet connection available. I have configured 5 other client machines to use internet through squid proxy server. Now I want to restrict the total data usage/transfer (upload+download) to say 1 GB during a calender month. How can I achieve this setting.
I'm confused about "hard link" feature. I've been learning from my UNIX Academy DVDs training that hard links to a file can be many and each of them is an effective filename for the associated data. So let assume that we have some very sensitive data in a file and we want it to be deleted and file has 20 links. I "delete" a file, but in fact I deleted only one "name" of it. My understanding from the training that data is still there until we delete the last associated hard link. But how can I find the names of all of them? If we have the names, they can be removed one by one. Or may be there's command that can trace all the "names" and remove them at once?
I have a router that is 1000 Base T and two computers each with ethernet cards that support 1000 Base T. All are equipped with Cat 5e cable. Before I had a router that only went up to 100 Base T and I would setup one box with linux running proftpd. On the other box,I would use win xp pro and use firefox to ftp into the other box and download a file. Download speeds went up to 11.2 MB/sec. Now when I switched routers, I expected something like 120 MB/sec but I'm only getting 5.3 MB/sec. What do I need to change?