I am running following commands and ist runs fine for me.cp -r /mnt/datasource/* /opt/IBM/WebSphere/DataSource/But if I put this command in script, it does not work.
vi cpydata
cp -r /mnt/datasource/* /opt/IBM/WebSphere/DataSource/
#save it
I am trying to install iscsi-target on centos 5.4 i386. When I go into the iscsi-target directory and type make I get the following error: [root@cluster-storage iscsitarget-1.4.19]# make cc: /lib/modules/2.6.18-164.el5/build/include/linux/version.h: Datei oder Verzeichnis nicht gefunden cc: keine Eingabedateien /bin/sh: line 0: [: -le: unary operator expected /bin/sh: line 0: [: -le: unary operator expected /bin/sh: line 0: [: -le: unary operator expected /bin/sh: line 0: [: -le: unary operator expected /bin/sh: line 0: [: -le: unary operator expected /bin/sh: line 0: [: -le: unary operator expected /bin/sh: line 0: [: -le: unary operator expected /bin/sh: line 0: [: -le: unary operator expected /bin/sh: line 0: [: -lt: unary operator expected make -C usr make[1]: Entering directory `/opt/iscsitarget-1.4.19/usr' ..... make: *** [kernel] Fehler 2
I've got a few large, unsorted folders where I dump mp3s. I manage it all through rhythmbox and I've decided to separate the wheat from the chaff by paring down stuff I don't want and throwing all the desirable stuff into a big (~1,800 file) playlist file saved to my desktop.My question to you is, is there a hack that'll let me cp all the file entries in this playlist from their various original directories and into one target directory? Playlist format is:
There is a symlink from /var/www to a personal directory. FollowSymlink and chmod 755 are all set. It works perfectly until each morning I will get a "Symbolic link not allowed or link target not accessible" error. When I do a "sudo service apache2 restart", the problem will go away.
I have synology nas box DS710+ that has intel atom processor inside. i've installed ubuntu under virtual box, downloaded needed toolchain for x86 processors, and decided to cross-compile openvpn.
I 've started from compiling lzo library, but first question that appeared in my head - how i have to move all compiled binaries to my synology?
1. configure - okay 2. make - okay 3. make install - this is not necessary?
Will --prefix parameter helps me? or do i have to use chroot? cause there is a plenty amount of files that i need to move to my synology.
I'm using the command below to sync two directories. Problem is insted of deleting the files on the target directory it simply appends a ~ character at the end of the file name. Not sure why this is happening?I'd like to have all deletes on the source replicated on target.
it is not clear to me, by reading different postings/documentations for ntfsclone, what restrictions and/or requirements exist for the target partition (TARGET). Can the TARGET be bigger in size than the SOURCE, or must it be identical? Must the TARGET be formatted as an NTFS filesystem, or must it left unformatted?
I am trying to insert a .ko into my kernel and I am getting the 'invalid module format' error. But the kernel object is being generated from or is included in a binary which sets up and populates a driver path when run. How can I tell what architecture/target/anything that .ko is meant for?
I know that I can use the uname and /proc/cpuinfo to see my current info, but I am also not sure what I might need to do to get them to sync up. I am running on a minimal, cli-install of Ubuntu8.10 right now. It's freshly installed (onto a 2GB USB drive) & imaged, so bring on the drastic changes.
I need to use xdelta tool in my shell script for creating and patching diff files. My PC Linux environment (Ubuntu) has a package installer, and when I try typing xdelta on my commandline, it prompted to try "sudo apt-get install".. after which it directly accessed the packages from the Internet and installed it. So I am able to use xdelta from the command-line in my Ubuntu environment.
I need to also use xdelta on my target. I am using a NFS. The bootloader and kernel are on the target and the RFS is referred from the PC. How can I install the xdelta package on my target so that it recognises the command xdelta. On Ubuntu, "sudo apt-get install", fetched the below packages:[URL]..
What I am trying to do is edit the link target to force ip=xxxx.xxx.xxx string at end. I have a software program which access 4 different servers running the software but with different configs. In xp I can copy links and modify as above to correctly force the program to the various servers.
I have read the various how tos on hard vs sym links which I get. Playing around with hardlinks and sysmlinks (the examples I find) does not seem to be what I need. Feel like this is pretty basic stuff but I am a bit stumped.
marek@marek$ ls -al /usr/share/solr/ razem 36 drwxr-xr-x 5 root root 4096 2010-11-30 08:25 .
[code]....
i want to copy it to ~/solrTest but i want to copy files from symlink as well when i try to cp -r /usr/share/solr/ ~/solrTest i will have symlink here:
I have installed Ubuntu 10.04 .. normally and I installed also some programs; I had problems with the monitor and many times I shutted dow the CPU when I changed the monitor and turned on angain and appears a black screen:after
mount: mounting /dev on root/dev failed: no such such file or dir mount: mounting /sys on root/sys failed: no such such file or dir mount: mounting /proc on root/proc failed: no such such file or dir
AFAIK, in addition to implementing the standard C library, glibc provides wrappers for system calls, threading support, and basic application facilities. So because of that, glibc that will be used on my target system should be built based on the kernel version running on my target, right?
Based on the above, I am trying to build a glibc version for my target machine's kernel. However, I don't know to to build a glibc library for a target system and also where on the target's filesystem should be put? In which location Linux will start looking for the libraries required for a program to run (should I create an /etc/ld.so.conf file)?
Well, I am facing problem when doing lab questions.
I must use DLXLinux bundled in Bochs (bochs.sourceforge.net).
I am required to use the /usr/local directory.
In /usr directory, there is no directory named 'local' but there is one thing called 'local@'. So, when I try to use mkdir command to create 'local' directory in /usr , there are error "cannot make directory.....".
I want to run a cronjob every 15 minutes that checks a directory for files. If the directory contains more than ten files I want it to send an email to me.
All I have is this...
*/15 * * * * ls -l | wc -l | [filename] | mail -s "This is just a test" [email address]
I would rather not write a bash script. Is there an easier way to do this? I was looking into some commands like find and grep.
I'm quite new to linux but I have configured a simple ftp server and it's working great. I have a FTP-Shared folder with upload and download subfolders. Under upload's and download's I have identical category subfolders like mp3's, movies, software etc. in both. As the guy's upload, I would like to create a line crontab where I can move all the content under /FTP-Shared/upload/mp3/* older than 14 day's to FTP-Shared/downloads/mp3/ recursively (Like in cp command), but the timestamp must be searched on the first directory and not sub files example: /mp3/Club Dance/CD1/Hallo world.mp3This is how far I got:[root@clients ~]# /usr/bin/find /FTP_Shared/upload/Mp3s/ -depth -mindepth 1 -mtime +14 -type d -exec mv -f {} /FTP_Shared/download/Mp3s/ ;This command moves the directory and files, but it is not recursively
1. Every Sunday2. Find all files older than 1 day3. Gzip these file4. Tar up the gzipped files into one tar file.5. Name the tarball with a date stamp indicating what day it was created, so we know that week's files are in the file
I've just installed Ubuntu 10.04, trying out linux for the first time. I'm wondering if it's possible to link my documents folder in Ubuntu to my documents folder in Windows, so that when I open the folder in Ubuntu, I am linked to the folder in Windows.
I'd like to mount the directory /var/www/mysite to the directory /home/daniel/mysite, but also have the user of the mounted files mapped from the original user (www-data) to my own user (daniel). So that the file /var/www/mysite/index.php who's user is www-data will appear in the mounted directory as /home/daniel/mysite/index.php and be owned by daniel - and alternatively, if I create a file /home/daniel/mysite/test.php with my own user, it will be created in the original directory under the user www-data Is it possible? If not, what alternatives do I have so I can use an IDE and still make sure all the files belong to the HTTP server's user?