I observed a problem with the Yahoo Mail Attachments.If I want to attach a file from a drive that's mounted (one of my windows partitions), the file who I attach to mail seems attached but after I send the mail the recipients do not receive my file. First I thought it was a mistake, but then it happened several times. Does anyone know any workarounds except copying the files who you want to send to an Linux drive and then sending it from that location?
I have a question about NFS and CentOS 6.0. Here is the environment: Server running VMware's esxi 3.5 (Fully patched) 10.1.1.50 Server running CentOS 5.5 with NFS. 10.1.1.51 With the systems in the above config everything is working fine. Install a new hard drive in the CentOS server and install CentOS 6.0. Use the same IP 10.1.1.51. Configure NFS, the VM server can't attach to the NFS share.
Here is the config of the NFS server: Firewall Disabled SELinux Disabled NFS installed and running. NFS-Utils installed. Contents of /etc/exports
I have confirmed I can telnet to the ports listed. Any clue why the setup with CentOS 6.0 doesn't work? Is this a NFS4 vs. NFS3 thing? Is there a way to force CentOS to use NFS3? I am pretty sure the CentOS 5.5 server is using NFS4.
I am trying to ptrace a vsftpd server process on linux to be able to get control whenever vsftpd process makes a system call. I start the vsftpd process and pass this process id as command line to the following program which traces vsftpd. however, when I run the following program it just hangs and does not print anything.
Recently installed debian lenny.Running very well.WM: ratpoison.Suddenly, today, the internet runs much slower.ONLY ON THIS MACHINE - the parallel windows machine, off of which I was successfully weaning myself, is not having this problem, so it's neither the router nor the actual internet connection.Specifically: I can't attach a document in yahoo mail.I tried it in iceweasel, same problem, so it's not the browser.I piped the output of "top" for a minute or so to a text file, which I'm attaching.
i have recently been having random problems with mostly firefox. the 2 most annoying problems (that work in other browsers) were: no sound on various sites, like videos and amazon mp3 previews. the mailto command would not work.
the fix is 2 part: i installed the standard release of thunderbird 3.1.1 and the firefox mailto problem is fixed. i installed the standard release of firefox 3.6.7 and sound now works fine.
additionally i have removed thunderbird, firefox and related ubuntu brandng from the package manager for the following reasons:above fixes, time and frustration involved in debug.eliminates the chronic application directory rename during updates, /usr/lib/... and softlink change in /usr/bin.allows me to keep up with t'bird and firefox security patches as they are released, i.e. immmediately.
I'm using my Linux (SLES 10) server as a File Server at this point. I need to set File Permissions to nested folders differently to different groups. For example:
homesharedengineering* should be read only for groupA homesharedengineeringadmin should be read & write for groupB Plus read only for groupA homesharedengineeringautocad should be read & write for groupC Plus read only for groupA
I've been using Webmin and Putty to set permissions but Putty only allows me the Default Group, it won't allow me to set several groups on the same directory. Webmin seems to allow me to add multiple groups (Webmin --> Others --> File Manager --> Info & ACL tab will provide extended abilities) but when I add multiple groups, they don't seem to take effect? I'm wondering if my setup at the 'Share' level or at the hierarchy of my folder structure (unix based) needs to be set specifically?
I'm trying to transfer a directory, and all it's sub-directories and their contents, to another FTP server via SSH. I'm using CentOS Server 4.4. I can 'put' a single file, but can't figure out how to do everything at once, which I need to do.
I have a file with 5 columns. Column 4 contains numbers.Is it possible to split the file into multiple files using a condition for the contents of column 4 i.e if column 4 contains a value between 0-10 then print the lines to a new file called less_than_10.txt
I am using nessus to scan systems, and it generates the report file without any specific delimiter. Shown below is how a typical line in the file reads (It might show up as a very long line.
xx.xx.xx.xx|cpq-wbem (2301/tcp)|49272|REPORT|Synopsis :;;The remote web server is affected by multiple vulnerabilities.;;Description :;;According to its self-reported version number, the HP System;Management Homepage install on the remote host is earlier than 6.2. ;Such versions are reportedly affected by the following ;vulnerabilities :;; - Session renegotiations are not handled properly, which; could be exploited to insert
The field of interest (for me) are IP ADDRESS, Synopsys, Description, Solution and Risk Factor. I tried the following command to parse this:
But this would not give desired results as ':;;' is not a standard delimiter and hence gives incorrect results. how to grab everything that occurs between the strings 'Synopsys', 'Description', 'Solution' & 'Risk factor' in the above example? In other words i may be asking to use multiple delimiters.. in the order of occurance.. is there a tool that supports multiple delimiters?
Im new to the forum and fairly new to Linux as well. my apologies if this is posted in the wrong section. My problem is How can you search for a file in multiple machines (like a server farm) ? For example i log onto machine num1 and want to search for a file named "xxx.yyy" which may be in one of 4 machnies. What i do right now is manually log into each machine and run the find command. However i have heard that it is possible to do it via a couple of simple commands. I have looked into pssh and cssh as well as ssh tunneling (along with public key authentication to stop the machines requesting a password every time i log in) and unfortunately i was unable to find an answer.
One of the most common qs i see in forums is >>> How to convert multiple jpg files to one one pdf file by one click . I have got 2 solutions that I consider solve this situation to the best ....
*1.* Install SCRIBUS from synaptic package manager . It can convert jpegs to pdf without any issues . F Spot Photo Manager can also be used .
*2.* This 2nd step I find much better .... I have been using it myself for some time now ... and its flaw less ...
A) Install WINE using the terminal . For beginers .... Its a software that can run selected windows applications on a linux destro like ubuntu. Installing is very easy >>>> open synaptic manager and in quick search box type Wine. Once the wine files show in the search mark them for installation. For details on easy manual for wine installation check here .. [URL]
B) Now go to the page below .... Its a free software.
Down load and save the software from the link given . Open the file using wine ( Right click on the file and choose the option open with wine.) The installer will run and the programme will be installed . And Viola ... Its ready .... Import ur jpg files and merge them all in to 1 pdf file .. Very Useful In Merging Comics together.
I work for a school consulting company.We helped a school deploy about 1500 computers.The computers have windows XP but we have been using G4L for the restore partition on the drives.So far the software works great. We did however run into a problem in that many of the computers we deployed are missing the restore partition. The reason they are missing is long and convoluted and not really that important. What I have been charged to do is try and fix the restore partition problem. One solution that I had, which im not even sure if it will work, was to backup the recovery file, that g4l created, to DVD and write a basic script to recreate the partition and then copy the file over. This process would need to be as automated as possible since this disc will be inserted by the end user(the students). The backup file that g4l created is 5.9GB so it wont fit on just one disc and Dual layer discs are too expensive to use for this project, so the file will either need to be compressed again (not sure if that's a good idea or not) or split across two DVD's.
I have searched the forums here and I was not able to find anything to fix this problem. I was able to find some info on splitting files across two discs but im not sure how to use that to fix my problem.
I'm not sure if this is possible or even where to start. I assume that this can be done with an sh script using tar or similar.I have several very large zip files that contain images for all of the products in my online store. Each image is named after its 13 digit SKU (for example, 9987788000012.jpg). In order to import products into my store, all images are placed into a media directory. Unfortunately, there are over 100,000 images.
So I would like to break the images into sub-folders based on file name. For example, when I extract store_images.zip (or tar or whatever), my extract script would create directories (if they don't already exist) based on the first three digits of each image name, placing each image into the appropriate bottom level directory. For example, "9987788000012.jpg" would be placed in the following directory "media/9/9/8", with media as the root and "8" as the directory that holds any images that start with "998". Perhaps two sub-folders would be less cumbersome.Assuming this requires a script, particularly since it involves scanning image names, creating folders, and saving images to specific directories, which language would serve my needs best? PHP? Has anyone had to do something similar?
Despite the history file is unique (~/.bash_history) I see that the multiple bash processes run on different windows are not all updating that file. I presume that the bash is not taking into account this possibility (multiple bashes on multiple windows) and writes thus the history file in a simple straightforward way. This would mean that a number of history entries are lost. I've tried to find information but had no luck so far.