I got an task assigned to me, i have to create new ssl key, csr & crt files using openssl. But the file name must be of this kind (*.aaa.xx.aa). When I tried the file name starting with * its not accepting the file name. But when I tried with the file name starting with . its getting generated.
I've just discovered that crontab is creating a new file in the root directory every time it executes a cronjob, and it doesn't erase over the old file so there are thousands of files in the root directory, they have the same name as the script file (appended with a numeral) but are all blank.here is what one of the cronjob's looks like[URL]
I want to recompile Apache in order to be SNI supportive because I need to use ssl named based virtual host:
I referred to the following links:
[URL]
I installed the latest version of openssl which is now openssl-1.0.0 I ran the following commands:
./config enable-tlsext --prefix=/usr/local2 --openssldir=/usr/local2/openssl make make test make install
then to recompile apache with new SNI support I ran the following:
./configure --enable-so --enable-ssl --enable-rewrite --enable-unique-id --with-ssl=/usr/local2/openssl make make install
After that when I start Apache: /usr/local/apache2/bin/apachectl -k start I get the following error:
SSLStrictSNIVHostCheck failed; OpenSSL is not built with support for TLS extensions and SNI indication. Refer to the documentation, and build a compatible version of OpenSSL.
I would like to encrypt and decrypt zip file using OpenSSL keys. I have generated the keys and can encrypt normal text files but if I try to encrypt the zip file, I get error: "Error reading input Data" Following is what I have done.
I must use public/ private key pair (without any password) and I must use OpenSSL. But I can use any algorithm other than RSA (not sure which one to use and how).
For some years now I have been able to use openssl (apache-mod_ssl) to process encrypted traffic because I had, in effect, only one host - the main server - as the sole entry in our ssl_vhost.conf file.
Now we are working toward serving a couple of more secure sites for closely related organizations, but with their own distinct identities. This, in the past, would have meant additional static IPs with matching nic cards for starters. But my understanding is that since 2007/8 we have been able to use gnutls (apache-mod_gnutls) which gets around the old problem of Apache not being able to direct name-based traffic because that would not yet have been decoded. This is referred to as SNI - Server Name Indication.
Here my confusion begins. Is there an overlap between SSL and TLS? For instance, I would have generated RSA keys and a self-signed certificate with the genrsa command. Is this sufficient for gnutls or does it need to generate its own keys and certificates? I realize gnutls is relatively a new kid on the block but it is appealing and I'd like to give it a try.
I am working with the Mandriva/Mageia cooker with an x86_64 architecture so all packages are up-to-the-minute.
I am using fetchmail (with ssl support) to download my e-mail. About a week ago, the downloading from two servers stopped, among them gmail.com, due to an ssl error. Usually that means updating the ssl fingerprint in fetchmail.conf, and when I wanted to do this I discovered that there seems to a issue with openssl.
When I do openssl s_client -connect pop.gmail.com:995 -showcerts I now get an error
I've been scooting around Google to try and figure out how to create ackage files.pparently their created by a utility (which is already installed on my PC) called 'autopackage'. I've been to their website (www.autopackage.org) but I can't find anything on how to actually create the .package files
I'm trying to create a torrent file via the command line. I used "transmisson-create", but since the .torrent file is binary I can't read it using "less" or "nano". Is there a way of verifying that the torrent I created was correctly done? I know I could probably upload the .torrent file to my public server and then use another torrent client to test, but I simply want to see what's inside the .torrent file to verify that it got all the files I wanted. I cd'd into the directory with the files that I want to torrent and ran the following command.
I'm using openSUSE 11.4 (x86_64) with KDE: 4.6.00 (4.6.0) "release 6"Everytime I visit a direcotorya hidden file ".directory" is created. How to disable that? Is There a possibility to disable that behaviour only for public_html directory
I like to create files using the GUI context menu. I right click on windows and one of the options in the menu is "new" and then I have many options such as Word 2007 file etc. I create a lot of files and so this method is more convenient by far for me than opening the application, saving and then navigating to where I want the file to go. Because when I want to create the file, I will probably be in the directory in which I want to create the file, the process takes only a few gestures to complete which is not the case if I have to go through the file system in the "save as" window. Is it possible to have this functionality in Ubuntu?
rying to find a way to generate a PDF file from a text file in a command mode (X is not installed). Is there a simple way to do that?I don't need anything fancy - no special formatting and no images to include, just simple text converted to a PDF format
you can make a file that runs the contents as if typed out by the user in command prompt by saving the file as either .bat or .cmd . When using BASH in ubuntu, is it possible to save terminal executable files (from a text editor, like GEDIT), What file ending does it take, can i lay it out with just commands? (what is the syntax)
full snapshot of ArchLinux Repo for x86_64I want to use this as my restore backup should I need to reinstall Arch without network support.How do I build several *tar volumes of my /mount/my_repo to fit into 4.5GB DVDs ... the thing is 18 GB size...How do I extract all the *tar created to a folder later on...? is it the same as extracting a single *tar, will tar find all volumes in the same directory level so as to continue extracing or do I need to merge them in some way
how I can randomly write / create a 1 GB file in bash to test disk / network i/o? I was told I could use the 'dd' command but I don't know if there are some better ways and or what the 'dd' command looks like.
Concatenating two files without creating a newline between them? So how is it possible to do that? I've tried the following: Code: echo 123 > file1 echo 456 > file2 cat file2 file2 > file3
I have read alot about exporting or printing to PDF using open office. Haven t tried it but its pretty standard to be able to do that from a word processor so I'm not worried about it.here is an example of what I need to do. I just put together my wifes resume. I had the cover sheet in .doc I had the resume in .doc I had a pdf of some ground plans for large scale gardens she had done and I had a few new pix in jpeg. I was able to go into Lightening PDF hit create new, pull all those files in, sort them in the order I needed and it created a single mutipage .pdf from those files.
Any luck on finding something like that for linux? That would be just about the last piece of software (If I can make wings work for me) that I need to start testing ubuntu as a primary operating system.
Basically I download to main PC and rsync to my pi as the pi has a fraction of the speed for some reason when downloading form a source outside of my network.
The first few files will sync fine. Then I start getting errors like this
Code: Select allrsync: rename "/data/temp/.File.ext.y1716M" -> "File.ext": No such file or directory (2)
The directory on the host has user and group ownership as pi. Rsync has been setup to login without a password.
How would I go about creating a script to upload files to my webserver? I want it to only upload files that have been modified on the current date. So if I make changes to a file or touch a file, I run the shell (or perl) script to upload those files. This is what I have so far:
Code:
#/bin/bash FTPU="username" # ftp login name FTPP="password" # ftp password
I am using ubuntu 8.04 right now but I'd like to install ubuntu 10.10 or ubuntuStudio. Well I did it for trying as a dual boot just to see if it works. But I have files to back up. I tried to have them on dvds or usb but it takes time. I want to know if I can create a new partition where to stock my files and then erase the rest to install ubuntu 10.10.
When I am in gparted I see: /dev/sda 111.79GB: /dev/sda1 ext3 107GB /dev/sda2 extended 4.58GB -> I cant use it /dev/sda5 ext3 4.3GB /dev/sdb fat32 15GB -> I think its my ubuntustudio Is there any chance to create a new partition (40GB) to stock my files?
I am directly creating "qf" & "df" files into the sendmail queue folder and then processing this queue by command line. This is the only way to export data and email from this old application I am stuck with.This works quite well in my test enviornment but I am really new to linux/sendmail and just looking for any feedback on this process. Is this direct creation of queue files safe and any pitfalls that I should be aware of?
I have been browsing trough forums, manuals and people for a few days now wihout success so I thought i give this a go, any help will be very appricieted!
I'm using Matlab 2010a in Ubuntu 10.10. I also have a AVT Marlin F033C firewire camera and want to use the two together. Since most things i will do with the images will be processed in matlab I decided that the best would be if i could do everything in matlab.
The code for capturing a picture is however in c and uses the library dc1394 (usr/lib). Compiling this with gcc works good: $ gcc grab_gray_image.c -ldc1394
To be able to create a mex-file of it (that matlab can run) I have included "mex.h" in the headers and added the function mexFunction in the very top, this function should do more once it works...:
Choosing "Template Options file for building gcc MEX-files".
When trying to compile i get a few errors but these are easily removed by taking away all the //-comments and all the inline for a few functions (this should do anything for the functionality right?). After that I am still stuck with this error:
I should probably also note that I needed to rollback to gcc 4.1 since that is the latest one matlab 2010a can handle and now don't get that warning anymore.
There is a syntactical feature I would like to see in Vim, and so I thought to write myself a syntax file for my needs, but I am having trouble finding the exact type of syntax highlighting I want to incorporate'emulate. Currently in Vim, from the terminal instead of say gvim, when I place my cursor over a parenthesis or brace, the corresponding brace or parenthesis will highlight, or block its character. In my case, instead of a single character they will be full words specified within the syntax file.
I want to do something similar. For example if my document has three words: Group Element1 Element2
I want the syntax file to tell Vim that when I have my cursor over the word Element1 or Element2 that Group will become highlighted. In the same, to have the cursor over Group will cause both Element1 and Element2 to highlight. I have looked here but understanding still eludes me: [URL]. I would also be happy with creating some sort of command input, where I type in the group name and the elements highlight. Like a very specific search function that searches for only and all instances that are contained within the group named.
Here is the deal: my school purchased a test prep which they distributed to us, the only problem is that it is for windows. When I try to run the exe in wine the program windows pops up but the screen is blank (i can see the menu bar and and the title bar). It appears that most of the information is located in a folder as html files (separated into chapters for each section) with links to all the figures/tables/questions (also html files). This morning I have been trying to find a way to compile the html files into a format that I can read without having to hunt for each individual html file. I tried navigating with firefox but had to change each individual link before I could do that (very time consuming), I tried using doxygen (the GUI front end) to convert it to Latex/rtf/html/xml (all the formats) but did not have any success (It could be that I am using the app wrong). I tried opening some java applications found within the application data folder in hopes that it would bring up the navigator without success. And even tried installing java in wine (unsuccessfully) to see if that was the missing key to run the program in wine.
My question is, do I just have to fork out the $500+ that it would cost me to purchase the books instead of using the school provided program (free to us)? or is there a relatively simple way to compile these files into a format that could be readable.
tl;dr - have multiple html files for a windows native program that won't run on wine that I would like to convert to a more easily readable format (i.e. chm, pdf, etc), is it possible?the program was legally acquired through my school, I did not download or pirate it. I am not looking to distribute it, just looking to use it for my own studying purposes without spending $500+ or buying a windows license/computer.
Actuaaly i am creating watch on one directory in which files are continuously coming.Is there any command which can give listing of all files who have come in last 24 hrs.