I must have not read through the tutorial first.. I made changes to a project managed with git. A local copy of the repository exists on my machine (as is supposed to be with git). No branch was created for my additions. I've been committing changes to my local repository and doing pulls/merges with the remote repository. I do not have permissions to push changes to the remote repository forcing me to create a patch to email to someone who does.
$ git status
# On branch devel
# Your branch is ahead of 'origin/devel' by 6 commits.
No tracked files are now modified. I cannot figure out how to create a patch that reflects changes between the remote repository and the commits of my repository. All tutorials I've read online state that you make a patch before committing to local. Is there a way to do this with git? Or must I download a separate copy and manually perform a diff?
I m new to setup the CVS in linux , i have tried lots but every time whenever i try to commit the change in module or file this error has been generated : cvs [commit aborted]: 'root' is not allowed to commit files.
I've a friend who would like to install openSUSE 11.3. But due to a bad Internet connectivity he cannot download the updates and the extra softwares. So I was thinking if it was possible to create a patch and a software cd for him which could then be used as a repository in YaST ?
I just want to upgrade my Slackware 13.1 kernel (184.108.40.206) to the latest stable kernel from kernel.org (220.127.116.11). I have never done anything like this and I am a Linux newbie, so I would appreciate a "Kernel Patching for Dummies" version if possible. I did do a search on this forum and most of what I read was over my head. I found an FAQ on kernelnewbies.org on "How To Apply A Patch" but when I attempted what they suggested, it said it couldn't find the file to patch at line 5 and asked me which file to patch. So I CTRL-Z'd out of there and came here. Here's what I tried:
I'd like to share my notes with all on how I create local yum repositories. By local I mean that the rpm files are on a computer here at home, not somewhere out in cyberspace. I've done this since Fedora 9 with both the "Fedora" repository and the install DVD.This how to was written with Fedora 14 in mind but should work equally as well with any earlier release.Hopefully, if we did everything right, there will be no errors and you will see your new repository listed amongst the normal Fedora repositories. Your new repository is now ready for use.This completes this section on how to create a local yum repository for the installation DVD rpm files.
I am trying to create a local mirror/repository for CentOS 5.6 for updates and network installs. I have read the page here on the site about how to do it over and over again, and I still can't figure it out. I have already created the directory, but right after that, I can't go any further. I know this is much easier than I probably think it is. Would someone mind telling me, (in beginners terms) on how to do this? I just want the 'os' package and nothing else.
I'm trying to install a patch but when I copy it into terminal I get message " /home/john/patch-modules_v62-opensuse.sh 'vmware-7.1.3-2.6.37-rc5.patch' not found. copy it to the current '/home/john' directory. Exiting" But I have it in my home directory!
Well my local LUG has developed a cooperation with the university. Their major is music. So we would like to develop a local repository to help the updates to be faster. I checked the How to set up a local Yum Repository on Fedora 8. I guess it's same with 12. My questions are:
1. How much space do we need?
2. Since we might setup repository for another distro, is it a way to be done with other distro? The above address describes only how to make it in Fedora.
3. Since we'll have it installed and will be usefull for the university, can we make it available to the rest of the world? Maybe contact someone for this.
We have several linux hosts at our site. I would like to set one of the servers to offer a local update service rather than have each host contact the public servers for update information. This would greatly speed the updates for each host since we only have a DSL connection (about 1.2Mbps transfer rate). Is there a tutorial somewhere that describes how to set up a "shadow" update server?
I am using SuSE 11.2's built in Samba and LDAP server (that comes with Yast) on our network. Everything connects fine, I can access shares, I can authenticate correctly, but I just need help on how to create a local admins group so Windows users can install their own software and such.
I'm in the process of creating local repos for our company servers (CentOS 5.5) and laptops (Fedora 13). And while the CentOS part went perfect the Fedora part is causing major trouble.
But first things first, here's the setup: a central CentOS 5.5 server is running Apache2 and has a VirtualHost listening on Port 8080 for both CentOS and Fedora. The DocumentRoot for this VirtualHost is /data/repo wherein two directories, centos and fedora, reside.
This is the .repo-file for CentOS that works like a charm:
[local] name=CentOS-$releasever - local packages for $basearch baseurl=url enabled=1
After the CentOS repository was up and running I follwed the exact same steps and was expecting to find a Fedora repo up and running (I know, I know, silly me...). However, yum complains it cannot download the repomd.xml. The .repo-File is as follows:
I'm new here and hope to profit from your immense linux knowledge and of course to share my own experience where I can.
I'm in a student organization and we use a file server that runs linux. I can log in through ssh and copy using scp using login and password (no rsa/dsa keys because most users are windows users using winSCP and they're lacking in computer knowledge so we don't require them to mess around with keys)
However, I don't have network access everywhere, so I'd like to make a copy on my laptop harddisk of some of the folders I use most frequently. Note that I don't need it to copy files from my pc back to the remote server so I don't need two-way sync. Deleting the local copy every time and downloading a new full copy is not an option as we are talking about several gigabytes and the download speed is limited. Normally I would use Unison, however, this requires unison to be installed on both pc's and I can't install any software on the file server so this is not an option.
Any ideas on how to achieve this? I'm reasonable knowledgable about linux so I don't mind tinkering with some config files and using command line applications.
I dont have a internet connection and I want to install VLC. I want to create a local repo for a vlc by downloading all the required packages. I tried to do by downloading all the packages from from VLC repo and create a local repo but failed. Its say this dependency failed and that dependency failed. i tried to manually install each and every dependency but still the problem exists.
I am currently interning at a place and my job is to essentially learn UNIX. My supervisor gives me problems here and there to help guide me with my learning but for the most part I'm doing this all by self-teaching myself. Needless to say I have run into a few obstacles...for instance-Create a *one* line command that, using tar, will collect the full /usr/local directory (you need to run this as root again) and copy the whole /usr/local structure under /optFor example /usr/local/bin/hello will become /opt/local/bin/hello, etc. I want this as follows:1. /usr/local is collected by tar, but the output of this tar command is its stdout.. what you get from the previous stdout, you compress with gzip and send it to stdout again 3. get this output and decompress with gzip.. get this output and pipe to tar in a way that will extract the tree under /opt.If anyone knows how I could go about doing this, please let me know, or at the very least point me in the right direction. What I've got so far (which could be completely wrong) is:tar cvf - usr/local/ | gzip -c - | gunzip -c - | tar xvf -in theory I feel like this should work (except for extracting the tree under /opt...i'm kinda stuck there)
I'm in the process of creating local repos for our company servers (CentOS 5.5) and laptops (Fedora 13). And while the CentOS part went perfect the Fedora part is causing major trouble.But first things first, here's the setup: a central CentOS 5.5 server is running Apache2 and has a VirtualHost listening on Port 8080 for both CentOS and Fedora. The DocumentRoot for this VirtualHost is /data/repo wherein two directories, centos and fedora, reside.
This is the .repo-file for CentOS that works like a charm:
Quote:[local] name=CentOS-$releasever - local packages for $basearch
I'm working with a dual-boot laptop running Ubuntu 10.0/Windows 7 and a Debian 5 VPS while the OS's shouldn't have much impact on my question.
What I would like to do is create a html page that I can upload to my VPS which lists all of the files/folders on my local 2TB hard drive (Specifically media such as Movies, Music, TV Shows...). The media obviously will not reside on the server, but I would like to at least have a list which will allow me to select, for instance, a bands artist so that it redirects me to the albums in the directory below.
Ultimately, I'm looking for Open Directory Browsing without actually having the media on my server. I have been attempting to create something to this effect using lynx, however, I'm not sure if it can be done with this command or if it's even possible for that matter.
I opened nautilus from gnome-terminal, then (With the options in context menu of right click) I realized a checkout, and when I click on the option "commit" nothing happens. But on terminal I receive the error shown in this screenshot...
I have installed SVN and running it fine. Update and commit and all are running without issues and running it under DAV from apache on Ubuntu server.
Now I need to run post-commit hook to auto update a testing staging server to test the changes. Without post-commit hook, everything works fine. The post-commit hook if run individually as a script manually runs fine without issues and it updates the said working copy. But if it is run by SVN, it somehow does not run and gives error to the client from where commit was done. This is the error I get for the post-commit script:
I have been searching EVERYWHERE for a solution to my problem, I have been sent to the red-bean "answer" so many times I think I know it word by word.Here is my problem, I run a WHM/cPanel server and want to make a repository (/home/<user>/svn) update a working copy (/home/<user>/public_html/dev) to work on a dev. subdomain on each site we have.I have the the SVN part working but can not get anything about the post commit to work. I even simplified it to the point that all the /hooks/post-commit file contained was.
I'm attempting to use svn (client-side) on an Ubuntu 9.10 virtual machine and, although I am able to check out without issue, I seem unable to commit changes to the repository and get the error:
svn: Commit failed (details follow): svn: Server sent unexpected return value (403 Forbidden) in response to MKACTIVITY request for '/svn/rockoon_repos/!svn/act/8ed27527-fbfc-4dd5-b2aa-5a4d4b3e4e14'
I'm pretty sure this has something to do with the fact that the svn server uses ssl authentication (the server is running subversion+apache on a windows server box in case that matters).
The strange thing is that I am able to commit to the same repository on another computer using Tortise SVN (ie on Windows) without issue, so it seems that the issue must be client side (most resolutions I found for this issue online seemed to be server side fixes). I've also checked to make sure that I'm using "https" in the checkout URL.