Debian Configuration :: Rsync From Linux Server To Planet NAS
Jan 5, 2016
After hours of struggles I'm still not able to configure a simple backup from a Debian Jessie server to a Planet branded NAS.
Configuration is as follows:
NAS
user name: user
pass: user12
share name: user. Inside that directory, there is bkp directory where stuff should go. bkp was created using Dolphin.
The NAS offers smb service and is accessible through Dolphin file manager and from Windows stations.
Server also has a user named "user" and a password "1"..
Fstab looks like this:
//192.168.2.3/user /mnt/NAS cifs username=user,password=user12,iocharset=utf8 0 0
The command is: Code: Select allrsync -ahv /home/user/Desktop/ /mnt/NAS/bkp/
Errors are: Code: Select allrsync: chgrp "/mnt/NAS/bkp/." failed: Operation not permitted (1)
and
Code: Select allrsync: mkstemp "/mnt/NAS/bkp/.teszt.YEmVM3" failed: Operation not permitted (1) where teszt is a file created by me and YEmVM3 seems to be randomly generated character string.
Has anyone had any experience on using SUA(Services for UNIX Applications) rsync to "pull" files down to the Win2k3R2 server from a linux rsync host?I was trying to use cygwin rsync before until I found out from cygwin that the cygwin port of rsync was "flakey" and would fail intermittently for no apparent reason. cygwin suggested I use SUA or SFU for rsync services.
I've looked for/ am looking for any experience using SUA rsync to copy files down from a linux rsync host to the Windows host via rsync on the Windows host. Also, if you have done this successfully, do you have any pointers/caveats you can share on how you got it working? What I am basically looking to do is copy files and subdirectories of files from a linux host using rsync to some static location on a Windows server on a scheduled basis so that I can backup the windows server to tape using Symantec's Backup Exec application.
I'm doing it this way to avoid deploying the Remote Agents for either linux or Windows on the target hosts. As an alternative I've seen reference to a product called DeltaCopy that uses a native Windows rsync port with the native linux port of rsync to do what I need also.I realize this is not a strictly linux question, but more of a hybrid as I'm moving data to and from Windows and linux hosts. So, if this is too Windows-y a question, please say so and I'll withdraw my question.
Right now i have a HP DL 180 Server with 130 Gb Hard Disk & 8 Gb ram after Raiding0+1. i want to configure Domain Controller Server for my office for 200 to 300 Users. what should the partition size must be mentioned in my 130 Gb Hard Disk, is that going to be Sufficient for ME ?
i am bit confused about /Usr /Var /Boot partitions, as i need to manage perfectly in 130 GB
if i go with 4 Gb swap and remaining for " / " is that will be fine ? should i need to specify partition sizes separately for / tmp /var / usr ..
Basically I download to main PC and rsync to my pi as the pi has a fraction of the speed for some reason when downloading form a source outside of my network.
The first few files will sync fine. Then I start getting errors like this
Code: Select allrsync: rename "/data/temp/.File.ext.y1716M" -> "File.ext": No such file or directory (2)
The directory on the host has user and group ownership as pi. Rsync has been setup to login without a password.
I use openvpn to connect otherwise isolated machines, and use samba to share filesystems across the vpn, which works just fine.But I recently discovered that copying files using rsync -e ssh is so much faster than copying from a mounted filesystem - like about 5 times faster.I've got comp-lzo enabled in both server and the client, at least I think I have, the directive is there in both the server.conf and the client.conf files, but how do I check that it's active?Does anyone know if I can make openvpn behave more like rsync, because copying is easier than rsyncing?
I cannot see my debian linux machine in my virgin media router>>attached devices. I can see everything else including my raspberry pi.
The box is headless but I connected up to tv temporary and did ifconfig and got the ip address. I can ssh to it and connect to minecraft server I have running on it but I would like to know why I cannot see it in the router.
FLAGS=" --archive --one-file-system --acls --xattrs --numeric-ids --fake-super --progress --human-readable --ignore-existing --log-file="backup.log" "                                                        Â
FLAGS+=" ${DIRS} ${BACKUP_FILE}"
rsync ${FLAGS}Â Â
I then used iptraf to check the performance and I get miserable results (from the PC), even below 1 MB:
I have Debian 8 with VBoxGuestAdditions (4.3.18). I can connect to this box from Win7 using RDP (mstsc.exe). When I connect from Windows 10 using mstsc.exe I've got error code 2308. I tried NoMachine, 2XClient, RoyalTS none worked.
W10 connecting with RDP to Linux server? If I would install xrdp onto the Linux server would it work for me?
Is it possible to connect from Win10 to Linux using RDP?
Newly trying cluster configuration setup on RHEL5.3_64 bit machine.Basic Requirement :Need to Configuration GFS file systemHerewith I have shared details:
System : > I have 2 HP proliant Dl385 server. > Both system are connecting on Public network. (eth0) > I have connected eth1 - directly each other system like a Private Network (May be I am
I have a laptop connected to internet via wlan0. I also have eth0 interface and with it I share internet. I want to modify/filter all the traffic passing by the first laptop, something like this:
Code: Select all          *---------------------------*           |    LAPTOP 1      |   *--------------* ?           |-----*  *------*  *----*   |       |   INTERNET<------>|wlan0|<-->|MY_APP|<-->|eth0|<---->|ANOTHER LAPTOP|           |-----*  *------*  *----|   |       |           *---------------------------*   *--------------*
I know that in FreeBSD it is possible to use ipfw for that purpose, because it build-in into kernel. We set for example rule Code: Select allipfw add divert 2000 ip from any to 1.0.1.1
and we can use our own application to process those packets, reinject them forward etc. It will work also fast, because as I said, it build into kernel.
Is there any standart Linux-based solution to do the same? I found some info about netmap-ipfw. Is this a correct solution? Or I have to use for example IP-aliases and iptables to do that?
I need to process all the IP-packets, not only TCP/UDP/etc-protocol. Solution also must be very fast.
I have some old linux based games and applications that require old libraries to run.But in the mists of time, I've lost the scripts on how to set up a sandbox environment to run them. Any template for that type of fiddle? I should be able to resource libraries contemporary to the games from my collection of old distros, but my web fu on this matter(jargon) is cracked. It would have to be easier that continually trying to fiddle the MS OS version to run under Wine/PlayOnLinux. I've got Myst II, Civilisation CTP and Railway Tycoon II, although I think the last app I ran was Word Perfect 8.
I've a Lenovo G50-80T with W8.1. I want to install Debian 8.1 in dualbooting mode. I've done this other times without problems. But this time I want encrypt the Linux partition (not the Windows partition). I'll use dm-crypt to do that. I want to know if this way is secure for protect the data on Linux partition or if I need encrypt the entire drive.
Following VMware recommendations on how to change linux I/O scheduler for guests, I'm trying to do it on my VM machine running Debian Wheezy. At the moment I'm running kernel backports:
Code: Select all$ uname -a Linux brutus2 3.16.0-0.bpo.4-amd64 #1 SMP Debian 3.16.7-ckt20-1+deb8u3~bpo70+1 (2016-01-19) x86_64 GNU/Linux
The default I/O scheduler at the moment for device /dev/sda is CFQ, and I can change it on the fly to NOOP:
My card Wifi network doesn't turn on, even at loading configuration of the system is failed. I have to wait a long time until the system itself to load.
And when I write ifconfig or iwconfig is displayed in the console interface wireless card.
And when I come in the panel on the network card, it immediately pops up a window that you can not load the network card modules.
I don't know what could have happened, I do not recall that I recently messed with the system.
Here are the pictures: [URL]....
And here I join dmesg.log file, at the end of a failed: [URL] ....
I have a tiny shell script to rsync files between two servers and remove the source files.
This script works fine, when it has been initiated manually or even when the rsync command is executed on the command line.
But the same script doesn't work, when I try to automate it through crontab.
I am using 'abc' user to execute this rsync, instead of root, as root login to servers are restricted in all of our servers, by us.
As I mentioned earlier, manual execution works like charm!
When this rsync.sh is initiated through crontab, it runs the first command(chown abc.abc ...) perfectly without any issues. But the second line is not at all executed, and there is no log entry i can find at /mnt/xyz/folder/rsync.log.
I switched last summer from Windows (used it since Windows 95) to Debian. I'm using Debian Jessie for a couple of months now and I'm getting used a little.
There are problems here and there, but I can solved them with some reading on the web. Not really a big problem...till now
I run Debian 8.2 om my PC (PC1). Bought an older PC (PC2) that I want to use as a backup server.
I'm using PC2 only for making backups, after the backup I switch it off again.
So I installed Debian 8.2 (net-install without DE and with SSH) on PC2 and tried to configure it to let it work as my backup location. Made a public SSH key and exported it to the root account (no problem) and to the user account (sensdeb), but there was an error "Access Denied"
Gave the user (sensdeb) sudo-rights via visudo file
# User privilege specification root ALL=(ALL:ALL) ALL sensdeb ALL=(ALL:ALL) ALL
I installed rsync.
The problem is that Rsync only works when I use the root account.
I don know how to give the user sensdeb the rights so that I can use that account for my backup tasks. Now it's possible to sync with the root account, but that should not be the way to do it, I read many times.
I am using Centos 5. I got one error unexpectedly after some configuration for rsync and folder settings. The error i am getting is "Serve Authorization directory (daemon/ServAuthDir) is set to /var/gdm but does not exist. Please correct the configuration and restart GDM." after rebooting.
I tried to resolve the issue by setting permission for folder /var/gdm to 755. but the problem is still not resolved.
I'm using Ubuntu 10.04 LTS server and Postgresql 8.4. I have a .sh script that is run by cron every other hour. That works fine. The .sh script includes an rsync command that copies a postgresql dump .tar file to a remote archive location via ssh. That fails when run by cron; I think because it is (quietly) asking for the remote user's password (and not getting it). I set up the public/private ssh key arrangement. The script succeeds when run manually as the same user that the cron job uses, and does not ask for the password. I am able to ssh to the remote server from the source server (using the same username) and not get the password prompt (both directions), so why doesn't rsync work? I even put a .pgpass file in the root of that user's directory with that user's password, and the user/password are identical on both servers.
I think the problem is rsync is not able to use the ssh key correctly. I tried adding this to my script but it didn't help.
Code:
Here is the rsync command embedding in the .sh script.
I am *finally* getting around to rebuilding my file-sharing computer. I'll be sharing files with both Linux and Windoze machines. It's a home network, so there's nothing fancy needed. I know I have to tweak my smb.conf file until I'm satisfied with the features and security. I'm using SWAT and I'm starting with a bare-bones conf file. It's not secure but I can see the server and selected files/directories from my other Linux box.
My really dumb question is, do I have to reboot both the server and the client machines every time I change the SAMBA configuration? I thought I just had to stop and restart the SAMBA service in the SWAT software - but then the server disappears from my client. It looks like I need to reboot both machines for the client to see the server.
Program for simulating planetary movements, especially the movement of a planet in a binary system. It would be nice if the program has some kind of visualisation but I could do that myself. I searched sourceforge but only found several windows programs and Linux programs still in alpha phase, which didn't work for me.
I will be relocating to a permanent residence sometime in the next year or two. I've recently begun thinking about the best way to implement a home-based network. It occurred to me that the most elegant solution might be the use of VM technology to eliminate as much hardware and wiring as possible.My thinking is this: Install a multi-core system and configure it to run several VMs, one each for a firewall, a caching proxy server, a mail server, a web server. Additionally, I would like to run 2-4 VMs as remote (RDP)workstations, using diskless workstations to boot the VMs over powerline ethernet.The latest powerline technology (available later this year) will allow multiple devices on a residential circuit operating at near gigabit speed, just like legacy wired networks.
In theory, the above would allow me to consolidate everything but the disklessworkstations on a single server and eliminate all wired (and wireless) connections except the broadband connection to the Internet and the cabling to the nearest power outlets. It appears technically possible, but I'm not sure about the various virtual connections among VMs. In theory, each VM should be able to communicate with the other as if it was on the same network via the server data bus, but what about setting up firewall zones? Any internal I/O bandwidth bottlenecks? Any other potential "gotchas", caveats, issues? (Other than the obvious requirement of having enough CPU and RAM).Any thoughts or observations welcome, especially if they are from real world experience in a VM environment. BTW--in case you're wondering why I'm posting here, it's because I run Debian on all my workstations/servers (running VirtualBox as a VM for Windows XP on one workstation).
I'm trying to install supercollider following this guide: [url]
And I'm completely stuck on installing Planet CCRMA at Home. The guide says update your computer with PackageKit, KPackageKit. I believe I have both installed but cannot see them in the applications menu. Where can check?
Next the guide says: Run the following command in a terminal window: su -c 'rpm -Uvh
iam trying to sync file server data into backup server machine by command- rsync -avu path/of/data ipaddress-of-backup-server:/path/where/to/save after running it ask for root password and manually it is successful.but i want to make it automatic.for that i also tried cronjob and also generated authentication key but iam not successful in login automatically..anybody know how to authenticate root to login for storing data in backup server.
I'm running the current release of Debian with the 2.6.26-2 kernel. This is an upgrade from an older (2.4 kernel) series redhat release. One of the things I had working in the older system was a dns server with accompanying monthly update of the root hints file. I tried working through a dns how-to to set this up again, but it seems much has moved around since I last played with this. All of the files listed in the how-to are not where it says they should be. I am looking for a better reference on keeping the dns server running with current server information.
I have installed nfs-kernel-server on the server and nfs-common on the client. Assumeserver 192.168.1.1client 192.168.1.3
content of /etc/exports is: /home 192.168.1.0/255.255.255.0(rw,sync,no_root_squash) /home/nfsroot 192.168.1.0/255.255.255.0(rw,no_subtree_check,async,no_root_squash)
I wanted to setup my server as a router/gateway just for educational purposes and also when it succeeded i will keep the gateway and place it on the front line right after my modem.
My current setup as it is now :
The GATEWAY/ROUTER will be doing the main work to act as a dhcp server and firewall.
Now my setup as it is now :
So now what i ment to do was :
Let the clients connect to the gateway via ETH1 and all the stuff being router to ETH0 which is the NIC that is connected to the internet.
But now i have two huge problems :
1. The DHCP configuration seems to be fine the client machines get their ip adresses and /etc/resolve.conf looks fine to. But somehow the GATEWAY/ROUTER wont let me use the specified external DNs server so i cant do anything on these machines.
2. The firewall script [url] at line 27 should be DROP but if i use this line of code the GATEWAY/ROUTER cannot use his dns server settings so when i try to browse the internet or ping a machine outside my local network it would came up with nothing.
I ran into this problem myself, and I've seen that others have had the same problem when updating from Karmic to Lucid via apt-get. There is a bug in apt for Lucid, and it has been noted on one of the Ubuntu lists
This is a simple how-to and a work-around to the upgrading with apt-get:
1) Update your sources.list file to replace all instances of 'karmic" to 'lucid' : $ sudo sed -i 's/karmic/lucid/g' /etc/apt/sources.list
2) Run apt-get update: $ sudo apt-get update
3) Run a regular upgrade; $ sudo apt-get upgrade -y
Here is where the problem comes in. Apt is looking for the util-linux package, which is part of upstart-job, but someone apparenlty forgot to write the code to tell apt-about this (whoops!)
So, you now need to install upstart-job by itself. By skipping this step you will get the error message when you try to run step 5: "E: Could not perform immediate configuration on 'util-linux'.Please see man 5 apt"
4) $ sudo apt-get install upstart-job
5) Now you can run apt-get dist-upgrade: $ sudo apt-get dist-upgrade -y
I've been struggling for a couple of days to understand how configuration files work in Linux but haven't really found any clear information regarding that.My main question, as there is not a standard configuration file format in Linux, is how the format of a configuration file is understood by a Linux program? How the program reads its configuration file and how it understands its syntax (file format)? Also, a further question is how the program knows where its configuration file is being located?
For example inittab has its own "Identifier:RunLevel:Action:Command" format. How this is recognised by the "init" process?