General :: Uploading A *.php File To Real Server It Don't Transfer?
Nov 13, 2010Whenever I am uploading a *.php file to real server It don't transfer.
View 1 RepliesWhenever I am uploading a *.php file to real server It don't transfer.
View 1 RepliesI got an error while uploading a file named "xxxxxxxx.php"to linux server
Error in detail
Upload of file 'xxxxxxxx.php' was successful, but error occurred while setting the permissions and/or timestamp. If the problem persists, turn on 'Ignore permission errors' option.
Permission denied.
Error code: 3
Error message from server: Permission denied
Request code: 9
i am trying to transfer a file from my live linux machine to remote linux machine it is a mail server and single .tar.gz file include all data. but during transfer it stop working. how can i work and trouble shooot the matter. is there any better way then this to transfer huge 14 gb file over network,vpn,wan transfer. the speed is 1mbps,rest of the file it copy it.
rsync -avz --stats bkup_1.tar.gz root@10.1.1.22:/var/opt/bkup
[root@sa1 logs_os_backup]# less remote.log
Wed Mar 10 09:12:01 AST 2010
building file list ... done
bkup_1.tar.gz
deflate on token returned 0 (87164 bytes left)
rsync error: error in rsync protocol data stream (code 12) at token.c(274)
building file list ... done
code....
I have login of two ftp servers and I want to transfer the data from one ftp server to another ftp server. How can I do that, without downloading to local and then upload to other ftp?
View 7 Replies View Relatedi am new in tcp/ip.i want to write a program using c for file transfer where FTP client and FTP server will be used.and also this program should work for ipv4 as well as ipv6.and muiltple client can be connect simultaneously.i dont know how to start program.should i use shell script or socket programming for file transfer?can we use FTP client and FTP server in socket programming?
View 2 Replies View RelatedHaving a bit of a issue with Debian Squeeze and transferring files to the Sony PSP..Hook up PSP to USB port and Debian mounts it..I go to drag a 125 meg mp4 to video folder..Copy windows takes about 10 seconds to transfer it..Exit USB mode and there is no video there. Go back into USB mode and look at video folder on the PSP memory stick and there is no video..It vanished. From another after copy progress closed I right clicked PSP and unmounted it..
It error-ed saying device was busy and could not unmount..Looking at light on PSP i see memory stick is still being written to..i wait for light to stop flashing..About a minute or so..Then am able to unmount it..Go to PSP video and theres the video ready to be watched. Debian isnt accurately showing the copy progress...Its showing complete when it isnt..I have to watch the light on PSP to know when it is truly finished.
Let's say we have machine 1 and machine 2. Machine 2, which is linux based, is receiving a file (let's say cat.jpg sent via SCP from machine 1 (doesn't really matter which operating system..) How can I know in machine 2, that the file has finished uploading from machine 1 ? Please note that by "knowing" I mean by running a command (via a cron that runs every minute), and not by an event driven way ("make the scp issue a command when it is finished uploading"). Another limitation I have is that I can't change anything in machine 1 or in the way it SCP files.
If it really matters: the files being transferred are video file, MP4 extension. machine 2 is ubuntu 10.10
authors recommended to upload pages of a site in a condensed form to the server to save in uploads and downloads by visitors. This involved removing all unwanted spaces and making one single string whenever possible.Is this approach still encouraged?Is there software that does that and respects the line-oriented scripts like Javascript embeded in HTML code?
View 1 Replies View RelatedI'm having difficulties with uploading files to a CentOS-server with vsftpd. I have the exact same configuration on a Fedora10 and there I have no problems...
[Code]...
I have 2 computers on the same network that i need to link together to transfer files 1 is a web server the other is a minecraft server. the problem is that the file transfer will be constant as the minecraft server will constantly updates files on the web server and I dont want it to go to the router then to come back to the web server. I want to add a second network card to each computer and link them together and use this second connection to transfer the files is it possible?
View 2 Replies View RelatedI have a log file that I would like to examine during some changes under process that writes to this log.
Is there some way to open this file and read in real time changes written to it ?
setting up a basic apache webpage that would allow users to view files and folders on a server and transfer it from one directory to another? I know it doesn't sound very secure but this is an internal server so security is not a big issue as it is pretty secured in its network already. We have users that need to move files from one directory to another and we'd like to give them some intuitive interface to do this rather than having them log in to the linux system and run mv/cp commands. They are not very linux saavy.
PS: The system is running CENTOS 5.4.
I know this command exists I just can't seem to find it. I want to see the last few lines of a file as more are added in real time. Can someone point me in the right direction?
View 2 Replies View RelatedI was wondering if there is a command to show a real-time creation of files. I basically executed a command that will created thousands of files and takes a long time. I want to check if it is still creating additional files or if ti got frozen.
View 3 Replies View RelatedI have a few servers that are exposed to the internet. When someone tried to brute force hack in to the ssh, ossec adds their IP to the hosts.deny. Then the hacker (read: script kiddie) moves to the next IP up the line and hits my next server, etc, etc.
I end up getting 20 emails for all the servers that they hit.
My question, is there anyway to sync the hosts.deny file across multiple servers so that if they are locked out of one, they are locked out of all?
I'm trying to transfer a large .tgz file from a CentOS dedicated server to a linux webhost (unknown OS). The problem is the webhost will not allow a 1.1gb file to be uploaded, however it will allow the upload in 149MB chunks. I used the split command to segment my tgz into 7 segments under 150mb. I then uploaded all segments via FTP which worked. Then I tried to join the segments to create the original tgz. The join appears to work with no issues. However, when I try to extract the tgz it appears there is a problem, most, but not all files are extracted and there is this error message:
Code: gzip: stdin: Input/output error
tar: Unexpected EOF in archive
tar: Unexpected EOF in archive
tar: Error is not recoverable: exiting now It appears the join did not work and the tgz is slightly corrupt. What am I doing wrong? Here's the commands I'm using:
1. Create the original tgz on the dedicated server
Code: tar -czf mysite.tgz ./myfolder
2. Split the tgz into segments
Code: split -b 149m -d mysite.tgz seg
# using the -d switch so the segment files use a numerical suffix
# I now have these files:
seg00
seg01
seg02
seg03
seg04
seg05
seg06
seg07
3. Transfer segments to the other webhost using FTP
Code: # hand typing (not a script)
ftp ftp.mysite.com
myusername
mypassword
binary
cd somefolder
put seg00
put seg01
put seg02
# through to seg07
4. Join up the segments on the new webhost
Code: # this is in a .sh script file
cd /full/path/to/somefolder
cat seg* > mysite.tgz 5. Extract the new tgz
Code: # this is in a .sh script file
cd /full/path/to/somefolder
tar -xzf mysite.tgz
# the above error is now thrown.
That's it. What am I doing wrong that's causing the above error?
if connecting to my server for file transfer using gFtp is secure. I told gFtp to connect to the server using SSH2 and it works. It says it uses this command "ssh -e none -l wordpress -p 1883 IPADDRESS -s sftp." Is this more or less secure then using ftpes or ftps? What I thought was weird was that I could shutdown vsftpd and still connect. Does SSH2 SFTP use its own ftp server?
View 4 Replies View RelatedI have two questions. 1- How can I set up FTP server for the first time on the Centos? 2-I want to give the ftp user full root access in the directory of /var/www/html so he would be able to upload or download files and folders without getting "FTP Critical file transfer error". From command prompt how can I give the user test root access in the /var/www/html with all the folders, sub folders and files in one shut?
View 9 Replies View Relatedi find no solution to play .rm file which open in real player.
real player does not support linux.
how can i play .rm file in linux or altrnatively how can i install ral player in ubuntu 11.04
I have a ubuntu server 10.04 with LAMP installed. I also have ubuntu 10.10 on a laptop and can copy files to the server fine. To keep my website uptodate, I usually use Filezilla without any problems. I have just installed Fedora 14 on an old desktop and set up "my stall" ok. The problem is that I cannot copy any files from Ferdoa to the server due to:-
Response: 550 Permission denied.
Error: Critical file transfer error
I have tried to change the directory on the server "/var/www" using chmod -R 775* and chmod -R 777*, but it makes no difference, the file transfer still fails.
I have a centos server installation running, and have installed and configured vsftpd.
FileZilla works great. I am able to connect and transfer files both ways. I used this just for testing purposes.
What I need to do is get Fling File Transfer working. I can connect to vsftpd with Fling, but that is as far as it goes.Sep 20 11:18:44 ftp vsftpd[28286]: warning: can't get client address: Socket operation on non-socketSep 20 11:31:03 ftp avahi-daemon[2240]:
Fling's logfile shows this:
11:25:57 Connect: 192.168.40.41
11:25:57 Connect: 192.168.40.41 Ok
11:25:57 Upload: C:UsersPublicDocumentsFling 1
[code]....
I am newbie to linux, I am dealing with Fedora core 13. I have already succeedded in building an ad hoc network and a WLAN network. My next step is to get some networking statistics such delay, throughput, etc. So, I need traffic to flow through the network for both multimedia traffic and bulk data traffic. How to send files between nodes on the network (I found commands ftp, rcp but do not know whether they are the correct choice for this). ID their any application to perform this?
Also, I want to perform some video streaming between two nodes to study the behavior of real time data transfer. Is thier any application to perform this?
i have installed LAMP server on my ubuntu 10.10 i am having problem in uploading files
upload form
<html>
<body>
<form action="tmp.php" method="post"
enctype="multipart/form*data">
[Code]....
So yeah I want to transfer my system to the real Ubuntu. Problem is I can't boot into my Wubi Ubuntu cause I lost boot.
View 5 Replies View RelatedI'm setting up a htpc system (Zotac IONITX-F based) based upon a minimal install of ubuntu 9.10, with no GUI other than xbmc. It's connected to my router (d-link dir-615) over a wifi connection configured for static IP (ath9k driver), with the following /etc/network/interfaces:
Code:
auto lo
iface lo inet loopback
# The primary network interface
#auto eth0
[code]....
Network is fine, samba share to the media direction works, until I try to upload a large file to it from my desktop system. Then it downloads a couple of percents at a really nice speed, but then it stalls and the box becomes unpingable (Destination Host Unreachable), even after canceling the transfer, requiring a restart of the network.
Same thing when I scp the file from my desktop system to the htpc, same thing when I ssh into the htpc, and scp the file from there. Occasionally (rarely) the file does pass through, but most of the time the problem repeats itself. Transfer of small text files causes no problems, and the same goes for the fanart downloads done by xbmc. I tried the solution proposed in this thread, and set mtu to 800 in the interfaces file, but the problem persists.
I'm trying to check my server's bandwidth usage in real time, installed the following programs but none worked so far.
Iptraf - No results even when using iptraf -u
Tcptrack - Error : pcap_loop: cooked-mode frame doesn't have room for sll header
Iftop - No results, everything 0b
Are there any programs that displays bandwidth usage in real time and actually works on VPSes? Or getting real time bandwidth usage on a VPS is simply impossible?
I have a linux server and I have SSH Access to it.
How do I transfer my files to my desktop.
how can we transfer the file from putty session to windows host which resides remotely....
View 1 Replies View RelatedWe are having few computers in ms-dos 6.22 , we want to download file from Dos machine to redhat linux machine.
View 4 Replies View RelatedI have set up a local server for testing on my home network and installed openSSH. I can login using filezilla and SFTP and can even download files. Error messages saying cannot find directory (the directory I am trying to upload)?
Do I need to configure openSSH to allow this. I am using my usual ubuntu login. Maybe I need to set up another user for SFTP.