General :: File Transfer Program In C Using FTP Client And FTP Server?
Jun 8, 2010
i am new in tcp/ip.i want to write a program using c for file transfer where FTP client and FTP server will be used.and also this program should work for ipv4 as well as ipv6.and muiltple client can be connect simultaneously.i dont know how to start program.should i use shell script or socket programming for file transfer?can we use FTP client and FTP server in socket programming?
I need help to do a program that emulates the following shell command line:"sort </etc/passwd | less".Using a socket in a connection-oriented setup.First the contents of the file /etc/password are being fed into the sort program. The sort program is then sorting the contents of the file and sending it to another programm called less. Less is paging the results so that you can see one page of the results at a time.I know that i can solve the problem this way:
1. Create a server program that will stream the contents of a file to any client that connects to it 2. Create a client application that will connect to the server and download the contents of the file. The program should then sort the file. After sorting the file the client should connect to the "less" program and send the sorted results 3. Create a server program that will listen for connections and download whatever is sent to it. It will then display that data one page at a time.
i am trying to transfer a file from my live linux machine to remote linux machine it is a mail server and single .tar.gz file include all data. but during transfer it stop working. how can i work and trouble shooot the matter. is there any better way then this to transfer huge 14 gb file over network,vpn,wan transfer. the speed is 1mbps,rest of the file it copy it.
[root@sa1 logs_os_backup]# less remote.log Wed Mar 10 09:12:01 AST 2010 building file list ... done bkup_1.tar.gz deflate on token returned 0 (87164 bytes left) rsync error: error in rsync protocol data stream (code 12) at token.c(274) building file list ... done code....
I am trying to write a chat program with sockets as part of my project. I made a socket server, using fork ,I tried to connect it with different clients.Each client will send messages to others through server.So server has a shared memory storing every clients username and client id. But the problem facing is first client is able to sent messages to every other client formed after that.But second client onwards it is not able to sent message to previously connected ones. Server is not able to write on those client id's.
When accessing an NFS mount for a large (200MB+) file transfer, the transfer starts rapidly, then becomes slower and slower until it hangs. On several occasions, it has frozen the client machine. Both client and server are set to default to nfs version 3. Slowdown and hang also occur when connecting to FreeBSD NFS mounts.
Presumably (I hope), there is some sort of configuration for the client that needs to be set. what should be changed in the configuration? This worked out of the box in OpenSUSE 11.0.
I'm carrying out a project for my university (CIT in Cork, Ireland) and I'm using CentOS running over WMware. I have a server and a client. The server has no GUI (command line UI) while the client has a UI. I need to install a Simple Forum Machine application and I'm told to FTP the files into the server. I figured out that the best option is to load the files in the client via the GUI and then ftp them in the server. How do I transfer the files from a the client t o the server using FTP? I'm totally new to Linux so the more details the better. Also I'm trying to mount a USB key on the server but have had no luck.
I was trying to copy a few files to my phone via bluetooth when the file transfer window popped up with the progress bar stating "Connecting" and an error: the name org.openbox.client was not provided by any .service files..
I am using: openSUSE 11.3 Gnome 2.32, Kernel 2.6.38, obex-data-server 0.4.5-12.2,
This is a recent problem, and I can't pinpoint any change/upgrade that would cause this. Rsync transfer from Client to Server: sent 11756196 bytes received 1032741 bytes 138258.78 bytes/sec total size is 144333466390 speedup is 11285.81 Pinging back and forth from each machine is fine. No Ifconfig errors Client, but Server has RX packet errors.
Having a bit of a issue with Debian Squeeze and transferring files to the Sony PSP..Hook up PSP to USB port and Debian mounts it..I go to drag a 125 meg mp4 to video folder..Copy windows takes about 10 seconds to transfer it..Exit USB mode and there is no video there. Go back into USB mode and look at video folder on the PSP memory stick and there is no video..It vanished. From another after copy progress closed I right clicked PSP and unmounted it..
It error-ed saying device was busy and could not unmount..Looking at light on PSP i see memory stick is still being written to..i wait for light to stop flashing..About a minute or so..Then am able to unmount it..Go to PSP video and theres the video ready to be watched. Debian isnt accurately showing the copy progress...Its showing complete when it isnt..I have to watch the light on PSP to know when it is truly finished.
Is there a way I can install a windows program to webspace and run it from a linux client computer? To specify, I'm a student at the University of Minnesota. I have access to linux machines running Ubuntu, however, the space allocated to us is too small for the program I would like to run on these computers. I do have webspace I can use though. The thing I'd like to be running on the linux computers is a windows application requiring installation. So is there a way to put/install/(whatever else it might be called) this program onto my webspace and be able to run it from the linux computers? I know it's probably unlikely, but maybe?
I have 2 computers on the same network that i need to link together to transfer files 1 is a web server the other is a minecraft server. the problem is that the file transfer will be constant as the minecraft server will constantly updates files on the web server and I dont want it to go to the router then to come back to the web server. I want to add a second network card to each computer and link them together and use this second connection to transfer the files is it possible?
setting up a basic apache webpage that would allow users to view files and folders on a server and transfer it from one directory to another? I know it doesn't sound very secure but this is an internal server so security is not a big issue as it is pretty secured in its network already. We have users that need to move files from one directory to another and we'd like to give them some intuitive interface to do this rather than having them log in to the linux system and run mv/cp commands. They are not very linux saavy.
I'm trying to transfer a large .tgz file from a CentOS dedicated server to a linux webhost (unknown OS). The problem is the webhost will not allow a 1.1gb file to be uploaded, however it will allow the upload in 149MB chunks. I used the split command to segment my tgz into 7 segments under 150mb. I then uploaded all segments via FTP which worked. Then I tried to join the segments to create the original tgz. The join appears to work with no issues. However, when I try to extract the tgz it appears there is a problem, most, but not all files are extracted and there is this error message:
Code: gzip: stdin: Input/output error tar: Unexpected EOF in archive tar: Unexpected EOF in archive tar: Error is not recoverable: exiting now It appears the join did not work and the tgz is slightly corrupt. What am I doing wrong? Here's the commands I'm using:
1. Create the original tgz on the dedicated server Code: tar -czf mysite.tgz ./myfolder 2. Split the tgz into segments Code: split -b 149m -d mysite.tgz seg
# using the -d switch so the segment files use a numerical suffix # I now have these files: seg00 seg01 seg02 seg03 seg04 seg05 seg06 seg07
3. Transfer segments to the other webhost using FTP Code: # hand typing (not a script) ftp ftp.mysite.com myusername mypassword binary cd somefolder put seg00 put seg01 put seg02 # through to seg07
4. Join up the segments on the new webhost Code: # this is in a .sh script file cd /full/path/to/somefolder cat seg* > mysite.tgz 5. Extract the new tgz Code: # this is in a .sh script file cd /full/path/to/somefolder tar -xzf mysite.tgz # the above error is now thrown.
That's it. What am I doing wrong that's causing the above error?
if connecting to my server for file transfer using gFtp is secure. I told gFtp to connect to the server using SSH2 and it works. It says it uses this command "ssh -e none -l wordpress -p 1883 IPADDRESS -s sftp." Is this more or less secure then using ftpes or ftps? What I thought was weird was that I could shutdown vsftpd and still connect. Does SSH2 SFTP use its own ftp server?
I have two questions. 1- How can I set up FTP server for the first time on the Centos? 2-I want to give the ftp user full root access in the directory of /var/www/html so he would be able to upload or download files and folders without getting "FTP Critical file transfer error". From command prompt how can I give the user test root access in the /var/www/html with all the folders, sub folders and files in one shut?
In case that NFS client of fedora is set to cache nfs file handle, where is the cache stored such as /var/lib/nfs?If it's stored in memory instead of directory tree, what command can I know the combination of file names and file handles with?OS in my test environment: Fedora Core 6
I have a ubuntu server 10.04 with LAMP installed. I also have ubuntu 10.10 on a laptop and can copy files to the server fine. To keep my website uptodate, I usually use Filezilla without any problems. I have just installed Fedora 14 on an old desktop and set up "my stall" ok. The problem is that I cannot copy any files from Ferdoa to the server due to:-
Response: 550 Permission denied. Error: Critical file transfer error I have tried to change the directory on the server "/var/www" using chmod -R 775* and chmod -R 777*, but it makes no difference, the file transfer still fails.
I have a centos server installation running, and have installed and configured vsftpd. FileZilla works great. I am able to connect and transfer files both ways. I used this just for testing purposes.
What I need to do is get Fling File Transfer working. I can connect to vsftpd with Fling, but that is as far as it goes.Sep 20 11:18:44 ftp vsftpd: warning: can't get client address: Socket operation on non-socketSep 20 11:31:03 ftp avahi-daemon:
I have a CentOS + Samba server and Windows XP client machines. Users, passwords and permissions are entered on the server machine.users and passwords ( same as on the server ) are entered in the XP client machine.When attempting to access a public file on the server using a XP client machine and the IP address of my server, I am asked a user name and password and none of the already entered seem to work. I cannot access the server file (prompted again and again to enter user name and password). What did i miss
I'm setting up a htpc system (Zotac IONITX-F based) based upon a minimal install of ubuntu 9.10, with no GUI other than xbmc. It's connected to my router (d-link dir-615) over a wifi connection configured for static IP (ath9k driver), with the following /etc/network/interfaces:
auto lo iface lo inet loopback # The primary network interface #auto eth0
Network is fine, samba share to the media direction works, until I try to upload a large file to it from my desktop system. Then it downloads a couple of percents at a really nice speed, but then it stalls and the box becomes unpingable (Destination Host Unreachable), even after canceling the transfer, requiring a restart of the network.
Same thing when I scp the file from my desktop system to the htpc, same thing when I ssh into the htpc, and scp the file from there. Occasionally (rarely) the file does pass through, but most of the time the problem repeats itself. Transfer of small text files causes no problems, and the same goes for the fanart downloads done by xbmc. I tried the solution proposed in this thread, and set mtu to 800 in the interfaces file, but the problem persists.
I am transferring a file from a linux pc going to a windows ftp server using "ftp put". My file is a zip file that includes .txt files inside it. Here is what's happening when I transferred this file :I used ftp put for transferring and found that my transferred zip file was corrupted and couldn't be opened on the ftp server.I found the solution for this on the internet. I needed to use 'binary' to make it right.
I transferred again using binary and then ftp put the zip file onto the other end. Yes, it worked. My zip file wasn't corrupted anymore and I could already opened it on the ftp server. But the problem remains on the .txt files inside it. Converting the file into binary made my .txt files to be distorted and unreadable. I read from the internet that .txt files need to use Ascii instead of Binary to be readable, but if I use ascii it would cause my zip file to be corrupted again. I need to successfully transfer a zip file that contains .txt files using ftp put.
I am trying to transfer a file using tftp between two PC's. I have installed tftp-server, tftp, xinetd on both PC's. I am unable to transfer the file between the two systems. although the self transfer is working on both the system. I am getting the error "Transfer timed out"
cat /etc/xinetd.d/tftp # default: off # description: The tftp server serves files using the trivial file transfer