Fedora Hardware :: Low Transfer Rate When Copying Large Files Over Wireless
Jan 11, 2010
I just bought a HP 3085dx laptop with an intel 5100 agn wireless card.
The problem: copying a big file over the wireless to a gigabit hardwired to the router computer only gives an average 3.5MB/Second transfer rate. If I do the same copy from my wireless-n macbook pro to the same computer. I get a transfer rate of about 11MB/sec. Why the big difference? I noticed the HP always connects to the 2.4 GHZ band instead of the 5GHZ bands...
[jerry@bigbox ~]$ iwconfig wlan0
wlan0 IEEE 802.11abgn ESSID:"<censored>"
Mode: Managed Frequency:2.412 GHz
Access Point: 00:24:36:A7:27:A3
Bit Rate=0 kb/s Tx-Power=15 dBm
Retry long limit: 7 RTS thr: off Fragment thr:off
Power Management: off
Link Quality=70/70 Signal level=-8 dBm Noise level=-87 dBm
Rx invalid nwid:0 Rx invalid crypt:0 Rx invalid frag:0
Tx excessive retries:0 Invalid misc:0 Missed beacon:0
I am not getting any errors. I don't know why the bit rate is not known. My airport extreme base station typically reports that the 'rate' for the hp is typically 250~300MBi and about the same for the MacBook Pro. The hp is about 6 inchs away from the base station. Is there anyway to get the rascal to go mo'faster? Is there anyway to get the rascal to use the 5GHZ band.
I have Fedora 12 (with all the latest patches, including the 2.6.31.6-162 kernel) installed on a new Supermicro SYS-5015A-H 1U Server [Intel Atom 330 (1.6GHz) CPU, Intel 945GC NB, Intel ICH7R SB, 2x Realtek RTL8111C-GR Gigabit Ethernet, Onboard GMA950 video]. This all works great until I try to transfer a large file over the network, then the computer hard locks, forcing a power-off reset.
Some info about my setup:
[root@Epsilon ~]# uname -a Linux Epsilon 2.6.31.6-162.fc12.i686.PAE #1 SMP Fri Dec 4 00:43:59 EST 2009 i686 i686 i386 GNU/Linux [root@Epsilon ~]# dmesg | grep r8169 r8169 Gigabit Ethernet driver 2.3LK-NAPI loaded
[code]....
I'm pretty sure this is an issue with the r8169 driver (what I'm seeing is somewhat reminiscent of the bug reported here). The computer will operate fine for days as a (low volume) web server, and is reasonably stable transferring small files, but as when as I try to transfer a large file (say during a backup to a NAS or a NFS share), the computer will hard lock (no keyboard, mouse, etc.) at some point into the transfer of the file. It doesn't seem to matter how the file is transferred (sftp, rsync to NFS share, etc.).
I've discovered that Dolphin seems to lose random files when copying many large folders.
I first noticed this a few months ago when I tried to copy my music library from one folder to another on the same HDD. It consisted of around 600 folders and 6500 files. During the copy there were no errors but after the copy I found that some of the newly copied folders were missing files. I put it down to human error or a glitch.
Yesterday I tried to copy 13 folders containing rips of some of my DVDs. Each folder basically had one film of either 700MB or 1.4GB. Again no errors showed up during the copy but I found 3 of the newly copied folders were empty.
It's not so critical with music or films but I can't afford to lose work data like this.
Has anyone experienced or seen a similar problem with Dolphin? I'm going to have to do some more extensive testing but this is not good.
The first time I noticed the problem I was running KDE4.3.4 (I think) and now the latest was with KDE4.4.0.
I've a directory containing around 2.8 lacs of files. I want to move them to another directory.If I use cp or mv then I get an error 'argument list too long'. If I write a script like
for file in ls *; do cp {source} to {destination} done
then because of ls command , its performance degrades.How can I do this?
I am facing problem in copying a large number of file 18 lakh (18,000,000) files from my personal hardisk to another hardisk each file is very small and size of folder is around 3.95 GB copying files using copy given by Windows is frustrating and I am not even able to compress file its giving me error that its not readable.And problem is I am not able to open this drive in Linux it showing me error there saying do diskchk in Windows and Windows disk check is also not able to repair this drive and goes into some mode unsolvable.Is there any way to open disk with error to open in Windows and if not any way I can copy data faster?ERROR: Disk labled EDU is corrupt go to windows and chkdsk /f there and reboot into window 2 times.
My friend bought an old hard drive. He noticed something with the hard drive that it was just replaced with a SATA socket. So meaning, the SATA socket was soldered to the PATA hard disk to replace the PATA socket to SATA socket to make it a SATA.
Now the question is:
1. Does the Transfer Rate of the harddisk (that has been replaced from PATA socket to SATA socket) would be SATA transfer rate? OR would still be PATA transfer rate?
For my research I have some very large files that are basically millions of lines of ten columns of numbers. These files can be up to 5 GB in size. Recently I noticed that when I made a copy of one of my files, some exclamation points appeared in it where there should not be any: in front of random numbers throughout the file. Making another copy of the file would result in exclamation points in front of different numbers in different parts of the file. Doing this many times has given me up to four exclamation points in different parts of the file. Sometimes the file copies just fine without producing any extraneous exclamation points.Additionally, I have occasionally seen a "^K" where there should be a newline (the data that should have been on the next line was instead on the previous line with a ^K in front of it) in copies that I have made of my files. I don't know if this is related or not.
I would like to transfer my music library and movie collection from my Desktop computer running Windows Vista and my laptop running Debian Squeeze. I have the laptop connected via wireless but it's possible to connect the two either directly with a CAT5e cable or through the router. I'm just wondering what the best way to do this would be.
how to transfer large files from my laptop to external hard drive. Problem occurs when I'm sending Blu-ray films (4.4GB) to external, gets to 4GB and then comes up with error. Is there any way of breaking it up and then merging when it reaches the hard drive or is there a way of sending it as one whole file.
I'm setting up a htpc system (Zotac IONITX-F based) based upon a minimal install of ubuntu 9.10, with no GUI other than xbmc. It's connected to my router (d-link dir-615) over a wifi connection configured for static IP (ath9k driver), with the following /etc/network/interfaces:
Code:
auto lo iface lo inet loopback # The primary network interface #auto eth0
[code]....
Network is fine, samba share to the media direction works, until I try to upload a large file to it from my desktop system. Then it downloads a couple of percents at a really nice speed, but then it stalls and the box becomes unpingable (Destination Host Unreachable), even after canceling the transfer, requiring a restart of the network.
Same thing when I scp the file from my desktop system to the htpc, same thing when I ssh into the htpc, and scp the file from there. Occasionally (rarely) the file does pass through, but most of the time the problem repeats itself. Transfer of small text files causes no problems, and the same goes for the fanart downloads done by xbmc. I tried the solution proposed in this thread, and set mtu to 800 in the interfaces file, but the problem persists.
I have a problem that I can't seem to fix.When I try to transfer a large file lets say 700Mb or so my wireless shuts down and i have to restart my ubuntu computer the other computer is vista.ubuntu is on a wusb54gver4 and vista is going through a wrt54g tm running dd-wrt mega.I have tried every thing i know with no luck.
I have a FC13 box that has both Gnome and KDE sessions installed.
I have noticed on the KDE session that data transfer rates are slower than when I use Gnome.
In Gnome, I can transfer files between my FC13 machine and my Ubuntu 10.04 pc at a rate of 6.5 MB/s (52 Mb/s if my maths is correct), but in KDE the rate is only 3.5 MB/s (28 Mb/s).
"ethtool eth0" shows my NIC speed as 100 Mb/s. Obviously I am not hitting anywhere near that speed in either session, (a separate article may be happen in the future to address that), but I am curious as to why KDE is that much slower for file transfer.
I am dual booting XP and Ubuntu 10.04, but in the future I will be getting a new machine and I will only be running Ubuntu and won't have access to iTunes. Because I have an iPod Touch, I have been trying to find workarounds for syncing everything that iTunes took care of in the past. One problem I have is managing movies. I have looked through various media players/iPod management tools (Amarok, Rhythmbox, gtkpod) and I am using Rhythmbox to sync my music and and attempting to use gtkpod to sync my movies.
gtkpod is able to sync songs (Tested with a few minute test clip) and short *.mp4 files (15mb I know for sure from test). I am unable, however, to get it to sync a movie (~700mb) I am able to drag it onto my iPod in gtkpod, but when I try to save the changes and write the files, it hangs at "Copying Tracks" at 0%. It eventually crashes during the couple times I have tried to wait it out. So this being my situation, my question is, is there a size limit to the *.mp4 files I can sync to my iPod Touch via gtkpod? is there any other tools that you know of that I can sync videos to my iPod with?
I had used DELL 1950 with 300 GB raid disk. Now, I purchased Dell 2950 with 450 GB (6 disk - 3 pairs of raid). I wanted to pull out old 300 GB from 1950 and put it in 2950 (temporarily) to copy all contents to the new system. How do I know which HDDs I need to pull out from 2950 so that I can replace them with 300 GB HDD to mount. I do not know how raid setup (I know unix alone - not raid commands). Is this possible? How to do it?
I have two servers, one has an empty / and the other has a subdirectory with a large number (4 gig) with many, many files. I need a way to transfer the files en masse from the server with the large number of files to the one that is essentially blank.I don't have space on the used host to simply gzip all the files. I've googled this and see that there may be some combination of tar and/or gzip that will let me do this with some sort of redirection.
I really need and example line of how this can be accomplished. If my explanation seems rather sparse, I can supply more details.
I have a LaCie NAS which is mounted on my main linux machine over a wifi LAN using the cifs file system. I would stupidly expect the transfer rate between my hard drive and the NAS to be limited by the Wifi speed (54 Mbps) but when I transfer files, the speed tops at 1.9 Mb/s which is roughly 15.2 Mbps. The most puzzling thing is that when I do multiple simultaneous transfers, I reach approximately 3MB/s in total but none of the individual transfers goes beyond 1.8. Does anyone have an idea about what is keeping the transfer rate so low?
Iam using public ftp server,the server file transfer rate is very very slow & the server is also very slow,In which way can i check the server to make it fast,can any one give sugession on this.
Sd card transfer rate is 133mb/s and ide transfer rate is same 133mb/s. Using no extra wire and direct connection, why is that sd card is much gradual than ide?
We have a Linux box which acts a a file server. Currently, files and directories are exported using NFS.At the moment, we are a bit concern on its data transfer performance. FYI, we are using a embedded Gigabit Ethernet port on the file server. We ran a few simple write tests between NFS client (also utilizes GigE port) and the NFS server. In these tests, both NFS server and client are both connected directly to each other with a Cat5E cable. Unfortunately, the write/transfer speed results are not as per our expectation. It scores roughly about 11-12MByte/s, where as theoretically Gigabit Ethernet transfer rate is able to reach up to approximately 120MByte/s.I wouldn't expect to reach the theoretical max transfer rate (it would be great if we can , but I would appreciate if you guys can share with us in terms of the following :
1) What's the practical max data transfer rate which you guys managed to observe in a normal Gigabit based connection? What about jumbo frames configuration?
2) Is there any additional tuning/configuration we need to do within the OS to reach those practical max data transfer rate figure?
3) Does PCI-e / system bus plays a role in achieving this speed? For example, we are using the embedded GigE port and we heard some people says embedded ports are actually sharing the system bus and resources with other devices, which might adds into performance issue. Correct me if I'm wrong.
4) Does converting to Cat6 cabling will guarantee an increase in the data transfer performance?
5) In the future (once we are clear on how much single GigE transfer rate we can go) , we are looking into doing bonding since that the NFS server's shared directory/volume read-write speed is way much higher (i.e 400-600MByte/s). Will bonding allow us to achieve higher NFS read/write speed? What are the bonding modes best used for this purposes? Appreciate if anybody who has experience in doing bonding for NFS can share their experience.
I would like to know if it is normal to experience 10MB/s data transfer rates during copying between partitions on my local hard drives (Toshiba 250GB 5400rpm SATA) while having three times faster (30MB/s) transfer rates between local partitions and USB drives (Kingston 8GB).
I am struggling trying to understand the reason for a fairly slow data transfer rate between two machines. ( tried point to point and also via a 1 gb switch ) One is nfs/http/ftp server ( with raid1 and lvm on top ), the other one my desktop pc. Both OS with default options, no changes to kernel in proc or other sort of thing.
Hardware is full recognized and perfectly working: The server has 4gb ram, Intel Core 2 Duo CPUE6850 @ 3.00GHz, 1000Mb/s NIC card and Lucid 10.04 64 bit, 250Giga Hard disk. The client has 3gb ram, Intel Core 2 CPU 6320 @ 1.86GHz, 1000Mb/s NIC card and Ubuntu Maverick 32bit , 150Gb Hard disk.
Raw data is good: gettons@gettons-desktop:~$ iperf -c MYSERVER ------------------------------------------------------------ Client connecting to MYSERVER, TCP port 5001
I have not tried this, and I am only wondering about the result.Let's say that I have a PC/Laptop with two network devices: an ethernet and a wireless. Can I connect both of the to the same network (if this network allows both connection) to increase the transfer rate between the PC/Laptop and the server???
I have 2 10.04 machines connected through a switch, both with gigabit on board ethernet. Both machines show 1000 Mb/s connections.When I transfer large files (gig plus up to multi-gig) the maximum I get according to ftp 11472 kB/s.I did rough computer school math in my head and that seems low but I'll admit I know very little about network transfer rates.My question is what transfer rates should I expect to get between the 2?
I am trying to copy a file from a network resource on my Local Area Network, which is about 4.5 GB. I copy this file through GNOME copying utilities by first going to Places --> Network and then selecting the Windows Share on another computer on my network. I open it and start copying the file to my FAT32 drive by Ctrl+C and Ctrl+V. It copies well up-to 4 GB and then it hangs.
After trying it almost half a dozen times I got really annoyed and left it hung and went to bed. Next morning when I checked a message box saying "file too large to write" has appeared.
I am very annoyed. I desperately need that file. It's an ISO image and it is not damaged at all, it copies well to any windows system. Also, I have sufficient space on the drive in which I am trying to copy that file.
Same computer and flash. Two results on different OS:
Write speed under debian: 0.4 MB/s Write speed under Windows: 6.6 MB/s
From what i've been reading on the net I guess it's something to do with the mount - sync option. I've notice that there are couple of workarounds but all of them are pretty tedious.
Is there any simple (GUI - one/two clicks) tool that allow to mount/unmount flash disk easily using the correct options?
I'm using Squeeze 6.0.2 (Kernel 2.6.32-5-amd64), gnome 2.30.2
I recently purchased an iRiver P7 media player which I am using with Rhythmbox under Ubuntu 9.04. The transfer speed when uploading files to the player is phenomenally slow - the first 50 or so MB reported virtually instantly then the speed immediately drops to about 12-15KB/sec yes KB/sec so a 3-400MB video file takes hours to transfer to the player. This happens with the internal memory and the added memory (SD card) too, althoufg it is faster ror the SD card. No other devices have ever behaved like this on this system (and I have used 2 different Creative players, a different iRiver player and numerous memory sticks). The computer is dual booting with windows (for once that was useful) and under windows it doesn't happen, the speed is as expected so it isn't the hardware. I don't want to have to resort to damned windoze just for this. dmesg after plugging in the player shows it identifying both the internal storage and the added SD card but I don't understand the rest of it:
[code]....
This last message gets repeated lots of times, and during transfers it seems to be issued continuously and I suppose this is why the system is so slow because it is always retrying loads of times before a success.
I'm planing to copy a productive mysql innodb file from one server to another, and the file size is around 300GB. As the file is keeping changing all the time, I have to shutdown mysql instance and copy the large data file to other server as quickly as possible.I should have to find a way to speed up file copying ... I'm wondering whether there's a way to copy file block by block.If the destination side block has same content, then bypass it.