Slackware :: Get Ftp To Download Large Files?

May 29, 2010

I am using xfce desktop and terminal.

1) How do i check to see what is the default FTP Client used by Slackware12.2 when i type:
$ftp

2) How do i get ftp to download large files? > 2gb from ftp server elektroni.phys.tut.fi or any ftp server.

View 14 Replies


ADVERTISEMENT

OpenSUSE Network :: Large File Download Will Start And Then Freezes

Mar 26, 2010

I've started having problems with large file downloads. A download will start and after a while freeze. The downloads window reports the correct connection speed and gives an estimated time to complete, but it stays frozen. Small downloads, torrents and surfing are not affected. I can do everything else normally even when the download is frozen. I've checked with my ISP and everything with my equipment checks out.

View 9 Replies View Related

General :: Split A Large File To Download On A Windows Machine?

Apr 16, 2011

I am removing some old graphics from my server and one of the gallery programs have created two enormous directories that I cannot even open with FTP.

I tried to tar each directory and the first came out to about 37gb and the second keeps failing (its bigger one would assume).

How can I archive and split these into smaller files?

View 13 Replies View Related

Ubuntu :: Command With The -r Option To Compare A Large Number Of Files And Files In Subdirectories

Jun 16, 2011

I am using the diff command with the -r option, to compare a large number of files and files in subdirectories. My main interest is to find out which files have been changed, and not what the actual changes are, and since a lot of files has been changed, it would be a lot easier to view the file names only. Is there and option for diff that might do this, or does there exist a similar tool/command that could do the job?

View 1 Replies View Related

OpenSUSE :: Dolphin Losing Files When Copying Many Files Or Large Folders?

Feb 14, 2010

I've discovered that Dolphin seems to lose random files when copying many large folders.

I first noticed this a few months ago when I tried to copy my music library from one folder to another on the same HDD. It consisted of around 600 folders and 6500 files. During the copy there were no errors but after the copy I found that some of the newly copied folders were missing files. I put it down to human error or a glitch.

Yesterday I tried to copy 13 folders containing rips of some of my DVDs. Each folder basically had one film of either 700MB or 1.4GB. Again no errors showed up during the copy but I found 3 of the newly copied folders were empty.

It's not so critical with music or films but I can't afford to lose work data like this.

Has anyone experienced or seen a similar problem with Dolphin? I'm going to have to do some more extensive testing but this is not good.

The first time I noticed the problem I was running KDE4.3.4 (I think) and now the latest was with KDE4.4.0.

View 9 Replies View Related

Software :: Differentiate Two Large Text Files Using Shell Script / Files Are Like Below?

Jan 20, 2009

I want to automate this using script.How to automate it?

File1:
s.no# 1 name:aaaaaa
city:abcd

[code]...

View 1 Replies View Related

Slackware :: Slapt-get - Failed To Download: Incomplete Download

May 1, 2009

I've been getting this error when trying to upgrade packages with slapt-get:

Failed to download: Incomplete download

Not sure what's going on. I tried to use --clean and tried to use a different mirror, but still getting the error.

View 7 Replies View Related

Slackware :: Freeze On Large File Copy With NFS?

Mar 22, 2010

i don't know if this is a slackware related issue but i have the following problem.I'm running a slackware64-current on my system. For my private data I'm using a QNAP NAS (Some ARM CPU with linux kernel 2.6.22), the file shares provided by NFS. I mount them withmount -t nfs 192.168.0.2:/Public /mnt/qnapWorks fine, no problems.But now, if i try to copy some large files ( > 1GiB) to the NAS share, sometimes the systems completely freezes during the copy process. I have to do a hard reset to bring the system back to work

View 5 Replies View Related

Slackware :: Is The Next 32B - 14 - Going To Have Large Memory Capability Enabled?

Jan 4, 2011

Just wondering now I see computers everywhere with 6G of ram or more if the new 32b kernels are going to be setup for large amounts of ram or will they still need to be recompiled? Also will there be better ssd support for things like "trim"? I know you edit the fstab to get the Linux version of trim but I am just wondering if stuff like that will just be automatic or if I will continue to need to tweak. I like to tweak but sometimes its the lack of need for tweaking that makes me like slack.

View 14 Replies View Related

Debian :: Jigdo Download Of DVD-7 ALWAYS Fails With Just 5 Files Left To Download

Mar 9, 2011

I haved tried 3 times to download DVD-7 from http://cdimage.debian.org/debian-cd/...md64/jigdo-dvd, and every time it has failed with just 5 files left to download.

It says:
I cannot begin to describe. All those hours of downloading for nothing! What the heck is happening here? When I try to just continue on, I get error code 3 aborts and have to just start all over.

View 3 Replies View Related

Slackware :: Large Partition Fsck On Shutdown Instead Of Boot?

Sep 15, 2010

Got tired of long waits for fsck on very large partitions.Here's a script to fsck selected partitions every 'N' shutdowns. No more boot delays for fsck (unless something is really wrong

Update1: On my system '/usr/libexec/gam_server' (gamin component used by xfce) prevented /home from being unmounted. I changed
Code:

[code]...

View 4 Replies View Related

Slackware :: Amarok Hangs At 10% When Scanning Large Music Collection?

Jun 6, 2010

After installing slackware 13.1 I start up amarok and when I go in and configure the settings and it starts to scan the folder and it either hangs at 10%, stops responding all together or crashes, the library is about 130 gigs of mp3s. I do not know where to start on this one. Amarok version 2.3.0

View 4 Replies View Related

OpenSUSE :: Burn Large Files With K3b Udf

Nov 3, 2010

Problem is with files greater than 4G onto dual layer or BD disks.mkisofs crashes. I believe the problem to be when this requires udf filesystem and disk does not get written to. As yesterday version is suse11.3 patched to date. k3b writes standard dvd disks ok. I am seeing a lot of searches saying this is because of cdrkit rather than cdrtools, can I replace this easily? Is this the case? Tried setting the k3b options when burning to udf same error. Other searches show k3b has fixed probs with these issues, so appears to point to underlying mkisofs stuff full error log in yesterdays post if it helps. I will also try on ubuntu to see if it works there

View 1 Replies View Related

Software :: Cp To Have A Progress Bar For Large Files?

Mar 10, 2010

I always wanted cp to have a progress bar for large files. I came across this:[URL]... I just wonder, how could you install it as an Arch package? Is it possible?

View 9 Replies View Related

Software :: RPM Not Supporting Large Files?

Jan 5, 2010

I currently using RPM version 4.4.2.3. It fails to build with files larger than 2 GB and file sets larger than 3 GB. I have files larger than 2 GB and sets larger than 3 GB.Is there a work around, possibly a switch or option for RPM, that will ignore this limitation?

View 8 Replies View Related

Slackware :: NFS - How To Speed Up Large File/folder Transfer - Write - Speeds?

Dec 13, 2010

About NFS.

Server:

Client(s):

Code:

I have followed Robbie Workmans' HowTo [url]

Reading and writing works absolutely fine with small files but large files are tediously slow in writing to the server. (rw,no_subtree_check) are options in exported directories.

What is your experience with NFS and how can I speed up large file/folder transfer(write) speeds?

View 5 Replies View Related

General :: Downloading Very Large Files Via SFTP

Apr 1, 2011

I need to download some very large files (circa 75 GB) from a remote server via SFTP. I've been using SFTP via the command line on my Linux netbook. Around halfway through, the transfer stops and says "stalled." Can anybody recommend a reliable way to download these files?

View 1 Replies View Related

Ubuntu :: Rsync Really Slow On Large Files?

Mar 1, 2010

I have Ubuntu on both my laptop and desktop machines, both are connected to the same network. I back up the laptop to the desktop by running the following on the laptop:

rsync -avv --stats /home/alisdt alisdt@xxx.xxx.xxx.xxx:/home/alisdt/laptop_backup (with the IP address of the desktop instead of the many x, obviously). Whenever rsync hits a large file (greater than a few MB), the network use rapidly drops to ~60KB/s (that's kilobytes not bits). When I copy the same file to the same place using scp, I get > 500KB/s throughout the transfer. Things I've tried:

* mounting the desktop home dir on the laptop using SSHFS -- a simple file copy is fast, rsync is still slow
* ditto with NFS
* rsync --whole-file option, in case the delta-transfer algorithm was choking on large files
* rsync --inplace option
* HPN-SSH (http://www.psc.edu/networking/projects/hpn-ssh/) to enable dynamic window and unencrypted bulk transfer, just in case it was some ssh bottleneck I think it's either an rsync application problem, or a network problem that is only affecting rsync. Any ideas, or other ideas of what I can try to debug? In case it's relevant, I'm using 9.04 on both machines. (A standing bug prevents me from upgrading the laptop, and I haven't bothered to upgrade the desktop).

View 3 Replies View Related

Ubuntu :: Moving Large Amounts Of Files

Mar 6, 2010

I am trying to move a large amount of files (over 30k and 86GB) to another HDD but I get a Augment list too large error?? I tried rsync, cp, mv and still the same error

View 1 Replies View Related

Ubuntu :: Cups: Large Files Do Not Print?

Nov 29, 2010

i have a problem with cups on a lucid/64 machine.

"Unable to write print data: Broken pipe"

The pdf-file to print has a size of 4,7MB. After sending the file to the printer the size of the file is more than 18 MB.

We use a Xerox WorkCentre 7232 which is via

socket://ip_adress:9100

connected to cups. The same configuration had been working fine for several years with hardy.

Cups refuses to print large files. When splitting up the print-file all works fine.

View 4 Replies View Related

General :: Creating Random Large Files?

Aug 27, 2010

how I can randomly write / create a 1 GB file in bash to test disk / network i/o? I was told I could use the 'dd' command but I don't know if there are some better ways and or what the 'dd' command looks like.

View 7 Replies View Related

General :: Tar Not Working With Large Number Of Files?

Dec 6, 2010

in my middle of script i am using tar command to tar some 1000 images and size of each image is 5MB.

all the images are provided as argument as tar -cvf file.tar <all images as argument>

but my tar file file.tar does not contain all the images.

View 4 Replies View Related

Slackware :: Compiling Large Slackbuilds - Error-message - Not Enough Space Left On Device

Aug 20, 2010

I am currently running slackware 12.2 on a 25 GB partition. I like to use slackbuilds, but when I try to compile larger tarballs (like abiword, or a patched version of Ghostscript as I did today) I receive an error-message: 'Not enough space left on device'. I think the size of the partition must be big enough (I never got this message when compiling with Linux From Scratch). I think it has something to do with the size of my /tmp directory, but I don't know how to fix this. Is there a way to solve the problem, so that I could be able to use slackbuild-scripts?

View 4 Replies View Related

Debian Configuration :: Transfer Large Files Between Windows ?

Jul 17, 2011

I would like to transfer my music library and movie collection from my Desktop computer running Windows Vista and my laptop running Debian Squeeze. I have the laptop connected via wireless but it's possible to connect the two either directly with a CAT5e cable or through the router. I'm just wondering what the best way to do this would be.

View 8 Replies View Related

Fedora Networking :: 12 Hangs On Transfer Of Large Files?

Dec 9, 2009

I have Fedora 12 (with all the latest patches, including the 2.6.31.6-162 kernel) installed on a new Supermicro SYS-5015A-H 1U Server [Intel Atom 330 (1.6GHz) CPU, Intel 945GC NB, Intel ICH7R SB, 2x Realtek RTL8111C-GR Gigabit Ethernet, Onboard GMA950 video]. This all works great until I try to transfer a large file over the network, then the computer hard locks, forcing a power-off reset.

Some info about my setup:

[root@Epsilon ~]# uname -a
Linux Epsilon 2.6.31.6-162.fc12.i686.PAE #1 SMP Fri Dec 4 00:43:59 EST 2009 i686 i686 i386 GNU/Linux
[root@Epsilon ~]# dmesg | grep r8169
r8169 Gigabit Ethernet driver 2.3LK-NAPI loaded

[code]....

I'm pretty sure this is an issue with the r8169 driver (what I'm seeing is somewhat reminiscent of the bug reported here). The computer will operate fine for days as a (low volume) web server, and is reasonably stable transferring small files, but as when as I try to transfer a large file (say during a backup to a NAS or a NFS share), the computer will hard lock (no keyboard, mouse, etc.) at some point into the transfer of the file. It doesn't seem to matter how the file is transferred (sftp, rsync to NFS share, etc.).

View 10 Replies View Related

OpenSUSE :: Burn Large Files With K3b Mkisofs Crashed

Nov 2, 2010

I have a problem with burning some disks. I believe the problem to be when the file sizes are large. using suse 11.3 x 64 and updates done. k3b version 2.0.1-1pm 2.7 burned normal dvd yesterday with3.5 Gb + 1.5Gb files. Tried burning to BD-R disk with 11.5 Gb file. I have burned blu ray disks before with my machine but used nero, I would prefer to use k3b as should do this. I get error of mkisofs crashed and following error log

[code]...

View 3 Replies View Related

General :: Copying Large Number Of Files From One Directory To Another

Feb 10, 2010

I've a directory containing around 2.8 lacs of files. I want to move them to another directory.If I use cp or mv then I get an error 'argument list too long'. If I write a script like

for file in ls *; do
cp {source} to {destination}
done

then because of ls command , its performance degrades.How can I do this?

View 7 Replies View Related

General :: Transferring Large Files Using Scp With CPU And Memory Considerations?

Oct 5, 2010

I want to transfer an arbitrarily large file (say >20GB) between 2 servers. I have several considerations:

Must use port 22 (ssh) because of firewall restrictions
Cannot tax the CPU (production server)
Memory efficiency
Would prefer a checksum check but that could be done manually
Time is not of the essence

Server A and Server B are on the same private network (sharing a switch) and data security is not a concern, Server A and Server B are not on the same network and transfer will be via the public internet so data security is a concern, My first thought was using nice on an scp command with a non-CPU-intensive cypher (blowfish?). But I thought I'll refer to the SU community for recommendations.

View 1 Replies View Related

General :: Copying Large Number Of Files In Windows?

Mar 15, 2011

I am facing problem in copying a large number of file 18 lakh (18,000,000) files from my personal hardisk to another hardisk each file is very small and size of folder is around 3.95 GB copying files using copy given by Windows is frustrating and I am not even able to compress file its giving me error that its not readable.And problem is I am not able to open this drive in Linux it showing me error there saying do diskchk in Windows and Windows disk check is also not able to repair this drive and goes into some mode unsolvable.Is there any way to open disk with error to open in Windows and if not any way I can copy data faster?ERROR: Disk labled EDU is corrupt go to windows and chkdsk /f there and reboot into window 2 times.

View 3 Replies View Related

Ubuntu :: File System Format For Mac OSX 10.5 For Large Files?

Sep 19, 2010

Is there a file system that both Mac OSX 10.5 and linux can read/write for large files (like 4gb files)? My desktop is Ubuntu and I run most from there, but I want to back up my MacBook and linux box on the same external hard drive. Seems there are some (paid) apps for Mac that will mount NTFS but I'm wondering if there is just a shared files ystem that will work for both.

View 9 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved