These forums are locked and archived, but all topics have been migrated to the new forum. You can search for this topic on the new forum: Search for Large file stops FTP working on the new forum.
I uploaded a large file over 1GB and it has stopped FTP uploading any new files across the entire server. I can change and upload existing files. There is plenty of space left for that particular owners virtual server. I did ulimit -a and the fsize is set to unlimited. This is running on CentOS 5. Any ideas as to why this is happening and how to rectify this?
core file size (blocks, -c) 0
data seg size (kbytes, -d) unlimited
scheduling priority (-e) 0
file size (blocks, -f) unlimited
pending signals (-i) 128362
max locked memory (kbytes, -l) 32
max memory size (kbytes, -m) unlimited
open files (-n) 1024
pipe size (512 bytes, -p) 8
POSIX message queues (bytes, -q) 819200
real-time priority (-r) 0
stack size (kbytes, -s) 10240
cpu time (seconds, -t) unlimited
max user processes (-u) 128362
virtual memory (kbytes, -v) unlimited
file locks (-x) unlimited
Howdy,
Hmm, I hadn't heard of that happening before. Do you see any errors or related messages in either /var/log/messages, or in any of the logs in /var/log/proftpd?
-Eric
sftp-server[17993]: error: process_write: write failed
It writes an empty file. I found that if I change the group owner of the large file then ftp works. The ftp user was the same owner.