Maxim D. wrote:
On Fri, Apr 10, 2009 at 02:04:09PM +0200, Robert G. wrote:
I have nginx 0.6.36 with php-fastcgi. Im using SquirrelMail and have a
mail that is 25M, no attachement, just a text mail so big. Im trying to
read it, but fastcgi ends-up in
2009/04/10 13:55:35 [error] 22626#0: *537 recv() failed (104: Connection
reset by peer) while reading response header from upstream, client:
From the error message it seems that php died even before it was
able to send header to nginx.
Before the problem was that php didnt have enough memory or
max_execution_time was too low. I modified that and set
keepalive_timeout to 32, but it just die even like this.
keepalive_timeout is completely unrelated thing, it’s only used by
nginx for client connections.
Is it possible
fastcgi is limited to how big is the request or something.
No, at least not the protocol itself.
How could I setup up nginx and/or php to be able to read that mail?
Try looking into php. It seems that it just dies due to errors or
you haven’t tuned limits enough.
Like I said before, in the logs at the begining I got this:
PHP Fatal error: Allowed memory size of 67108864 bytes exhausted
I modified this to 128M and didnt complain anymore, but it did complain
about the max_execution_time and it looked like this:
PHP Fatal error: Maximum execution time of 60 seconds exceeded
I modified this to 240 seconds and after this didnt get any errors
anymore from php, but got the erros in the error_log from nginx and it
just didnt do anything… also the php-cgi process just hanged in the
backgroud with 100% CPU usage. I had to stop fastcgi and kill then the
process which was using 100% CPU.
So I really dont know what else to tune in php.ini