mirror of
https://github.com/caddyserver/caddy.git
synced 2024-12-16 21:56:40 -05:00
parent
b3ce260389
commit
197c564f20
2 changed files with 18 additions and 5 deletions
|
@ -401,6 +401,9 @@ func (app *App) Provision(ctx caddy.Context) error {
|
|||
if srv.IdleTimeout == 0 {
|
||||
srv.IdleTimeout = defaultIdleTimeout
|
||||
}
|
||||
if srv.ReadHeaderTimeout == 0 {
|
||||
srv.ReadHeaderTimeout = defaultReadHeaderTimeout // see #6663
|
||||
}
|
||||
}
|
||||
ctx.Context = oldContext
|
||||
return nil
|
||||
|
@ -770,11 +773,20 @@ func (app *App) httpsPort() int {
|
|||
return app.HTTPSPort
|
||||
}
|
||||
|
||||
// defaultIdleTimeout is the default HTTP server timeout
|
||||
// for closing idle connections; useful to avoid resource
|
||||
// exhaustion behind hungry CDNs, for example (we've had
|
||||
// several complaints without this).
|
||||
const defaultIdleTimeout = caddy.Duration(5 * time.Minute)
|
||||
const (
|
||||
// defaultIdleTimeout is the default HTTP server timeout
|
||||
// for closing idle connections; useful to avoid resource
|
||||
// exhaustion behind hungry CDNs, for example (we've had
|
||||
// several complaints without this).
|
||||
defaultIdleTimeout = caddy.Duration(5 * time.Minute)
|
||||
|
||||
// defaultReadHeaderTimeout is the default timeout for
|
||||
// reading HTTP headers from clients. Headers are generally
|
||||
// small, often less than 1 KB, so it shouldn't take a
|
||||
// long time even on legitimately slow connections or
|
||||
// busy servers to read it.
|
||||
defaultReadHeaderTimeout = caddy.Duration(time.Minute)
|
||||
)
|
||||
|
||||
// Interface guards
|
||||
var (
|
||||
|
|
|
@ -61,6 +61,7 @@ type Server struct {
|
|||
ReadTimeout caddy.Duration `json:"read_timeout,omitempty"`
|
||||
|
||||
// ReadHeaderTimeout is like ReadTimeout but for request headers.
|
||||
// Default is 1 minute.
|
||||
ReadHeaderTimeout caddy.Duration `json:"read_header_timeout,omitempty"`
|
||||
|
||||
// WriteTimeout is how long to allow a write to a client. Note
|
||||
|
|
Loading…
Reference in a new issue