mirror of
https://github.com/caddyserver/caddy.git
synced 2024-12-16 21:56:40 -05:00
4b1a9b6cc1
* caddy adapt for listen_protocols * adapt listen_socket * allow multiple listen sockets for port ranges and readd socket fd listen logic * readd logic to start servers according to listener protocols * gofmt * adapt caddytest * gosec * fmt and rename listen to listenWithSocket * fmt and rename listen to listenWithSocket * more consistent error msg * non unix listenReusableWithSocketFile * remove unused func * doc comment typo * nonosec * commit * doc comments * more doc comments * comment was misleading, cardinality did not change * addressesWithProtocols * update test * fd/ and fdgram/ * rm addr * actually write... * i guess we doin' "skip": now * wrong var in placeholder * wrong var in placeholder II * update param name in comment * dont save nil file pointers * windows * key -> parsedKey * osx * multiple default_bind with protocols * check for h1 and h2 listener netw
318 lines
9.5 KiB
Go
318 lines
9.5 KiB
Go
// Copyright 2015 Matthew Holt and The Caddy Authors
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
//go:build !unix || solaris
|
|
|
|
package caddy
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"net"
|
|
"os"
|
|
"slices"
|
|
"strconv"
|
|
"sync"
|
|
"sync/atomic"
|
|
"time"
|
|
|
|
"go.uber.org/zap"
|
|
)
|
|
|
|
func reuseUnixSocket(network, addr string) (any, error) {
|
|
return nil, nil
|
|
}
|
|
|
|
func listenReusable(ctx context.Context, lnKey string, network, address string, config net.ListenConfig) (any, error) {
|
|
var socketFile *os.File
|
|
|
|
fd := slices.Contains([]string{"fd", "fdgram"}, network)
|
|
if fd {
|
|
socketFd, err := strconv.ParseUint(address, 0, strconv.IntSize)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("invalid file descriptor: %v", err)
|
|
}
|
|
|
|
func() {
|
|
socketFilesMu.Lock()
|
|
defer socketFilesMu.Unlock()
|
|
|
|
socketFdWide := uintptr(socketFd)
|
|
var ok bool
|
|
|
|
socketFile, ok = socketFiles[socketFdWide]
|
|
|
|
if !ok {
|
|
socketFile = os.NewFile(socketFdWide, lnKey)
|
|
if socketFile != nil {
|
|
socketFiles[socketFdWide] = socketFile
|
|
}
|
|
}
|
|
}()
|
|
|
|
if socketFile == nil {
|
|
return nil, fmt.Errorf("invalid socket file descriptor: %d", socketFd)
|
|
}
|
|
}
|
|
|
|
datagram := slices.Contains([]string{"udp", "udp4", "udp6", "unixgram", "fdgram"}, network)
|
|
if datagram {
|
|
sharedPc, _, err := listenerPool.LoadOrNew(lnKey, func() (Destructor, error) {
|
|
var (
|
|
pc net.PacketConn
|
|
err error
|
|
)
|
|
if fd {
|
|
pc, err = net.FilePacketConn(socketFile)
|
|
} else {
|
|
pc, err = config.ListenPacket(ctx, network, address)
|
|
}
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &sharedPacketConn{PacketConn: pc, key: lnKey}, nil
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &fakeClosePacketConn{sharedPacketConn: sharedPc.(*sharedPacketConn)}, nil
|
|
}
|
|
|
|
sharedLn, _, err := listenerPool.LoadOrNew(lnKey, func() (Destructor, error) {
|
|
var (
|
|
ln net.Listener
|
|
err error
|
|
)
|
|
if fd {
|
|
ln, err = net.FileListener(socketFile)
|
|
} else {
|
|
ln, err = config.Listen(ctx, network, address)
|
|
}
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &sharedListener{Listener: ln, key: lnKey}, nil
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &fakeCloseListener{sharedListener: sharedLn.(*sharedListener), keepAlivePeriod: config.KeepAlive}, nil
|
|
}
|
|
|
|
// fakeCloseListener is a private wrapper over a listener that
|
|
// is shared. The state of fakeCloseListener is not shared.
|
|
// This allows one user of a socket to "close" the listener
|
|
// while in reality the socket stays open for other users of
|
|
// the listener. In this way, servers become hot-swappable
|
|
// while the listener remains running. Listeners should be
|
|
// re-wrapped in a new fakeCloseListener each time the listener
|
|
// is reused. This type is atomic and values must not be copied.
|
|
type fakeCloseListener struct {
|
|
closed int32 // accessed atomically; belongs to this struct only
|
|
*sharedListener // embedded, so we also become a net.Listener
|
|
keepAlivePeriod time.Duration
|
|
}
|
|
|
|
type canSetKeepAlive interface {
|
|
SetKeepAlivePeriod(d time.Duration) error
|
|
SetKeepAlive(bool) error
|
|
}
|
|
|
|
func (fcl *fakeCloseListener) Accept() (net.Conn, error) {
|
|
// if the listener is already "closed", return error
|
|
if atomic.LoadInt32(&fcl.closed) == 1 {
|
|
return nil, fakeClosedErr(fcl)
|
|
}
|
|
|
|
// call underlying accept
|
|
conn, err := fcl.sharedListener.Accept()
|
|
if err == nil {
|
|
// if 0, do nothing, Go's default is already set
|
|
// and if the connection allows setting KeepAlive, set it
|
|
if tconn, ok := conn.(canSetKeepAlive); ok && fcl.keepAlivePeriod != 0 {
|
|
if fcl.keepAlivePeriod > 0 {
|
|
err = tconn.SetKeepAlivePeriod(fcl.keepAlivePeriod)
|
|
} else { // negative
|
|
err = tconn.SetKeepAlive(false)
|
|
}
|
|
if err != nil {
|
|
Log().With(zap.String("server", fcl.sharedListener.key)).Warn("unable to set keepalive for new connection:", zap.Error(err))
|
|
}
|
|
}
|
|
return conn, nil
|
|
}
|
|
|
|
// since Accept() returned an error, it may be because our reference to
|
|
// the listener (this fakeCloseListener) may have been closed, i.e. the
|
|
// server is shutting down; in that case, we need to clear the deadline
|
|
// that we set when Close() was called, and return a non-temporary and
|
|
// non-timeout error value to the caller, masking the "true" error, so
|
|
// that server loops / goroutines won't retry, linger, and leak
|
|
if atomic.LoadInt32(&fcl.closed) == 1 {
|
|
// we dereference the sharedListener explicitly even though it's embedded
|
|
// so that it's clear in the code that side-effects are shared with other
|
|
// users of this listener, not just our own reference to it; we also don't
|
|
// do anything with the error because all we could do is log it, but we
|
|
// explicitly assign it to nothing so we don't forget it's there if needed
|
|
_ = fcl.sharedListener.clearDeadline()
|
|
|
|
if netErr, ok := err.(net.Error); ok && netErr.Timeout() {
|
|
return nil, fakeClosedErr(fcl)
|
|
}
|
|
}
|
|
|
|
return nil, err
|
|
}
|
|
|
|
// Close stops accepting new connections without closing the
|
|
// underlying listener. The underlying listener is only closed
|
|
// if the caller is the last known user of the socket.
|
|
func (fcl *fakeCloseListener) Close() error {
|
|
if atomic.CompareAndSwapInt32(&fcl.closed, 0, 1) {
|
|
// There are two ways I know of to get an Accept()
|
|
// function to return to the server loop that called
|
|
// it: close the listener, or set a deadline in the
|
|
// past. Obviously, we can't close the socket yet
|
|
// since others may be using it (hence this whole
|
|
// file). But we can set the deadline in the past,
|
|
// and this is kind of cheating, but it works, and
|
|
// it apparently even works on Windows.
|
|
_ = fcl.sharedListener.setDeadline()
|
|
_, _ = listenerPool.Delete(fcl.sharedListener.key)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// sharedListener is a wrapper over an underlying listener. The listener
|
|
// and the other fields on the struct are shared state that is synchronized,
|
|
// so sharedListener structs must never be copied (always use a pointer).
|
|
type sharedListener struct {
|
|
net.Listener
|
|
key string // uniquely identifies this listener
|
|
deadline bool // whether a deadline is currently set
|
|
deadlineMu sync.Mutex
|
|
}
|
|
|
|
func (sl *sharedListener) clearDeadline() error {
|
|
var err error
|
|
sl.deadlineMu.Lock()
|
|
if sl.deadline {
|
|
switch ln := sl.Listener.(type) {
|
|
case *net.TCPListener:
|
|
err = ln.SetDeadline(time.Time{})
|
|
}
|
|
sl.deadline = false
|
|
}
|
|
sl.deadlineMu.Unlock()
|
|
return err
|
|
}
|
|
|
|
func (sl *sharedListener) setDeadline() error {
|
|
timeInPast := time.Now().Add(-1 * time.Minute)
|
|
var err error
|
|
sl.deadlineMu.Lock()
|
|
if !sl.deadline {
|
|
switch ln := sl.Listener.(type) {
|
|
case *net.TCPListener:
|
|
err = ln.SetDeadline(timeInPast)
|
|
}
|
|
sl.deadline = true
|
|
}
|
|
sl.deadlineMu.Unlock()
|
|
return err
|
|
}
|
|
|
|
// Destruct is called by the UsagePool when the listener is
|
|
// finally not being used anymore. It closes the socket.
|
|
func (sl *sharedListener) Destruct() error {
|
|
return sl.Listener.Close()
|
|
}
|
|
|
|
// fakeClosePacketConn is like fakeCloseListener, but for PacketConns,
|
|
// or more specifically, *net.UDPConn
|
|
type fakeClosePacketConn struct {
|
|
closed int32 // accessed atomically; belongs to this struct only
|
|
*sharedPacketConn // embedded, so we also become a net.PacketConn; its key is used in Close
|
|
}
|
|
|
|
func (fcpc *fakeClosePacketConn) ReadFrom(p []byte) (n int, addr net.Addr, err error) {
|
|
// if the listener is already "closed", return error
|
|
if atomic.LoadInt32(&fcpc.closed) == 1 {
|
|
return 0, nil, &net.OpError{
|
|
Op: "readfrom",
|
|
Net: fcpc.LocalAddr().Network(),
|
|
Addr: fcpc.LocalAddr(),
|
|
Err: errFakeClosed,
|
|
}
|
|
}
|
|
|
|
// call underlying readfrom
|
|
n, addr, err = fcpc.sharedPacketConn.ReadFrom(p)
|
|
if err != nil {
|
|
// this server was stopped, so clear the deadline and let
|
|
// any new server continue reading; but we will exit
|
|
if atomic.LoadInt32(&fcpc.closed) == 1 {
|
|
if netErr, ok := err.(net.Error); ok && netErr.Timeout() {
|
|
if err = fcpc.SetReadDeadline(time.Time{}); err != nil {
|
|
return
|
|
}
|
|
}
|
|
}
|
|
return
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
// Close won't close the underlying socket unless there is no more reference, then listenerPool will close it.
|
|
func (fcpc *fakeClosePacketConn) Close() error {
|
|
if atomic.CompareAndSwapInt32(&fcpc.closed, 0, 1) {
|
|
_ = fcpc.SetReadDeadline(time.Now()) // unblock ReadFrom() calls to kick old servers out of their loops
|
|
_, _ = listenerPool.Delete(fcpc.sharedPacketConn.key)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (fcpc *fakeClosePacketConn) Unwrap() net.PacketConn {
|
|
return fcpc.sharedPacketConn.PacketConn
|
|
}
|
|
|
|
// sharedPacketConn is like sharedListener, but for net.PacketConns.
|
|
type sharedPacketConn struct {
|
|
net.PacketConn
|
|
key string
|
|
}
|
|
|
|
// Destruct closes the underlying socket.
|
|
func (spc *sharedPacketConn) Destruct() error {
|
|
return spc.PacketConn.Close()
|
|
}
|
|
|
|
// Unwrap returns the underlying socket
|
|
func (spc *sharedPacketConn) Unwrap() net.PacketConn {
|
|
return spc.PacketConn
|
|
}
|
|
|
|
// Interface guards (see https://github.com/caddyserver/caddy/issues/3998)
|
|
var (
|
|
_ (interface {
|
|
Unwrap() net.PacketConn
|
|
}) = (*fakeClosePacketConn)(nil)
|
|
)
|
|
|
|
// socketFiles is a fd -> *os.File map used to make a FileListener/FilePacketConn from a socket file descriptor.
|
|
var socketFiles = map[uintptr]*os.File{}
|
|
|
|
// socketFilesMu synchronizes socketFiles insertions
|
|
var socketFilesMu sync.Mutex
|