* Speed up testing * make notification run in the background, this recudes the test_readme time from 18s to 0.10s * reduce time for zone reload * TestServeDNSConcurrent remove entirely. This took a whopping 58s for ... ? A few minutes staring didn't reveal wth it is actually testing. Making values smaller revealed race conditions in the tests. Remove entirely. * Move many interval values to variables so we can reset them to short values for the tests. * test_large_axfr: make the zone smaller. The number used 64K has no rational, make it 64/10 to speed up. * TestProxyThreeWay: use client with shorter timeout A few random tidbits in other tests. Total time saved: 177s (almost 3m) - which makes it worthwhile again to run the test locally: this branch: ~~~ ok github.com/coredns/coredns/test 10.437s cd plugin; time go t ./... 5,51s user 7,51s system 11,15s elapsed 744%CPU ( ~~~ master: ~~~ ok github.com/coredns/coredns/test 35.252s cd plugin; time go t ./... 157,64s user 15,39s system 50,05s elapsed 345%CPU () ~~~ tests/ -25s plugins/ -40s This brings the total on 20s, and another 10s can be saved by fixing dnstapio. Moving this to 5s would be even better, but 10s is also nice. Signed-off-by: Miek Gieben <miek@miek.nl> * Also 0.01 Signed-off-by: Miek Gieben <miek@miek.nl>
91 lines
2.1 KiB
Go
91 lines
2.1 KiB
Go
package forward
|
|
|
|
import (
|
|
"crypto/tls"
|
|
"sync/atomic"
|
|
"time"
|
|
|
|
"github.com/coredns/coredns/plugin/pkg/transport"
|
|
|
|
"github.com/miekg/dns"
|
|
)
|
|
|
|
// HealthChecker checks the upstream health.
|
|
type HealthChecker interface {
|
|
Check(*Proxy) error
|
|
SetTLSConfig(*tls.Config)
|
|
SetRecursionDesired(bool)
|
|
GetRecursionDesired() bool
|
|
}
|
|
|
|
// dnsHc is a health checker for a DNS endpoint (DNS, and DoT).
|
|
type dnsHc struct {
|
|
c *dns.Client
|
|
recursionDesired bool
|
|
}
|
|
|
|
var (
|
|
hcReadTimeout = 1 * time.Second
|
|
hcWriteTimeout = 1 * time.Second
|
|
)
|
|
|
|
// NewHealthChecker returns a new HealthChecker based on transport.
|
|
func NewHealthChecker(trans string, recursionDesired bool) HealthChecker {
|
|
switch trans {
|
|
case transport.DNS, transport.TLS:
|
|
c := new(dns.Client)
|
|
c.Net = "udp"
|
|
c.ReadTimeout = hcReadTimeout
|
|
c.WriteTimeout = hcWriteTimeout
|
|
|
|
return &dnsHc{c: c, recursionDesired: recursionDesired}
|
|
}
|
|
|
|
log.Warningf("No healthchecker for transport %q", trans)
|
|
return nil
|
|
}
|
|
|
|
func (h *dnsHc) SetTLSConfig(cfg *tls.Config) {
|
|
h.c.Net = "tcp-tls"
|
|
h.c.TLSConfig = cfg
|
|
}
|
|
|
|
func (h *dnsHc) SetRecursionDesired(recursionDesired bool) {
|
|
h.recursionDesired = recursionDesired
|
|
}
|
|
func (h *dnsHc) GetRecursionDesired() bool {
|
|
return h.recursionDesired
|
|
}
|
|
|
|
// For HC we send to . IN NS +[no]rec message to the upstream. Dial timeouts and empty
|
|
// replies are considered fails, basically anything else constitutes a healthy upstream.
|
|
|
|
// Check is used as the up.Func in the up.Probe.
|
|
func (h *dnsHc) Check(p *Proxy) error {
|
|
err := h.send(p.addr)
|
|
if err != nil {
|
|
HealthcheckFailureCount.WithLabelValues(p.addr).Add(1)
|
|
atomic.AddUint32(&p.fails, 1)
|
|
return err
|
|
}
|
|
|
|
atomic.StoreUint32(&p.fails, 0)
|
|
return nil
|
|
}
|
|
|
|
func (h *dnsHc) send(addr string) error {
|
|
ping := new(dns.Msg)
|
|
ping.SetQuestion(".", dns.TypeNS)
|
|
ping.MsgHdr.RecursionDesired = h.recursionDesired
|
|
|
|
m, _, err := h.c.Exchange(ping, addr)
|
|
// If we got a header, we're alright, basically only care about I/O errors 'n stuff.
|
|
if err != nil && m != nil {
|
|
// Silly check, something sane came back.
|
|
if m.Response || m.Opcode == dns.OpcodeQuery {
|
|
err = nil
|
|
}
|
|
}
|
|
|
|
return err
|
|
}
|