1288 lines
31 KiB
Go
1288 lines
31 KiB
Go
package torrent
|
|
|
|
import (
|
|
"bufio"
|
|
"bytes"
|
|
"context"
|
|
"crypto/rand"
|
|
"encoding/binary"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"net"
|
|
"strconv"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/anacrolix/dht"
|
|
"github.com/anacrolix/dht/krpc"
|
|
"github.com/anacrolix/log"
|
|
"github.com/anacrolix/missinggo"
|
|
"github.com/anacrolix/missinggo/perf"
|
|
"github.com/anacrolix/missinggo/pproffd"
|
|
"github.com/anacrolix/missinggo/pubsub"
|
|
"github.com/anacrolix/missinggo/slices"
|
|
"github.com/anacrolix/sync"
|
|
"github.com/davecgh/go-spew/spew"
|
|
"github.com/dustin/go-humanize"
|
|
"github.com/google/btree"
|
|
|
|
"github.com/anacrolix/torrent/bencode"
|
|
"github.com/anacrolix/torrent/iplist"
|
|
"github.com/anacrolix/torrent/metainfo"
|
|
"github.com/anacrolix/torrent/mse"
|
|
pp "github.com/anacrolix/torrent/peer_protocol"
|
|
"github.com/anacrolix/torrent/storage"
|
|
)
|
|
|
|
// Clients contain zero or more Torrents. A Client manages a blocklist, the
|
|
// TCP/UDP protocol ports, and DHT as desired.
|
|
type Client struct {
|
|
// An aggregate of stats over all connections. First in struct to ensure
|
|
// 64-bit alignment of fields. See #262.
|
|
stats ConnStats
|
|
mu sync.RWMutex
|
|
event sync.Cond
|
|
closed missinggo.Event
|
|
|
|
config *ClientConfig
|
|
logger *log.Logger
|
|
|
|
halfOpenLimit int
|
|
peerID PeerID
|
|
defaultStorage *storage.Client
|
|
onClose []func()
|
|
conns []socket
|
|
dhtServers []*dht.Server
|
|
ipBlockList iplist.Ranger
|
|
// Our BitTorrent protocol extension bytes, sent in our BT handshakes.
|
|
extensionBytes pp.PeerExtensionBits
|
|
|
|
// Set of addresses that have our client ID. This intentionally will
|
|
// include ourselves if we end up trying to connect to our own address
|
|
// through legitimate channels.
|
|
dopplegangerAddrs map[string]struct{}
|
|
badPeerIPs map[string]struct{}
|
|
torrents map[InfoHash]*Torrent
|
|
|
|
acceptLimiter map[ipStr]int
|
|
}
|
|
|
|
type ipStr string
|
|
|
|
func (cl *Client) BadPeerIPs() []string {
|
|
cl.mu.RLock()
|
|
defer cl.mu.RUnlock()
|
|
return cl.badPeerIPsLocked()
|
|
}
|
|
|
|
func (cl *Client) badPeerIPsLocked() []string {
|
|
return slices.FromMapKeys(cl.badPeerIPs).([]string)
|
|
}
|
|
|
|
func (cl *Client) PeerID() PeerID {
|
|
return cl.peerID
|
|
}
|
|
|
|
type torrentAddr string
|
|
|
|
func (torrentAddr) Network() string { return "" }
|
|
|
|
func (me torrentAddr) String() string { return string(me) }
|
|
|
|
func (cl *Client) LocalPort() (port int) {
|
|
cl.eachListener(func(l socket) bool {
|
|
_port := missinggo.AddrPort(l.Addr())
|
|
if _port == 0 {
|
|
panic(l)
|
|
}
|
|
if port == 0 {
|
|
port = _port
|
|
} else if port != _port {
|
|
panic("mismatched ports")
|
|
}
|
|
return true
|
|
})
|
|
return
|
|
}
|
|
|
|
func writeDhtServerStatus(w io.Writer, s *dht.Server) {
|
|
dhtStats := s.Stats()
|
|
fmt.Fprintf(w, "\t# Nodes: %d (%d good, %d banned)\n", dhtStats.Nodes, dhtStats.GoodNodes, dhtStats.BadNodes)
|
|
fmt.Fprintf(w, "\tServer ID: %x\n", s.ID())
|
|
fmt.Fprintf(w, "\tAnnounces: %d\n", dhtStats.SuccessfulOutboundAnnouncePeerQueries)
|
|
fmt.Fprintf(w, "\tOutstanding transactions: %d\n", dhtStats.OutstandingTransactions)
|
|
}
|
|
|
|
// Writes out a human readable status of the client, such as for writing to a
|
|
// HTTP status page.
|
|
func (cl *Client) WriteStatus(_w io.Writer) {
|
|
cl.mu.RLock()
|
|
defer cl.mu.RUnlock()
|
|
w := bufio.NewWriter(_w)
|
|
defer w.Flush()
|
|
fmt.Fprintf(w, "Listen port: %d\n", cl.LocalPort())
|
|
fmt.Fprintf(w, "Peer ID: %+q\n", cl.PeerID())
|
|
fmt.Fprintf(w, "Announce key: %x\n", cl.announceKey())
|
|
fmt.Fprintf(w, "Banned IPs: %d\n", len(cl.badPeerIPsLocked()))
|
|
cl.eachDhtServer(func(s *dht.Server) {
|
|
fmt.Fprintf(w, "%s DHT server at %s:\n", s.Addr().Network(), s.Addr().String())
|
|
writeDhtServerStatus(w, s)
|
|
})
|
|
spew.Fdump(w, cl.stats)
|
|
fmt.Fprintf(w, "# Torrents: %d\n", len(cl.torrentsAsSlice()))
|
|
fmt.Fprintln(w)
|
|
for _, t := range slices.Sort(cl.torrentsAsSlice(), func(l, r *Torrent) bool {
|
|
return l.InfoHash().AsString() < r.InfoHash().AsString()
|
|
}).([]*Torrent) {
|
|
if t.name() == "" {
|
|
fmt.Fprint(w, "<unknown name>")
|
|
} else {
|
|
fmt.Fprint(w, t.name())
|
|
}
|
|
fmt.Fprint(w, "\n")
|
|
if t.info != nil {
|
|
fmt.Fprintf(w, "%f%% of %d bytes (%s)", 100*(1-float64(t.bytesMissingLocked())/float64(t.info.TotalLength())), t.length, humanize.Bytes(uint64(t.info.TotalLength())))
|
|
} else {
|
|
w.WriteString("<missing metainfo>")
|
|
}
|
|
fmt.Fprint(w, "\n")
|
|
t.writeStatus(w)
|
|
fmt.Fprintln(w)
|
|
}
|
|
}
|
|
|
|
const debugLogValue = "debug"
|
|
|
|
func (cl *Client) debugLogFilter(m *log.Msg) bool {
|
|
if !cl.config.Debug {
|
|
_, ok := m.Values()[debugLogValue]
|
|
return !ok
|
|
}
|
|
return true
|
|
}
|
|
|
|
func (cl *Client) initLogger() {
|
|
cl.logger = log.Default.Clone().AddValue(cl).AddFilter(log.NewFilter(cl.debugLogFilter))
|
|
}
|
|
|
|
func (cl *Client) announceKey() int32 {
|
|
return int32(binary.BigEndian.Uint32(cl.peerID[16:20]))
|
|
}
|
|
|
|
func NewClient(cfg *ClientConfig) (cl *Client, err error) {
|
|
if cfg == nil {
|
|
cfg = NewDefaultClientConfig()
|
|
}
|
|
defer func() {
|
|
if err != nil {
|
|
cl = nil
|
|
}
|
|
}()
|
|
cl = &Client{
|
|
halfOpenLimit: cfg.HalfOpenConnsPerTorrent,
|
|
config: cfg,
|
|
dopplegangerAddrs: make(map[string]struct{}),
|
|
torrents: make(map[metainfo.Hash]*Torrent),
|
|
}
|
|
go cl.acceptLimitClearer()
|
|
cl.initLogger()
|
|
defer func() {
|
|
if err == nil {
|
|
return
|
|
}
|
|
cl.Close()
|
|
}()
|
|
cl.extensionBytes = defaultPeerExtensionBytes()
|
|
cl.event.L = &cl.mu
|
|
storageImpl := cfg.DefaultStorage
|
|
if storageImpl == nil {
|
|
// We'd use mmap but HFS+ doesn't support sparse files.
|
|
storageImpl = storage.NewFile(cfg.DataDir)
|
|
cl.onClose = append(cl.onClose, func() {
|
|
if err := storageImpl.Close(); err != nil {
|
|
log.Printf("error closing default storage: %s", err)
|
|
}
|
|
})
|
|
}
|
|
cl.defaultStorage = storage.NewClient(storageImpl)
|
|
if cfg.IPBlocklist != nil {
|
|
cl.ipBlockList = cfg.IPBlocklist
|
|
}
|
|
|
|
if cfg.PeerID != "" {
|
|
missinggo.CopyExact(&cl.peerID, cfg.PeerID)
|
|
} else {
|
|
o := copy(cl.peerID[:], cfg.Bep20)
|
|
_, err = rand.Read(cl.peerID[o:])
|
|
if err != nil {
|
|
panic("error generating peer id")
|
|
}
|
|
}
|
|
|
|
cl.conns, err = listenAll(cl.enabledPeerNetworks(), cl.config.ListenHost, cl.config.ListenPort, cl.config.ProxyURL)
|
|
if err != nil {
|
|
return
|
|
}
|
|
// Check for panics.
|
|
cl.LocalPort()
|
|
|
|
for _, s := range cl.conns {
|
|
if peerNetworkEnabled(s.Addr().Network(), cl.config) {
|
|
go cl.acceptConnections(s)
|
|
}
|
|
}
|
|
|
|
go cl.forwardPort()
|
|
if !cfg.NoDHT {
|
|
for _, s := range cl.conns {
|
|
if pc, ok := s.(net.PacketConn); ok {
|
|
ds, err := cl.newDhtServer(pc)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
cl.dhtServers = append(cl.dhtServers, ds)
|
|
}
|
|
}
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
func (cl *Client) enabledPeerNetworks() (ns []string) {
|
|
for _, n := range allPeerNetworks {
|
|
if peerNetworkEnabled(n, cl.config) {
|
|
ns = append(ns, n)
|
|
}
|
|
}
|
|
return
|
|
}
|
|
|
|
func (cl *Client) newDhtServer(conn net.PacketConn) (s *dht.Server, err error) {
|
|
cfg := dht.ServerConfig{
|
|
IPBlocklist: cl.ipBlockList,
|
|
Conn: conn,
|
|
OnAnnouncePeer: cl.onDHTAnnouncePeer,
|
|
PublicIP: func() net.IP {
|
|
if connIsIpv6(conn) && cl.config.PublicIp6 != nil {
|
|
return cl.config.PublicIp6
|
|
}
|
|
return cl.config.PublicIp4
|
|
}(),
|
|
StartingNodes: cl.config.DhtStartingNodes,
|
|
}
|
|
s, err = dht.NewServer(&cfg)
|
|
if err == nil {
|
|
go func() {
|
|
if _, err := s.Bootstrap(); err != nil {
|
|
log.Printf("error bootstrapping dht: %s", err)
|
|
}
|
|
}()
|
|
}
|
|
return
|
|
}
|
|
|
|
func firstNonEmptyString(ss ...string) string {
|
|
for _, s := range ss {
|
|
if s != "" {
|
|
return s
|
|
}
|
|
}
|
|
return ""
|
|
}
|
|
|
|
func (cl *Client) Closed() <-chan struct{} {
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
return cl.closed.C()
|
|
}
|
|
|
|
func (cl *Client) eachDhtServer(f func(*dht.Server)) {
|
|
for _, ds := range cl.dhtServers {
|
|
f(ds)
|
|
}
|
|
}
|
|
|
|
func (cl *Client) closeSockets() {
|
|
cl.eachListener(func(l socket) bool {
|
|
l.Close()
|
|
return true
|
|
})
|
|
cl.conns = nil
|
|
}
|
|
|
|
// Stops the client. All connections to peers are closed and all activity will
|
|
// come to a halt.
|
|
func (cl *Client) Close() {
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
cl.closed.Set()
|
|
cl.eachDhtServer(func(s *dht.Server) { s.Close() })
|
|
cl.closeSockets()
|
|
for _, t := range cl.torrents {
|
|
t.close()
|
|
}
|
|
for _, f := range cl.onClose {
|
|
f()
|
|
}
|
|
cl.event.Broadcast()
|
|
}
|
|
|
|
func (cl *Client) ipBlockRange(ip net.IP) (r iplist.Range, blocked bool) {
|
|
if cl.ipBlockList == nil {
|
|
return
|
|
}
|
|
return cl.ipBlockList.Lookup(ip)
|
|
}
|
|
|
|
func (cl *Client) ipIsBlocked(ip net.IP) bool {
|
|
_, blocked := cl.ipBlockRange(ip)
|
|
return blocked
|
|
}
|
|
|
|
func (cl *Client) waitAccept() {
|
|
for {
|
|
for _, t := range cl.torrents {
|
|
if t.wantConns() {
|
|
return
|
|
}
|
|
}
|
|
if cl.closed.IsSet() {
|
|
return
|
|
}
|
|
cl.event.Wait()
|
|
}
|
|
}
|
|
|
|
func (cl *Client) rejectAccepted(conn net.Conn) bool {
|
|
ra := conn.RemoteAddr()
|
|
rip := missinggo.AddrIP(ra)
|
|
if cl.config.DisableIPv4Peers && rip.To4() != nil {
|
|
return true
|
|
}
|
|
if cl.config.DisableIPv4 && len(rip) == net.IPv4len {
|
|
return true
|
|
}
|
|
if cl.config.DisableIPv6 && len(rip) == net.IPv6len && rip.To4() == nil {
|
|
return true
|
|
}
|
|
if cl.rateLimitAccept(rip) {
|
|
return true
|
|
}
|
|
return cl.badPeerIPPort(rip, missinggo.AddrPort(ra))
|
|
}
|
|
|
|
func (cl *Client) acceptConnections(l net.Listener) {
|
|
for {
|
|
conn, err := l.Accept()
|
|
conn = pproffd.WrapNetConn(conn)
|
|
cl.mu.RLock()
|
|
closed := cl.closed.IsSet()
|
|
reject := false
|
|
if conn != nil {
|
|
reject = cl.rejectAccepted(conn)
|
|
}
|
|
cl.mu.RUnlock()
|
|
if closed {
|
|
if conn != nil {
|
|
conn.Close()
|
|
}
|
|
return
|
|
}
|
|
if err != nil {
|
|
log.Print(err)
|
|
// I think something harsher should happen here? Our accept
|
|
// routine just fucked off.
|
|
return
|
|
}
|
|
go func() {
|
|
if reject {
|
|
torrent.Add("rejected accepted connections", 1)
|
|
conn.Close()
|
|
} else {
|
|
go cl.incomingConnection(conn)
|
|
}
|
|
log.Fmsg("accepted %s connection from %s", conn.RemoteAddr().Network(), conn.RemoteAddr()).AddValue(debugLogValue).Log(cl.logger)
|
|
torrent.Add(fmt.Sprintf("accepted conn remote IP len=%d", len(missinggo.AddrIP(conn.RemoteAddr()))), 1)
|
|
torrent.Add(fmt.Sprintf("accepted conn network=%s", conn.RemoteAddr().Network()), 1)
|
|
torrent.Add(fmt.Sprintf("accepted on %s listener", l.Addr().Network()), 1)
|
|
}()
|
|
}
|
|
}
|
|
|
|
func (cl *Client) incomingConnection(nc net.Conn) {
|
|
defer nc.Close()
|
|
if tc, ok := nc.(*net.TCPConn); ok {
|
|
tc.SetLinger(0)
|
|
}
|
|
c := cl.newConnection(nc, false)
|
|
c.Discovery = peerSourceIncoming
|
|
cl.runReceivedConn(c)
|
|
}
|
|
|
|
// Returns a handle to the given torrent, if it's present in the client.
|
|
func (cl *Client) Torrent(ih metainfo.Hash) (t *Torrent, ok bool) {
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
t, ok = cl.torrents[ih]
|
|
return
|
|
}
|
|
|
|
func (cl *Client) torrent(ih metainfo.Hash) *Torrent {
|
|
return cl.torrents[ih]
|
|
}
|
|
|
|
type dialResult struct {
|
|
Conn net.Conn
|
|
}
|
|
|
|
func countDialResult(err error) {
|
|
if err == nil {
|
|
torrent.Add("successful dials", 1)
|
|
} else {
|
|
torrent.Add("unsuccessful dials", 1)
|
|
}
|
|
}
|
|
|
|
func reducedDialTimeout(minDialTimeout, max time.Duration, halfOpenLimit int, pendingPeers int) (ret time.Duration) {
|
|
ret = max / time.Duration((pendingPeers+halfOpenLimit)/halfOpenLimit)
|
|
if ret < minDialTimeout {
|
|
ret = minDialTimeout
|
|
}
|
|
return
|
|
}
|
|
|
|
// Returns whether an address is known to connect to a client with our own ID.
|
|
func (cl *Client) dopplegangerAddr(addr string) bool {
|
|
_, ok := cl.dopplegangerAddrs[addr]
|
|
return ok
|
|
}
|
|
|
|
func ipNetworkSuffix(allowIpv4, allowIpv6 bool) string {
|
|
switch {
|
|
case allowIpv4 && allowIpv6:
|
|
return ""
|
|
case allowIpv4 && !allowIpv6:
|
|
return "4"
|
|
case !allowIpv4 && allowIpv6:
|
|
return "6"
|
|
default:
|
|
panic("unhandled ip network combination")
|
|
}
|
|
}
|
|
|
|
func dialUTP(ctx context.Context, addr string, sock utpSocket) (c net.Conn, err error) {
|
|
return sock.DialContext(ctx, "", addr)
|
|
}
|
|
|
|
var allPeerNetworks = []string{"tcp4", "tcp6", "udp4", "udp6"}
|
|
|
|
func peerNetworkEnabled(network string, cfg *ClientConfig) bool {
|
|
c := func(s string) bool {
|
|
return strings.Contains(network, s)
|
|
}
|
|
if cfg.DisableUTP {
|
|
if c("udp") || c("utp") {
|
|
return false
|
|
}
|
|
}
|
|
if cfg.DisableTCP && c("tcp") {
|
|
return false
|
|
}
|
|
if cfg.DisableIPv6 && c("6") {
|
|
return false
|
|
}
|
|
return true
|
|
}
|
|
|
|
// Returns a connection over UTP or TCP, whichever is first to connect.
|
|
func (cl *Client) dialFirst(ctx context.Context, addr string) net.Conn {
|
|
ctx, cancel := context.WithCancel(ctx)
|
|
// As soon as we return one connection, cancel the others.
|
|
defer cancel()
|
|
left := 0
|
|
resCh := make(chan dialResult, left)
|
|
dial := func(f func(_ context.Context, addr string) (net.Conn, error)) {
|
|
left++
|
|
go func() {
|
|
c, err := f(ctx, addr)
|
|
// This is a bit optimistic, but it looks non-trivial to thread
|
|
// this through the proxy code. Set it now in case we close the
|
|
// connection forthwith.
|
|
if tc, ok := c.(*net.TCPConn); ok {
|
|
tc.SetLinger(0)
|
|
}
|
|
countDialResult(err)
|
|
resCh <- dialResult{c}
|
|
}()
|
|
}
|
|
func() {
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
cl.eachListener(func(s socket) bool {
|
|
if peerNetworkEnabled(s.Addr().Network(), cl.config) {
|
|
dial(s.dial)
|
|
}
|
|
return true
|
|
})
|
|
}()
|
|
var res dialResult
|
|
// Wait for a successful connection.
|
|
func() {
|
|
defer perf.ScopeTimer()()
|
|
for ; left > 0 && res.Conn == nil; left-- {
|
|
res = <-resCh
|
|
}
|
|
}()
|
|
// There are still incompleted dials.
|
|
go func() {
|
|
for ; left > 0; left-- {
|
|
conn := (<-resCh).Conn
|
|
if conn != nil {
|
|
conn.Close()
|
|
}
|
|
}
|
|
}()
|
|
if res.Conn != nil {
|
|
go torrent.Add(fmt.Sprintf("network dialed first: %s", res.Conn.RemoteAddr().Network()), 1)
|
|
}
|
|
return res.Conn
|
|
}
|
|
|
|
func (cl *Client) noLongerHalfOpen(t *Torrent, addr string) {
|
|
if _, ok := t.halfOpen[addr]; !ok {
|
|
panic("invariant broken")
|
|
}
|
|
delete(t.halfOpen, addr)
|
|
t.openNewConns()
|
|
}
|
|
|
|
// Performs initiator handshakes and returns a connection. Returns nil
|
|
// *connection if no connection for valid reasons.
|
|
func (cl *Client) handshakesConnection(ctx context.Context, nc net.Conn, t *Torrent, encryptHeader bool) (c *connection, err error) {
|
|
c = cl.newConnection(nc, true)
|
|
c.headerEncrypted = encryptHeader
|
|
ctx, cancel := context.WithTimeout(ctx, cl.config.HandshakesTimeout)
|
|
defer cancel()
|
|
dl, ok := ctx.Deadline()
|
|
if !ok {
|
|
panic(ctx)
|
|
}
|
|
err = nc.SetDeadline(dl)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
ok, err = cl.initiateHandshakes(c, t)
|
|
if !ok {
|
|
c = nil
|
|
}
|
|
return
|
|
}
|
|
|
|
// Returns nil connection and nil error if no connection could be established
|
|
// for valid reasons.
|
|
func (cl *Client) establishOutgoingConnEx(t *Torrent, addr string, ctx context.Context, obfuscatedHeader bool) (c *connection, err error) {
|
|
nc := cl.dialFirst(ctx, addr)
|
|
if nc == nil {
|
|
return
|
|
}
|
|
defer func() {
|
|
if c == nil || err != nil {
|
|
nc.Close()
|
|
}
|
|
}()
|
|
return cl.handshakesConnection(ctx, nc, t, obfuscatedHeader)
|
|
}
|
|
|
|
// Returns nil connection and nil error if no connection could be established
|
|
// for valid reasons.
|
|
func (cl *Client) establishOutgoingConn(t *Torrent, addr string) (c *connection, err error) {
|
|
ctx, cancel := context.WithTimeout(context.Background(), time.Minute)
|
|
defer cancel()
|
|
obfuscatedHeaderFirst := !cl.config.DisableEncryption && !cl.config.PreferNoEncryption
|
|
c, err = cl.establishOutgoingConnEx(t, addr, ctx, obfuscatedHeaderFirst)
|
|
if err != nil {
|
|
return
|
|
}
|
|
if c != nil {
|
|
torrent.Add("initiated conn with preferred header obfuscation", 1)
|
|
return
|
|
}
|
|
if cl.config.ForceEncryption {
|
|
// We should have just tried with an obfuscated header. A plaintext
|
|
// header can't result in an encrypted connection, so we're done.
|
|
if !obfuscatedHeaderFirst {
|
|
panic(cl.config.EncryptionPolicy)
|
|
}
|
|
return
|
|
}
|
|
// Try again with encryption if we didn't earlier, or without if we did.
|
|
c, err = cl.establishOutgoingConnEx(t, addr, ctx, !obfuscatedHeaderFirst)
|
|
if c != nil {
|
|
torrent.Add("initiated conn with fallback header obfuscation", 1)
|
|
}
|
|
return
|
|
}
|
|
|
|
// Called to dial out and run a connection. The addr we're given is already
|
|
// considered half-open.
|
|
func (cl *Client) outgoingConnection(t *Torrent, addr string, ps peerSource) {
|
|
c, err := cl.establishOutgoingConn(t, addr)
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
// Don't release lock between here and addConnection, unless it's for
|
|
// failure.
|
|
cl.noLongerHalfOpen(t, addr)
|
|
if err != nil {
|
|
if cl.config.Debug {
|
|
log.Printf("error establishing outgoing connection: %s", err)
|
|
}
|
|
return
|
|
}
|
|
if c == nil {
|
|
return
|
|
}
|
|
defer c.Close()
|
|
c.Discovery = ps
|
|
cl.runHandshookConn(c, t)
|
|
}
|
|
|
|
// The port number for incoming peer connections. 0 if the client isn't
|
|
// listening.
|
|
func (cl *Client) incomingPeerPort() int {
|
|
return cl.LocalPort()
|
|
}
|
|
|
|
func (cl *Client) initiateHandshakes(c *connection, t *Torrent) (ok bool, err error) {
|
|
if c.headerEncrypted {
|
|
var rw io.ReadWriter
|
|
rw, c.cryptoMethod, err = mse.InitiateHandshake(
|
|
struct {
|
|
io.Reader
|
|
io.Writer
|
|
}{c.r, c.w},
|
|
t.infoHash[:],
|
|
nil,
|
|
func() mse.CryptoMethod {
|
|
switch {
|
|
case cl.config.ForceEncryption:
|
|
return mse.CryptoMethodRC4
|
|
case cl.config.DisableEncryption:
|
|
return mse.CryptoMethodPlaintext
|
|
default:
|
|
return mse.AllSupportedCrypto
|
|
}
|
|
}(),
|
|
)
|
|
c.setRW(rw)
|
|
if err != nil {
|
|
return
|
|
}
|
|
}
|
|
ih, ok, err := cl.connBTHandshake(c, &t.infoHash)
|
|
if ih != t.infoHash {
|
|
ok = false
|
|
}
|
|
return
|
|
}
|
|
|
|
// Calls f with any secret keys.
|
|
func (cl *Client) forSkeys(f func([]byte) bool) {
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
for ih := range cl.torrents {
|
|
if !f(ih[:]) {
|
|
break
|
|
}
|
|
}
|
|
}
|
|
|
|
// Do encryption and bittorrent handshakes as receiver.
|
|
func (cl *Client) receiveHandshakes(c *connection) (t *Torrent, err error) {
|
|
defer perf.ScopeTimerErr(&err)()
|
|
var rw io.ReadWriter
|
|
rw, c.headerEncrypted, c.cryptoMethod, err = handleEncryption(c.rw(), cl.forSkeys, cl.config.EncryptionPolicy)
|
|
c.setRW(rw)
|
|
if err != nil {
|
|
if err == mse.ErrNoSecretKeyMatch {
|
|
err = nil
|
|
}
|
|
return
|
|
}
|
|
if cl.config.ForceEncryption && !c.headerEncrypted {
|
|
err = errors.New("connection not encrypted")
|
|
return
|
|
}
|
|
ih, ok, err := cl.connBTHandshake(c, nil)
|
|
if err != nil {
|
|
err = fmt.Errorf("error during bt handshake: %s", err)
|
|
return
|
|
}
|
|
if !ok {
|
|
return
|
|
}
|
|
cl.mu.Lock()
|
|
t = cl.torrents[ih]
|
|
cl.mu.Unlock()
|
|
return
|
|
}
|
|
|
|
// Returns !ok if handshake failed for valid reasons.
|
|
func (cl *Client) connBTHandshake(c *connection, ih *metainfo.Hash) (ret metainfo.Hash, ok bool, err error) {
|
|
res, ok, err := pp.Handshake(c.rw(), ih, cl.peerID, cl.extensionBytes)
|
|
if err != nil || !ok {
|
|
return
|
|
}
|
|
ret = res.Hash
|
|
c.PeerExtensionBytes = res.PeerExtensionBits
|
|
c.PeerID = res.PeerID
|
|
c.completedHandshake = time.Now()
|
|
return
|
|
}
|
|
|
|
func (cl *Client) runReceivedConn(c *connection) {
|
|
err := c.conn.SetDeadline(time.Now().Add(cl.config.HandshakesTimeout))
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
t, err := cl.receiveHandshakes(c)
|
|
if err != nil {
|
|
log.Fmsg(
|
|
"error receiving handshakes: %s", err,
|
|
).AddValue(
|
|
debugLogValue,
|
|
).Add(
|
|
"network", c.remoteAddr().Network(),
|
|
).Log(cl.logger)
|
|
torrent.Add("error receiving handshake", 1)
|
|
cl.mu.Lock()
|
|
cl.onBadAccept(c.remoteAddr())
|
|
cl.mu.Unlock()
|
|
return
|
|
}
|
|
if t == nil {
|
|
torrent.Add("received handshake for unloaded torrent", 1)
|
|
cl.mu.Lock()
|
|
cl.onBadAccept(c.remoteAddr())
|
|
cl.mu.Unlock()
|
|
return
|
|
}
|
|
torrent.Add("received handshake for loaded torrent", 1)
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
cl.runHandshookConn(c, t)
|
|
}
|
|
|
|
func (cl *Client) runHandshookConn(c *connection, t *Torrent) {
|
|
c.setTorrent(t)
|
|
if c.PeerID == cl.peerID {
|
|
if c.outgoing {
|
|
connsToSelf.Add(1)
|
|
addr := c.conn.RemoteAddr().String()
|
|
cl.dopplegangerAddrs[addr] = struct{}{}
|
|
} else {
|
|
// Because the remote address is not necessarily the same as its
|
|
// client's torrent listen address, we won't record the remote address
|
|
// as a doppleganger. Instead, the initiator can record *us* as the
|
|
// doppleganger.
|
|
}
|
|
return
|
|
}
|
|
c.conn.SetWriteDeadline(time.Time{})
|
|
c.r = deadlineReader{c.conn, c.r}
|
|
completedHandshakeConnectionFlags.Add(c.connectionFlags(), 1)
|
|
if connIsIpv6(c.conn) {
|
|
torrent.Add("completed handshake over ipv6", 1)
|
|
}
|
|
if err := t.addConnection(c); err != nil {
|
|
log.Fmsg("error adding connection: %s", err).AddValues(c, debugLogValue).Log(t.logger)
|
|
return
|
|
}
|
|
defer t.dropConnection(c)
|
|
go c.writer(time.Minute)
|
|
cl.sendInitialMessages(c, t)
|
|
err := c.mainReadLoop()
|
|
if err != nil && cl.config.Debug {
|
|
log.Printf("error during connection main read loop: %s", err)
|
|
}
|
|
}
|
|
|
|
func (cl *Client) sendInitialMessages(conn *connection, torrent *Torrent) {
|
|
func() {
|
|
if conn.fastEnabled() {
|
|
if torrent.haveAllPieces() {
|
|
conn.Post(pp.Message{Type: pp.HaveAll})
|
|
conn.sentHaves.AddRange(0, conn.t.NumPieces())
|
|
return
|
|
} else if !torrent.haveAnyPieces() {
|
|
conn.Post(pp.Message{Type: pp.HaveNone})
|
|
conn.sentHaves.Clear()
|
|
return
|
|
}
|
|
}
|
|
conn.PostBitfield()
|
|
}()
|
|
if conn.PeerExtensionBytes.SupportsExtended() && cl.extensionBytes.SupportsExtended() {
|
|
conn.Post(pp.Message{
|
|
Type: pp.Extended,
|
|
ExtendedID: pp.HandshakeExtendedID,
|
|
ExtendedPayload: func() []byte {
|
|
d := map[string]interface{}{
|
|
"m": func() (ret map[string]int) {
|
|
ret = make(map[string]int, 2)
|
|
ret["ut_metadata"] = metadataExtendedId
|
|
if !cl.config.DisablePEX {
|
|
ret["ut_pex"] = pexExtendedId
|
|
}
|
|
return
|
|
}(),
|
|
"v": cl.config.ExtendedHandshakeClientVersion,
|
|
// No upload queue is implemented yet.
|
|
"reqq": 64,
|
|
}
|
|
if !cl.config.DisableEncryption {
|
|
d["e"] = 1
|
|
}
|
|
if torrent.metadataSizeKnown() {
|
|
d["metadata_size"] = torrent.metadataSize()
|
|
}
|
|
if p := cl.incomingPeerPort(); p != 0 {
|
|
d["p"] = p
|
|
}
|
|
yourip, err := addrCompactIP(conn.remoteAddr())
|
|
if err != nil {
|
|
log.Printf("error calculating yourip field value in extension handshake: %s", err)
|
|
} else {
|
|
d["yourip"] = yourip
|
|
}
|
|
// log.Printf("sending %v", d)
|
|
b, err := bencode.Marshal(d)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
return b
|
|
}(),
|
|
})
|
|
}
|
|
if conn.PeerExtensionBytes.SupportsDHT() && cl.extensionBytes.SupportsDHT() && cl.haveDhtServer() {
|
|
conn.Post(pp.Message{
|
|
Type: pp.Port,
|
|
Port: cl.dhtPort(),
|
|
})
|
|
}
|
|
}
|
|
|
|
func (cl *Client) dhtPort() (ret uint16) {
|
|
cl.eachDhtServer(func(s *dht.Server) {
|
|
ret = uint16(missinggo.AddrPort(s.Addr()))
|
|
})
|
|
return
|
|
}
|
|
|
|
func (cl *Client) haveDhtServer() (ret bool) {
|
|
cl.eachDhtServer(func(_ *dht.Server) {
|
|
ret = true
|
|
})
|
|
return
|
|
}
|
|
|
|
// Process incoming ut_metadata message.
|
|
func (cl *Client) gotMetadataExtensionMsg(payload []byte, t *Torrent, c *connection) error {
|
|
var d map[string]int
|
|
err := bencode.Unmarshal(payload, &d)
|
|
if _, ok := err.(bencode.ErrUnusedTrailingBytes); ok {
|
|
} else if err != nil {
|
|
return fmt.Errorf("error unmarshalling bencode: %s", err)
|
|
}
|
|
msgType, ok := d["msg_type"]
|
|
if !ok {
|
|
return errors.New("missing msg_type field")
|
|
}
|
|
piece := d["piece"]
|
|
switch msgType {
|
|
case pp.DataMetadataExtensionMsgType:
|
|
if !c.requestedMetadataPiece(piece) {
|
|
return fmt.Errorf("got unexpected piece %d", piece)
|
|
}
|
|
c.metadataRequests[piece] = false
|
|
begin := len(payload) - metadataPieceSize(d["total_size"], piece)
|
|
if begin < 0 || begin >= len(payload) {
|
|
return fmt.Errorf("data has bad offset in payload: %d", begin)
|
|
}
|
|
t.saveMetadataPiece(piece, payload[begin:])
|
|
c.allStats(add(1, func(cs *ConnStats) *Count { return &cs.ChunksReadUseful }))
|
|
c.lastUsefulChunkReceived = time.Now()
|
|
return t.maybeCompleteMetadata()
|
|
case pp.RequestMetadataExtensionMsgType:
|
|
if !t.haveMetadataPiece(piece) {
|
|
c.Post(t.newMetadataExtensionMessage(c, pp.RejectMetadataExtensionMsgType, d["piece"], nil))
|
|
return nil
|
|
}
|
|
start := (1 << 14) * piece
|
|
c.Post(t.newMetadataExtensionMessage(c, pp.DataMetadataExtensionMsgType, piece, t.metadataBytes[start:start+t.metadataPieceSize(piece)]))
|
|
return nil
|
|
case pp.RejectMetadataExtensionMsgType:
|
|
return nil
|
|
default:
|
|
return errors.New("unknown msg_type value")
|
|
}
|
|
}
|
|
|
|
func (cl *Client) badPeerIPPort(ip net.IP, port int) bool {
|
|
if port == 0 {
|
|
return true
|
|
}
|
|
if cl.dopplegangerAddr(net.JoinHostPort(ip.String(), strconv.FormatInt(int64(port), 10))) {
|
|
return true
|
|
}
|
|
if _, ok := cl.ipBlockRange(ip); ok {
|
|
return true
|
|
}
|
|
if _, ok := cl.badPeerIPs[ip.String()]; ok {
|
|
return true
|
|
}
|
|
return false
|
|
}
|
|
|
|
// Return a Torrent ready for insertion into a Client.
|
|
func (cl *Client) newTorrent(ih metainfo.Hash, specStorage storage.ClientImpl) (t *Torrent) {
|
|
// use provided storage, if provided
|
|
storageClient := cl.defaultStorage
|
|
if specStorage != nil {
|
|
storageClient = storage.NewClient(specStorage)
|
|
}
|
|
|
|
t = &Torrent{
|
|
cl: cl,
|
|
infoHash: ih,
|
|
peers: prioritizedPeers{
|
|
om: btree.New(32),
|
|
getPrio: func(p Peer) peerPriority {
|
|
return bep40PriorityIgnoreError(cl.publicAddr(p.IP), p.addr())
|
|
},
|
|
},
|
|
conns: make(map[*connection]struct{}, 2*cl.config.EstablishedConnsPerTorrent),
|
|
|
|
halfOpen: make(map[string]Peer),
|
|
pieceStateChanges: pubsub.NewPubSub(),
|
|
|
|
storageOpener: storageClient,
|
|
maxEstablishedConns: cl.config.EstablishedConnsPerTorrent,
|
|
|
|
networkingEnabled: true,
|
|
requestStrategy: 3,
|
|
metadataChanged: sync.Cond{
|
|
L: &cl.mu,
|
|
},
|
|
duplicateRequestTimeout: 1 * time.Second,
|
|
}
|
|
t.logger = cl.logger.Clone().AddValue(t)
|
|
t.setChunkSize(defaultChunkSize)
|
|
return
|
|
}
|
|
|
|
// A file-like handle to some torrent data resource.
|
|
type Handle interface {
|
|
io.Reader
|
|
io.Seeker
|
|
io.Closer
|
|
io.ReaderAt
|
|
}
|
|
|
|
func (cl *Client) AddTorrentInfoHash(infoHash metainfo.Hash) (t *Torrent, new bool) {
|
|
return cl.AddTorrentInfoHashWithStorage(infoHash, nil)
|
|
}
|
|
|
|
// Adds a torrent by InfoHash with a custom Storage implementation.
|
|
// If the torrent already exists then this Storage is ignored and the
|
|
// existing torrent returned with `new` set to `false`
|
|
func (cl *Client) AddTorrentInfoHashWithStorage(infoHash metainfo.Hash, specStorage storage.ClientImpl) (t *Torrent, new bool) {
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
t, ok := cl.torrents[infoHash]
|
|
if ok {
|
|
return
|
|
}
|
|
new = true
|
|
|
|
t = cl.newTorrent(infoHash, specStorage)
|
|
cl.eachDhtServer(func(s *dht.Server) {
|
|
go t.dhtAnnouncer(s)
|
|
})
|
|
cl.torrents[infoHash] = t
|
|
cl.clearAcceptLimits()
|
|
t.updateWantPeersEvent()
|
|
// Tickle Client.waitAccept, new torrent may want conns.
|
|
cl.event.Broadcast()
|
|
return
|
|
}
|
|
|
|
// Add or merge a torrent spec. If the torrent is already present, the
|
|
// trackers will be merged with the existing ones. If the Info isn't yet
|
|
// known, it will be set. The display name is replaced if the new spec
|
|
// provides one. Returns new if the torrent wasn't already in the client.
|
|
// Note that any `Storage` defined on the spec will be ignored if the
|
|
// torrent is already present (i.e. `new` return value is `true`)
|
|
func (cl *Client) AddTorrentSpec(spec *TorrentSpec) (t *Torrent, new bool, err error) {
|
|
t, new = cl.AddTorrentInfoHashWithStorage(spec.InfoHash, spec.Storage)
|
|
if spec.DisplayName != "" {
|
|
t.SetDisplayName(spec.DisplayName)
|
|
}
|
|
if spec.InfoBytes != nil {
|
|
err = t.SetInfoBytes(spec.InfoBytes)
|
|
if err != nil {
|
|
return
|
|
}
|
|
}
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
if spec.ChunkSize != 0 {
|
|
t.setChunkSize(pp.Integer(spec.ChunkSize))
|
|
}
|
|
t.addTrackers(spec.Trackers)
|
|
t.maybeNewConns()
|
|
return
|
|
}
|
|
|
|
func (cl *Client) dropTorrent(infoHash metainfo.Hash) (err error) {
|
|
t, ok := cl.torrents[infoHash]
|
|
if !ok {
|
|
err = fmt.Errorf("no such torrent")
|
|
return
|
|
}
|
|
err = t.close()
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
delete(cl.torrents, infoHash)
|
|
return
|
|
}
|
|
|
|
func (cl *Client) allTorrentsCompleted() bool {
|
|
for _, t := range cl.torrents {
|
|
if !t.haveInfo() {
|
|
return false
|
|
}
|
|
if !t.haveAllPieces() {
|
|
return false
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
// Returns true when all torrents are completely downloaded and false if the
|
|
// client is stopped before that.
|
|
func (cl *Client) WaitAll() bool {
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
for !cl.allTorrentsCompleted() {
|
|
if cl.closed.IsSet() {
|
|
return false
|
|
}
|
|
cl.event.Wait()
|
|
}
|
|
return true
|
|
}
|
|
|
|
// Returns handles to all the torrents loaded in the Client.
|
|
func (cl *Client) Torrents() []*Torrent {
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
return cl.torrentsAsSlice()
|
|
}
|
|
|
|
func (cl *Client) torrentsAsSlice() (ret []*Torrent) {
|
|
for _, t := range cl.torrents {
|
|
ret = append(ret, t)
|
|
}
|
|
return
|
|
}
|
|
|
|
func (cl *Client) AddMagnet(uri string) (T *Torrent, err error) {
|
|
spec, err := TorrentSpecFromMagnetURI(uri)
|
|
if err != nil {
|
|
return
|
|
}
|
|
T, _, err = cl.AddTorrentSpec(spec)
|
|
return
|
|
}
|
|
|
|
func (cl *Client) AddTorrent(mi *metainfo.MetaInfo) (T *Torrent, err error) {
|
|
T, _, err = cl.AddTorrentSpec(TorrentSpecFromMetaInfo(mi))
|
|
var ss []string
|
|
slices.MakeInto(&ss, mi.Nodes)
|
|
cl.AddDHTNodes(ss)
|
|
return
|
|
}
|
|
|
|
func (cl *Client) AddTorrentFromFile(filename string) (T *Torrent, err error) {
|
|
mi, err := metainfo.LoadFromFile(filename)
|
|
if err != nil {
|
|
return
|
|
}
|
|
return cl.AddTorrent(mi)
|
|
}
|
|
|
|
func (cl *Client) DhtServers() []*dht.Server {
|
|
return cl.dhtServers
|
|
}
|
|
|
|
func (cl *Client) AddDHTNodes(nodes []string) {
|
|
for _, n := range nodes {
|
|
hmp := missinggo.SplitHostMaybePort(n)
|
|
ip := net.ParseIP(hmp.Host)
|
|
if ip == nil {
|
|
log.Printf("won't add DHT node with bad IP: %q", hmp.Host)
|
|
continue
|
|
}
|
|
ni := krpc.NodeInfo{
|
|
Addr: krpc.NodeAddr{
|
|
IP: ip,
|
|
Port: hmp.Port,
|
|
},
|
|
}
|
|
cl.eachDhtServer(func(s *dht.Server) {
|
|
s.AddNode(ni)
|
|
})
|
|
}
|
|
}
|
|
|
|
func (cl *Client) banPeerIP(ip net.IP) {
|
|
if cl.badPeerIPs == nil {
|
|
cl.badPeerIPs = make(map[string]struct{})
|
|
}
|
|
cl.badPeerIPs[ip.String()] = struct{}{}
|
|
}
|
|
|
|
func (cl *Client) newConnection(nc net.Conn, outgoing bool) (c *connection) {
|
|
c = &connection{
|
|
conn: nc,
|
|
outgoing: outgoing,
|
|
Choked: true,
|
|
PeerChoked: true,
|
|
PeerMaxRequests: 250,
|
|
writeBuffer: new(bytes.Buffer),
|
|
}
|
|
c.writerCond.L = &cl.mu
|
|
c.setRW(connStatsReadWriter{nc, c})
|
|
c.r = &rateLimitedReader{
|
|
l: cl.config.DownloadRateLimiter,
|
|
r: c.r,
|
|
}
|
|
return
|
|
}
|
|
|
|
func (cl *Client) onDHTAnnouncePeer(ih metainfo.Hash, p dht.Peer) {
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
t := cl.torrent(ih)
|
|
if t == nil {
|
|
return
|
|
}
|
|
t.addPeers([]Peer{{
|
|
IP: p.IP,
|
|
Port: p.Port,
|
|
Source: peerSourceDHTAnnouncePeer,
|
|
}})
|
|
}
|
|
|
|
func firstNotNil(ips ...net.IP) net.IP {
|
|
for _, ip := range ips {
|
|
if ip != nil {
|
|
return ip
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (cl *Client) eachListener(f func(socket) bool) {
|
|
for _, s := range cl.conns {
|
|
if !f(s) {
|
|
break
|
|
}
|
|
}
|
|
}
|
|
|
|
func (cl *Client) findListener(f func(net.Listener) bool) (ret net.Listener) {
|
|
cl.eachListener(func(l socket) bool {
|
|
ret = l
|
|
return !f(l)
|
|
})
|
|
return
|
|
}
|
|
|
|
func (cl *Client) publicIp(peer net.IP) net.IP {
|
|
// TODO: Use BEP 10 to determine how peers are seeing us.
|
|
if peer.To4() != nil {
|
|
return firstNotNil(
|
|
cl.config.PublicIp4,
|
|
cl.findListenerIp(func(ip net.IP) bool { return ip.To4() != nil }),
|
|
)
|
|
} else {
|
|
return firstNotNil(
|
|
cl.config.PublicIp6,
|
|
cl.findListenerIp(func(ip net.IP) bool { return ip.To4() == nil }),
|
|
)
|
|
}
|
|
}
|
|
|
|
func (cl *Client) findListenerIp(f func(net.IP) bool) net.IP {
|
|
return missinggo.AddrIP(cl.findListener(func(l net.Listener) bool {
|
|
return f(missinggo.AddrIP(l.Addr()))
|
|
}).Addr())
|
|
}
|
|
|
|
// Our IP as a peer should see it.
|
|
func (cl *Client) publicAddr(peer net.IP) ipPort {
|
|
return ipPort{cl.publicIp(peer), uint16(cl.incomingPeerPort())}
|
|
}
|
|
|
|
func (cl *Client) ListenAddrs() (ret []net.Addr) {
|
|
cl.mu.Lock()
|
|
defer cl.mu.Unlock()
|
|
cl.eachListener(func(l socket) bool {
|
|
ret = append(ret, l.Addr())
|
|
return true
|
|
})
|
|
return
|
|
}
|
|
|
|
func (cl *Client) onBadAccept(addr net.Addr) {
|
|
ip := maskIpForAcceptLimiting(missinggo.AddrIP(addr))
|
|
if cl.acceptLimiter == nil {
|
|
cl.acceptLimiter = make(map[ipStr]int)
|
|
}
|
|
cl.acceptLimiter[ipStr(ip.String())]++
|
|
}
|
|
|
|
func maskIpForAcceptLimiting(ip net.IP) net.IP {
|
|
if ip4 := ip.To4(); ip4 != nil {
|
|
return ip4.Mask(net.CIDRMask(24, 32))
|
|
}
|
|
return ip
|
|
}
|
|
|
|
func (cl *Client) clearAcceptLimits() {
|
|
cl.acceptLimiter = nil
|
|
}
|
|
|
|
func (cl *Client) acceptLimitClearer() {
|
|
for {
|
|
select {
|
|
case <-cl.closed.LockedChan(&cl.mu):
|
|
return
|
|
case <-time.After(15 * time.Minute):
|
|
cl.mu.Lock()
|
|
cl.clearAcceptLimits()
|
|
cl.mu.Unlock()
|
|
}
|
|
}
|
|
}
|
|
|
|
func (cl *Client) rateLimitAccept(ip net.IP) bool {
|
|
if cl.config.DisableAcceptRateLimiting {
|
|
return false
|
|
}
|
|
return cl.acceptLimiter[ipStr(maskIpForAcceptLimiting(ip).String())] > 0
|
|
}
|