FedP2P/webseed/client.go

185 lines
4.7 KiB
Go
Raw Normal View History

2020-05-31 19:00:44 +08:00
package webseed
import (
2020-06-01 16:25:45 +08:00
"bytes"
"context"
"fmt"
"io"
"log"
2020-05-31 19:00:44 +08:00
"net/http"
"strings"
2020-05-31 19:00:44 +08:00
"github.com/RoaringBitmap/roaring"
"github.com/anacrolix/torrent/common"
2020-06-01 16:25:45 +08:00
"github.com/anacrolix/torrent/metainfo"
"github.com/anacrolix/torrent/segments"
2020-05-31 19:00:44 +08:00
)
2020-06-01 16:25:45 +08:00
type RequestSpec = segments.Extent
type requestPartResult struct {
2020-06-01 16:25:45 +08:00
resp *http.Response
err error
}
type requestPart struct {
req *http.Request
e segments.Extent
result chan requestPartResult
start func()
2020-06-01 16:25:45 +08:00
}
type Request struct {
2020-06-01 16:25:45 +08:00
cancel func()
Result chan RequestResult
}
func (r Request) Cancel() {
r.cancel()
2020-06-01 16:25:45 +08:00
}
2020-05-31 19:00:44 +08:00
type Client struct {
HttpClient *http.Client
Url string
fileIndex segments.Index
info *metainfo.Info
// The pieces we can request with the Url. We're more likely to ban/block at the file-level
// given that's how requests are mapped to webseeds, but the torrent.Client works at the piece
// level. We can map our file-level adjustments to the pieces here. This probably need to be
// private in the future, if Client ever starts removing pieces.
Pieces roaring.Bitmap
}
func (me *Client) SetInfo(info *metainfo.Info) {
if !strings.HasSuffix(me.Url, "/") && info.IsDir() {
// In my experience, this is a non-conforming webseed. For example the
// http://ia600500.us.archive.org/1/items URLs in archive.org torrents.
return
}
me.fileIndex = segments.NewIndex(common.LengthIterFromUpvertedFiles(info.UpvertedFiles()))
me.info = info
me.Pieces.AddRange(0, uint64(info.NumPieces()))
2020-06-01 16:25:45 +08:00
}
type RequestResult struct {
Bytes []byte
Err error
2020-06-01 16:25:45 +08:00
}
func (ws *Client) NewRequest(r RequestSpec) Request {
2020-06-01 16:25:45 +08:00
ctx, cancel := context.WithCancel(context.Background())
var requestParts []requestPart
if !ws.fileIndex.Locate(r, func(i int, e segments.Extent) bool {
req, err := NewRequest(ws.Url, i, ws.info, e.Start, e.Length)
2020-06-01 16:25:45 +08:00
if err != nil {
panic(err)
}
req = req.WithContext(ctx)
part := requestPart{
req: req,
result: make(chan requestPartResult, 1),
2020-06-01 16:25:45 +08:00
e: e,
}
part.start = func() {
go func() {
resp, err := ws.HttpClient.Do(req)
part.result <- requestPartResult{
resp: resp,
err: err,
}
}()
}
2020-06-01 16:25:45 +08:00
requestParts = append(requestParts, part)
return true
}) {
panic("request out of file bounds")
}
req := Request{
cancel: cancel,
Result: make(chan RequestResult, 1),
2020-06-01 16:25:45 +08:00
}
go func() {
b, err := readRequestPartResponses(ctx, requestParts)
req.Result <- RequestResult{
Bytes: b,
Err: err,
2020-06-01 16:25:45 +08:00
}
}()
return req
2020-06-01 16:25:45 +08:00
}
type ErrBadResponse struct {
Msg string
Response *http.Response
}
func (me ErrBadResponse) Error() string {
return me.Msg
}
func recvPartResult(ctx context.Context, buf io.Writer, part requestPart) error {
2020-06-01 16:25:45 +08:00
result := <-part.result
// Make sure there's no further results coming, it should be a one-shot channel.
close(part.result)
2020-06-01 16:25:45 +08:00
if result.err != nil {
return result.err
}
defer result.resp.Body.Close()
if ctx.Err() != nil {
return ctx.Err()
}
switch result.resp.StatusCode {
case http.StatusPartialContent:
copied, err := io.Copy(buf, result.resp.Body)
if err != nil {
return err
}
if copied != part.e.Length {
return fmt.Errorf("got %v bytes, expected %v", copied, part.e.Length)
}
return nil
case http.StatusOK:
// This number is based on
// https://archive.org/download/BloodyPitOfHorror/BloodyPitOfHorror.asr.srt. It seems that
// archive.org might be using a webserver implementation that refuses to do partial
// responses to small files.
if part.e.Start < 48<<10 {
if part.e.Start != 0 {
log.Printf("resp status ok but requested range [url=%q, range=%q]",
part.req.URL,
part.req.Header.Get("Range"))
}
// Instead of discarding, we could try receiving all the chunks present in the response
// body. I don't know how one would handle multiple chunk requests resulting in an OK
// response for the same file. The request algorithm might be need to be smarter for
// that.
discarded, _ := io.CopyN(io.Discard, result.resp.Body, part.e.Start)
if discarded != 0 {
log.Printf("discarded %v bytes in webseed request response part", discarded)
}
_, err := io.CopyN(buf, result.resp.Body, part.e.Length)
return err
} else {
return ErrBadResponse{"resp status ok but requested range", result.resp}
}
default:
return ErrBadResponse{
fmt.Sprintf("unhandled response status code (%v)", result.resp.StatusCode),
result.resp,
}
2020-06-01 16:25:45 +08:00
}
2020-05-31 19:00:44 +08:00
}
func readRequestPartResponses(ctx context.Context, parts []requestPart) (_ []byte, err error) {
2020-06-01 16:25:45 +08:00
var buf bytes.Buffer
for _, part := range parts {
part.start()
err = recvPartResult(ctx, &buf, part)
if err != nil {
err = fmt.Errorf("reading %q at %q: %w", part.req.URL, part.req.Header.Get("Range"), err)
break
}
}
return buf.Bytes(), err
2020-05-31 19:00:44 +08:00
}