| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371 |
- package torrent
- import (
- "context"
- "encoding/gob"
- "fmt"
- "reflect"
- "runtime/pprof"
- "time"
- "unsafe"
- "github.com/RoaringBitmap/roaring"
- g "github.com/anacrolix/generics"
- "github.com/anacrolix/generics/heap"
- "github.com/anacrolix/log"
- "github.com/anacrolix/multiless"
- requestStrategy "github.com/anacrolix/torrent/request-strategy"
- typedRoaring "github.com/anacrolix/torrent/typed-roaring"
- )
- type (
- // Since we have to store all the requests in memory, we can't reasonably exceed what could be
- // indexed with the memory space available.
- maxRequests = int
- )
- func (t *Torrent) requestStrategyPieceOrderState(i int) requestStrategy.PieceRequestOrderState {
- return requestStrategy.PieceRequestOrderState{
- Priority: t.piece(i).purePriority(),
- Partial: t.piecePartiallyDownloaded(i),
- Availability: t.piece(i).availability(),
- }
- }
- func init() {
- gob.Register(peerId{})
- }
- type peerId struct {
- *Peer
- ptr uintptr
- }
- func (p peerId) Uintptr() uintptr {
- return p.ptr
- }
- func (p peerId) GobEncode() (b []byte, _ error) {
- *(*reflect.SliceHeader)(unsafe.Pointer(&b)) = reflect.SliceHeader{
- Data: uintptr(unsafe.Pointer(&p.ptr)),
- Len: int(unsafe.Sizeof(p.ptr)),
- Cap: int(unsafe.Sizeof(p.ptr)),
- }
- return
- }
- func (p *peerId) GobDecode(b []byte) error {
- if uintptr(len(b)) != unsafe.Sizeof(p.ptr) {
- panic(len(b))
- }
- ptr := unsafe.Pointer(&b[0])
- p.ptr = *(*uintptr)(ptr)
- log.Printf("%p", ptr)
- dst := reflect.SliceHeader{
- Data: uintptr(unsafe.Pointer(&p.Peer)),
- Len: int(unsafe.Sizeof(p.Peer)),
- Cap: int(unsafe.Sizeof(p.Peer)),
- }
- copy(*(*[]byte)(unsafe.Pointer(&dst)), b)
- return nil
- }
- type (
- RequestIndex = requestStrategy.RequestIndex
- chunkIndexType = requestStrategy.ChunkIndex
- )
- type desiredPeerRequests struct {
- requestIndexes []RequestIndex
- peer *Peer
- pieceStates []g.Option[requestStrategy.PieceRequestOrderState]
- }
- func (p *desiredPeerRequests) lessByValue(leftRequest, rightRequest RequestIndex) bool {
- t := p.peer.t
- leftPieceIndex := t.pieceIndexOfRequestIndex(leftRequest)
- rightPieceIndex := t.pieceIndexOfRequestIndex(rightRequest)
- ml := multiless.New()
- // Push requests that can't be served right now to the end. But we don't throw them away unless
- // there's a better alternative. This is for when we're using the fast extension and get choked
- // but our requests could still be good when we get unchoked.
- if p.peer.peerChoking {
- ml = ml.Bool(
- !p.peer.peerAllowedFast.Contains(leftPieceIndex),
- !p.peer.peerAllowedFast.Contains(rightPieceIndex),
- )
- }
- leftPiece := p.pieceStates[leftPieceIndex].UnwrapPtr()
- rightPiece := p.pieceStates[rightPieceIndex].UnwrapPtr()
- // Putting this first means we can steal requests from lesser-performing peers for our first few
- // new requests.
- priority := func() PiecePriority {
- // Technically we would be happy with the cached priority here, except we don't actually
- // cache it anymore, and Torrent.PiecePriority just does another lookup of *Piece to resolve
- // the priority through Piece.purePriority, which is probably slower.
- leftPriority := leftPiece.Priority
- rightPriority := rightPiece.Priority
- ml = ml.Int(
- -int(leftPriority),
- -int(rightPriority),
- )
- if !ml.Ok() {
- if leftPriority != rightPriority {
- panic("expected equal")
- }
- }
- return leftPriority
- }()
- if ml.Ok() {
- return ml.MustLess()
- }
- leftRequestState := t.requestState[leftRequest]
- rightRequestState := t.requestState[rightRequest]
- leftPeer := leftRequestState.peer
- rightPeer := rightRequestState.peer
- // Prefer chunks already requested from this peer.
- ml = ml.Bool(rightPeer == p.peer, leftPeer == p.peer)
- // Prefer unrequested chunks.
- ml = ml.Bool(rightPeer == nil, leftPeer == nil)
- if ml.Ok() {
- return ml.MustLess()
- }
- if leftPeer != nil {
- // The right peer should also be set, or we'd have resolved the computation by now.
- ml = ml.Uint64(
- rightPeer.requestState.Requests.GetCardinality(),
- leftPeer.requestState.Requests.GetCardinality(),
- )
- // Could either of the lastRequested be Zero? That's what checking an existing peer is for.
- leftLast := leftRequestState.when
- rightLast := rightRequestState.when
- if leftLast.IsZero() || rightLast.IsZero() {
- panic("expected non-zero last requested times")
- }
- // We want the most-recently requested on the left. Clients like Transmission serve requests
- // in received order, so the most recently-requested is the one that has the longest until
- // it will be served and therefore is the best candidate to cancel.
- ml = ml.CmpInt64(rightLast.Sub(leftLast).Nanoseconds())
- }
- ml = ml.Int(
- leftPiece.Availability,
- rightPiece.Availability)
- if priority == PiecePriorityReadahead {
- // TODO: For readahead in particular, it would be even better to consider distance from the
- // reader position so that reads earlier in a torrent don't starve reads later in the
- // torrent. This would probably require reconsideration of how readahead priority works.
- ml = ml.Int(leftPieceIndex, rightPieceIndex)
- } else {
- ml = ml.Int(t.pieceRequestOrder[leftPieceIndex], t.pieceRequestOrder[rightPieceIndex])
- }
- return ml.Less()
- }
- type desiredRequestState struct {
- Requests desiredPeerRequests
- Interested bool
- }
- func (p *Peer) getDesiredRequestState() (desired desiredRequestState) {
- t := p.t
- if !t.haveInfo() {
- return
- }
- if t.closed.IsSet() {
- return
- }
- if t.dataDownloadDisallowed.Bool() {
- return
- }
- input := t.getRequestStrategyInput()
- requestHeap := desiredPeerRequests{
- peer: p,
- pieceStates: t.requestPieceStates,
- requestIndexes: t.requestIndexes,
- }
- clear(requestHeap.pieceStates)
- // Caller-provided allocation for roaring bitmap iteration.
- var it typedRoaring.Iterator[RequestIndex]
- requestStrategy.GetRequestablePieces(
- input,
- t.getPieceRequestOrder(),
- func(ih InfoHash, pieceIndex int, pieceExtra requestStrategy.PieceRequestOrderState) bool {
- if ih != *t.canonicalShortInfohash() {
- return false
- }
- if !p.peerHasPiece(pieceIndex) {
- return false
- }
- requestHeap.pieceStates[pieceIndex].Set(pieceExtra)
- allowedFast := p.peerAllowedFast.Contains(pieceIndex)
- t.iterUndirtiedRequestIndexesInPiece(&it, pieceIndex, func(r requestStrategy.RequestIndex) {
- if !allowedFast {
- // We must signal interest to request this. TODO: We could set interested if the
- // peers pieces (minus the allowed fast set) overlap with our missing pieces if
- // there are any readers, or any pending pieces.
- desired.Interested = true
- // We can make or will allow sustaining a request here if we're not choked, or
- // have made the request previously (presumably while unchoked), and haven't had
- // the peer respond yet (and the request was retained because we are using the
- // fast extension).
- if p.peerChoking && !p.requestState.Requests.Contains(r) {
- // We can't request this right now.
- return
- }
- }
- cancelled := &p.requestState.Cancelled
- if !cancelled.IsEmpty() && cancelled.Contains(r) {
- // Can't re-request while awaiting acknowledgement.
- return
- }
- requestHeap.requestIndexes = append(requestHeap.requestIndexes, r)
- })
- return true
- },
- )
- t.assertPendingRequests()
- desired.Requests = requestHeap
- return
- }
- func (p *Peer) maybeUpdateActualRequestState() {
- if p.closed.IsSet() {
- return
- }
- if p.needRequestUpdate == "" {
- return
- }
- if p.needRequestUpdate == peerUpdateRequestsTimerReason {
- since := time.Since(p.lastRequestUpdate)
- if since < updateRequestsTimerDuration {
- panic(since)
- }
- }
- pprof.Do(
- context.Background(),
- pprof.Labels("update request", string(p.needRequestUpdate)),
- func(_ context.Context) {
- next := p.getDesiredRequestState()
- p.applyRequestState(next)
- p.t.cacheNextRequestIndexesForReuse(next.Requests.requestIndexes)
- },
- )
- }
- func (t *Torrent) cacheNextRequestIndexesForReuse(slice []RequestIndex) {
- // The incoming slice can be smaller when getDesiredRequestState short circuits on some
- // conditions.
- if cap(slice) > cap(t.requestIndexes) {
- t.requestIndexes = slice[:0]
- }
- }
- // Whether we should allow sending not interested ("losing interest") to the peer. I noticed
- // qBitTorrent seems to punish us for sending not interested when we're streaming and don't
- // currently need anything.
- func (p *Peer) allowSendNotInterested() bool {
- // Except for caching, we're not likely to lose pieces very soon.
- if p.t.haveAllPieces() {
- return true
- }
- all, known := p.peerHasAllPieces()
- if all || !known {
- return false
- }
- // Allow losing interest if we have all the pieces the peer has.
- return roaring.AndNot(p.peerPieces(), &p.t._completedPieces).IsEmpty()
- }
- // Transmit/action the request state to the peer.
- func (p *Peer) applyRequestState(next desiredRequestState) {
- current := &p.requestState
- // Make interest sticky
- if !next.Interested && p.requestState.Interested {
- if !p.allowSendNotInterested() {
- next.Interested = true
- }
- }
- if !p.setInterested(next.Interested) {
- return
- }
- more := true
- orig := next.Requests.requestIndexes
- requestHeap := heap.InterfaceForSlice(
- &next.Requests.requestIndexes,
- next.Requests.lessByValue,
- )
- heap.Init(requestHeap)
- t := p.t
- originalRequestCount := current.Requests.GetCardinality()
- for {
- if requestHeap.Len() == 0 {
- break
- }
- numPending := maxRequests(current.Requests.GetCardinality() + current.Cancelled.GetCardinality())
- if numPending >= p.nominalMaxRequests() {
- break
- }
- req := heap.Pop(requestHeap)
- if cap(next.Requests.requestIndexes) != cap(orig) {
- panic("changed")
- }
- // don't add requests on reciept of a reject - because this causes request back
- // to potentially permanently unresponive peers - which just adds network noise. If
- // the peer can handle more requests it will send an "unchoked" message - which
- // will cause it to get added back to the request queue
- if p.needRequestUpdate == peerUpdateRequestsRemoteRejectReason {
- continue
- }
- existing := t.requestingPeer(req)
- if existing != nil && existing != p {
- // don't steal on cancel - because this is triggered by t.cancelRequest below
- // which means that the cancelled can immediately try to steal back a request
- // it has lost which can lead to circular cancel/add processing
- if p.needRequestUpdate == peerUpdateRequestsPeerCancelReason {
- continue
- }
- // Don't steal from the poor.
- diff := int64(current.Requests.GetCardinality()) + 1 - (int64(existing.uncancelledRequests()) - 1)
- // Steal a request that leaves us with one more request than the existing peer
- // connection if the stealer more recently received a chunk.
- if diff > 1 || (diff == 1 && !p.lastUsefulChunkReceived.After(existing.lastUsefulChunkReceived)) {
- continue
- }
- t.cancelRequest(req)
- }
- more = p.mustRequest(req)
- if !more {
- break
- }
- }
- if !more {
- // This might fail if we incorrectly determine that we can fit up to the maximum allowed
- // requests into the available write buffer space. We don't want that to happen because it
- // makes our peak requests dependent on how much was already in the buffer.
- panic(fmt.Sprintf(
- "couldn't fill apply entire request state [newRequests=%v]",
- current.Requests.GetCardinality()-originalRequestCount))
- }
- newPeakRequests := maxRequests(current.Requests.GetCardinality() - originalRequestCount)
- // log.Printf(
- // "requests %v->%v (peak %v->%v) reason %q (peer %v)",
- // originalRequestCount, current.Requests.GetCardinality(), p.peakRequests, newPeakRequests, p.needRequestUpdate, p)
- p.peakRequests = newPeakRequests
- p.needRequestUpdate = ""
- p.lastRequestUpdate = time.Now()
- if enableUpdateRequestsTimer {
- p.updateRequestsTimer.Reset(updateRequestsTimerDuration)
- }
- }
- // This could be set to 10s to match the unchoke/request update interval recommended by some
- // specifications. I've set it shorter to trigger it more often for testing for now.
- const (
- updateRequestsTimerDuration = 3 * time.Second
- enableUpdateRequestsTimer = false
- )
|