Merge scrape branch:

Automatically scrape trackers as needed.
  If tracker supplies a trackerid then use it (untested).
  Use tracker's min interval, clamped to the same range as interval.
  Show total completed downloads in the MacOS X frontend.
This commit is contained in:
Josh Elsasser
2006-12-02 01:46:54 +00:00
parent 5ac768644e
commit 44f8325312
13 changed files with 380 additions and 166 deletions

View File

@@ -263,7 +263,7 @@
4DAB87C40ABE1F730081CF7E /* http.c */ = {isa = PBXFileReference; fileEncoding = 30; lastKnownFileType = sourcecode.c.c; name = http.c; path = libtransmission/http.c; sourceTree = "<group>"; };
4DCCBB3C09C3D71100D3CABF /* TorrentCell.m */ = {isa = PBXFileReference; fileEncoding = 30; lastKnownFileType = sourcecode.c.objc; name = TorrentCell.m; path = macosx/TorrentCell.m; sourceTree = "<group>"; };
4DCCBB3D09C3D71100D3CABF /* TorrentCell.h */ = {isa = PBXFileReference; fileEncoding = 30; lastKnownFileType = sourcecode.c.h; name = TorrentCell.h; path = macosx/TorrentCell.h; sourceTree = "<group>"; };
4DDBB71909E16BAE00284745 /* transmissioncli */ = {isa = PBXFileReference; includeInIndex = 0; lastKnownFileType = "compiled.mach-o.executable"; path = transmissioncli; sourceTree = BUILT_PRODUCTS_DIR; };
4DDBB71909E16BAE00284745 /* transmissioncli */ = {isa = PBXFileReference; explicitFileType = "compiled.mach-o.executable"; includeInIndex = 0; path = transmissioncli; sourceTree = BUILT_PRODUCTS_DIR; };
4DDBB71B09E16BF100284745 /* transmissioncli.c */ = {isa = PBXFileReference; fileEncoding = 30; lastKnownFileType = sourcecode.c.c; name = transmissioncli.c; path = cli/transmissioncli.c; sourceTree = "<group>"; };
4DDFDD20099A5D8E00189D81 /* DownloadBadge.png */ = {isa = PBXFileReference; lastKnownFileType = image.png; name = DownloadBadge.png; path = macosx/Images/DownloadBadge.png; sourceTree = "<group>"; };
4DDFDD21099A5D8E00189D81 /* UploadBadge.png */ = {isa = PBXFileReference; lastKnownFileType = image.png; name = UploadBadge.png; path = macosx/Images/UploadBadge.png; sourceTree = "<group>"; };

View File

@@ -146,15 +146,16 @@ int main( int argc, char ** argv )
if( showScrape )
{
int seeders, leechers;
int seeders, leechers, downloaded;
if( tr_torrentScrape( tor, &seeders, &leechers ) )
if( tr_torrentScrape( tor, &seeders, &leechers, &downloaded ) )
{
printf( "Scrape failed.\n" );
}
else
{
printf( "%d seeder(s), %d leecher(s).\n", seeders, leechers );
printf( "%d seeder(s), %d leecher(s), %d download(s).\n",
seeders, leechers, downloaded );
}
goto cleanup;

View File

@@ -158,7 +158,7 @@ struct tr_torrent_s
int * bindPort;
/* An escaped string used to include the hash in HTTP queries */
char hashString[3*SHA_DIGEST_LENGTH+1];
char escapedHashString[3*SHA_DIGEST_LENGTH+1];
char scrape[MAX_PATH_LENGTH];

View File

@@ -29,58 +29,74 @@ struct tr_tracker_s
tr_torrent_t * tor;
char * id;
char * trackerid;
char started;
char completed;
char stopped;
int interval;
int minInterval;
int scrapeInterval;
int seeders;
int leechers;
int hasManyPeers;
int complete;
uint64_t dateTry;
uint64_t dateOk;
uint64_t dateScrape;
int lastScrapeFailed;
#define TC_ATTEMPT_NOREACH 1
#define TC_ATTEMPT_ERROR 2
#define TC_ATTEMPT_OK 4
char lastAttempt;
int scrapeNeeded;
tr_http_t * http;
tr_http_t * httpScrape;
int bindPort;
int newPort;
};
static tr_http_t * getQuery ( tr_tracker_t * tc );
static void readAnswer ( tr_tracker_t * tc, const char *, int );
static tr_http_t * getQuery ( tr_tracker_t * tc );
static tr_http_t * getScrapeQuery ( tr_tracker_t * tc );
static void readAnswer ( tr_tracker_t * tc, const char *, int );
static void readScrapeAnswer ( tr_tracker_t * tc, const char *, int );
static void killHttp ( tr_http_t ** http, tr_fd_t * fdlimit );
tr_tracker_t * tr_trackerInit( tr_torrent_t * tor )
{
tr_tracker_t * tc;
tc = calloc( 1, sizeof( tr_tracker_t ) );
tc->tor = tor;
tc->id = tor->id;
tc = calloc( 1, sizeof( tr_tracker_t ) );
tc->tor = tor;
tc->id = tor->id;
tc->started = 1;
tc->started = 1;
tc->interval = 300;
tc->seeders = -1;
tc->leechers = -1;
tc->interval = 300;
tc->scrapeInterval = 600;
tc->seeders = -1;
tc->leechers = -1;
tc->complete = -1;
tc->lastAttempt = TC_ATTEMPT_NOREACH;
tc->lastAttempt = TC_ATTEMPT_NOREACH;
tc->bindPort = *(tor->bindPort);
tc->newPort = -1;
tc->bindPort = *(tor->bindPort);
tc->newPort = -1;
return tc;
}
static int shouldConnect( tr_tracker_t * tc )
{
uint64_t now = tr_date();
tr_torrent_t * tor = tc->tor;
uint64_t now;
now = tr_date();
/* Unreachable tracker, try 10 seconds before trying again */
if( tc->lastAttempt == TC_ATTEMPT_NOREACH &&
@@ -112,25 +128,59 @@ static int shouldConnect( tr_tracker_t * tc )
/* If there is quite a lot of people on this torrent, stress
the tracker a bit until we get a decent number of peers */
if( tc->hasManyPeers )
if( tc->hasManyPeers && !tr_cpIsSeeding( tor->completion ) )
{
if( tc->tor->peerCount < 5 && now > tc->dateOk + 10000 )
/* reannounce in 10 seconds if we have less than 5 peers */
if( tor->peerCount < 5 )
{
return 1;
if( now > tc->dateOk + 1000 * MAX( 10, tc->minInterval ) )
{
return 1;
}
}
if( tc->tor->peerCount < 10 && now > tc->dateOk + 20000 )
/* reannounce in 20 seconds if we have less than 10 peers */
else if( tor->peerCount < 10 )
{
return 1;
if( now > tc->dateOk + 1000 * MAX( 20, tc->minInterval ) )
{
return 1;
}
}
if( tc->tor->peerCount < 15 && now > tc->dateOk + 30000 )
/* reannounce in 30 seconds if we have less than 15 peers */
else if( tor->peerCount < 15 )
{
return 1;
if( now > tc->dateOk + 1000 * MAX( 30, tc->minInterval ) )
{
return 1;
}
}
}
return 0;
}
static int shouldScrape( tr_tracker_t * tc )
{
uint64_t now, interval;
/* scrape not supported */
if( !tc->tor->scrape[0] )
{
return 0;
}
now = tr_date();
interval = 1000 * MAX( tc->scrapeInterval, 60 );
/* scrape half as often if there is no need to */
if( !tc->scrapeNeeded && !tc->lastScrapeFailed )
{
interval *= 2;
}
return now > tc->dateScrape + interval;
}
void tr_trackerChangePort( tr_tracker_t * tc, int port )
{
tc->newPort = port;
@@ -166,20 +216,47 @@ int tr_trackerPulse( tr_tracker_t * tc )
switch( tr_httpPulse( tc->http, &data, &len ) )
{
case TR_WAIT:
return 0;
break;
case TR_ERROR:
tr_httpClose( tc->http );
tr_fdSocketClosed( tor->fdlimit, 1 );
tc->http = NULL;
killHttp( &tc->http, tor->fdlimit );
tc->dateTry = tr_date();
return 0;
break;
case TR_OK:
readAnswer( tc, data, len );
tr_httpClose( tc->http );
tc->http = NULL;
tr_fdSocketClosed( tor->fdlimit, 1 );
killHttp( &tc->http, tor->fdlimit );
break;
}
}
if( ( NULL == tc->httpScrape ) && shouldScrape( tc ) )
{
if( tr_fdSocketWillCreate( tor->fdlimit, 1 ) )
{
return 0;
}
tc->dateScrape = tr_date();
tc->httpScrape = getScrapeQuery( tc );
tr_inf( "Scrape: sent http request to %s:%d",
inf->trackerAddress, inf->trackerPort );
}
if( NULL != tc->httpScrape )
{
switch( tr_httpPulse( tc->httpScrape, &data, &len ) )
{
case TR_WAIT:
break;
case TR_ERROR:
killHttp( &tc->httpScrape, tor->fdlimit );
tc->lastScrapeFailed = 1;
break;
case TR_OK:
readScrapeAnswer( tc, data, len );
killHttp( &tc->httpScrape, tor->fdlimit );
break;
}
}
@@ -198,14 +275,9 @@ void tr_trackerStopped( tr_tracker_t * tc )
{
tr_torrent_t * tor = tc->tor;
if( NULL != tc->http )
{
/* If we are already sendy a query at the moment, we need to
reconnect */
tr_httpClose( tc->http );
tc->http = NULL;
tr_fdSocketClosed( tor->fdlimit, 1 );
}
/* If we are already sending a query at the moment, we need to
reconnect */
killHttp( &tc->http, tor->fdlimit );
tc->started = 0;
tc->completed = 0;
@@ -219,11 +291,9 @@ void tr_trackerClose( tr_tracker_t * tc )
{
tr_torrent_t * tor = tc->tor;
if( NULL != tc->http )
{
tr_httpClose( tc->http );
tr_fdSocketClosed( tor->fdlimit, 1 );
}
killHttp( &tc->http, tor->fdlimit );
killHttp( &tc->httpScrape, tor->fdlimit );
free( tc->trackerid );
free( tc );
}
@@ -232,24 +302,25 @@ static tr_http_t * getQuery( tr_tracker_t * tc )
tr_torrent_t * tor = tc->tor;
tr_info_t * inf = &tor->info;
char * event;
char * event, * trackerid, * idparam;
uint64_t left;
uint64_t down;
uint64_t up;
char * start;
char start;
int numwant = 50;
down = tor->downloadedCur;
up = tor->uploadedCur;
up = tor->uploadedCur;
if( tc->started )
{
event = "&event=started";
down = up = 0;
down = 0;
up = 0;
if( 0 < tc->newPort )
{
tc->bindPort = tc->newPort;
tc->newPort = -1;
tc->newPort = -1;
}
}
else if( tc->completed )
@@ -266,20 +337,22 @@ static tr_http_t * getQuery( tr_tracker_t * tc )
event = "";
}
if( NULL == strchr( inf->trackerAnnounce, '?' ) )
if( NULL == tc->trackerid )
{
start = "?";
trackerid = "";
idparam = "";
}
else
{
start = "&";
trackerid = tc->trackerid;
idparam = "&trackerid=";
}
left = tr_cpLeftBytes( tor->completion );
start = ( strchr( inf->trackerAnnounce, '?' ) ? '&' : '?' );
left = tr_cpLeftBytes( tor->completion );
return tr_httpClient( TR_HTTP_GET, inf->trackerAddress,
inf->trackerPort,
"%s%s"
return tr_httpClient( TR_HTTP_GET, inf->trackerAddress, inf->trackerPort,
"%s%c"
"info_hash=%s&"
"peer_id=%s&"
"port=%d&"
@@ -289,9 +362,26 @@ static tr_http_t * getQuery( tr_tracker_t * tc )
"compact=1&"
"numwant=%d&"
"key=%s"
"%s%s"
"%s",
inf->trackerAnnounce, start, tor->hashString, tc->id,
tc->bindPort, up, down, left, numwant, tor->key, event );
inf->trackerAnnounce, start, tor->escapedHashString,
tc->id, tc->bindPort, up, down, left, numwant,
tor->key, idparam, trackerid, event );
}
static tr_http_t * getScrapeQuery( tr_tracker_t * tc )
{
tr_torrent_t * tor = tc->tor;
tr_info_t * inf = &tor->info;
char start;
start = ( strchr( tor->scrape, '?' ) ? '&' : '?' );
return tr_httpClient( TR_HTTP_GET, inf->trackerAddress, inf->trackerPort,
"%s%c"
"info_hash=%s",
tor->scrape, start, tor->escapedHashString );
}
static void readAnswer( tr_tracker_t * tc, const char * data, int len )
@@ -302,8 +392,7 @@ static void readAnswer( tr_tracker_t * tc, const char * data, int len )
benc_val_t beAll;
benc_val_t * bePeers, * beFoo;
const uint8_t * body;
int bodylen;
int shouldfree;
int bodylen, shouldfree, scrapeNeeded;
tc->dateTry = tr_date();
code = tr_httpResponseCode( data, len );
@@ -331,7 +420,7 @@ static void readAnswer( tr_tracker_t * tc, const char * data, int len )
tc->lastAttempt = TC_ATTEMPT_NOREACH;
return;
}
bodylen = len - (body - (const uint8_t*)data);
bodylen = len - ( body - (const uint8_t*)data );
/* Find and load the dictionary */
shouldfree = 0;
@@ -371,40 +460,82 @@ static void readAnswer( tr_tracker_t * tc, const char * data, int len )
tor->error &= ~TR_ETRACKER;
tc->lastAttempt = TC_ATTEMPT_OK;
if( !tc->interval )
/* Get the tracker interval, force to between
10 sec and 5 mins */
beFoo = tr_bencDictFind( &beAll, "interval" );
if( !beFoo || TYPE_INT != beFoo->type )
{
/* Get the tracker interval, ignore it if it is not between
10 sec and 5 mins */
if( !( beFoo = tr_bencDictFind( &beAll, "interval" ) ) ||
!( beFoo->type & TYPE_INT ) )
{
tr_err( "Tracker: no 'interval' field" );
goto cleanup;
}
tc->interval = beFoo->val.i;
tc->interval = MIN( tc->interval, 300 );
tc->interval = MAX( 10, tc->interval );
tr_inf( "Tracker: interval = %d seconds", tc->interval );
tr_err( "Tracker: no 'interval' field" );
goto cleanup;
}
if( ( beFoo = tr_bencDictFind( &beAll, "complete" ) ) &&
( beFoo->type & TYPE_INT ) )
tc->interval = beFoo->val.i;
tr_inf( "Tracker: interval = %d seconds", tc->interval );
tc->interval = MIN( tc->interval, 300 );
tc->interval = MAX( 10, tc->interval );
/* Get the tracker minimum interval, force to between
10 sec and 5 mins */
beFoo = tr_bencDictFind( &beAll, "min interval" );
if( beFoo && TYPE_INT == beFoo->type )
{
tc->minInterval = beFoo->val.i;
tr_inf( "Tracker: min interval = %d seconds", tc->minInterval );
tc->minInterval = MIN( tc->minInterval, 300 );
tc->minInterval = MAX( 10, tc->minInterval );
if( tc->interval < tc->minInterval )
{
tc->interval = tc->minInterval;
tr_inf( "Tracker: 'interval' less than 'min interval', "
"using 'min interval'" );
}
}
else
{
tc->minInterval = 0;
tr_inf( "Tracker: no 'min interval' field" );
}
scrapeNeeded = 0;
beFoo = tr_bencDictFind( &beAll, "complete" );
if( beFoo && TYPE_INT == beFoo->type )
{
tc->seeders = beFoo->val.i;
}
if( ( beFoo = tr_bencDictFind( &beAll, "incomplete" ) ) &&
( beFoo->type & TYPE_INT ) )
else
{
scrapeNeeded = 1;
}
beFoo = tr_bencDictFind( &beAll, "incomplete" );
if( beFoo && TYPE_INT == beFoo->type )
{
tc->leechers = beFoo->val.i;
}
if( tc->seeders + tc->leechers >= 50 )
else
{
tc->hasManyPeers = 1;
scrapeNeeded = 1;
}
if( !( bePeers = tr_bencDictFind( &beAll, "peers" ) ) )
tc->scrapeNeeded = scrapeNeeded;
if( !scrapeNeeded )
{
tc->hasManyPeers = ( tc->seeders + tc->leechers >= 50 );
}
beFoo = tr_bencDictFind( &beAll, "tracker id" );
if( beFoo )
{
free( tc->trackerid );
tc->trackerid = strdup( beFoo->val.s.s );
tr_inf( "Tracker: tracker id = %s", tc->trackerid );
}
bePeers = tr_bencDictFind( &beAll, "peers" );
if( !bePeers )
{
if( tc->stopped || 0 < tc->newPort )
{
@@ -493,115 +624,104 @@ cleanup:
}
}
int tr_trackerScrape( tr_torrent_t * tor, int * seeders, int * leechers )
static void readScrapeAnswer( tr_tracker_t * tc, const char * data, int len )
{
tr_info_t * inf = &tor->info;
int code;
const uint8_t * body;
int bodylen, ii;
benc_val_t scrape, * val1, * val2;
tr_http_t * http;
const char * data, * body;
int datalen, bodylen;
int code, ii;
benc_val_t scrape, * val1, * val2;
if( !tor->scrape[0] )
code = tr_httpResponseCode( data, len );
if( 0 > code )
{
/* scrape not supported */
return 1;
/* We don't have a valid HTTP status line */
tr_inf( "Scrape: invalid HTTP status line" );
tc->lastScrapeFailed = 1;
return;
}
http = tr_httpClient( TR_HTTP_GET, inf->trackerAddress, inf->trackerPort,
"%s?info_hash=%s", tor->scrape, tor->hashString );
data = NULL;
while( NULL == data )
{
switch( tr_httpPulse( http, &data, &datalen ) )
{
case TR_WAIT:
break;
case TR_ERROR:
tr_httpClose( http );
return 1;
case TR_OK:
if( NULL == data || 0 >= datalen )
{
tr_httpClose( http );
return 1;
}
break;
}
tr_wait( 10 );
}
code = tr_httpResponseCode( data, datalen );
if( !TR_HTTP_STATUS_OK( code ) )
{
tr_httpClose( http );
return 1;
/* we didn't get a 2xx status code */
tr_err( "Scrape: invalid HTTP status code: %i", code );
tc->lastScrapeFailed = 1;
return;
}
body = tr_httpParse( data, datalen , NULL );
/* find the end of the http headers */
body = (uint8_t *) tr_httpParse( data, len, NULL );
if( NULL == body )
{
tr_httpClose( http );
return 1;
tr_err( "Scrape: could not find end of HTTP headers" );
tc->lastScrapeFailed = 1;
return;
}
bodylen = datalen - ( body - data );
for( ii = 0; ii < bodylen - 8; ii++ )
tc->lastScrapeFailed = 0;
bodylen = len - ( body - (const uint8_t*)data );
for( ii = 0; ii < bodylen; ii++ )
{
if( !memcmp( body + ii, "d5:files", 8 ) )
if( !tr_bencLoad( body + ii, bodylen - ii, &scrape, NULL ) )
{
break;
}
}
if( ii >= bodylen - 8 )
if( ii >= bodylen )
{
tr_httpClose( http );
return 1;
}
if( tr_bencLoad( body + ii, bodylen - ii, &scrape, NULL ) )
{
tr_httpClose( http );
return 1;
return;
}
val1 = tr_bencDictFind( &scrape, "files" );
if( !val1 )
{
tr_bencFree( &scrape );
tr_httpClose( http );
return 1;
return;
}
val1 = &val1->val.l.vals[1];
if( !val1 )
{
tr_bencFree( &scrape );
tr_httpClose( http );
return 1;
return;
}
val2 = tr_bencDictFind( val1, "complete" );
if( !val2 )
{
tr_bencFree( &scrape );
tr_httpClose( http );
return 1;
return;
}
*seeders = val2->val.i;
tc->seeders = val2->val.i;
val2 = tr_bencDictFind( val1, "incomplete" );
if( !val2 )
{
tr_bencFree( &scrape );
tr_httpClose( http );
return 1;
return;
}
*leechers = val2->val.i;
tc->leechers = val2->val.i;
val2 = tr_bencDictFind( val1, "downloaded" );
if( !val2 )
{
tr_bencFree( &scrape );
return;
}
tc->complete = val2->val.i;
val2 = tr_bencDictFind( val1, "flags" );
if( val2 )
{
val2 = tr_bencDictFind( val2, "min_request_interval" );
if( val2 )
{
tc->scrapeInterval = val2->val.i;
}
}
tc->hasManyPeers = ( tc->seeders + tc->leechers >= 50 );
tr_bencFree( &scrape );
tr_httpClose( http );
return 0;
}
int tr_trackerSeeders( tr_tracker_t * tc )
@@ -621,3 +741,71 @@ int tr_trackerLeechers( tr_tracker_t * tc )
}
return tc->leechers;
}
int tr_trackerDownloaded( tr_tracker_t * tc )
{
if( !tc )
{
return -1;
}
return tc->complete;
}
/* Blocking version */
int tr_trackerScrape( tr_torrent_t * tor, int * s, int * l, int * d )
{
tr_tracker_t * tc;
tr_http_t * http;
const char * data;
int len;
int ret;
if( !tor->scrape[0] )
{
return 1;
}
tc = tr_trackerInit( tor );
http = getScrapeQuery( tc );
for( data = NULL; !data; tr_wait( 10 ) )
{
switch( tr_httpPulse( http, &data, &len ) )
{
case TR_WAIT:
break;
case TR_ERROR:
goto scrapeDone;
case TR_OK:
readScrapeAnswer( tc, data, len );
goto scrapeDone;
}
}
scrapeDone:
tr_httpClose( http );
ret = 1;
if( tc->seeders > -1 && tc->leechers > -1 && tc->complete > -1 )
{
*s = tc->seeders;
*l = tc->leechers;
*d = tc->complete;
ret = 0;
}
tr_trackerClose( tc );
return ret;
}
static void killHttp( tr_http_t ** http, tr_fd_t * fdlimit )
{
if( NULL != *http )
{
tr_httpClose( *http );
tr_fdSocketClosed( fdlimit, 1 );
*http = NULL;
}
}

View File

@@ -37,17 +37,31 @@ void tr_trackerClose ( tr_tracker_t * );
/***********************************************************************
* tr_trackerSeeders
***********************************************************************
* Looks for the seeders/leechers as returned by the tracker.
* Looks for the seeders as returned by the tracker.
**********************************************************************/
int tr_trackerSeeders ( tr_tracker_t * );
int tr_trackerSeeders ( tr_tracker_t * );
/***********************************************************************
* tr_trackerLeechers
***********************************************************************
* Looks for the seeders/leechers as returned by the tracker.
* Looks for the leechers as returned by the tracker.
**********************************************************************/
int tr_trackerLeechers ( tr_tracker_t * );
int tr_trackerLeechers ( tr_tracker_t * );
int tr_trackerScrape ( tr_torrent_t *, int *, int * );
/***********************************************************************
* tr_trackerDownloaded
***********************************************************************
* Looks for number of completed downloads as returned by the tracker
* (from scrape).
**********************************************************************/
int tr_trackerDownloaded( tr_tracker_t * tc );
/***********************************************************************
* tr_trackerScrape
***********************************************************************
* Attempt a blocking scrape and return the seeders, leechers, and
* completed downloads if successful.
**********************************************************************/
int tr_trackerScrape( tr_torrent_t * tor, int * s, int * l, int * d );
#endif

View File

@@ -322,7 +322,7 @@ static tr_torrent_t * torrentRealInit( tr_handle_t * h, tr_torrent_t * tor,
/* Escaped info hash for HTTP queries */
for( i = 0; i < SHA_DIGEST_LENGTH; i++ )
{
sprintf( &tor->hashString[3*i], "%%%02x", inf->hash[i] );
sprintf( &tor->escapedHashString[3*i], "%%%02x", inf->hash[i] );
}
/* Block size: usually 16 ko, or less if we have to */
@@ -366,11 +366,11 @@ tr_info_t * tr_torrentInfo( tr_torrent_t * tor )
}
/***********************************************************************
* tr_torrentScrape
* tr_torrentScrape
**********************************************************************/
int tr_torrentScrape( tr_torrent_t * tor, int * s, int * l )
int tr_torrentScrape( tr_torrent_t * tor, int * s, int * l, int * d )
{
return tr_trackerScrape( tor, s, l );
return tr_trackerScrape( tor, s, l, d );
}
void tr_torrentSetFolder( tr_torrent_t * tor, const char * path )
@@ -539,6 +539,7 @@ tr_stat_t * tr_torrentStat( tr_torrent_t * tor )
s->seeders = tr_trackerSeeders(tor->tracker);
s->leechers = tr_trackerLeechers(tor->tracker);
s->completedFromTracker = tr_trackerDownloaded(tor->tracker);
s->swarmspeed = tr_rcRate( tor->swarmspeed );

View File

@@ -210,7 +210,7 @@ tr_info_t * tr_torrentInfo( tr_torrent_t * );
* replied with some error. tr_torrentScrape may block up to 20 seconds
* before returning.
**********************************************************************/
int tr_torrentScrape( tr_torrent_t *, int * s, int * l );
int tr_torrentScrape( tr_torrent_t *, int * s, int * l, int * d );
/***********************************************************************
* tr_torrentStart
@@ -368,6 +368,7 @@ struct tr_stat_s
int peersDownloading;
int seeders;
int leechers;
int completedFromTracker;
uint64_t downloaded;
uint64_t uploaded;

View File

@@ -19,6 +19,7 @@
LANGUAGE = ObjC;
OUTLETS = {
fCommentView = NSTextView;
fCompletedFromTrackerField = NSTextField;
fConnectedPeersField = NSTextField;
fCreatorField = NSTextField;
fDataLocationField = NSTextField;

View File

@@ -49,7 +49,7 @@
IBOutlet NSTableView * fPeerTable;
IBOutlet NSTextField * fSeedersField, * fLeechersField, * fConnectedPeersField,
* fDownloadingFromField, * fUploadingToField;
* fDownloadingFromField, * fUploadingToField, * fCompletedFromTrackerField;
IBOutlet NSTableView * fFileTable;
IBOutlet NSTextField * fFileTableStatusField;

View File

@@ -181,6 +181,7 @@
[fSeedersField setStringValue: @""];
[fLeechersField setStringValue: @""];
[fCompletedFromTrackerField setStringValue: @""];
[fConnectedPeersField setStringValue: @""];
[fDownloadingFromField setStringValue: @""];
[fUploadingToField setStringValue: @""];
@@ -326,9 +327,10 @@
return;
Torrent * torrent = [fTorrents objectAtIndex: 0];
int seeders = [torrent seeders], leechers = [torrent leechers];
int seeders = [torrent seeders], leechers = [torrent leechers], downloaded = [torrent completedFromTracker];
[fSeedersField setStringValue: seeders < 0 ? @"" : [NSString stringWithInt: seeders]];
[fLeechersField setStringValue: leechers < 0 ? @"" : [NSString stringWithInt: leechers]];
[fCompletedFromTrackerField setStringValue: downloaded < 0 ? @"" : [NSString stringWithInt: downloaded]];
BOOL active = [torrent isActive];
[fConnectedPeersField setStringValue: active ? [NSString stringWithFormat: NSLocalizedString(@"%d (%d incoming)",

View File

@@ -142,6 +142,7 @@
- (int) seeders;
- (int) leechers;
- (int) completedFromTracker;
- (int) totalPeers;
- (int) totalPeersIncoming;
- (int) totalPeersOutgoing;

View File

@@ -725,7 +725,7 @@ static uint32_t kRed = BE(0xFF6450FF), //255, 100, 80
- (NSDate *) dateCreated
{
int date = fInfo->dateCreated;
return date > 0 ? [NSDate dateWithTimeIntervalSince1970: fInfo->dateCreated] : nil;
return date > 0 ? [NSDate dateWithTimeIntervalSince1970: date] : nil;
}
- (int) pieceSize
@@ -883,6 +883,11 @@ static uint32_t kRed = BE(0xFF6450FF), //255, 100, 80
return fStat->leechers;
}
- (int) completedFromTracker
{
return fStat->completedFromTracker;
}
- (int) totalPeers
{
return fStat->peersTotal;