[Author Prev][Author Next][Thread Prev][Thread Next][Author Index][Thread Index]
[tor-commits] [torflow/master] Fix some confusion with bandwidth recording events.
commit 56d397f203f4e3a1593817e0e4cb585e516ad3af
Author: Mike Perry <mikeperry-git@xxxxxxxxxx>
Date: Sun Oct 23 02:51:11 2011 -0700
Fix some confusion with bandwidth recording events.
---
bwauth-spec.txt | 26 ++++++++++++--------------
1 files changed, 12 insertions(+), 14 deletions(-)
diff --git a/bwauth-spec.txt b/bwauth-spec.txt
index 9c4dc75..8c638cd 100644
--- a/bwauth-spec.txt
+++ b/bwauth-spec.txt
@@ -8,7 +8,8 @@
Karsten Loesing
- Mike Perry
+ Mike Perry
+ Aaron Gibson
0. Preliminaries
@@ -155,7 +156,7 @@
- Nodes are selected uniformly among those with the lowest measurement
count for the current slice. Otherwise, there is no preference for
- relays, e.g., based on bandwidth.
+ relays.
- Relays in the paths must come from different /16 subnets.
@@ -202,13 +203,16 @@
Unfinished downloads are aborted after 30 minutes.
- For each download, the bandwidth scanners process STREAM_BW events with
- a StreamListener (SQLSupport.py). The throughput for each stream is
- defined as the ratio of total read bytes over the time delta between the
- stream start timestamp and the newest STREAM_BW event received timestamp:
+ For each download, the bandwidth scanners process STREAM and STREAM_BW events
+ with a StreamListener (in TorCtl/SQLSupport.py). The throughput for each
+ stream is defined as the ratio of total read bytes over the time delta between
+ the STREAM NEW timestamp and the STREAM CLOSED event received timestamp:
- (stream read bytes / (event received timestamp - stream start timestamp)
+ bandwidth = (STREAM_BW bytes / (CLOSED timestamp - NEW timestamp)
+ We store both read and write bandwidths in the SQL tables, but only use
+ the read bytes for results.
+
1.6. Writing measurement results
Once a bandwidth scanner has completed a slice of relays, it writes the
@@ -259,7 +263,7 @@
The filt_bw field is computed similarly, but only the streams equal to
or greater than the strm_bw are counted in order to filter very slow
- streams.
+ streams due to slow node pairings.
The nickname field is entirely informational and may change between
measurements.
@@ -268,12 +272,6 @@
filtered stream bandwidth, and non-negative stream bandwidth are
included in the output file.
-# Starting to count slices at 0 whenever we start at the lower end of our
-# percentile range seems error-prone. What if the number of slices
-# changes while we're only half through with all slices? Isn't there a
-# potential from overlooking results? Or do we not care about the slice
-# number when aggregating results? -KL
-
2. Aggregating scanner results
Once per hour (via cron), the bandwidth scanner results are aggregated
_______________________________________________
tor-commits mailing list
tor-commits@xxxxxxxxxxxxxxxxxxxx
https://lists.torproject.org/cgi-bin/mailman/listinfo/tor-commits