[Author Prev][Author Next][Thread Prev][Thread Next][Author Index][Thread Index]
[tor-commits] [metrics-tasks/master] Remove outdated censorship detector code (#2718).
commit 4426cf9cfa1d8965057c3fe34184c5ce459fa390
Author: Karsten Loesing <karsten.loesing@xxxxxxx>
Date: Tue Aug 9 15:52:38 2011 +0200
Remove outdated censorship detector code (#2718).
---
task-2718/README | 27 --------------------
task-2718/convert-dirreqs-sql.R | 8 ------
task-2718/detect-censorship.R | 52 ---------------------------------------
3 files changed, 0 insertions(+), 87 deletions(-)
diff --git a/task-2718/README b/task-2718/README
deleted file mode 100644
index 8c60847..0000000
--- a/task-2718/README
+++ /dev/null
@@ -1,27 +0,0 @@
-Here's how you run the censorship detector prototype:
-
- $ wget https://metrics.torproject.org/csv/direct-users.csv
- $ R --slave -f detect-censorship.R
-
--------------------------------------------------------------------------
-
-Extracting raw directory requests from the metrics-web database:
-
-- Export dirreq_stats table from the metrics-web database via psql:
-
- # \f ','
- # \a
- # \t
- # \o dirreqs-sql.csv
- # SELECT * FROM dirreq_stats ORDER BY statsend;
- # \o
- # \t
- # \a
-
-- Transform the huge (!) CSV file (104M) from long to wide format. Note
- that this takes a while:
-
- $ R --slave -f convert-dirreqs-sql.R
-
-- The result is in dirreqs.csv (8.8M).
-
diff --git a/task-2718/convert-dirreqs-sql.R b/task-2718/convert-dirreqs-sql.R
deleted file mode 100644
index e330307..0000000
--- a/task-2718/convert-dirreqs-sql.R
+++ /dev/null
@@ -1,8 +0,0 @@
-library(ggplot2)
-data <- read.csv("dirreqs-sql.csv", header = FALSE)
-data <- data.frame(fingerprint = data$V1, statsend = data$V2,
- seconds = data$V3, country = data$V4, requests = data$V5)
-data <- cast(data, fingerprint + statsend + seconds ~ country,
- value = "requests")
-write.csv(data, file = "dirreqs.csv", quote = FALSE, row.names = FALSE)
-
diff --git a/task-2718/detect-censorship.R b/task-2718/detect-censorship.R
deleted file mode 100644
index f27e863..0000000
--- a/task-2718/detect-censorship.R
+++ /dev/null
@@ -1,52 +0,0 @@
-# Tor Censorship Detector
-# Usage: R --slave < detect-censorship.R
-
-# Read CSV file containing daily user number estimates.
-direct <- read.csv("direct-users.csv")
-
-# Start plotting everything to a single PDF (with multiple pages).
-pdf("detect-censorship.pdf")
-
-# Convert the column containing daily Iranian users to a time series
-# object, starting on the 263th day of 2009 with a frequency of 365 days.
-# We're probably off by a day or two, but this should be fine for now.
-all <- ts(direct$ir, start = c(2009, 263), frequency = 365)
-
-# Uncomment to print the time series values.
-#print(all)
-
-# Let's try our approach for the last 365 days to see if we detect any
-# blocking in that time period. In the final version of this script, we'd
-# only have a single run with i = 1.
-for (i in 365:1) {
- idx <- length(direct$date) - i
-
- # Convert the daily Iranian users until i days in the past to a time
- # series object.
- x <- ts(direct$ir[1:idx], start = c(2009, 263), frequency = 365)
-
- # Apply an ARIMA(1, 0, 1) model to the time series.
- x.fit = arima(x, order = c(1, 0, 1))
-
- # Predict 10 dates ahead.
- x.fore=predict(x.fit, n.ahead=10)
-
- # Calculate a lower bound. Here we use the predicted value minus three
- # standard errors.
- L = x.fore$pred - 3*x.fore$se
-
- # If the observed daily user number is lower than our predicted lower
- # bound, plot the data and lower bound.
- if (direct$ir[idx + 1] < L[1]) {
-
- # Plot the full time series.
- ts.plot(all)
-
- # Add a line for the ten predicted values.
- lines(L, col = "red", lwd = 2)
- }
-}
-
-# Close the PDF device.
-dev.off()
-
_______________________________________________
tor-commits mailing list
tor-commits@xxxxxxxxxxxxxxxxxxxx
https://lists.torproject.org/cgi-bin/mailman/listinfo/tor-commits