[Author Prev][Author Next][Thread Prev][Thread Next][Author Index][Thread Index]
[tor-commits] [depictor/master] Unify files on tabs, not spaces
commit d53048d3d5cc26cda821acbd366903e90e7ce25f
Author: Tom Ritter <tom@xxxxxxxxx>
Date: Sun Mar 8 13:46:51 2015 -0500
Unify files on tabs, not spaces
---
write_website.py | 161 +++++++++++++++++++++++++++---------------------------
1 file changed, 80 insertions(+), 81 deletions(-)
diff --git a/write_website.py b/write_website.py
index 415c7ea..9c5463e 100755
--- a/write_website.py
+++ b/write_website.py
@@ -27,120 +27,119 @@ from website import WebsiteWriter
DIRECTORY_AUTHORITIES = stem.descriptor.remote.get_authorities()
CONFIG = stem.util.conf.config_dict('consensus', {
- 'ignored_authorities': [],
- 'bandwidth_authorities': [],
- 'known_params': [],
+ 'ignored_authorities': [],
+ 'bandwidth_authorities': [],
+ 'known_params': [],
})
downloader = stem.descriptor.remote.DescriptorDownloader(
- timeout = 60,
- fall_back_to_authority = False,
- document_handler = stem.descriptor.DocumentHandler.DOCUMENT,
+ timeout = 60,
+ fall_back_to_authority = False,
+ document_handler = stem.descriptor.DocumentHandler.DOCUMENT,
)
@lru_cache()
def directory_authorities():
- return dict((k, v) for (k, v) in DIRECTORY_AUTHORITIES.items() if k not in CONFIG['ignored_authorities'])
+ return dict((k, v) for (k, v) in DIRECTORY_AUTHORITIES.items() if k not in CONFIG['ignored_authorities'])
def main():
- # loads configuration data
-
- config = stem.util.conf.get_config("consensus")
- config.load(os.path.join(os.path.dirname(__file__), 'data', 'consensus.cfg'))
-
- consensuses, consensus_fetching_issues, consensus_fetching_runtimes = get_consensuses()
- votes, vote_fetching_issues, vote_fetching_runtimes = get_votes()
-
- # updates the download statistics file
- f = open(os.path.join(os.path.dirname(__file__), 'out', 'download-stats.csv'), 'a')
- for ds in consensus_fetching_runtimes:
- f.write("%s,%i,%i\n" % (ds, time.time() * 1000, int(consensus_fetching_runtimes[ds] * 1000)))
- f.close()
-
- # great for debugging
- #import pickle
- #pickle.dump(consensuses, open('consensus.p', 'wb'))
- #pickle.dump(votes, open('votes.p', 'wb'))
-
- # produces the website
- w = WebsiteWriter()
- w.set_consensuses(consensuses)
- w.set_votes(votes)
- w.set_config(CONFIG)
- w.write_website(os.path.join(os.path.dirname(__file__), 'out', 'consensus-health.html'), True)
- w.write_website(os.path.join(os.path.dirname(__file__), 'out', 'index.html'), False)
-
- archived = os.path.join(os.path.dirname(__file__), 'out', \
- 'consensus-health-' + w.get_consensus_time().strftime("%Y-%m-%d-%H-%M") + '.html')
- subprocess.call(["cp", os.path.join(os.path.dirname(__file__), 'out', 'consensus-health.html'), archived])
- subprocess.call(["gzip", "-9", archived])
+ # loads configuration data
+ config = stem.util.conf.get_config("consensus")
+ config.load(os.path.join(os.path.dirname(__file__), 'data', 'consensus.cfg'))
+
+ consensuses, consensus_fetching_issues, consensus_fetching_runtimes = get_consensuses()
+ votes, vote_fetching_issues, vote_fetching_runtimes = get_votes()
+
+ # updates the download statistics file
+ f = open(os.path.join(os.path.dirname(__file__), 'out', 'download-stats.csv'), 'a')
+ for ds in consensus_fetching_runtimes:
+ f.write("%s,%i,%i\n" % (ds, time.time() * 1000, int(consensus_fetching_runtimes[ds] * 1000)))
+ f.close()
+
+ # great for debugging
+ #import pickle
+ #pickle.dump(consensuses, open('consensus.p', 'wb'))
+ #pickle.dump(votes, open('votes.p', 'wb'))
+
+ # produces the website
+ w = WebsiteWriter()
+ w.set_consensuses(consensuses)
+ w.set_votes(votes)
+ w.set_config(CONFIG)
+ w.write_website(os.path.join(os.path.dirname(__file__), 'out', 'consensus-health.html'), True)
+ w.write_website(os.path.join(os.path.dirname(__file__), 'out', 'index.html'), False)
+
+ archived = os.path.join(os.path.dirname(__file__), 'out', \
+ 'consensus-health-' + w.get_consensus_time().strftime("%Y-%m-%d-%H-%M") + '.html')
+ subprocess.call(["cp", os.path.join(os.path.dirname(__file__), 'out', 'consensus-health.html'), archived])
+ subprocess.call(["gzip", "-9", archived])
def get_consensuses():
- """
- Provides a mapping of directory authority nicknames to their present consensus.
+ """
+ Provides a mapping of directory authority nicknames to their present consensus.
- :returns: tuple of the form ({authority => consensus}, issues, runtimes)
- """
+ :returns: tuple of the form ({authority => consensus}, issues, runtimes)
+ """
- return _get_documents('consensus', '/tor/status-vote/current/consensus')
+ return _get_documents('consensus', '/tor/status-vote/current/consensus')
def get_votes():
- """
- Provides a mapping of directory authority nicknames to their present vote.
+ """
+ Provides a mapping of directory authority nicknames to their present vote.
- :returns: tuple of the form ({authority => vote}, issues, runtimes)
- """
+ :returns: tuple of the form ({authority => vote}, issues, runtimes)
+ """
- return _get_documents('vote', '/tor/status-vote/current/authority')
+ return _get_documents('vote', '/tor/status-vote/current/authority')
def _get_documents(label, resource):
- queries, documents, issues, runtimes = {}, {}, [], {}
+ queries, documents, issues, runtimes = {}, {}, [], {}
- for authority in directory_authorities().values():
- if authority.v3ident is None:
- continue # not a voting authority
+ for authority in directory_authorities().values():
+ if authority.v3ident is None:
+ continue # not a voting authority
- queries[authority.nickname] = downloader.query(
- resource,
- endpoints = [(authority.address, authority.dir_port)],
- default_params = False,
- )
+ queries[authority.nickname] = downloader.query(
+ resource,
+ endpoints = [(authority.address, authority.dir_port)],
+ default_params = False,
+ )
- for authority, query in queries.items():
- try:
- result = query.run()
- documents[authority] = result[0]
- runtimes[authority] = query.runtime
- except Exception, exc:
- if label == 'vote':
- # try to download the vote via the other authorities
+ for authority, query in queries.items():
+ try:
+ result = query.run()
+ documents[authority] = result[0]
+ runtimes[authority] = query.runtime
+ except Exception, exc:
+ if label == 'vote':
+ # try to download the vote via the other authorities
- v3ident = directory_authorities()[authority].v3ident
+ v3ident = directory_authorities()[authority].v3ident
- query = downloader.query(
- '/tor/status-vote/current/%s' % v3ident,
- default_params = False,
- )
+ query = downloader.query(
+ '/tor/status-vote/current/%s' % v3ident,
+ default_params = False,
+ )
- query.run(True)
+ query.run(True)
- if not query.error:
- documents[authority] = list(query)[0]
- continue
+ if not query.error:
+ documents[authority] = list(query)[0]
+ continue
- issues.append(('AUTHORITY_UNAVAILABLE', label, authority, query.download_url, exc))
+ issues.append(('AUTHORITY_UNAVAILABLE', label, authority, query.download_url, exc))
- return documents, issues, runtimes
+ return documents, issues, runtimes
if __name__ == '__main__':
- try:
- main()
- except:
- msg = "%s failed with:\n\n%s" % (sys.argv[0], traceback.format_exc())
- print "Error: %s" % msg
+ try:
+ main()
+ except:
+ msg = "%s failed with:\n\n%s" % (sys.argv[0], traceback.format_exc())
+ print "Error: %s" % msg
_______________________________________________
tor-commits mailing list
tor-commits@xxxxxxxxxxxxxxxxxxxx
https://lists.torproject.org/cgi-bin/mailman/listinfo/tor-commits