This commit is contained in:
unknown 2016-07-09 21:38:49 -07:00
parent 5e88950156
commit ef15e1a644
6 changed files with 367 additions and 116 deletions

View file

@ -7,7 +7,7 @@
<body> <body>
<div id="control_panel"> <div id="control_panel">
<input type="text" id="subreddit_field" placeholder="learnpython"> <input type="text" id="subreddit_field" placeholder="/r/learnpython">
<button id="start_button" onclick="start()">Start</button> <button id="start_button" onclick="start()">Start</button>
<a id="browser_link"></a> <a id="browser_link"></a>
<button id="clear_button" onclick="clear_workspace()">Clear workspace</button> <button id="clear_button" onclick="clear_workspace()">Clear workspace</button>
@ -32,14 +32,14 @@ body
margin-right: 10%; margin-right: 10%;
padding: 5px; padding: 5px;
} }
#control_panel #control_panel
{ {
background-color: #284142; background-color: #284142;
padding: 5px; padding: 5px;
} }
#workspace
{
}
a a
{ {
@ -53,11 +53,26 @@ a:hover
text-decoration: underline; text-decoration: underline;
} }
.delete_button
{
position: absolute;
top: 0px;
right: 0px;
height: 100%;
width: 5%;
border: 0px;
background-color: #b53030;
}
.submission .submission
{ {
position: relative;
padding: 10px; padding: 10px;
padding-top: 20px; padding-top: 20px;
padding-bottom: 20px; padding-bottom: 20px;
padding-right: 6%; /*for the delete button*/
margin: 10px; margin: 10px;
margin-top: 20px; margin-top: 20px;
@ -76,8 +91,21 @@ a:hover
*/ */
var CHECK_DELAY = 30 * 1000; var CHECK_DELAY = 30 * 1000;
var CONTROL_PANEL = document.getElementById("control_panel");
CONTROL_PANEL.default_color = CONTROL_PANEL.style.backgroundColor;
var WORKSPACE = document.getElementById("workspace"); var WORKSPACE = document.getElementById("workspace");
var id_cache = [];
var id_cache_size = 20;
var first_loop = true;
var unread_notification_count = 0;
var subreddit = "";
var check_timer = null;
var page_focused_cached;
var HTTPClient = function() var HTTPClient = function()
{ {
/* Thanks ttgagne http://stackoverflow.com/a/22076667 */ /* Thanks ttgagne http://stackoverflow.com/a/22076667 */
@ -86,10 +114,20 @@ var HTTPClient = function()
{ {
var request = new XMLHttpRequest(); var request = new XMLHttpRequest();
request.onreadystatechange = function() request.onreadystatechange = function()
{ {
if (request.readyState == 4 && request.status == 200) // console.log(request.readyState);
// console.log(request.status);
if (request.readyState == 4)
{ {
callback(request.responseText); if (request.status == 200)
{
CONTROL_PANEL.style.backgroundColor = CONTROL_PANEL.default_color;
callback(request.responseText);
}
else
{
CONTROL_PANEL.style.backgroundColor = "#f00";
}
} }
} }
request.open("GET", url, asynchronous); request.open("GET", url, asynchronous);
@ -98,21 +136,21 @@ var HTTPClient = function()
} }
} }
function apply_to_page(text) function apply_to_page(response_json)
{ {
var j = JSON.parse(text); var j = JSON.parse(response_json);
var submissions = j["data"]["children"]; var submissions = j["data"]["children"];
submissions.reverse(); // newest last submissions.reverse(); // newest last
var new_items = 0; var new_items = 0;
for (var index = 0; index < submissions.length; index += 1) for (var index = 0; index < submissions.length; index += 1)
{ {
var submission = submissions[index]["data"]; var submission = submissions[index]["data"];
if (done_ids.has(submission["id"])) if (id_cache.includes(submission["id"]))
{ {
continue; continue;
} }
done_ids.add(submission["id"]); id_cache.push(submission["id"]);
if (first_loop) if (first_loop)
{ {
@ -128,11 +166,21 @@ function apply_to_page(text)
anchor.href = "https://reddit.com/r/" + submission["subreddit"] + "/comments/" + submission["id"]; anchor.href = "https://reddit.com/r/" + submission["subreddit"] + "/comments/" + submission["id"];
anchor.target = "_blank"; anchor.target = "_blank";
var delete_button = document.createElement("button")
delete_button.className = "delete_button";
delete_button.div = div;
delete_button.innerHTML = "X";
delete_button.onclick = function()
{
this.div.parentElement.removeChild(this.div);
}
var timestamp = document.createElement("span"); var timestamp = document.createElement("span");
var submission_time = new Date(submission["created_utc"]) var submission_time = new Date(submission["created_utc"])
timestamp.innerHTML = "" + submission_time.getHours() + ":" + submission_time.getMinutes(); timestamp.innerHTML = "" + submission_time.getHours() + ":" + submission_time.getMinutes();
div.appendChild(anchor); div.appendChild(anchor);
div.appendChild(delete_button);
//WORKSPACE.insertBefore(div, WORKSPACE.firstChild); //WORKSPACE.insertBefore(div, WORKSPACE.firstChild);
WORKSPACE.appendChild(div); WORKSPACE.appendChild(div);
} }
@ -142,6 +190,12 @@ function apply_to_page(text)
unread_notification_count += new_items; unread_notification_count += new_items;
update_title(); update_title();
} }
while (id_cache.length < id_cache_size)
{
id_cache.shift();
}
first_loop = false; first_loop = false;
} }
@ -160,7 +214,7 @@ function check_once()
console.log("no subreddit"); console.log("no subreddit");
return; return;
} }
var url = "https://api.reddit.com/r/" + subreddit + "/new.json"; var url = "https://api.reddit.com" + subreddit + "/new.json";
session.get(url, apply_to_page); session.get(url, apply_to_page);
} }
@ -196,6 +250,18 @@ function page_focused_fresh()
return page_focused_cached; return page_focused_cached;
} }
function sort_submission_comparator(submission_1, submission_2)
{
created_1 = submission_1["created_utc"];
created_2 = submission_2["created_utc"];
if (created_1 < created_2)
{return -1;}
if (created_1 > created_2)
{return 1;}
return 0;
}
function start() function start()
{ {
console.log("start"); console.log("start");
@ -203,10 +269,14 @@ function start()
clear_workspace(); clear_workspace();
var field = document.getElementById("subreddit_field"); var field = document.getElementById("subreddit_field");
var text = field.value; var text = field.value;
text = text.replace("/r/", "").replace("r/", ""); text = text.replace("/u/", "/user/");
if (text.indexOf("/") == -1)
{
text = "/r/" + text;
}
subreddit = text; subreddit = text;
var link = document.getElementById("browser_link"); var link = document.getElementById("browser_link");
var url = "https://reddit.com/r/" + subreddit + "/new"; var url = "https://reddit.com" + subreddit + "/new";
link.href = url; link.href = url;
link.innerHTML = url; link.innerHTML = url;
update_title(); update_title();
@ -242,14 +312,7 @@ function visibility_property()
return null; return null;
} }
var done_ids = new Set();
var first_loop = true;
var unread_notification_count = 0;
var subreddit = "";
var check_timer = null;
var page_focused_cached;
page_focused_fresh(); page_focused_fresh();
var my_visibility_property = visibility_property(); var my_visibility_property = visibility_property();

View file

@ -1,6 +1,9 @@
Open Dir DL Open Dir DL
=========== ===========
- 2016 07 08
- Fixed bug in which trees wouldn't generate on server:port urls.
- 2016 07 04 - 2016 07 04
- Added new argparse command "tree" - Added new argparse command "tree"

View file

@ -80,9 +80,14 @@ measure:
flags: flags:
-f | --fullscan: -f | --fullscan:
When included, perform HEAD requests when a file's size is not known. When included, perform HEAD requests on all files to update their size.
If this flag is not included, and some file's size is unkown, you will
receive a printed note. -n | --new_only:
When included, perform HEAD requests only on files that haven't gotten one
yet.
If a file's size is not known by the time this operation completes, you will
receive a printed note.
tree: tree:
Print the file / folder tree. Print the file / folder tree.
@ -120,6 +125,7 @@ import requests
import shutil import shutil
import sqlite3 import sqlite3
## ~tkinter ## ~tkinter
import traceback
import urllib.parse import urllib.parse
FILENAME_BADCHARS = '/\\:*?"<>|' FILENAME_BADCHARS = '/\\:*?"<>|'
@ -166,6 +172,7 @@ SKIPPABLE_FILETYPES = [
'.tar', '.tar',
'.ttf', '.ttf',
'.txt', '.txt',
'.wav',
'.webm', '.webm',
'.wma', '.wma',
'.zip', '.zip',
@ -229,6 +236,11 @@ SQL_CONTENT_TYPE = 3
SQL_DO_DOWNLOAD = 4 SQL_DO_DOWNLOAD = 4
UNMEASURED_WARNING = '''
Note: %d files do not have a stored Content-Length.
Run `measure` with `-f`|`--fullscan` or `-n`|`--new_only` to HEAD request those files.
'''.strip()
## DOWNLOADER ###################################################################################### ## DOWNLOADER ######################################################################################
## ## ## ##
class Downloader: class Downloader:
@ -293,6 +305,12 @@ class Generic:
setattr(self, kwarg, kwargs[kwarg]) setattr(self, kwarg, kwargs[kwarg])
class TreeExistingChild(Exception):
pass
class TreeInvalidIdentifier(Exception):
pass
class TreeNode: class TreeNode:
def __init__(self, identifier, data, parent=None): def __init__(self, identifier, data, parent=None):
assert isinstance(identifier, str) assert isinstance(identifier, str)
@ -329,9 +347,9 @@ class TreeNode:
def check_child_availability(self, identifier): def check_child_availability(self, identifier):
if ':' in identifier: if ':' in identifier:
raise Exception('Only roots may have a colon') raise TreeInvalidIdentifier('Only roots may have a colon')
if identifier in self.children: if identifier in self.children:
raise Exception('Node %s already has child %s' % (self.identifier, identifier)) raise TreeExistingChild('Node %s already has child %s' % (self.identifier, identifier))
def detach(self): def detach(self):
del self.parent.children[self.identifier] del self.parent.children[self.identifier]
@ -363,8 +381,11 @@ class TreeNode:
for node in self.walk(customsort): for node in self.walk(customsort):
print(node.abspath()) print(node.abspath())
def sorted_children(self): def sorted_children(self, customsort=None):
keys = sorted(self.children.keys()) if customsort:
keys = sorted(self.children.keys(), key=customsort)
else:
keys = sorted(self.children.keys())
for key in keys: for key in keys:
yield (key, self.children[key]) yield (key, self.children[key])
@ -388,7 +409,7 @@ class Walker:
if databasename is None or databasename == "": if databasename is None or databasename == "":
self.domain = url_to_filepath(walkurl)['root'] self.domain = url_to_filepath(walkurl)['root']
databasename = self.domain + '.db' databasename = self.domain + '.db'
databasename = databasename.replace(':', '') databasename = databasename.replace(':', '#')
self.databasename = databasename self.databasename = databasename
safeprint('Opening %s' % self.databasename) safeprint('Opening %s' % self.databasename)
@ -451,6 +472,10 @@ class Walker:
else: else:
url = urllib.parse.urljoin(self.walkurl, url) url = urllib.parse.urljoin(self.walkurl, url)
if url in self.seen_directories:
# We already picked this up at some point
return
if not url.startswith(self.walkurl): if not url.startswith(self.walkurl):
# Don't follow external links or parent directory. # Don't follow external links or parent directory.
safeprint('Skipping "%s" due to external url.' % url) safeprint('Skipping "%s" due to external url.' % url)
@ -480,11 +505,14 @@ class Walker:
return return
raise raise
content_type = head.headers.get('Content-Type', '?') content_type = head.headers.get('Content-Type', '?')
#print(content_type)
if content_type.startswith('text/html') and head.url.endswith('/'): if content_type.startswith('text/html'):# and head.url.endswith('/'):
# This is an index page, so extract links and queue them. # This is an index page, so extract links and queue them.
response = do_get(url) response = do_get(url)
hrefs = self.extract_hrefs(response) hrefs = self.extract_hrefs(response)
# Just in case the URL we used is different than the real one,
# such as missing a trailing slash, add both.
self.seen_directories.add(url)
self.seen_directories.add(head.url) self.seen_directories.add(head.url)
added = 0 added = 0
for href in hrefs: for href in hrefs:
@ -660,6 +688,7 @@ def smart_insert(sql, cur, url=None, head=None, commit=True):
if is_new: if is_new:
cur.execute('INSERT INTO urls VALUES(?, ?, ?, ?, ?)', data) cur.execute('INSERT INTO urls VALUES(?, ?, ?, ?, ?)', data)
else: else:
print(url)
command = ''' command = '''
UPDATE urls SET UPDATE urls SET
content_length = coalesce(?, content_length), content_length = coalesce(?, content_length),
@ -832,34 +861,37 @@ def list_basenames_argparse(args):
outputfile=args.outputfile, outputfile=args.outputfile,
) )
def measure(databasename, fullscan=False): def measure(databasename, fullscan=False, new_only=False):
''' '''
Given a database, print the sum of all Content-Lengths. Given a database, print the sum of all Content-Lengths.
If `fullscan`, then URLs with no Content-Length will be URLs will be HEAD requested if:
HEAD requested, and the result will be saved back into the file. `new_only` is True and the file has no stored content length, or
`fullscan` is True and `new_only` is False
''' '''
if isinstance(fullscan, str): if isinstance(fullscan, str):
fullscan = bool(fullscan) fullscan = bool(fullscan)
totalsize = 0 totalsize = 0
sql = sqlite3.connect(databasename) sql = sqlite3.connect(databasename)
cur1 = sql.cursor() cur = sql.cursor()
cur2 = sql.cursor()
cur2.execute('SELECT * FROM urls WHERE do_download == 1') if new_only:
cur.execute('SELECT * FROM urls WHERE do_download == 1 AND content_length IS NULL')
else:
cur.execute('SELECT * FROM urls WHERE do_download == 1')
items = cur.fetchall()
filecount = 0 filecount = 0
unmeasured_file_count = 0 unmeasured_file_count = 0
try: try:
while True: for fetch in items:
fetch = cur2.fetchone()
if fetch is None:
break
size = fetch[SQL_CONTENT_LENGTH] size = fetch[SQL_CONTENT_LENGTH]
if fullscan: if fullscan or new_only:
url = fetch[SQL_URL] url = fetch[SQL_URL]
head = do_head(url, raise_for_status=False) head = do_head(url, raise_for_status=False)
fetch = smart_insert(sql, cur1, head=head, commit=False) fetch = smart_insert(sql, cur, head=head, commit=True)
size = fetch[SQL_CONTENT_LENGTH] size = fetch[SQL_CONTENT_LENGTH]
if size is None: if size is None:
safeprint('"%s" is not revealing Content-Length' % url) safeprint('"%s" is not revealing Content-Length' % url)
@ -881,14 +913,14 @@ def measure(databasename, fullscan=False):
totalsize_string = '{} ({:,} bytes) in {:,} files'.format(short_string, totalsize, filecount) totalsize_string = '{} ({:,} bytes) in {:,} files'.format(short_string, totalsize, filecount)
print(totalsize_string) print(totalsize_string)
if unmeasured_file_count > 0: if unmeasured_file_count > 0:
print('Note: %d files do not have a stored Content-Length.' % unmeasured_file_count) print(UNMEASURED_WARNING % unmeasured_file_count)
print('Run `measure` with `-f` or `--fullscan` to HEAD request those files.')
return totalsize return totalsize
def measure_argparse(args): def measure_argparse(args):
return measure( return measure(
databasename=args.databasename, databasename=args.databasename,
fullscan=args.fullscan, fullscan=args.fullscan,
new_only=args.new_only,
) )
def remove_pattern(args): def remove_pattern(args):
@ -913,9 +945,13 @@ def tree(databasename, output_filename=None):
path_parts = url_to_filepath(items[0][SQL_URL]) path_parts = url_to_filepath(items[0][SQL_URL])
root_identifier = path_parts['root'] root_identifier = path_parts['root']
#print('Root', root_identifier) print('Root', root_identifier)
root_data = {'name': root_identifier, 'item_type': 'directory'} root_data = {'name': root_identifier, 'item_type': 'directory'}
tree = TreeNode(identifier=root_identifier, data=root_data) root_identifier = root_identifier.replace(':', '')
tree = TreeNode(
identifier=root_identifier,
data=root_data
)
node_map = {} node_map = {}
unmeasured_file_count = 0 unmeasured_file_count = 0
@ -923,20 +959,27 @@ def tree(databasename, output_filename=None):
for item in items: for item in items:
path = url_to_filepath(item[SQL_URL]) path = url_to_filepath(item[SQL_URL])
scheme = path['scheme'] scheme = path['scheme']
# I join and re-split because 'folder' may contain slashes of its own
# and I want to break all the pieces
path = '\\'.join([path['root'], path['folder'], path['filename']]) path = '\\'.join([path['root'], path['folder'], path['filename']])
parts = path.split('\\') parts = path.split('\\')
#print(path)
for (index, part) in enumerate(parts): for (index, part) in enumerate(parts):
index += 1 this_path = '/'.join(parts[:index + 1])
this_path = '/'.join(parts[:index]) parent_path = '/'.join(parts[:index])
parent_path = '/'.join(parts[:index-1])
#safeprint('this:' + this_path) #safeprint('this:' + this_path)
#safeprint('parent:' + parent_path) #safeprint('parent:' + parent_path)
#input() #input()
data = { data = {
'name': part, 'name': part,
'url': scheme + '://' + this_path, 'url': scheme + '://' + this_path,
} }
if index == len(parts): this_identifier = this_path.replace(':', '')
parent_identifier = parent_path.replace(':', '')
if (index + 1) == len(parts):
data['item_type'] = 'file' data['item_type'] = 'file'
if item[SQL_CONTENT_LENGTH]: if item[SQL_CONTENT_LENGTH]:
data['size'] = item[SQL_CONTENT_LENGTH] data['size'] = item[SQL_CONTENT_LENGTH]
@ -948,29 +991,29 @@ def tree(databasename, output_filename=None):
# Ensure this comment is in a node of its own # Ensure this comment is in a node of its own
this_node = node_map.get(this_path, None) this_node = node_map.get(this_identifier, None)
if this_node: if this_node:
# This ID was detected as a parent of a previous iteration # This ID was detected as a parent of a previous iteration
# Now we're actually filling it in. # Now we're actually filling it in.
this_node.data = data this_node.data = data
else: else:
this_node = TreeNode(this_path, data) this_node = TreeNode(this_identifier, data)
node_map[this_path] = this_node node_map[this_identifier] = this_node
# Attach this node to the parent. # Attach this node to the parent.
if parent_path == root_identifier: if parent_identifier == root_identifier:
try: try:
tree.add_child(this_node) tree.add_child(this_node)
except: except TreeExistingChild:
pass pass
else: else:
parent_node = node_map.get(parent_path, None) parent_node = node_map.get(parent_identifier, None)
if not parent_node: if not parent_node:
parent_node = TreeNode(parent_path, data=None) parent_node = TreeNode(parent_identifier, data=None)
node_map[parent_path] = parent_node node_map[parent_identifier] = parent_node
try: try:
parent_node.add_child(this_node) parent_node.add_child(this_node)
except: except TreeExistingChild:
pass pass
this_node.parent = parent_node this_node.parent = parent_node
#print(this_node.data) #print(this_node.data)
@ -997,7 +1040,7 @@ def tree(databasename, output_filename=None):
if node.data['item_type'] == 'directory': if node.data['item_type'] == 'directory':
div_id = hashit(node.identifier, 16) div_id = hashit(node.identifier, 16)
line = '<button onclick="collapse(\'{div_id}\')">{name} ({size})</button>' line = '<button onclick="collapse(\'{div_id}\')">{name} ({size})</button>'
line += '<div id="{div_id}">' line += '<div id="{div_id}" style="display:none">'
line = line.format( line = line.format(
div_id=div_id, div_id=div_id,
name=node.data['name'], name=node.data['name'],
@ -1020,7 +1063,8 @@ def tree(databasename, output_filename=None):
) )
write(line, outfile) write(line, outfile)
for (key, child) in node.sorted_children(): customsort = lambda x: (node.children[x].data['item_type'] == 'file', node.children[x].data['url'].lower())
for (key, child) in node.sorted_children(customsort=customsort):
recursive_print_node(child, depth+1, outfile=outfile) recursive_print_node(child, depth+1, outfile=outfile)
if node.data['item_type'] == 'directory': if node.data['item_type'] == 'directory':
@ -1030,19 +1074,22 @@ def tree(databasename, output_filename=None):
# This helps put some space between sibling directories # This helps put some space between sibling directories
write('| ' * (depth), outfile) write('| ' * (depth), outfile)
recursive_get_size(tree)
use_html = output_filename.lower().endswith('.html')
if output_filename is not None: if output_filename is not None:
output_file = open(output_filename, 'w', encoding='utf-8') output_file = open(output_filename, 'w', encoding='utf-8')
use_html = output_filename.lower().endswith('.html')
else:
output_file = None
use_html = False
if use_html: if use_html:
write(HTML_TREE_HEADER, outfile=output_file) write(HTML_TREE_HEADER, outfile=output_file)
recursive_get_size(tree)
recursive_print_node(tree, outfile=output_file) recursive_print_node(tree, outfile=output_file)
if unmeasured_file_count > 0: if unmeasured_file_count > 0:
write('Note: %d files do not have a stored Content-Length.' % unmeasured_file_count, outfile=output_file) write(UNMEASURED_WARNING % unmeasured_file_count, outfile=output_file)
write('Run `measure` with `-f` or `--fullscan` to HEAD request those files.', outfile=output_file)
if output_file is not None: if output_file is not None:
output_file.close() output_file.close()
@ -1091,6 +1138,7 @@ if __name__ == '__main__':
p_measure = subparsers.add_parser('measure') p_measure = subparsers.add_parser('measure')
p_measure.add_argument('databasename') p_measure.add_argument('databasename')
p_measure.add_argument('-f', '--fullscan', dest='fullscan', action='store_true') p_measure.add_argument('-f', '--fullscan', dest='fullscan', action='store_true')
p_measure.add_argument('-n', '--new_only', dest='new_only', action='store_true')
p_measure.set_defaults(func=measure_argparse) p_measure.set_defaults(func=measure_argparse)
p_remove_pattern = subparsers.add_parser('remove_pattern') p_remove_pattern = subparsers.add_parser('remove_pattern')

View file

@ -4,12 +4,13 @@ Generators
# What are they # What are they
Generators are a type of iterable that create their contents on-the-fly. Unlike a list, whose entire contents are available before beginning any loops or manipulations, generators don't know how many items they will produce or when they will stop. Generators are a type of iterable that create their contents on-the-fly. Unlike a list, whose entire contents are available before beginning any loops or manipulations, generators don't know how many items they will produce or when they will stop. They can even go on forever.
&nbsp;
# Writing one # Writing one
Writing a generator looks like writing a function, but instead of `return`, you use `yield`. The object which is yielded is what you'll get when you do a loop over the generator. This generator lets you count to a billion: Writing a generator looks like writing a function, but instead of `return`, you use `yield`. The object which is yielded is what you'll get when you do a loop over the generator. This one lets you count to a billion:
def billion(): def billion():
x = 0 x = 0
@ -17,41 +18,92 @@ Writing a generator looks like writing a function, but instead of `return`, you
yield x yield x
x += 1 x += 1
Note that, unlike a `return` statement, you can include more code after a `yield` statement. Also notice that generators keep track of their internal state. The `billion` generator has an `x` that it increments every time you loop over it. Note that, unlike a `return` statement, you can include more code after a `yield` statement. Also notice that generators keep track of their internal state -- the `billion` generator has an `x` that it increments every time you loop over it. You can imagine the code pausing after the `yield` line, and resuming when you come back for the next cycle. Try this with some extra print statements to help visualize.
Generators can also take arguments. Here's a generator that counts to a custom amount:
def count_to(y):
x = 0
while x < y:
yield x
x += 1
&nbsp;
# Using one # Using one
Although generators look like a function when you're writing them, they feel more like objects when using them. Remember that generators don't calculate their contents until they are actually used in a loop, so simply doing: Although generators look like functions when you're writing them, they feel more like objects when using them. Remember that generators don't calculate their contents until they are actually used in a loop, so simply doing:
numbers = billion() numbers = count_to(100)
does **not** create a list of a billion numbers. It creates a new instance of the generator that is ready to be iterated over, like this: does **not** create a list of 100 numbers. It creates a new instance of the generator that is ready to be iterated over, like this:
numbers = billion() numbers = count_to(100)
for number in numbers: for number in numbers:
print(number) print(number)
This might remind you of: or this:
for number in range(1000000000): for number in count_to(100):
print(number) print(number)
because `range` is simply a generator. This should remind you of:
for number in range(100):
print(number)
because the `range` class behaves a lot like a generator ([but not exactly](http://stackoverflow.com/a/13092317)).
Generators are excellent for cases where using a list is infeasible or unnecessary. If you wanted to count to a billion using a list, you would first have to create a list of every number, which is a huge waste of time and memory. With a generator, the item is created, used, and trashed. Generators are excellent for cases where using a list is infeasible or unnecessary. In order to loop over a list, you have to generate the entire thing first. If you're only going to use each item once, storing the entire list can be a big memory waste when a generator could take its place. With a generator, the items are created, used, and trashed, so memory can be recycled.
Since generators can go on forever, they're great for abstracting out ugly `while` loops, so you can get down to business faster.
To get a single item from a generator without looping, use `next(generator)`. To get a single item from a generator without looping, use `next(generator)`.
&nbsp;
# StopIteration # StopIteration
When a generator is all finished, it will raise a `StopIteration` exception every time you try to do `next()`. `for` loops will detect this automatically and stop themselves. Generators pause and resume a lot, but they still flow like normal functions. As long as there is no endless `while` loop inside, they'll come to an end at some point. When a generator is all finished, it will raise a `StopIteration` exception every time you try to do `next()`. Luckily, `for` loops will detect this automatically and stop themselves.
Earlier, I said that generators use `yield` instead of `return`, but in fact you can include a return statement. If it is encountered, it will raise a `StopIteration`, and the generator will not resume even if there is more code.
>>> def generator():
... yield 1
... return 2
... yield 3
...
>>>
>>> g = generator()
>>> next(g)
1
>>> next(g)
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
StopIteration: 2
>>> next(g)
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
StopIteration
>>>
In general, I don't like to use `return` in generators. I prefer to `break` from their internal loops and conclude naturally.
&nbsp;
# Minor notes
- You cannot access the items of a generator by index, because only one item exists at a time. Once you do a loop over a generator, those items are gone forever unless you kept them somewhere else.
&nbsp;
# More examples # More examples
#### Yielding individual items from batches
Suppose you're getting data from an imaginary website which sends you items in groups of 100. You want to let the user loop over every item without having to worry about the groups themselves. Suppose you're getting data from an imaginary website which sends you items in groups of 100. You want to let the user loop over every item without having to worry about the groups themselves.
def item_generator(url): def item_generator(url):
@ -60,7 +112,7 @@ Suppose you're getting data from an imaginary website which sends you items in g
# get_items is a pretend method that collects the 100 items from that page # get_items is a pretend method that collects the 100 items from that page
batch = get_items(url, page=page) batch = get_items(url, page=page)
if not batch: if len(batch) == 0:
# for this imaginary website, the batch will be empty when that page # for this imaginary website, the batch will be empty when that page
# doesn't have any items on it. # doesn't have any items on it.
break break
@ -73,9 +125,18 @@ Suppose you're getting data from an imaginary website which sends you items in g
page += 1 page += 1
# When the while loop breaks, we reach the end of the function body, # When the while loop breaks, we reach the end of the function body,
# and a StopIteration will be raised and handled automatically, # concluding the generator.
# ending the for-loop.
comments = item_generator('http://website.com/user/voussoir/comments') comments = item_generator('http://website.com/user/voussoir/comments')
for comment in comments: for comment in comments:
print(comment.body) print(comment.body)
&nbsp;
# Further reading
[Stack Overflow - What are the main uses for `yield from`?](http://stackoverflow.com/questions/9708902/in-practice-what-are-the-main-uses-for-the-new-yield-from-syntax-in-python-3) -- If you like recursive functions, how about recursive generators? The only time I've ever used this is to [iterate over a tree's nodes](https://github.com/voussoir/reddit/blob/2069c3bd731cc8f90401ee49a9fc4d0dbf436cfc/Prawtimestamps/timesearch.py#L756-L761).
[Stack Overflow - Python generator `send` function purpose?](http://stackoverflow.com/questions/19302530/python-generator-send-function-purpose) -- This quickly dives out of "quick tips" territory.

View file

@ -5,12 +5,8 @@ import urllib.parse
import random import random
import sys import sys
sys.path.append('C:\\git\\else\\Bytestring') sys.path.append('C:\\git\\else\\Bytestring'); import bytestring
import bytestring sys.path.append('C:\\git\\else\\Ratelimiter'); import ratelimiter
sys.path.append('C:\\git\\else\\Ratelimiter')
import ratelimiter
f = open('favicon.png', 'rb') f = open('favicon.png', 'rb')
FAVI = f.read() FAVI = f.read()
@ -60,7 +56,14 @@ class Path:
else: else:
# Diamond emoji, because there's not one for files. # Diamond emoji, because there's not one for files.
icon = '\U0001F48E' icon = '\U0001F48E'
return '<a href="{full}">{icon} {display}</a>'.format(full=self.path, icon=icon, display=display_name)
quoted_path = urllib.parse.quote(self.path)
a = '<a href="{full}">{icon} {display}</a>'.format(
full=quoted_path,
icon=icon,
display=display_name,
)
return a
@property @property
def basename(self): def basename(self):
@ -70,6 +73,10 @@ class Path:
def is_dir(self): def is_dir(self):
return os.path.isdir(self.os_path) return os.path.isdir(self.os_path)
@property
def is_file(self):
return os.path.isfile(self.os_path)
@property @property
def os_path(self): def os_path(self):
abspath = os.path.join(CWD, self.relative_path) abspath = os.path.join(CWD, self.relative_path)
@ -96,7 +103,11 @@ class Path:
form = '<tr style="background-color:#{bg}"><td>{anchor}</td><td>{size}</td></tr>' form = '<tr style="background-color:#{bg}"><td>{anchor}</td><td>{size}</td></tr>'
bg = 'ddd' if shaded else 'fff'; bg = 'ddd' if shaded else 'fff';
size = bytestring.bytestring(self.size) if self.size != -1 else '' size = bytestring.bytestring(self.size) if self.size != -1 else ''
row = form.format(bg=bg, anchor=self.anchor(display_name=display_name), size=size) row = form.format(
bg=bg,
anchor=self.anchor(display_name=display_name),
size=size,
)
return row return row
@ -124,12 +135,8 @@ class RequestHandler(http.server.BaseHTTPRequestHandler):
def do_GET(self): def do_GET(self):
#print(dir(self)) #print(dir(self))
path = self.path.lower() path = normalize_path(self.path)
path = urllib.parse.unquote(path).rstrip('/') if self.send_path_validation_error(path):
error = path_validation(path)
if error:
self.send_error(*error)
return return
path = Path(path) path = Path(path)
@ -139,11 +146,58 @@ class RequestHandler(http.server.BaseHTTPRequestHandler):
if mime is not None: if mime is not None:
#print(mime) #print(mime)
self.send_header('Content-type', mime) self.send_header('Content-type', mime)
self.end_headers()
if path.is_file:
self.send_header('Content-length', path.size)
d = self.read_filebytes(path) d = self.read_filebytes(path)
#print('write')
self.end_headers()
self.write(d) self.write(d)
def do_HEAD(self):
path = normalize_path(self.path)
if self.send_path_validation_error(path):
return
path = Path(path)
self.send_response(200)
if path.is_dir:
mime = 'text/html'
else:
mime = mimetypes.guess_type(path.path)[0]
if mime is not None:
self.send_header('Content-type', mime)
if path.is_file:
self.send_header('Content-length', path.size)
self.end_headers()
def path_validation(self, path):
path = path.lstrip('/')
absolute_path = os.path.join(CWD, path)
absolute_path = os.path.abspath(absolute_path)
path = absolute_path.replace(CWD, '')
path = path.lstrip('/')
path = path.replace('\\', '/')
#if '..' in path:
# return (403, 'I\'m not going to play games with you.')
#print(path)
print(path)
if not any(path.startswith(okay) for okay in OKAY_PATHS):
self.send_error(403, 'Stop that!')
return
def send_path_validation_error(self, path):
error = self.path_validation(path)
if error:
self.send_error(*error)
return True
return False
# def do_POST(self): # def do_POST(self):
# path = self.path.lower() # path = self.path.lower()
# path = urllib.parse.unquote(path).rstrip('/') # path = urllib.parse.unquote(path).rstrip('/')
@ -181,32 +235,34 @@ class RequestHandler(http.server.BaseHTTPRequestHandler):
# self.write('Thanks') # self.write('Thanks')
def generate_opendir(path): def generate_opendir(path):
print('Listdir:', path) #print('Listdir:', path)
items = os.listdir(path.relative_path) items = os.listdir(path.relative_path)
items = [os.path.join(path.relative_path, f) for f in items] items = [os.path.join(path.relative_path, f) for f in items]
# This places directories above files, each ordered alphabetically
items.sort(key=str.lower)
directories = [] directories = []
files = [] files = []
for item in items: for item in items:
#item = item.lstrip('/')
if os.path.isdir(item): if os.path.isdir(item):
directories.append(item) directories.append(item)
else: else:
files.append(item) files.append(item)
directories.sort(key=str.lower)
files.sort(key=str.lower) items = directories + files
files = directories + files items = [Path(f) for f in items]
#print(files)
files = [Path(f) for f in files]
entries = [] entries = []
if not any(okay == path.path for okay in OKAY_PATHS): if not any(okay == path.path for okay in OKAY_PATHS):
# If the path actually equals a okay_path, then we shouldn't # If the user is on one of the OKAY_PATHS, then he can't step up
# let them step up because that would be outisde the okay area. # because that would be outside the OKAY area.
entries.append(path.parent.table_row(display_name='up')) entries.append(path.parent.table_row(display_name='up'))
shaded = True shaded = True
for f in files: for item in items:
entry = f.table_row(shaded=shaded) entry = item.table_row(shaded=shaded)
entries.append(entry) entries.append(entry)
shaded = not shaded shaded = not shaded
entries = '\n'.join(entries) entries = '\n'.join(entries)
text = OPENDIR_TEMPLATE.format(entries=entries) text = OPENDIR_TEMPLATE.format(entries=entries)
return text return text
@ -218,12 +274,11 @@ def generate_random_filename(original_filename='', length=8):
identifier = '{:x}'.format(bits).rjust(length, '0') identifier = '{:x}'.format(bits).rjust(length, '0')
return identifier return identifier
def path_validation(path): def normalize_path(path):
if '..' in path: #path = path.lower()
return (403, 'I\'m not going to play games with you.') path = urllib.parse.unquote(path).rstrip('/')
if not any(path.startswith(okay) for okay in OKAY_PATHS): return path
self.send_error(403, 'Stop that!')
return
server = http.server.HTTPServer(('', 32768), RequestHandler) server = http.server.HTTPServer(('', 32768), RequestHandler)
print('server starting') print('server starting')

View file

@ -526,11 +526,32 @@ def get_path_casing(path):
''' '''
p = str_to_fp(path) p = str_to_fp(path)
path = p.path path = p.path
path = glob.escape(path)
(drive, subpath) = os.path.splitdrive(path) (drive, subpath) = os.path.splitdrive(path)
pattern = ["%s[%s]" % (piece[:-1], piece[-1]) for piece in subpath.split(os.sep)[1:]] subpath = subpath.lstrip(os.sep)
def patternize(piece):
'''
Create a pattern like "[u]ser" from "user", forcing glob to look up the
correct path name, and guaranteeing that the only result will be the correct path.
Special cases are:
!, because in glob syntax, [!x] tells glob to look for paths that don't contain
"x". [!] is invalid syntax, so we pick the first non-! character to put
in the brackets.
[, because this starts a capture group
'''
piece = glob.escape(piece)
for character in piece:
if character not in '!':
replacement = '[%s]' % character
piece = piece.replace(character, replacement, 1)
break
return piece
pattern = [patternize(piece) for piece in subpath.split(os.sep)]
pattern = os.sep.join(pattern) pattern = os.sep.join(pattern)
pattern = drive.upper() + os.sep + pattern pattern = drive.upper() + os.sep + pattern
print(pattern)
try: try:
return str_to_fp(glob.glob(pattern)[0]) return str_to_fp(glob.glob(pattern)[0])
except IndexError: except IndexError: