Skip to content

Instantly share code, notes, and snippets.

Ed Summers edsu

Block or report user

Report or block edsu

Hide content and notifications from this user.

Learn more about blocking users

Contact Support about this user’s behavior.

Learn more about reporting abuse

Report abuse
View GitHub Profile
View diffbot.json
"request": {
"pageUrl": "",
"api": "analyze",
"version": 3
"humanLanguage": "en",
"objects": [
"date": "Tue, 15 Oct 2019 00:00:00 GMT",
edsu /
Last active Oct 15, 2019
Panel proposal for DH2020.

Documenting Documenting the Now

Ed Summers & Bergis Jules

Over the past four years the Documenting the Now project has been working to help build a community of practice around social media archiving that centers the ethical concerns of content creators, rather than simply the interests of cultural heritage organizations or social media platforms. Starting in the aftermath of the killing of Michael Brown in Ferguson Missouri the project developed the Ferguson Principles to help guide memory workers who are interested in documenting activism and social movements.

The Ferguson Principles have been put to work in a set of workshops with activist communities in the United States, in order to generate new knowledge practices for memory work in the age of social media. In addition the project has also been actively developing a portfolio of tools for data collection, publishing and analysis and using existing web archiving tools to help cultivate new approaches, and relationships between archivists, researche

edsu /
Last active Oct 10, 2019
Convincing twint to not give up.
#!/usr/bin/env python3
import os
import csv
import time
import twint
import random
config = twint.Config()
config.Search = 'nodapl'
hours_worked = float(input("Enter hours worked: "))
hourly_rate = float(input("Enter hourly rate: "))
def salary(hours_worked, hourly_rate):
if hourly_rate < 15.0:
print("I'm Sorry " + str(hourly_rate) + " is lower than the minimum wage!")
pay = hours_worked * hourly_rate
print("Pay: " + str(pay))
View toni-morrison-tweeted-urls.csv
We can't make this file beautiful and searchable because it's too large.
View http-response.txt
% curl -i
HTTP/2 302
cache-control: no-cache, no-store, must-revalidate, pre-check=0, post-check=0
content-length: 122
content-type: text/html;charset=utf-8
date: Tue, 13 Aug 2019 11:10:07 GMT
expires: Tue, 31 Mar 1981 05:00:00 GMT
last-modified: Tue, 13 Aug 2019 11:10:07 GMT
edsu / feral-hogs-urls.csv
Last active Aug 6, 2019
% twarc search "feral hogs" | | | grep -v | sort | uniq -c | sort -rn | csv > feral-hogs-urls.csv
View feral-hogs-urls.csv
tweet_count url
# Configuration file for jupyterhub.
import os
# Configurable configuration
# LoggingConfigurable configuration
#!/usr/bin/env python3
import twarc
# This small script shows how to listen to the Twitter sample stream and
# deconstruct tweet ids into their various components. The tweet_components
# method accepts a tweet id and returns a dict object with key / values
# representing the various components of a tweet id. Each component has its own
# method detailing how values are extracted from the tweet id.
#!/usr/bin/env python
import csv
from xml.etree import ElementTree
from langdetect import detect_langs
from requests_html import HTMLSession
http = HTMLSession()
def langs(url):
You can’t perform that action at this time.