Skip to content

Instantly share code, notes, and snippets.

@plamere
Created June 25, 2011 13:05
Show Gist options
  • Star 10 You must be signed in to star a gist
  • Fork 9 You must be signed in to fork a gist
  • Save plamere/1046464 to your computer and use it in GitHub Desktop.
Save plamere/1046464 to your computer and use it in GitHub Desktop.
dedup.py - uses echoprint to find duplicates in a music collection
#!/usr/bin/python
import sys
import os
import pprint
import subprocess
import pickle
import atexit
import simplejson as json
sys.path.insert(0, "../API")
import fp
_codegen_path = "/Users/plamere/bin/echoprint-codegen"
done = {}
def codegen(filename, start=10, duration=30):
""" runs codegen on the given file, returns a tuple with
the track id, a code string for lookup, and a codegen
block for ingest
"""
if not os.path.exists(_codegen_path):
raise Exception("Codegen binary not found.")
command = _codegen_path + " \"" + filename + "\" "
if start >= 0:
command = command + str(start) + " "
if duration >= 0:
command = command + str(duration)
p = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
(json_block, errs) = p.communicate()
try:
return parse_json_block(json_block)
except ValueError:
logger.debug("No JSON object came out of codegen: error was %s" % (errs))
return None
def parse_json_block(json_block):
codes = json.loads(json_block)
code = codes[0]
if not 'code' in code:
return None
raw_code = code["code"]
m = code["metadata"]
if "track_id" in m:
trid = m["track_id"].encode("utf-8")
else:
trid = fp.new_track_id()
length = m["duration"]
version = m["version"]
artist = m.get("artist", None)
title = m.get("title", None)
release = m.get("release", None)
decoded = fp.decode_code_string(raw_code)
ingest_data = {
"track_id": trid,
"fp": decoded,
"length": length,
"codever": "%.2f" % version
}
if artist:
ingest_data["artist"] = artist
if release:
ingest_data["release"] = release
if title:
ingest_data["track"] = title
return trid, raw_code, ingest_data
def collect_mp3s(list, dir):
""" recursively collects all mp3s in a directory into
the given list.
"""
if os.path.isdir(dir):
files = os.listdir(dir)
for f in files:
fp = os.path.join(dir, f)
if os.path.isdir(fp):
collect_mp3s(list, fp)
elif fp.lower().endswith('.mp3'):
list.append(fp)
def save_state():
f = open("dedup.state.pkl", "w")
pickle.dump(done, f)
f.close()
def restore_state():
if os.path.exists("dedup.state.pkl"):
f = open("dedup.state.pkl")
dict = pickle.load(f)
f.close()
done.update(dict)
def delete_files():
if os.path.exists("dedup.state.pkl"):
os.remove("dedup.state.pkl")
if os.path.exists("dedup.dat"):
os.remove("dedup.dat")
def get_file(id):
iddone = dict (zip(done.values(),done.keys()))
return iddone[id]
if __name__ == "__main__":
if len(sys.argv) < 2:
print "Usage: dedup.py --reset"
print "Usage: dedup.py path-to-mp3s"
sys.exit(1)
if sys.argv[1] == '--reset':
fp.delete_database()
delete_files()
sys.exit(0)
restore_state()
atexit.register(save_state)
mp3s = []
S = '<sep>'
for dir in sys.argv[1:]:
collect_mp3s(mp3s, dir)
print 'processing', len(mp3s), 'files'
dups = open("dedup.dat", "a")
for count, mp3 in enumerate(mp3s):
if mp3 in done:
print ' skipping', mp3
continue
print " ", len(done), count, mp3
results = codegen(mp3, start=0, duration=60)
if not results:
print "can't process", mp3, "skipping"
continue
trid, raw_code, ingest_data = results
response = fp.best_match_for_query(raw_code)
if response.match():
print >>dups, 'duplicate', S, mp3, S, 'original', S, get_file(response.TRID)
print
print 'duplicate', mp3
print ' ', get_file(response.TRID)
print
else:
fp.ingest([ingest_data], do_commit=True)
done[mp3] = trid
dups.close()
@plamere
Copy link
Author

plamere commented Jun 25, 2011

note that I added fp.delete_database() to my local API/fp.py. Here's the definition:


def delete_database():
    i = _fp_solr.get()
    i.delete_query("*:*")
    i.commit()

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment