📄 btmakemetafile.py
字号:
#!/usr/bin/env python# Written by Bram Cohen# multitracker extensions by John Hoffman# see LICENSE.txt for license informationimport sysimport md5import zlibfrom os.path import getsize, split, join, abspath, isdir, normpathfrom os import listdirfrom sha import shafrom copy import copyfrom string import stripfrom threading import Eventfrom time import timefrom traceback import print_exc
try:
from sys import getfilesystemencoding
ENCODING = getfilesystemencoding()
except:
from sys import getdefaultencoding
ENCODING = getdefaultencoding()
from BitTornado.bencode import bencodefrom BitTornado.BT1.btformats import check_infodefaults = [ ('announce_list', '', 'a list of announce URLs - explained below'), ('httpseeds', '',
'a list of http seed URLs - explained below'),
('piece_size_pow2', 0, "which power of 2 to set the piece size to (0 = automatic)"), ('comment', '', "optional human-readable comment to put in .torrent"), ('filesystem_encoding', '',
"optional specification for filesystem encoding " +
"(set automatically in recent Python versions)"),
('target', '', "optional target file for the torrent"), ('created by', '', "optional information on who made the torrent") ]default_piece_len_exp = 18ignore = ['core', 'CVS']def print_announcelist_details(): print (' announce_list = optional list of redundant/backup tracker URLs, in the format:') print (' url[,url...][|url[,url...]...]') print (' where URLs separated by commas are all tried first') print (' before the next group of URLs separated by the pipe is checked.') print (" If none is given, it is assumed you don't want one in the metafile.") print (' If announce_list is given, clients which support it') print (' will ignore the <announce> value.') print (' Examples:') print (' http://tracker1.com|http://tracker2.com|http://tracker3.com') print (' (tries trackers 1-3 in order)') print (' http://tracker1.com,http://tracker2.com,http://tracker3.com') print (' (tries trackers 1-3 in a randomly selected order)') print (' http://tracker1.com|http://backup1.com,http://backup2.com') print (' (tries tracker 1 first, then tries between the 2 backups randomly)') print ('')
print (' httpseeds = optional list of http-seed URLs, in the format:')
print (' url[|url...]')
def make_meta_file(file, url, params = None, flag = Event(), progress = lambda x: None, progress_percent = 1, gethash = None):
if params is None: params = {} if 'piece_size_pow2' in params: piece_len_exp = params['piece_size_pow2'] else: piece_len_exp = default_piece_len_exp if 'target' in params and params['target']: f = join(params['target'], split(normpath(file))[1] + '.torrent') else: a, b = split(file) if b == '': f = a + '.torrent' else: f = join(a, b + '.torrent') if piece_len_exp == 0: # automatic size = calcsize(file) if size > 8L*1024*1024*1024: # > 8 gig =
piece_len_exp = 21 # 2 meg pieces elif size > 2*1024*1024*1024: # > 2 gig = piece_len_exp = 20 # 1 meg pieces elif size > 512*1024*1024: # > 512M = piece_len_exp = 19 # 512K pieces elif size > 64*1024*1024: # > 64M = piece_len_exp = 18 # 256K pieces elif size > 16*1024*1024: # > 16M = piece_len_exp = 17 # 128K pieces elif size > 4*1024*1024: # > 4M = piece_len_exp = 16 # 64K pieces else: # < 4M = piece_len_exp = 15 # 32K pieces piece_length = 2 ** piece_len_exp encoding = None
if 'filesystem_encoding' in params:
encoding = params['filesystem_encoding']
if not encoding:
encoding = ENCODING
if not encoding:
encoding = 'ascii'
info = makeinfo(file, piece_length, encoding, flag, progress, progress_percent, gethash = gethash)
if flag.isSet(): return check_info(info) h = open(f, 'wb') data = {'info': info, 'announce': strip(url), 'creation date': long(time())}
if 'comment' in params and params['comment']:
data['comment'] = params['comment'] if 'created by' in params and params['created by']: data['created by'] = params['created by'] if 'real_announce_list' in params: # shortcut for progs calling in from outside data['announce-list'] = params['real_announce_list'] elif 'announce_list' in params and params['announce_list']:
l = []
for tier in params['announce_list'].split('|'): l.append(tier.split(','))
data['announce-list'] = l
if 'real_httpseeds' in params: # shortcut for progs calling in from outside
data['httpseeds'] = params['real_httpseeds']
elif 'httpseeds' in params and params['httpseeds']:
data['httpseeds'] = params['httpseeds'].split('|')
h.write(bencode(data)) h.close()def calcsize(file): if not isdir(file): return getsize(file) total = 0L
for s in subfiles(abspath(file)): total += getsize(s[1]) return total
def uniconvertl(l, e):
r = []
try:
for s in l:
r.append(uniconvert(s, e))
except UnicodeError:
raise UnicodeError('bad filename: '+join(l))
return r
def uniconvert(s, e):
try:
s = unicode(s,e)
except UnicodeError:
raise UnicodeError('bad filename: '+s)
return s.encode('utf-8')
def makeinfo(file, piece_length, encoding, flag, progress, progress_percent=1, gethash = None): if gethash is None: gethash = {} if not 'md5' in gethash: gethash['md5'] = False if not 'crc32' in gethash: gethash['crc32'] = False if not 'sha1' in gethash: gethash['sha1'] = False
file = abspath(file) if isdir(file): subs = subfiles(file) subs.sort() pieces = [] sh = sha() done = 0L
fs = [] totalsize = 0.0 totalhashed = 0L
for p, f in subs: totalsize += getsize(f) for p, f in subs: pos = 0L
size = getsize(f) h = open(f, 'rb') if gethash['md5']: hash_md5 = md5.new() if gethash['sha1']: hash_sha1 = sha() if gethash['crc32']: hash_crc32 = zlib.crc32('') while pos < size: a = min(size - pos, piece_length - done) readpiece = h.read(a) # See if the user cancelled if flag.isSet(): return sh.update(readpiece) # See if the user cancelled if flag.isSet(): return if gethash['md5']: # Update MD5 hash_md5.update(readpiece) # See if the user cancelled if flag.isSet(): return if gethash['crc32']: # Update CRC32 hash_crc32 = zlib.crc32(readpiece, hash_crc32) # See if the user cancelled if flag.isSet(): return if gethash['sha1']: # Update SHA1 hash_sha1.update(readpiece) # See if the user cancelled if flag.isSet(): return done += a pos += a totalhashed += a if done == piece_length: pieces.append(sh.digest()) done = 0 sh = sha() if progress_percent: progress(totalhashed / totalsize) else: progress(a) newdict = {'length': size, 'path': uniconvertl(p, encoding) } if gethash['md5']: newdict['md5sum'] = hash_md5.hexdigest() if gethash['crc32']: newdict['crc32'] = "%08X" % hash_crc32 if gethash['sha1']: newdict['sha1'] = hash_sha1.digest() fs.append(newdict) h.close() if done > 0: pieces.append(sh.digest()) return {'pieces': ''.join(pieces), 'piece length': piece_length, 'files': fs, 'name': uniconvert(split(file)[1], encoding) }
else: size = getsize(file) pieces = [] p = 0L
h = open(file, 'rb') if gethash['md5']: hash_md5 = md5.new() if gethash['crc32']: hash_crc32 = zlib.crc32('') if gethash['sha1']: hash_sha1 = sha() while p < size: x = h.read(min(piece_length, size - p)) # See if the user cancelled if flag.isSet(): return if gethash['md5']: # Update MD5 hash_md5.update(x) # See if the user cancelled if flag.isSet(): return if gethash['crc32']: # Update CRC32 hash_crc32 = zlib.crc32(x, hash_crc32) # See if the user cancelled if flag.isSet(): return if gethash['sha1']: # Update SHA-1 hash_sha1.update(x) # See if the user cancelled if flag.isSet(): return pieces.append(sha(x).digest()) # See if the user cancelled if flag.isSet(): return p += piece_length if p > size: p = size if progress_percent: progress(float(p) / size) else: progress(min(piece_length, size - p)) h.close() newdict = {'pieces': ''.join(pieces), 'piece length': piece_length, 'length': size, 'name': uniconvert(split(file)[1], encoding) } if gethash['md5']: newdict['md5sum'] = hash_md5.hexdigest() if gethash['crc32']: newdict['crc32'] = "%08X" % hash_crc32 if gethash['sha1']: newdict['sha1'] = hash_sha1.digest() return newdict
def subfiles(d): r = [] stack = [([], d)] while stack: p, n = stack.pop() if isdir(n): for s in listdir(n): if s not in ignore and s[:1] != '.': stack.append((copy(p) + [s], join(n, s))) else: r.append((p, n)) return rdef completedir(dir, url, params = None, flag = Event(), vc = lambda x: None, fc = lambda x: None, getmd5 = False): if params is None: params = {} files = listdir(dir) files.sort() ext = '.torrent' if 'target' in params: target = params['target'] else: target = '' togen = [] for f in files: if f[-len(ext):] != ext and (f + ext) not in files: togen.append(join(dir, f)) total = 0 for i in togen: total += calcsize(i) subtotal = [0] def callback(x, subtotal = subtotal, total = total, vc = vc): subtotal[0] += x vc(float(subtotal[0]) / total) for i in togen: fc(i) try: t = split(i)[-1] if t not in ignore and t[0] != '.': if target != '': params['target'] = join(target,t+ext) make_meta_file(i, url, params, flag, progress = callback, progress_percent = 0, getmd5 = getmd5) except ValueError: print_exc()
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -