#!/usr/bin/env python3 """partyfuse: remote copyparty as a local filesystem""" __author__ = "ed " __copyright__ = 2019 __license__ = "MIT" __url__ = "https://github.com/9001/copyparty/" S_VERSION = "2.0" S_BUILD_DT = "2024-10-01" """ mount a copyparty server (local or remote) as a filesystem speeds: 1 GiB/s reading large files 27'000 files/sec: copy small files 700 folders/sec: copy small folders usage: python partyfuse.py http://192.168.1.69:3923/ ./music dependencies: python3 -m pip install --user fusepy # or grab it from the connect page + on Linux: sudo apk add fuse + on Macos: https://osxfuse.github.io/ + on Windows: https://github.com/billziss-gh/winfsp/releases/latest note: you probably want to run this on windows clients: https://github.com/9001/copyparty/blob/hovudstraum/contrib/explorer-nothumbs-nofoldertypes.reg get server cert: awk '/-BEGIN CERTIFICATE-/ {a=1} a; /-END CERTIFICATE-/{exit}' <(openssl s_client -connect 127.0.0.1:3923 /dev/null) >cert.pem """ import argparse import calendar import codecs import errno import json import os import platform import re import stat import struct import sys import threading import time import traceback import urllib.parse from datetime import datetime, timezone from urllib.parse import quote_from_bytes as quote from urllib.parse import unquote_to_bytes as unquote import builtins import http.client WINDOWS = sys.platform == "win32" MACOS = platform.system() == "Darwin" UTC = timezone.utc # !rm.yes> MON3S = "Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec" MON3 = {b: a for a, b in enumerate(MON3S.split(), 1)} # !rm.no> def print(*args, **kwargs): try: builtins.print(*list(args), **kwargs) except: builtins.print(termsafe(" ".join(str(x) for x in args)), **kwargs) print( "{} v{} @ {}".format( platform.python_implementation(), ".".join([str(x) for x in sys.version_info]), sys.executable, ) ) def nullfun(*a): pass info = dbg = nullfun is_dbg = False try: from fuse import FUSE, FuseOSError, Operations except: if WINDOWS: libfuse = "install https://github.com/billziss-gh/winfsp/releases/latest" elif MACOS: libfuse = "install https://osxfuse.github.io/" else: libfuse = "apt install libfuse3-3\n modprobe fuse" m = """\033[33m could not import fuse; these may help: {} -m pip install --user fusepy {} \033[0m""" print(m.format(sys.executable, libfuse)) raise def termsafe(txt): enc = sys.stdout.encoding try: return txt.encode(enc, "backslashreplace").decode(enc) except: return txt.encode(enc, "replace").decode(enc) def threadless_log(fmt, *a): fmt += "\n" print(fmt % a if a else fmt, end="") riced_tids = {} def rice_tid(): tid = threading.current_thread().ident try: return riced_tids[tid] except: c = struct.unpack(b"B" * 5, struct.pack(b">Q", tid)[-5:]) ret = "".join("\033[1;37;48;5;%dm%02x" % (x, x) for x in c) + "\033[0m" riced_tids[tid] = ret return ret def fancy_log(fmt, *a): msg = fmt % a if a else fmt print("%10.6f %s %s\n" % (time.time() % 900, rice_tid(), msg), end="") def register_wtf8(): def wtf8_enc(text): return str(text).encode("utf-8", "surrogateescape"), len(text) def wtf8_dec(binary): return bytes(binary).decode("utf-8", "surrogateescape"), len(binary) def wtf8_search(encoding_name): return codecs.CodecInfo(wtf8_enc, wtf8_dec, name="wtf-8") codecs.register(wtf8_search) bad_good = {} good_bad = {} def enwin(txt): return "".join([bad_good.get(x, x) for x in txt]) def dewin(txt): return "".join([good_bad.get(x, x) for x in txt]) class RecentLog(object): def __init__(self, ar): self.ar = ar self.mtx = threading.Lock() self.f = open(ar.logf, "wb") if ar.logf else None self.q = [] thr = threading.Thread(target=self.printer) thr.daemon = True thr.start() def put(self, fmt, *a): msg = fmt % a if a else fmt msg = "%10.6f %s %s\n" % (time.time() % 900, rice_tid(), msg) if self.f: zd = datetime.now(UTC) fmsg = "%d-%04d-%06d.%06d %s" % ( zd.year, zd.month * 100 + zd.day, (zd.hour * 100 + zd.minute) * 100 + zd.second, zd.microsecond, msg, ) self.f.write(fmsg.encode("utf-8")) with self.mtx: self.q.append(msg) if len(self.q) > 200: self.q = self.q[-50:] def printer(self): while True: time.sleep(0.05) with self.mtx: q = self.q if not q: continue self.q = [] print("".join(q), end="") # [windows/cmd/cpy3] python dev\copyparty\bin\partyfuse.py q: http://192.168.1.159:1234/ # [windows/cmd/msys2] C:\msys64\mingw64\bin\python3 dev\copyparty\bin\partyfuse.py q: http://192.168.1.159:1234/ # [windows/mty/msys2] /mingw64/bin/python3 /c/Users/ed/dev/copyparty/bin/partyfuse.py q: http://192.168.1.159:1234/ # # [windows] find /q/music/albums/Phant*24bit -printf '%s %p\n' | sort -n | tail -n 8 | sed -r 's/^[0-9]+ //' | while IFS= read -r x; do dd if="$x" of=/dev/null bs=4k count=8192 & done # [alpine] ll t; for x in t/2020_0724_16{2,3}*; do dd if="$x" of=/dev/null bs=4k count=10240 & done # # 72.4983 windows mintty msys2 fancy_log # 219.5781 windows cmd msys2 fancy_log # nope.avi windows cmd cpy3 fancy_log # 9.8817 windows mintty msys2 RecentLog 200 50 0.1 # 10.2241 windows cmd cpy3 RecentLog 200 50 0.1 # 9.8494 windows cmd msys2 RecentLog 200 50 0.1 # 7.8061 windows mintty msys2 fancy_log # 7.9961 windows mintty msys2 RecentLog # 4.2603 alpine xfce4 cpy3 RecentLog # 4.1538 alpine xfce4 cpy3 fancy_log # 3.1742 alpine urxvt cpy3 fancy_log def get_tid(): return threading.current_thread().ident def html_dec(txt): return ( txt.replace("<", "<") .replace(">", ">") .replace(""", '"') .replace(" ", "\r") .replace(" ", "\n") .replace("&", "&") ) class CacheNode(object): def __init__(self, tag, data): self.tag = tag self.data = data self.ts = time.time() class Gateway(object): def __init__(self, ar): zs = ar.base_url if "://" not in zs: zs = "http://" + zs self.base_url = zs self.password = ar.a ui = urllib.parse.urlparse(zs) self.web_root = ui.path.strip("/") self.SRS = "/%s/" % (self.web_root,) if self.web_root else "/" try: self.web_host, self.web_port = ui.netloc.split(":") self.web_port = int(self.web_port) except: self.web_host = ui.netloc if ui.scheme == "http": self.web_port = 80 elif ui.scheme == "https": self.web_port = 443 else: raise Exception("bad url?") self.ssl_context = None self.use_tls = ui.scheme.lower() == "https" if self.use_tls: import ssl if ar.td: self.ssl_context = ssl._create_unverified_context() elif ar.te: self.ssl_context = ssl.SSLContext(ssl.PROTOCOL_TLS) self.ssl_context.load_verify_locations(ar.te) self.conns = {} self.fsuf = "?raw" self.dsuf = "?ls<&dots" # !rm.yes> if not ar.html: self.parse_html = None elif ar.html == "cpp": self.parse_html = self.parse_cpp self.dsuf = "?lt&dots" self.re_row = re.compile( r'^(-|DIR|)]* href="([^"]+)"[^>]*>([^<]+)([^<]+).*([^<]+)$' ) elif ar.html == "nginx": self.parse_html = self.parse_nginx self.fsuf = "" self.dsuf = "" self.re_row = re.compile( r'^([^<]+) *([0-9]{2})-([A-Z][a-z]{2})-([0-9]{4}) ([0-9]{2}:[0-9]{2}) *(-|[0-9]+)\r?$' ) elif ar.html == "iis": self.parse_html = self.parse_iis self.fsuf = "" self.dsuf = "" self.re_2nl = re.compile(br"
|") self.re_row = re.compile( r'^ *([0-9]{1,2})/([0-9]{1,2})/([0-9]{4}) {1,2}([0-9]{1,2}:[0-9]{2}) ([AP]M) +(<dir>|[0-9]+) ([^<>]+)$' ) else: raise Exception("unknown HTML dialect: [%s]" % (ar.html,)) # !rm.no> def quotep(self, path): path = path.encode("wtf-8") return quote(path, safe="/") def getconn(self, tid=None): tid = tid or get_tid() try: return self.conns[tid] except: info("new conn [{}] [{}]".format(self.web_host, self.web_port)) args = {} if not self.use_tls: C = http.client.HTTPConnection else: C = http.client.HTTPSConnection if self.ssl_context: args = {"context": self.ssl_context} conn = C(self.web_host, self.web_port, timeout=260, **args) self.conns[tid] = conn return conn def closeconn(self, tid=None): tid = tid or get_tid() try: self.conns[tid].close() del self.conns[tid] except: pass def sendreq(self, meth, path, headers, **kwargs): tid = get_tid() if self.password: headers["Cookie"] = "=".join(["cppwd", self.password]) try: c = self.getconn(tid) c.request(meth, path, headers=headers, **kwargs) return c.getresponse() except Exception as ex: info("HTTP %r", ex) self.closeconn(tid) try: c = self.getconn(tid) c.request(meth, path, headers=headers, **kwargs) return c.getresponse() except: info("http connection failed:\n" + traceback.format_exc()) if self.use_tls and not self.ssl_context: import ssl cert = ssl.get_server_certificate((self.web_host, self.web_port)) info("server certificate probably not trusted:\n" + cert) raise def listdir(self, path): if bad_good: path = dewin(path) zs = "%s%s/" if path else "%s%s" web_path = self.quotep(zs % (self.SRS, path)) + self.dsuf r = self.sendreq("GET", web_path, {}) if r.status != 200: self.closeconn() info("http error %s reading dir %r", r.status, web_path) raise FuseOSError(errno.ENOENT) ctype = r.getheader("Content-Type", "") if ctype == "application/json": parser = self.parse_jls # !rm.yes> elif ctype.startswith("text/html"): parser = self.parse_html # !rm.no> else: info("listdir on file (%s): %r", ctype, path) raise FuseOSError(errno.ENOENT) try: return parser(r) except: info("parser: %r\n%s", path, traceback.format_exc()) raise FuseOSError(errno.EIO) def download_file_range(self, path, ofs1, ofs2): if bad_good: path = dewin(path) web_path = self.quotep("%s%s" % (self.SRS, path)) + self.fsuf hdr_range = "bytes=%d-%d" % (ofs1, ofs2 - 1) t = "DL %4.0fK\033[36m%9d-%-9d\033[0m%r" info(t, (ofs2 - ofs1) / 1024.0, ofs1, ofs2 - 1, path) r = self.sendreq("GET", web_path, {"Range": hdr_range}) if r.status != http.client.PARTIAL_CONTENT: t = "http error %d reading file %r range %s in %s" info(t, r.status, web_path, hdr_range, rice_tid()) self.closeconn() raise FuseOSError(errno.EIO) return r.read() def parse_jls(self, sck): rsp = b"" while True: buf = sck.read(1024 * 32) if not buf: break rsp += buf rsp = json.loads(rsp.decode("utf-8")) ret = {} for statfun, nodes in [ [self.stat_dir, rsp["dirs"]], [self.stat_file, rsp["files"]], ]: for n in nodes: fname = unquote(n["href"].split("?")[0]).rstrip(b"/").decode("wtf-8") if bad_good: fname = enwin(fname) ret[fname] = statfun(n["ts"], n["sz"]) return ret # !rm.yes> #################################################################### #################################################################### def parse_cpp(self, sck): # https://a.ocv.me/pub/ ret = {} rem = b"" ptn = self.re_row while True: buf = sck.read(1024 * 32) if not buf: break buf = rem + buf rem = b"" idx = buf.rfind(b"\n") if idx >= 0: rem = buf[idx + 1 :] buf = buf[:idx] lines = buf.decode("utf-8").split("\n") for line in lines: m = ptn.match(line) if not m: continue ftype, furl, fname, fsize, fdate = m.groups() fname = furl.rstrip("/").split("/")[-1] fname = unquote(fname) fname = fname.decode("wtf-8") if bad_good: fname = enwin(fname) sz = 1 ts = 60 * 60 * 24 * 2 try: sz = int(fsize) ts = calendar.timegm(time.strptime(fdate, "%Y-%m-%d %H:%M:%S")) except: info("bad HTML or OS %r %r\n%r", fdate, fsize, line) # python cannot strptime(1959-01-01) on windows if ftype != "DIR" and "zip=crc" not in ftype: ret[fname] = self.stat_file(ts, sz) else: ret[fname] = self.stat_dir(ts, sz) return ret def parse_nginx(self, sck): # https://ocv.me/stuff/ "06-Feb-2015 15:43" ret = {} rem = b"" re_row = self.re_row while True: buf = sck.read(1024 * 32) if not buf: break buf = rem + buf rem = b"" idx = buf.rfind(b"\n") if idx >= 0: rem = buf[idx + 1 :] buf = buf[:idx] fdate = "" lines = buf.decode("utf-8").split("\n") for line in lines: m = re_row.match(line) if not m: continue furl, fname, day, smon, year, hm, fsize = m.groups() fname = furl.rstrip("/").split("/")[-1] fname = unquote(fname) fname = fname.decode("wtf-8") if bad_good: fname = enwin(fname) sz = 1 ts = 60 * 60 * 24 * 2 try: fdate = "%s-%02d-%s %s" % (year, MON3[smon], day, hm) ts = calendar.timegm(time.strptime(fdate, "%Y-%m-%d %H:%M")) sz = -1 if fsize == "-" else int(fsize) except: info("bad HTML or OS %r %r\n%r", fdate, fsize, line) if sz == -1: ret[fname] = self.stat_dir(ts, 4096) else: ret[fname] = self.stat_file(ts, sz) return ret def parse_iis(self, sck): # https://nedlasting.miljodirektoratet.no/miljodata/ " 9/28/2024 5:24 AM" # https://grandcanyon.usgs.gov/photos/Foodbase/CISP/ " 6/29/2012 3:12 PM" ret = {} rem = b"" re_row = self.re_row re_2nl = self.re_2nl while True: buf = sck.read(1024 * 32) if not buf: break buf = rem + buf rem = b"" buf = re_2nl.sub(b"\n", buf) idx = buf.rfind(b"\n") if idx >= 0: rem = buf[idx + 1 :] buf = buf[:idx] lines = buf.decode("utf-8").split("\n") for line in lines: m = re_row.match(line) if not m: continue mon, day, year, hm, xm, fsize, furl, fname = m.groups() fname = furl.rstrip("/").split("/")[-1] fname = unquote(fname) fname = fname.decode("wtf-8") if bad_good: fname = enwin(fname) sz = 1 ts = 60 * 60 * 24 * 2 fdate = "%s-%s-%s %s %s" % (year, mon, day, hm, xm) try: ts = calendar.timegm(time.strptime(fdate, "%Y-%m-%d %H:%M %p")) sz = -1 if fsize == "<dir>" else int(fsize) except: info("bad HTML or OS %r %r\n%r", fdate, fsize, line) if sz == -1: ret[fname] = self.stat_dir(ts, 4096) else: ret[fname] = self.stat_file(ts, sz) return ret #################################################################### #################################################################### # !rm.no> def stat_dir(self, ts, sz): return { "st_mode": stat.S_IFDIR | 0o555, "st_uid": 1000, "st_gid": 1000, "st_size": sz, "st_atime": ts, "st_mtime": ts, "st_ctime": ts, "st_blocks": int((sz + 511) / 512), } def stat_file(self, ts, sz): return { "st_mode": stat.S_IFREG | 0o444, "st_uid": 1000, "st_gid": 1000, "st_size": sz, "st_atime": ts, "st_mtime": ts, "st_ctime": ts, "st_blocks": int((sz + 511) / 512), } class CPPF(Operations): def __init__(self, ar): self.gw = Gateway(ar) self.junk_fh_ctr = 3 self.t_dircache = ar.cds self.n_dircache = ar.cdn self.n_filecache = ar.cf self.dircache = [] self.dircache_mtx = threading.Lock() self.filecache = [] self.filecache_mtx = threading.Lock() info("up") def _describe(self): msg = [] with self.filecache_mtx: for n, cn in enumerate(self.filecache): cache_path, cache1 = cn.tag cache2 = cache1 + len(cn.data) t = "\n{:<2} {:>7} {:>10}:{:<9} {}".format( n, len(cn.data), cache1, cache2, cache_path.replace("\r", "\\r").replace("\n", "\\n"), ) msg.append(t) return "".join(msg) def clean_dircache(self): """not threadsafe""" now = time.time() cutoff = 0 for cn in self.dircache: if now - cn.ts <= self.t_dircache: break cutoff += 1 if cutoff > 0: self.dircache = self.dircache[cutoff:] elif len(self.dircache) > self.n_dircache: self.dircache.pop(0) def get_cached_dir(self, dirpath): with self.dircache_mtx: for cn in self.dircache: if cn.tag == dirpath: if time.time() - cn.ts <= self.t_dircache: return cn break return None # !rm.yes> """ ,-------------------------------, g1>=c1, g2<=c2 |cache1 cache2| buf[g1-c1:(g1-c1)+(g2-g1)] `-------------------------------' ,---------------, |get1 get2| `---------------' __________________________________________________________________________ ,-------------------------------, g2<=c2, (g2>=c1) |cache1 cache2| cdr=buf[:g2-c1] `-------------------------------' dl car; g1-512K:c1 ,---------------, |get1 get2| `---------------' __________________________________________________________________________ ,-------------------------------, g1>=c1, (g1<=c2) |cache1 cache2| car=buf[c2-g1:] `-------------------------------' dl cdr; c2:c2+1M ,---------------, |get1 get2| `---------------' """ # !rm.no> def get_cached_file(self, path, get1, get2, file_sz): car = None cdr = None ncn = -1 if is_dbg: dbg("cache request %d:%d |%d|%s", get1, get2, file_sz, self._describe()) with self.filecache_mtx: for cn in self.filecache: ncn += 1 cache_path, cache1 = cn.tag if cache_path != path: continue cache2 = cache1 + len(cn.data) if get2 <= cache1 or get1 >= cache2: # request does not overlap with cached area at all continue if get1 < cache1 and get2 > cache2: # cached area does overlap, but must specifically contain # either the first or last byte in the requested range continue if get1 >= cache1 and get2 <= cache2: # keep cache entry alive by moving it to the end self.filecache = ( self.filecache[:ncn] + self.filecache[ncn + 1 :] + [cn] ) buf_ofs = get1 - cache1 buf_end = buf_ofs + (get2 - get1) dbg( "found all (#%d %d:%d |%d|) [%d:%d] = %d", ncn, cache1, cache2, len(cn.data), buf_ofs, buf_end, buf_end - buf_ofs, ) return cn.data[buf_ofs:buf_end] if get2 <= cache2: x = cn.data[: get2 - cache1] if not cdr or len(cdr) < len(x): dbg( "found cdr (#%d %d:%d |%d|) [:%d-%d] = [:%d] = %d", ncn, cache1, cache2, len(cn.data), get2, cache1, get2 - cache1, len(x), ) cdr = x continue if get1 >= cache1: x = cn.data[-(max(0, cache2 - get1)) :] if not car or len(car) < len(x): dbg( "found car (#%d %d:%d |%d|) [-(%d-%d):] = [-%d:] = %d", ncn, cache1, cache2, len(cn.data), cache2, get1, cache2 - get1, len(x), ) car = x continue msg = "cache fallthrough\n%d %d %d\n%d %d %d\n%d %d --\n%s" % ( get1, get2, get2 - get1, cache1, cache2, cache2 - cache1, get1 - cache1, get2 - cache2, self._describe(), ) info(msg) raise FuseOSError(errno.EIO) if car and cdr and len(car) + len(cdr) == get2 - get1: dbg(" have both") return car + cdr elif cdr and (not car or len(car) < len(cdr)): h_end = get1 + (get2 - get1) - len(cdr) h_ofs = min(get1, h_end - 0x80000) # 512k if h_ofs < 0: h_ofs = 0 buf_ofs = get1 - h_ofs if dbg: t = " cdr %d, car %d:%d |%d| [%d:]" dbg(t, len(cdr), h_ofs, h_end, h_end - h_ofs, buf_ofs) buf = self.gw.download_file_range(path, h_ofs, h_end) if len(buf) == h_end - h_ofs: ret = buf[buf_ofs:] + cdr else: ret = buf[get1 - h_ofs :] t = "remote truncated %d:%d to |%d|, will return |%d|" info(t, h_ofs, h_end, len(buf), len(ret)) elif car: h_ofs = get1 + len(car) if get2 < 0x100000: # already cached from 0 to 64k, now do ~64k plus 1 MiB h_end = max(get2, h_ofs + 0x100000) # 1m else: # after 1 MiB, bump window to 8 MiB h_end = max(get2, h_ofs + 0x800000) # 8m if h_end > file_sz: h_end = file_sz buf_ofs = (get2 - get1) - len(car) t = " car %d, cdr %d:%d |%d| [:%d]" dbg(t, len(car), h_ofs, h_end, h_end - h_ofs, buf_ofs) buf = self.gw.download_file_range(path, h_ofs, h_end) ret = car + buf[:buf_ofs] else: if get2 - get1 < 0x500000: # 5m # unless the request is for the last n bytes of the file, # grow the start to cache some stuff around the range if get2 < file_sz - 1: h_ofs = get1 - 0x40000 # 256k else: h_ofs = get1 - 0x10000 # 64k # likewise grow the end unless start is 0 if get1 >= 0x100000: h_end = get2 + 0x400000 # 4m elif get1 > 0: h_end = get2 + 0x100000 # 1m else: h_end = get2 + 0x10000 # 64k else: # big enough, doesn't need pads h_ofs = get1 h_end = get2 if h_ofs < 0: h_ofs = 0 if h_end > file_sz: h_end = file_sz buf_ofs = get1 - h_ofs buf_end = buf_ofs + get2 - get1 t = " %d:%d |%d| [%d:%d]" dbg(t, h_ofs, h_end, h_end - h_ofs, buf_ofs, buf_end) buf = self.gw.download_file_range(path, h_ofs, h_end) ret = buf[buf_ofs:buf_end] cn = CacheNode([path, h_ofs], buf) with self.filecache_mtx: if len(self.filecache) >= self.n_filecache: self.filecache = self.filecache[1:] + [cn] else: self.filecache.append(cn) return ret def _readdir(self, path, fh=None): path = path.strip("/") dbg("readdir %r [%s]", path, fh) ret = self.gw.listdir(path) if not self.n_dircache: return ret with self.dircache_mtx: cn = CacheNode(path, ret) self.dircache.append(cn) self.clean_dircache() # import pprint; pprint.pprint(ret) return ret def readdir(self, path, fh=None): return [".", ".."] + list(self._readdir(path, fh)) def read(self, path, length, offset, fh=None): req_max = 1024 * 1024 * 8 cache_max = 1024 * 1024 * 2 if length > req_max: # windows actually doing 240 MiB read calls, sausage info("truncate |%d| to %dMiB", length, req_max >> 20) length = req_max path = path.strip("/") ofs2 = offset + length file_sz = self.getattr(path)["st_size"] dbg("read %r |%d| %d:%d max %d", path, length, offset, ofs2, file_sz) if ofs2 > file_sz: ofs2 = file_sz dbg("truncate to |%d| :%d", ofs2 - offset, ofs2) if file_sz == 0 or offset >= ofs2: return b"" if self.n_filecache and length <= cache_max: ret = self.get_cached_file(path, offset, ofs2, file_sz) else: ret = self.gw.download_file_range(path, offset, ofs2) return ret # !rm.yes> fn = "cppf-{}-{}-{}".format(time.time(), offset, length) if False: with open(fn, "wb", len(ret)) as f: f.write(ret) elif self.n_filecache: ret2 = self.gw.download_file_range(path, offset, ofs2) if ret != ret2: info(fn) for v in [ret, ret2]: try: info(len(v)) except: info("uhh " + repr(v)) with open(fn + ".bad", "wb") as f: f.write(ret) with open(fn + ".good", "wb") as f: f.write(ret2) raise Exception("cache bork") return ret # !rm.no> def getattr(self, path, fh=None): dbg("getattr %r", path) if WINDOWS: path = enwin(path) # windows occasionally decodes f0xx to xx path = path.strip("/") if not path: ret = self.gw.stat_dir(time.time(), 4096) dbg("/=%r", ret) return ret try: dirpath, fname = path.rsplit("/", 1) except: dirpath = "" fname = path cn = self.get_cached_dir(dirpath) if cn: dents = cn.data else: dbg("cache miss") dents = self._readdir(dirpath) try: ret = dents[fname] dbg("s=%r", ret) return ret except: pass fun = info if MACOS and path.split("/")[-1].startswith("._"): fun = dbg fun("=ENOENT %r", path) raise FuseOSError(errno.ENOENT) access = None flush = None getxattr = None listxattr = None open = None opendir = None release = None releasedir = None statfs = None # !rm.yes> if False: # incorrect semantics but good for debugging stuff like samba and msys2 def access(self, path, mode): dbg("@@ access [{}] [{}]".format(path, mode)) return 1 if self.getattr(path) else 0 def flush(self, path, fh): dbg("@@ flush [{}] [{}]".format(path, fh)) return True def getxattr(self, *args): dbg("@@ getxattr [{}]".format("] [".join(str(x) for x in args))) return False def listxattr(self, *args): dbg("@@ listxattr [{}]".format("] [".join(str(x) for x in args))) return False def open(self, path, flags): dbg("@@ open [{}] [{}]".format(path, flags)) return 42 def opendir(self, fh): dbg("@@ opendir [{}]".format(fh)) return 69 def release(self, ino, fi): dbg("@@ release [{}] [{}]".format(ino, fi)) return True def releasedir(self, ino, fi): dbg("@@ releasedir [{}] [{}]".format(ino, fi)) return True def statfs(self, path): dbg("@@ statfs [{}]".format(path)) return {} # !rm.no> if sys.platform == "win32": # quick compat for /mingw64/bin/python3 (msys2) def _open(self, path): try: x = self.getattr(path) if x["st_mode"] <= 0: raise Exception() self.junk_fh_ctr += 1 if self.junk_fh_ctr > 32000: # TODO untested self.junk_fh_ctr = 4 return self.junk_fh_ctr except Exception as ex: info("open ERR %r", ex) raise FuseOSError(errno.ENOENT) def open(self, path, flags): dbg("open %r [%s]", path, flags) return self._open(path) def opendir(self, path): dbg("opendir %r", path) return self._open(path) def flush(self, path, fh): dbg("flush %r [%s]", path, fh) def release(self, ino, fi): dbg("release %r [%s]", ino, fi) def releasedir(self, ino, fi): dbg("releasedir %r [%s]", ino, fi) def access(self, path, mode): dbg("access %r [%s]", path, mode) try: x = self.getattr(path) if x["st_mode"] <= 0: raise Exception() except: raise FuseOSError(errno.ENOENT) class TheArgparseFormatter( argparse.RawTextHelpFormatter, argparse.ArgumentDefaultsHelpFormatter ): pass def main(): global info, dbg, is_dbg time.strptime("19970815", "%Y%m%d") # python#7980 ver = "{0}, v{1}".format(S_BUILD_DT, S_VERSION) if "--version" in sys.argv: print("partyfuse", ver) return # filecache helps for reads that are ~64k or smaller; # windows likes to use 4k and 64k so cache is important, # linux generally does 128k so the cache is still nice, # value is numChunks (1~8M each) to keep in the cache nf = 12 # dircache is always a boost, # only want to disable it for tests etc, cdn = 24 # max num dirs; keep larger than max dir depth; 0=disable cds = 1 # numsec until an entry goes stale where = "local directory" if WINDOWS: where += " or DRIVE:" ex_pre = "\n " + os.path.basename(__file__) + " " examples = ["http://192.168.1.69:3923/music/ ./music"] if WINDOWS: examples.append("http://192.168.1.69:3923/music/ M:") ap = argparse.ArgumentParser( formatter_class=TheArgparseFormatter, description="mount a copyparty server as a local filesystem -- " + ver, epilog="example:" + ex_pre + ex_pre.join(examples), ) # fmt: off ap.add_argument("base_url", type=str, help="remote copyparty URL to mount") ap.add_argument("local_path", type=str, help=where + " to mount it on") ap.add_argument("-a", metavar="PASSWORD", help="password or $filepath") # !rm.yes> ap.add_argument("--html", metavar="TYPE", default="", help="which HTML parser to use; cpp, nginx, iis") # !rm.no> ap2 = ap.add_argument_group("https/TLS") ap2.add_argument("-te", metavar="PEMFILE", help="certificate to expect/verify") ap2.add_argument("-td", action="store_true", help="disable certificate check") ap2 = ap.add_argument_group("cache/perf") ap2.add_argument("-cdn", metavar="DIRS", type=float, default=cdn, help="directory-cache, max num dirs; 0=disable") ap2.add_argument("-cds", metavar="SECS", type=float, default=cds, help="directory-cache, expiration time") ap2.add_argument("-cf", metavar="BLOCKS", type=int, default=nf, help="file cache; each block is <= 1 MiB") ap2 = ap.add_argument_group("logging") ap2.add_argument("-q", action="store_true", help="quiet") ap2.add_argument("-d", action="store_true", help="debug/verbose") ap2.add_argument("--slowterm", action="store_true", help="only most recent msgs; good for windows") ap2.add_argument("--logf", metavar="FILE", type=str, default="", help="log to FILE; enables --slowterm") ap2 = ap.add_argument_group("fuse") ap2.add_argument("--oth", action="store_true", help="tell FUSE to '-o allow_other'") ap2.add_argument("--nonempty", action="store_true", help="tell FUSE to '-o nonempty'") ar = ap.parse_args() # fmt: on if ar.logf: ar.slowterm = True # windows terminals are slow (cmd.exe, mintty) # otoh fancy_log beats RecentLog on linux logger = RecentLog(ar).put if ar.slowterm else fancy_log if ar.d: info = logger dbg = logger is_dbg = True elif not ar.q: info = logger if ar.a and ar.a.startswith("$"): fn = ar.a[1:] info("reading password from file %r", fn) with open(fn, "rb") as f: ar.a = f.read().decode("utf-8").strip() if WINDOWS: os.system("rem") for ch in '<>:"\\|?*': # microsoft maps illegal characters to f0xx # (e000 to f8ff is basic-plane private-use) bad_good[ch] = chr(ord(ch) + 0xF000) for n in range(0, 0x100): # map surrogateescape to another private-use area bad_good[chr(n + 0xDC00)] = chr(n + 0xF100) for k, v in bad_good.items(): good_bad[v] = k register_wtf8() args = {"foreground": True, "nothreads": True} if ar.oth: args["allow_other"] = True if ar.nonempty: args["nonempty"] = True FUSE(CPPF(ar), ar.local_path, encoding="wtf-8", **args) if __name__ == "__main__": main()