mirror of
https://github.com/9001/copyparty.git
synced 2025-08-17 00:52:16 -06:00
dircache size should exceed max dir depth, because the OS may periodically listdir all parents of current folder
1226 lines
37 KiB
Python
Executable file
1226 lines
37 KiB
Python
Executable file
#!/usr/bin/env python3
|
|
|
|
"""partyfuse: remote copyparty as a local filesystem"""
|
|
__author__ = "ed <copyparty@ocv.me>"
|
|
__copyright__ = 2019
|
|
__license__ = "MIT"
|
|
__url__ = "https://github.com/9001/copyparty/"
|
|
|
|
S_VERSION = "2.0"
|
|
S_BUILD_DT = "2024-10-01"
|
|
|
|
"""
|
|
mount a copyparty server (local or remote) as a filesystem
|
|
|
|
speeds:
|
|
1 GiB/s reading large files
|
|
27'000 files/sec: copy small files
|
|
700 folders/sec: copy small folders
|
|
|
|
usage:
|
|
python partyfuse.py http://192.168.1.69:3923/ ./music
|
|
|
|
dependencies:
|
|
python3 -m pip install --user fusepy # or grab it from the connect page
|
|
+ on Linux: sudo apk add fuse
|
|
+ on Macos: https://osxfuse.github.io/
|
|
+ on Windows: https://github.com/billziss-gh/winfsp/releases/latest
|
|
|
|
note:
|
|
you probably want to run this on windows clients:
|
|
https://github.com/9001/copyparty/blob/hovudstraum/contrib/explorer-nothumbs-nofoldertypes.reg
|
|
|
|
get server cert:
|
|
awk '/-BEGIN CERTIFICATE-/ {a=1} a; /-END CERTIFICATE-/{exit}' <(openssl s_client -connect 127.0.0.1:3923 </dev/null 2>/dev/null) >cert.pem
|
|
"""
|
|
|
|
|
|
import argparse
|
|
import calendar
|
|
import codecs
|
|
import errno
|
|
import json
|
|
import os
|
|
import platform
|
|
import re
|
|
import stat
|
|
import struct
|
|
import sys
|
|
import threading
|
|
import time
|
|
import traceback
|
|
import urllib.parse
|
|
from datetime import datetime, timezone
|
|
from urllib.parse import quote_from_bytes as quote
|
|
from urllib.parse import unquote_to_bytes as unquote
|
|
|
|
import builtins
|
|
import http.client
|
|
|
|
WINDOWS = sys.platform == "win32"
|
|
MACOS = platform.system() == "Darwin"
|
|
UTC = timezone.utc
|
|
|
|
# !rm.yes>
|
|
MON3S = "Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec"
|
|
MON3 = {b: a for a, b in enumerate(MON3S.split(), 1)}
|
|
# !rm.no>
|
|
|
|
|
|
def print(*args, **kwargs):
|
|
try:
|
|
builtins.print(*list(args), **kwargs)
|
|
except:
|
|
builtins.print(termsafe(" ".join(str(x) for x in args)), **kwargs)
|
|
|
|
|
|
print(
|
|
"{} v{} @ {}".format(
|
|
platform.python_implementation(),
|
|
".".join([str(x) for x in sys.version_info]),
|
|
sys.executable,
|
|
)
|
|
)
|
|
|
|
|
|
def nullfun(*a):
|
|
pass
|
|
|
|
|
|
info = dbg = nullfun
|
|
is_dbg = False
|
|
|
|
|
|
try:
|
|
from fuse import FUSE, FuseOSError, Operations
|
|
except:
|
|
if WINDOWS:
|
|
libfuse = "install https://github.com/billziss-gh/winfsp/releases/latest"
|
|
elif MACOS:
|
|
libfuse = "install https://osxfuse.github.io/"
|
|
else:
|
|
libfuse = "apt install libfuse3-3\n modprobe fuse"
|
|
|
|
m = """\033[33m
|
|
could not import fuse; these may help:
|
|
{} -m pip install --user fusepy
|
|
{}
|
|
\033[0m"""
|
|
print(m.format(sys.executable, libfuse))
|
|
raise
|
|
|
|
|
|
def termsafe(txt):
|
|
enc = sys.stdout.encoding
|
|
try:
|
|
return txt.encode(enc, "backslashreplace").decode(enc)
|
|
except:
|
|
return txt.encode(enc, "replace").decode(enc)
|
|
|
|
|
|
def threadless_log(fmt, *a):
|
|
fmt += "\n"
|
|
print(fmt % a if a else fmt, end="")
|
|
|
|
|
|
riced_tids = {}
|
|
|
|
|
|
def rice_tid():
|
|
tid = threading.current_thread().ident
|
|
try:
|
|
return riced_tids[tid]
|
|
except:
|
|
c = struct.unpack(b"B" * 5, struct.pack(b">Q", tid)[-5:])
|
|
ret = "".join("\033[1;37;48;5;%dm%02x" % (x, x) for x in c) + "\033[0m"
|
|
riced_tids[tid] = ret
|
|
return ret
|
|
|
|
|
|
def fancy_log(fmt, *a):
|
|
msg = fmt % a if a else fmt
|
|
print("%10.6f %s %s\n" % (time.time() % 900, rice_tid(), msg), end="")
|
|
|
|
|
|
def register_wtf8():
|
|
def wtf8_enc(text):
|
|
return str(text).encode("utf-8", "surrogateescape"), len(text)
|
|
|
|
def wtf8_dec(binary):
|
|
return bytes(binary).decode("utf-8", "surrogateescape"), len(binary)
|
|
|
|
def wtf8_search(encoding_name):
|
|
return codecs.CodecInfo(wtf8_enc, wtf8_dec, name="wtf-8")
|
|
|
|
codecs.register(wtf8_search)
|
|
|
|
|
|
bad_good = {}
|
|
good_bad = {}
|
|
|
|
|
|
def enwin(txt):
|
|
return "".join([bad_good.get(x, x) for x in txt])
|
|
|
|
|
|
def dewin(txt):
|
|
return "".join([good_bad.get(x, x) for x in txt])
|
|
|
|
|
|
class RecentLog(object):
|
|
def __init__(self, ar):
|
|
self.ar = ar
|
|
self.mtx = threading.Lock()
|
|
self.f = open(ar.logf, "wb") if ar.logf else None
|
|
self.q = []
|
|
|
|
thr = threading.Thread(target=self.printer)
|
|
thr.daemon = True
|
|
thr.start()
|
|
|
|
def put(self, fmt, *a):
|
|
msg = fmt % a if a else fmt
|
|
msg = "%10.6f %s %s\n" % (time.time() % 900, rice_tid(), msg)
|
|
if self.f:
|
|
zd = datetime.now(UTC)
|
|
fmsg = "%d-%04d-%06d.%06d %s" % (
|
|
zd.year,
|
|
zd.month * 100 + zd.day,
|
|
(zd.hour * 100 + zd.minute) * 100 + zd.second,
|
|
zd.microsecond,
|
|
msg,
|
|
)
|
|
self.f.write(fmsg.encode("utf-8"))
|
|
|
|
with self.mtx:
|
|
self.q.append(msg)
|
|
if len(self.q) > 200:
|
|
self.q = self.q[-50:]
|
|
|
|
def printer(self):
|
|
while True:
|
|
time.sleep(0.05)
|
|
with self.mtx:
|
|
q = self.q
|
|
if not q:
|
|
continue
|
|
|
|
self.q = []
|
|
|
|
print("".join(q), end="")
|
|
|
|
|
|
# [windows/cmd/cpy3] python dev\copyparty\bin\partyfuse.py q: http://192.168.1.159:1234/
|
|
# [windows/cmd/msys2] C:\msys64\mingw64\bin\python3 dev\copyparty\bin\partyfuse.py q: http://192.168.1.159:1234/
|
|
# [windows/mty/msys2] /mingw64/bin/python3 /c/Users/ed/dev/copyparty/bin/partyfuse.py q: http://192.168.1.159:1234/
|
|
#
|
|
# [windows] find /q/music/albums/Phant*24bit -printf '%s %p\n' | sort -n | tail -n 8 | sed -r 's/^[0-9]+ //' | while IFS= read -r x; do dd if="$x" of=/dev/null bs=4k count=8192 & done
|
|
# [alpine] ll t; for x in t/2020_0724_16{2,3}*; do dd if="$x" of=/dev/null bs=4k count=10240 & done
|
|
#
|
|
# 72.4983 windows mintty msys2 fancy_log
|
|
# 219.5781 windows cmd msys2 fancy_log
|
|
# nope.avi windows cmd cpy3 fancy_log
|
|
# 9.8817 windows mintty msys2 RecentLog 200 50 0.1
|
|
# 10.2241 windows cmd cpy3 RecentLog 200 50 0.1
|
|
# 9.8494 windows cmd msys2 RecentLog 200 50 0.1
|
|
# 7.8061 windows mintty msys2 fancy_log <info-only>
|
|
# 7.9961 windows mintty msys2 RecentLog <info-only>
|
|
# 4.2603 alpine xfce4 cpy3 RecentLog
|
|
# 4.1538 alpine xfce4 cpy3 fancy_log
|
|
# 3.1742 alpine urxvt cpy3 fancy_log
|
|
|
|
|
|
def get_tid():
|
|
return threading.current_thread().ident
|
|
|
|
|
|
def html_dec(txt):
|
|
return (
|
|
txt.replace("<", "<")
|
|
.replace(">", ">")
|
|
.replace(""", '"')
|
|
.replace(" ", "\r")
|
|
.replace(" ", "\n")
|
|
.replace("&", "&")
|
|
)
|
|
|
|
|
|
class CacheNode(object):
|
|
def __init__(self, tag, data):
|
|
self.tag = tag
|
|
self.data = data
|
|
self.ts = time.time()
|
|
|
|
|
|
class Gateway(object):
|
|
def __init__(self, ar):
|
|
zs = ar.base_url
|
|
if "://" not in zs:
|
|
zs = "http://" + zs
|
|
|
|
self.base_url = zs
|
|
self.password = ar.a
|
|
|
|
ui = urllib.parse.urlparse(zs)
|
|
self.web_root = ui.path.strip("/")
|
|
self.SRS = "/%s/" % (self.web_root,) if self.web_root else "/"
|
|
try:
|
|
self.web_host, self.web_port = ui.netloc.split(":")
|
|
self.web_port = int(self.web_port)
|
|
except:
|
|
self.web_host = ui.netloc
|
|
if ui.scheme == "http":
|
|
self.web_port = 80
|
|
elif ui.scheme == "https":
|
|
self.web_port = 443
|
|
else:
|
|
raise Exception("bad url?")
|
|
|
|
self.ssl_context = None
|
|
self.use_tls = ui.scheme.lower() == "https"
|
|
if self.use_tls:
|
|
import ssl
|
|
|
|
if ar.td:
|
|
self.ssl_context = ssl._create_unverified_context()
|
|
elif ar.te:
|
|
self.ssl_context = ssl.SSLContext(ssl.PROTOCOL_TLS)
|
|
self.ssl_context.load_verify_locations(ar.te)
|
|
|
|
self.conns = {}
|
|
|
|
self.fsuf = "?raw"
|
|
self.dsuf = "?ls<&dots"
|
|
|
|
# !rm.yes>
|
|
if not ar.html:
|
|
self.parse_html = None
|
|
|
|
elif ar.html == "cpp":
|
|
self.parse_html = self.parse_cpp
|
|
self.dsuf = "?lt&dots"
|
|
self.re_row = re.compile(
|
|
r'^<tr><td>(-|DIR|<a [^<]+</a>)</td><td><a[^>]* href="([^"]+)"[^>]*>([^<]+)</a></td><td>([^<]+)</td><td>.*</td><td>([^<]+)</td></tr>$'
|
|
)
|
|
|
|
elif ar.html == "nginx":
|
|
self.parse_html = self.parse_nginx
|
|
self.fsuf = ""
|
|
self.dsuf = ""
|
|
self.re_row = re.compile(
|
|
r'^<a href="([^"]+)">([^<]+)</a> *([0-9]{2})-([A-Z][a-z]{2})-([0-9]{4}) ([0-9]{2}:[0-9]{2}) *(-|[0-9]+)\r?$'
|
|
)
|
|
|
|
elif ar.html == "iis":
|
|
self.parse_html = self.parse_iis
|
|
self.fsuf = ""
|
|
self.dsuf = ""
|
|
self.re_2nl = re.compile(br"<br>|</pre>")
|
|
self.re_row = re.compile(
|
|
r'^ *([0-9]{1,2})/([0-9]{1,2})/([0-9]{4}) {1,2}([0-9]{1,2}:[0-9]{2}) ([AP]M) +(<dir>|[0-9]+) <A HREF="([^"]+)">([^<>]+)</A>$'
|
|
)
|
|
|
|
else:
|
|
raise Exception("unknown HTML dialect: [%s]" % (ar.html,))
|
|
# !rm.no>
|
|
|
|
def quotep(self, path):
|
|
path = path.encode("wtf-8")
|
|
return quote(path, safe="/")
|
|
|
|
def getconn(self, tid=None):
|
|
tid = tid or get_tid()
|
|
try:
|
|
return self.conns[tid]
|
|
except:
|
|
info("new conn [{}] [{}]".format(self.web_host, self.web_port))
|
|
|
|
args = {}
|
|
if not self.use_tls:
|
|
C = http.client.HTTPConnection
|
|
else:
|
|
C = http.client.HTTPSConnection
|
|
if self.ssl_context:
|
|
args = {"context": self.ssl_context}
|
|
|
|
conn = C(self.web_host, self.web_port, timeout=260, **args)
|
|
|
|
self.conns[tid] = conn
|
|
return conn
|
|
|
|
def closeconn(self, tid=None):
|
|
tid = tid or get_tid()
|
|
try:
|
|
self.conns[tid].close()
|
|
del self.conns[tid]
|
|
except:
|
|
pass
|
|
|
|
def sendreq(self, meth, path, headers, **kwargs):
|
|
tid = get_tid()
|
|
if self.password:
|
|
headers["Cookie"] = "=".join(["cppwd", self.password])
|
|
|
|
try:
|
|
c = self.getconn(tid)
|
|
c.request(meth, path, headers=headers, **kwargs)
|
|
return c.getresponse()
|
|
except Exception as ex:
|
|
info("HTTP %r", ex)
|
|
|
|
self.closeconn(tid)
|
|
try:
|
|
c = self.getconn(tid)
|
|
c.request(meth, path, headers=headers, **kwargs)
|
|
return c.getresponse()
|
|
except:
|
|
info("http connection failed:\n" + traceback.format_exc())
|
|
if self.use_tls and not self.ssl_context:
|
|
import ssl
|
|
|
|
cert = ssl.get_server_certificate((self.web_host, self.web_port))
|
|
info("server certificate probably not trusted:\n" + cert)
|
|
|
|
raise
|
|
|
|
def listdir(self, path):
|
|
if bad_good:
|
|
path = dewin(path)
|
|
|
|
zs = "%s%s/" if path else "%s%s"
|
|
web_path = self.quotep(zs % (self.SRS, path)) + self.dsuf
|
|
r = self.sendreq("GET", web_path, {})
|
|
if r.status != 200:
|
|
self.closeconn()
|
|
info("http error %s reading dir %r", r.status, web_path)
|
|
raise FuseOSError(errno.ENOENT)
|
|
|
|
ctype = r.getheader("Content-Type", "")
|
|
if ctype == "application/json":
|
|
parser = self.parse_jls
|
|
# !rm.yes>
|
|
elif ctype.startswith("text/html"):
|
|
parser = self.parse_html
|
|
# !rm.no>
|
|
else:
|
|
info("listdir on file (%s): %r", ctype, path)
|
|
raise FuseOSError(errno.ENOENT)
|
|
|
|
try:
|
|
return parser(r)
|
|
except:
|
|
info("parser: %r\n%s", path, traceback.format_exc())
|
|
raise FuseOSError(errno.EIO)
|
|
|
|
def download_file_range(self, path, ofs1, ofs2):
|
|
if bad_good:
|
|
path = dewin(path)
|
|
|
|
web_path = self.quotep("%s%s" % (self.SRS, path)) + self.fsuf
|
|
hdr_range = "bytes=%d-%d" % (ofs1, ofs2 - 1)
|
|
|
|
t = "DL %4.0fK\033[36m%9d-%-9d\033[0m%r"
|
|
info(t, (ofs2 - ofs1) / 1024.0, ofs1, ofs2 - 1, path)
|
|
|
|
r = self.sendreq("GET", web_path, {"Range": hdr_range})
|
|
if r.status != http.client.PARTIAL_CONTENT:
|
|
t = "http error %d reading file %r range %s in %s"
|
|
info(t, r.status, web_path, hdr_range, rice_tid())
|
|
self.closeconn()
|
|
raise FuseOSError(errno.EIO)
|
|
|
|
return r.read()
|
|
|
|
def parse_jls(self, sck):
|
|
rsp = b""
|
|
while True:
|
|
buf = sck.read(1024 * 32)
|
|
if not buf:
|
|
break
|
|
rsp += buf
|
|
|
|
rsp = json.loads(rsp.decode("utf-8"))
|
|
ret = {}
|
|
for statfun, nodes in [
|
|
[self.stat_dir, rsp["dirs"]],
|
|
[self.stat_file, rsp["files"]],
|
|
]:
|
|
for n in nodes:
|
|
fname = unquote(n["href"].split("?")[0]).rstrip(b"/").decode("wtf-8")
|
|
if bad_good:
|
|
fname = enwin(fname)
|
|
|
|
ret[fname] = statfun(n["ts"], n["sz"])
|
|
|
|
return ret
|
|
|
|
# !rm.yes>
|
|
####################################################################
|
|
####################################################################
|
|
|
|
def parse_cpp(self, sck):
|
|
# https://a.ocv.me/pub/
|
|
|
|
ret = {}
|
|
rem = b""
|
|
ptn = self.re_row
|
|
|
|
while True:
|
|
buf = sck.read(1024 * 32)
|
|
if not buf:
|
|
break
|
|
|
|
buf = rem + buf
|
|
rem = b""
|
|
idx = buf.rfind(b"\n")
|
|
if idx >= 0:
|
|
rem = buf[idx + 1 :]
|
|
buf = buf[:idx]
|
|
|
|
lines = buf.decode("utf-8").split("\n")
|
|
for line in lines:
|
|
m = ptn.match(line)
|
|
if not m:
|
|
continue
|
|
|
|
ftype, furl, fname, fsize, fdate = m.groups()
|
|
fname = furl.rstrip("/").split("/")[-1]
|
|
fname = unquote(fname)
|
|
fname = fname.decode("wtf-8")
|
|
if bad_good:
|
|
fname = enwin(fname)
|
|
|
|
sz = 1
|
|
ts = 60 * 60 * 24 * 2
|
|
try:
|
|
sz = int(fsize)
|
|
ts = calendar.timegm(time.strptime(fdate, "%Y-%m-%d %H:%M:%S"))
|
|
except:
|
|
info("bad HTML or OS %r %r\n%r", fdate, fsize, line)
|
|
# python cannot strptime(1959-01-01) on windows
|
|
|
|
if ftype != "DIR" and "zip=crc" not in ftype:
|
|
ret[fname] = self.stat_file(ts, sz)
|
|
else:
|
|
ret[fname] = self.stat_dir(ts, sz)
|
|
|
|
return ret
|
|
|
|
def parse_nginx(self, sck):
|
|
# https://ocv.me/stuff/ "06-Feb-2015 15:43"
|
|
|
|
ret = {}
|
|
rem = b""
|
|
re_row = self.re_row
|
|
|
|
while True:
|
|
buf = sck.read(1024 * 32)
|
|
if not buf:
|
|
break
|
|
|
|
buf = rem + buf
|
|
rem = b""
|
|
idx = buf.rfind(b"\n")
|
|
if idx >= 0:
|
|
rem = buf[idx + 1 :]
|
|
buf = buf[:idx]
|
|
|
|
fdate = ""
|
|
lines = buf.decode("utf-8").split("\n")
|
|
for line in lines:
|
|
m = re_row.match(line)
|
|
if not m:
|
|
continue
|
|
|
|
furl, fname, day, smon, year, hm, fsize = m.groups()
|
|
fname = furl.rstrip("/").split("/")[-1]
|
|
fname = unquote(fname)
|
|
fname = fname.decode("wtf-8")
|
|
if bad_good:
|
|
fname = enwin(fname)
|
|
|
|
sz = 1
|
|
ts = 60 * 60 * 24 * 2
|
|
try:
|
|
fdate = "%s-%02d-%s %s" % (year, MON3[smon], day, hm)
|
|
ts = calendar.timegm(time.strptime(fdate, "%Y-%m-%d %H:%M"))
|
|
sz = -1 if fsize == "-" else int(fsize)
|
|
except:
|
|
info("bad HTML or OS %r %r\n%r", fdate, fsize, line)
|
|
|
|
if sz == -1:
|
|
ret[fname] = self.stat_dir(ts, 4096)
|
|
else:
|
|
ret[fname] = self.stat_file(ts, sz)
|
|
|
|
return ret
|
|
|
|
def parse_iis(self, sck):
|
|
# https://nedlasting.miljodirektoratet.no/miljodata/ " 9/28/2024 5:24 AM"
|
|
# https://grandcanyon.usgs.gov/photos/Foodbase/CISP/ " 6/29/2012 3:12 PM"
|
|
|
|
ret = {}
|
|
rem = b""
|
|
re_row = self.re_row
|
|
re_2nl = self.re_2nl
|
|
|
|
while True:
|
|
buf = sck.read(1024 * 32)
|
|
if not buf:
|
|
break
|
|
|
|
buf = rem + buf
|
|
rem = b""
|
|
buf = re_2nl.sub(b"\n", buf)
|
|
idx = buf.rfind(b"\n")
|
|
if idx >= 0:
|
|
rem = buf[idx + 1 :]
|
|
buf = buf[:idx]
|
|
|
|
lines = buf.decode("utf-8").split("\n")
|
|
for line in lines:
|
|
m = re_row.match(line)
|
|
if not m:
|
|
continue
|
|
|
|
mon, day, year, hm, xm, fsize, furl, fname = m.groups()
|
|
fname = furl.rstrip("/").split("/")[-1]
|
|
fname = unquote(fname)
|
|
fname = fname.decode("wtf-8")
|
|
if bad_good:
|
|
fname = enwin(fname)
|
|
|
|
sz = 1
|
|
ts = 60 * 60 * 24 * 2
|
|
fdate = "%s-%s-%s %s %s" % (year, mon, day, hm, xm)
|
|
try:
|
|
ts = calendar.timegm(time.strptime(fdate, "%Y-%m-%d %H:%M %p"))
|
|
sz = -1 if fsize == "<dir>" else int(fsize)
|
|
except:
|
|
info("bad HTML or OS %r %r\n%r", fdate, fsize, line)
|
|
|
|
if sz == -1:
|
|
ret[fname] = self.stat_dir(ts, 4096)
|
|
else:
|
|
ret[fname] = self.stat_file(ts, sz)
|
|
|
|
return ret
|
|
|
|
####################################################################
|
|
####################################################################
|
|
# !rm.no>
|
|
|
|
def stat_dir(self, ts, sz):
|
|
return {
|
|
"st_mode": stat.S_IFDIR | 0o555,
|
|
"st_uid": 1000,
|
|
"st_gid": 1000,
|
|
"st_size": sz,
|
|
"st_atime": ts,
|
|
"st_mtime": ts,
|
|
"st_ctime": ts,
|
|
"st_blocks": int((sz + 511) / 512),
|
|
}
|
|
|
|
def stat_file(self, ts, sz):
|
|
return {
|
|
"st_mode": stat.S_IFREG | 0o444,
|
|
"st_uid": 1000,
|
|
"st_gid": 1000,
|
|
"st_size": sz,
|
|
"st_atime": ts,
|
|
"st_mtime": ts,
|
|
"st_ctime": ts,
|
|
"st_blocks": int((sz + 511) / 512),
|
|
}
|
|
|
|
|
|
class CPPF(Operations):
|
|
def __init__(self, ar):
|
|
self.gw = Gateway(ar)
|
|
self.junk_fh_ctr = 3
|
|
self.t_dircache = ar.cds
|
|
self.n_dircache = ar.cdn
|
|
self.n_filecache = ar.cf
|
|
|
|
self.dircache = []
|
|
self.dircache_mtx = threading.Lock()
|
|
|
|
self.filecache = []
|
|
self.filecache_mtx = threading.Lock()
|
|
|
|
info("up")
|
|
|
|
def _describe(self):
|
|
msg = []
|
|
with self.filecache_mtx:
|
|
for n, cn in enumerate(self.filecache):
|
|
cache_path, cache1 = cn.tag
|
|
cache2 = cache1 + len(cn.data)
|
|
t = "\n{:<2} {:>7} {:>10}:{:<9} {}".format(
|
|
n,
|
|
len(cn.data),
|
|
cache1,
|
|
cache2,
|
|
cache_path.replace("\r", "\\r").replace("\n", "\\n"),
|
|
)
|
|
msg.append(t)
|
|
return "".join(msg)
|
|
|
|
def clean_dircache(self):
|
|
"""not threadsafe"""
|
|
now = time.time()
|
|
cutoff = 0
|
|
for cn in self.dircache:
|
|
if now - cn.ts <= self.t_dircache:
|
|
break
|
|
cutoff += 1
|
|
|
|
if cutoff > 0:
|
|
self.dircache = self.dircache[cutoff:]
|
|
elif len(self.dircache) > self.n_dircache:
|
|
self.dircache.pop(0)
|
|
|
|
def get_cached_dir(self, dirpath):
|
|
with self.dircache_mtx:
|
|
for cn in self.dircache:
|
|
if cn.tag == dirpath:
|
|
if time.time() - cn.ts <= self.t_dircache:
|
|
return cn
|
|
break
|
|
return None
|
|
|
|
# !rm.yes>
|
|
"""
|
|
,-------------------------------, g1>=c1, g2<=c2
|
|
|cache1 cache2| buf[g1-c1:(g1-c1)+(g2-g1)]
|
|
`-------------------------------'
|
|
,---------------,
|
|
|get1 get2|
|
|
`---------------'
|
|
__________________________________________________________________________
|
|
|
|
,-------------------------------, g2<=c2, (g2>=c1)
|
|
|cache1 cache2| cdr=buf[:g2-c1]
|
|
`-------------------------------' dl car; g1-512K:c1
|
|
,---------------,
|
|
|get1 get2|
|
|
`---------------'
|
|
__________________________________________________________________________
|
|
|
|
,-------------------------------, g1>=c1, (g1<=c2)
|
|
|cache1 cache2| car=buf[c2-g1:]
|
|
`-------------------------------' dl cdr; c2:c2+1M
|
|
,---------------,
|
|
|get1 get2|
|
|
`---------------'
|
|
"""
|
|
# !rm.no>
|
|
|
|
def get_cached_file(self, path, get1, get2, file_sz):
|
|
car = None
|
|
cdr = None
|
|
ncn = -1
|
|
if is_dbg:
|
|
dbg("cache request %d:%d |%d|%s", get1, get2, file_sz, self._describe())
|
|
with self.filecache_mtx:
|
|
for cn in self.filecache:
|
|
ncn += 1
|
|
|
|
cache_path, cache1 = cn.tag
|
|
if cache_path != path:
|
|
continue
|
|
|
|
cache2 = cache1 + len(cn.data)
|
|
if get2 <= cache1 or get1 >= cache2:
|
|
# request does not overlap with cached area at all
|
|
continue
|
|
|
|
if get1 < cache1 and get2 > cache2:
|
|
# cached area does overlap, but must specifically contain
|
|
# either the first or last byte in the requested range
|
|
continue
|
|
|
|
if get1 >= cache1 and get2 <= cache2:
|
|
# keep cache entry alive by moving it to the end
|
|
self.filecache = (
|
|
self.filecache[:ncn] + self.filecache[ncn + 1 :] + [cn]
|
|
)
|
|
buf_ofs = get1 - cache1
|
|
buf_end = buf_ofs + (get2 - get1)
|
|
dbg(
|
|
"found all (#%d %d:%d |%d|) [%d:%d] = %d",
|
|
ncn,
|
|
cache1,
|
|
cache2,
|
|
len(cn.data),
|
|
buf_ofs,
|
|
buf_end,
|
|
buf_end - buf_ofs,
|
|
)
|
|
return cn.data[buf_ofs:buf_end]
|
|
|
|
if get2 <= cache2:
|
|
x = cn.data[: get2 - cache1]
|
|
if not cdr or len(cdr) < len(x):
|
|
dbg(
|
|
"found cdr (#%d %d:%d |%d|) [:%d-%d] = [:%d] = %d",
|
|
ncn,
|
|
cache1,
|
|
cache2,
|
|
len(cn.data),
|
|
get2,
|
|
cache1,
|
|
get2 - cache1,
|
|
len(x),
|
|
)
|
|
cdr = x
|
|
|
|
continue
|
|
|
|
if get1 >= cache1:
|
|
x = cn.data[-(max(0, cache2 - get1)) :]
|
|
if not car or len(car) < len(x):
|
|
dbg(
|
|
"found car (#%d %d:%d |%d|) [-(%d-%d):] = [-%d:] = %d",
|
|
ncn,
|
|
cache1,
|
|
cache2,
|
|
len(cn.data),
|
|
cache2,
|
|
get1,
|
|
cache2 - get1,
|
|
len(x),
|
|
)
|
|
car = x
|
|
|
|
continue
|
|
|
|
msg = "cache fallthrough\n%d %d %d\n%d %d %d\n%d %d --\n%s" % (
|
|
get1,
|
|
get2,
|
|
get2 - get1,
|
|
cache1,
|
|
cache2,
|
|
cache2 - cache1,
|
|
get1 - cache1,
|
|
get2 - cache2,
|
|
self._describe(),
|
|
)
|
|
info(msg)
|
|
raise FuseOSError(errno.EIO)
|
|
|
|
if car and cdr and len(car) + len(cdr) == get2 - get1:
|
|
dbg("<cache> have both")
|
|
return car + cdr
|
|
|
|
elif cdr and (not car or len(car) < len(cdr)):
|
|
h_end = get1 + (get2 - get1) - len(cdr)
|
|
h_ofs = min(get1, h_end - 0x80000) # 512k
|
|
|
|
if h_ofs < 0:
|
|
h_ofs = 0
|
|
|
|
buf_ofs = get1 - h_ofs
|
|
|
|
if dbg:
|
|
t = "<cache> cdr %d, car %d:%d |%d| [%d:]"
|
|
dbg(t, len(cdr), h_ofs, h_end, h_end - h_ofs, buf_ofs)
|
|
|
|
buf = self.gw.download_file_range(path, h_ofs, h_end)
|
|
if len(buf) == h_end - h_ofs:
|
|
ret = buf[buf_ofs:] + cdr
|
|
else:
|
|
ret = buf[get1 - h_ofs :]
|
|
t = "remote truncated %d:%d to |%d|, will return |%d|"
|
|
info(t, h_ofs, h_end, len(buf), len(ret))
|
|
|
|
elif car:
|
|
h_ofs = get1 + len(car)
|
|
if get2 < 0x100000:
|
|
# already cached from 0 to 64k, now do ~64k plus 1 MiB
|
|
h_end = max(get2, h_ofs + 0x100000) # 1m
|
|
else:
|
|
# after 1 MiB, bump window to 8 MiB
|
|
h_end = max(get2, h_ofs + 0x800000) # 8m
|
|
|
|
if h_end > file_sz:
|
|
h_end = file_sz
|
|
|
|
buf_ofs = (get2 - get1) - len(car)
|
|
|
|
t = "<cache> car %d, cdr %d:%d |%d| [:%d]"
|
|
dbg(t, len(car), h_ofs, h_end, h_end - h_ofs, buf_ofs)
|
|
|
|
buf = self.gw.download_file_range(path, h_ofs, h_end)
|
|
ret = car + buf[:buf_ofs]
|
|
|
|
else:
|
|
if get2 - get1 < 0x500000: # 5m
|
|
# unless the request is for the last n bytes of the file,
|
|
# grow the start to cache some stuff around the range
|
|
if get2 < file_sz - 1:
|
|
h_ofs = get1 - 0x40000 # 256k
|
|
else:
|
|
h_ofs = get1 - 0x10000 # 64k
|
|
|
|
# likewise grow the end unless start is 0
|
|
if get1 >= 0x100000:
|
|
h_end = get2 + 0x400000 # 4m
|
|
elif get1 > 0:
|
|
h_end = get2 + 0x100000 # 1m
|
|
else:
|
|
h_end = get2 + 0x10000 # 64k
|
|
else:
|
|
# big enough, doesn't need pads
|
|
h_ofs = get1
|
|
h_end = get2
|
|
|
|
if h_ofs < 0:
|
|
h_ofs = 0
|
|
|
|
if h_end > file_sz:
|
|
h_end = file_sz
|
|
|
|
buf_ofs = get1 - h_ofs
|
|
buf_end = buf_ofs + get2 - get1
|
|
|
|
t = "<cache> %d:%d |%d| [%d:%d]"
|
|
dbg(t, h_ofs, h_end, h_end - h_ofs, buf_ofs, buf_end)
|
|
|
|
buf = self.gw.download_file_range(path, h_ofs, h_end)
|
|
ret = buf[buf_ofs:buf_end]
|
|
|
|
cn = CacheNode([path, h_ofs], buf)
|
|
with self.filecache_mtx:
|
|
if len(self.filecache) >= self.n_filecache:
|
|
self.filecache = self.filecache[1:] + [cn]
|
|
else:
|
|
self.filecache.append(cn)
|
|
|
|
return ret
|
|
|
|
def _readdir(self, path, fh=None):
|
|
path = path.strip("/")
|
|
dbg("readdir %r [%s]", path, fh)
|
|
|
|
ret = self.gw.listdir(path)
|
|
if not self.n_dircache:
|
|
return ret
|
|
|
|
with self.dircache_mtx:
|
|
cn = CacheNode(path, ret)
|
|
self.dircache.append(cn)
|
|
self.clean_dircache()
|
|
|
|
# import pprint; pprint.pprint(ret)
|
|
return ret
|
|
|
|
def readdir(self, path, fh=None):
|
|
return [".", ".."] + list(self._readdir(path, fh))
|
|
|
|
def read(self, path, length, offset, fh=None):
|
|
req_max = 1024 * 1024 * 8
|
|
cache_max = 1024 * 1024 * 2
|
|
if length > req_max:
|
|
# windows actually doing 240 MiB read calls, sausage
|
|
info("truncate |%d| to %dMiB", length, req_max >> 20)
|
|
length = req_max
|
|
|
|
path = path.strip("/")
|
|
ofs2 = offset + length
|
|
file_sz = self.getattr(path)["st_size"]
|
|
dbg("read %r |%d| %d:%d max %d", path, length, offset, ofs2, file_sz)
|
|
|
|
if ofs2 > file_sz:
|
|
ofs2 = file_sz
|
|
dbg("truncate to |%d| :%d", ofs2 - offset, ofs2)
|
|
|
|
if file_sz == 0 or offset >= ofs2:
|
|
return b""
|
|
|
|
if self.n_filecache and length <= cache_max:
|
|
ret = self.get_cached_file(path, offset, ofs2, file_sz)
|
|
else:
|
|
ret = self.gw.download_file_range(path, offset, ofs2)
|
|
|
|
return ret
|
|
|
|
# !rm.yes>
|
|
fn = "cppf-{}-{}-{}".format(time.time(), offset, length)
|
|
if False:
|
|
with open(fn, "wb", len(ret)) as f:
|
|
f.write(ret)
|
|
elif self.n_filecache:
|
|
ret2 = self.gw.download_file_range(path, offset, ofs2)
|
|
if ret != ret2:
|
|
info(fn)
|
|
for v in [ret, ret2]:
|
|
try:
|
|
info(len(v))
|
|
except:
|
|
info("uhh " + repr(v))
|
|
|
|
with open(fn + ".bad", "wb") as f:
|
|
f.write(ret)
|
|
with open(fn + ".good", "wb") as f:
|
|
f.write(ret2)
|
|
|
|
raise Exception("cache bork")
|
|
|
|
return ret
|
|
# !rm.no>
|
|
|
|
def getattr(self, path, fh=None):
|
|
dbg("getattr %r", path)
|
|
if WINDOWS:
|
|
path = enwin(path) # windows occasionally decodes f0xx to xx
|
|
|
|
path = path.strip("/")
|
|
if not path:
|
|
ret = self.gw.stat_dir(time.time(), 4096)
|
|
dbg("/=%r", ret)
|
|
return ret
|
|
|
|
try:
|
|
dirpath, fname = path.rsplit("/", 1)
|
|
except:
|
|
dirpath = ""
|
|
fname = path
|
|
|
|
cn = self.get_cached_dir(dirpath)
|
|
if cn:
|
|
dents = cn.data
|
|
else:
|
|
dbg("cache miss")
|
|
dents = self._readdir(dirpath)
|
|
|
|
try:
|
|
ret = dents[fname]
|
|
dbg("s=%r", ret)
|
|
return ret
|
|
except:
|
|
pass
|
|
|
|
fun = info
|
|
if MACOS and path.split("/")[-1].startswith("._"):
|
|
fun = dbg
|
|
|
|
fun("=ENOENT %r", path)
|
|
raise FuseOSError(errno.ENOENT)
|
|
|
|
access = None
|
|
flush = None
|
|
getxattr = None
|
|
listxattr = None
|
|
open = None
|
|
opendir = None
|
|
release = None
|
|
releasedir = None
|
|
statfs = None
|
|
|
|
# !rm.yes>
|
|
if False:
|
|
# incorrect semantics but good for debugging stuff like samba and msys2
|
|
def access(self, path, mode):
|
|
dbg("@@ access [{}] [{}]".format(path, mode))
|
|
return 1 if self.getattr(path) else 0
|
|
|
|
def flush(self, path, fh):
|
|
dbg("@@ flush [{}] [{}]".format(path, fh))
|
|
return True
|
|
|
|
def getxattr(self, *args):
|
|
dbg("@@ getxattr [{}]".format("] [".join(str(x) for x in args)))
|
|
return False
|
|
|
|
def listxattr(self, *args):
|
|
dbg("@@ listxattr [{}]".format("] [".join(str(x) for x in args)))
|
|
return False
|
|
|
|
def open(self, path, flags):
|
|
dbg("@@ open [{}] [{}]".format(path, flags))
|
|
return 42
|
|
|
|
def opendir(self, fh):
|
|
dbg("@@ opendir [{}]".format(fh))
|
|
return 69
|
|
|
|
def release(self, ino, fi):
|
|
dbg("@@ release [{}] [{}]".format(ino, fi))
|
|
return True
|
|
|
|
def releasedir(self, ino, fi):
|
|
dbg("@@ releasedir [{}] [{}]".format(ino, fi))
|
|
return True
|
|
|
|
def statfs(self, path):
|
|
dbg("@@ statfs [{}]".format(path))
|
|
return {}
|
|
|
|
# !rm.no>
|
|
|
|
if sys.platform == "win32":
|
|
# quick compat for /mingw64/bin/python3 (msys2)
|
|
def _open(self, path):
|
|
try:
|
|
x = self.getattr(path)
|
|
if x["st_mode"] <= 0:
|
|
raise Exception()
|
|
|
|
self.junk_fh_ctr += 1
|
|
if self.junk_fh_ctr > 32000: # TODO untested
|
|
self.junk_fh_ctr = 4
|
|
|
|
return self.junk_fh_ctr
|
|
|
|
except Exception as ex:
|
|
info("open ERR %r", ex)
|
|
raise FuseOSError(errno.ENOENT)
|
|
|
|
def open(self, path, flags):
|
|
dbg("open %r [%s]", path, flags)
|
|
return self._open(path)
|
|
|
|
def opendir(self, path):
|
|
dbg("opendir %r", path)
|
|
return self._open(path)
|
|
|
|
def flush(self, path, fh):
|
|
dbg("flush %r [%s]", path, fh)
|
|
|
|
def release(self, ino, fi):
|
|
dbg("release %r [%s]", ino, fi)
|
|
|
|
def releasedir(self, ino, fi):
|
|
dbg("releasedir %r [%s]", ino, fi)
|
|
|
|
def access(self, path, mode):
|
|
dbg("access %r [%s]", path, mode)
|
|
try:
|
|
x = self.getattr(path)
|
|
if x["st_mode"] <= 0:
|
|
raise Exception()
|
|
except:
|
|
raise FuseOSError(errno.ENOENT)
|
|
|
|
|
|
class TheArgparseFormatter(
|
|
argparse.RawTextHelpFormatter, argparse.ArgumentDefaultsHelpFormatter
|
|
):
|
|
pass
|
|
|
|
|
|
def main():
|
|
global info, dbg, is_dbg
|
|
time.strptime("19970815", "%Y%m%d") # python#7980
|
|
|
|
ver = "{0}, v{1}".format(S_BUILD_DT, S_VERSION)
|
|
if "--version" in sys.argv:
|
|
print("partyfuse", ver)
|
|
return
|
|
|
|
# filecache helps for reads that are ~64k or smaller;
|
|
# windows likes to use 4k and 64k so cache is important,
|
|
# linux generally does 128k so the cache is still nice,
|
|
# value is numChunks (1~8M each) to keep in the cache
|
|
nf = 12
|
|
|
|
# dircache is always a boost,
|
|
# only want to disable it for tests etc,
|
|
cdn = 24 # max num dirs; keep larger than max dir depth; 0=disable
|
|
cds = 1 # numsec until an entry goes stale
|
|
|
|
where = "local directory"
|
|
if WINDOWS:
|
|
where += " or DRIVE:"
|
|
|
|
ex_pre = "\n " + os.path.basename(__file__) + " "
|
|
examples = ["http://192.168.1.69:3923/music/ ./music"]
|
|
if WINDOWS:
|
|
examples.append("http://192.168.1.69:3923/music/ M:")
|
|
|
|
ap = argparse.ArgumentParser(
|
|
formatter_class=TheArgparseFormatter,
|
|
description="mount a copyparty server as a local filesystem -- " + ver,
|
|
epilog="example:" + ex_pre + ex_pre.join(examples),
|
|
)
|
|
# fmt: off
|
|
ap.add_argument("base_url", type=str, help="remote copyparty URL to mount")
|
|
ap.add_argument("local_path", type=str, help=where + " to mount it on")
|
|
ap.add_argument("-a", metavar="PASSWORD", help="password or $filepath")
|
|
|
|
# !rm.yes>
|
|
ap.add_argument("--html", metavar="TYPE", default="", help="which HTML parser to use; cpp, nginx, iis")
|
|
# !rm.no>
|
|
|
|
ap2 = ap.add_argument_group("https/TLS")
|
|
ap2.add_argument("-te", metavar="PEMFILE", help="certificate to expect/verify")
|
|
ap2.add_argument("-td", action="store_true", help="disable certificate check")
|
|
|
|
ap2 = ap.add_argument_group("cache/perf")
|
|
ap2.add_argument("-cdn", metavar="DIRS", type=float, default=cdn, help="directory-cache, max num dirs; 0=disable")
|
|
ap2.add_argument("-cds", metavar="SECS", type=float, default=cds, help="directory-cache, expiration time")
|
|
ap2.add_argument("-cf", metavar="BLOCKS", type=int, default=nf, help="file cache; each block is <= 1 MiB")
|
|
|
|
ap2 = ap.add_argument_group("logging")
|
|
ap2.add_argument("-q", action="store_true", help="quiet")
|
|
ap2.add_argument("-d", action="store_true", help="debug/verbose")
|
|
ap2.add_argument("--slowterm", action="store_true", help="only most recent msgs; good for windows")
|
|
ap2.add_argument("--logf", metavar="FILE", type=str, default="", help="log to FILE; enables --slowterm")
|
|
|
|
ap2 = ap.add_argument_group("fuse")
|
|
ap2.add_argument("--oth", action="store_true", help="tell FUSE to '-o allow_other'")
|
|
ap2.add_argument("--nonempty", action="store_true", help="tell FUSE to '-o nonempty'")
|
|
|
|
ar = ap.parse_args()
|
|
# fmt: on
|
|
|
|
if ar.logf:
|
|
ar.slowterm = True
|
|
|
|
# windows terminals are slow (cmd.exe, mintty)
|
|
# otoh fancy_log beats RecentLog on linux
|
|
logger = RecentLog(ar).put if ar.slowterm else fancy_log
|
|
if ar.d:
|
|
info = logger
|
|
dbg = logger
|
|
is_dbg = True
|
|
elif not ar.q:
|
|
info = logger
|
|
|
|
if ar.a and ar.a.startswith("$"):
|
|
fn = ar.a[1:]
|
|
info("reading password from file %r", fn)
|
|
with open(fn, "rb") as f:
|
|
ar.a = f.read().decode("utf-8").strip()
|
|
|
|
if WINDOWS:
|
|
os.system("rem")
|
|
|
|
for ch in '<>:"\\|?*':
|
|
# microsoft maps illegal characters to f0xx
|
|
# (e000 to f8ff is basic-plane private-use)
|
|
bad_good[ch] = chr(ord(ch) + 0xF000)
|
|
|
|
for n in range(0, 0x100):
|
|
# map surrogateescape to another private-use area
|
|
bad_good[chr(n + 0xDC00)] = chr(n + 0xF100)
|
|
|
|
for k, v in bad_good.items():
|
|
good_bad[v] = k
|
|
|
|
register_wtf8()
|
|
|
|
args = {"foreground": True, "nothreads": True}
|
|
if ar.oth:
|
|
args["allow_other"] = True
|
|
if ar.nonempty:
|
|
args["nonempty"] = True
|
|
|
|
FUSE(CPPF(ar), ar.local_path, encoding="wtf-8", **args)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
main()
|