2015-07-12 20:36:46 +02:00
|
|
|
import os
|
|
|
|
import re
|
|
|
|
import shutil
|
|
|
|
import json
|
|
|
|
import time
|
2015-11-15 11:13:57 +01:00
|
|
|
import sys
|
2015-07-12 20:36:46 +02:00
|
|
|
|
|
|
|
import sqlite3
|
2015-03-19 21:19:14 +01:00
|
|
|
import gevent.event
|
2015-07-12 20:36:46 +02:00
|
|
|
|
2015-03-19 21:19:14 +01:00
|
|
|
from Db import Db
|
version 0.3.0, rev187, Trusted authorization sites support, --publish option on signing, cryptSign command line option, OpenSSL enabled on OSX, Crypto verify allows list of valid addresses, Option for version 2 json DB tables, DbCursor SELECT parameters bugfix, Add peer to site on ListModified, Download blind includes when new site added, Publish command better messages, Multi-threaded announce, New http Torrent trackers, Wait for dbschema.json on query, Handle json import errors, More compact writeJson storage command, Testcase for signing and verifying, Workaround to make non target=_top links work, More clean UiWebsocket command route, Send cert_user_id on siteinfo, Notify other local clients on local file modify, Option to wait for file download before sql query, File rules websocket API command, Cert add and select, set websocket API command, Put focus on innerframe, innerloaded wrapper api command to add hashtag, Allow more file error on big sites, Keep worker running after stuked on done task, New more stable openSSL layer that works on OSX, Noparallel parameter bugfix, RateLimit allowed again interval bugfix, Updater skips non-writeable files, Try to close openssl dll before update
2015-05-25 01:26:33 +02:00
|
|
|
from Debug import Debug
|
2015-05-31 15:52:21 +02:00
|
|
|
from Config import config
|
Rev467, requirements.txt accept newer dependecies, Boost dbschema.json, Move getDirname getFilename to helper, Verify optional files, Includes not allowed in user files, Optional files rules, Peer hashfield functions, Test optional files signing, Test file info, Test verify file, Test helpers
2015-10-01 01:35:13 +02:00
|
|
|
from util import helper
|
version 0.3.0, rev187, Trusted authorization sites support, --publish option on signing, cryptSign command line option, OpenSSL enabled on OSX, Crypto verify allows list of valid addresses, Option for version 2 json DB tables, DbCursor SELECT parameters bugfix, Add peer to site on ListModified, Download blind includes when new site added, Publish command better messages, Multi-threaded announce, New http Torrent trackers, Wait for dbschema.json on query, Handle json import errors, More compact writeJson storage command, Testcase for signing and verifying, Workaround to make non target=_top links work, More clean UiWebsocket command route, Send cert_user_id on siteinfo, Notify other local clients on local file modify, Option to wait for file download before sql query, File rules websocket API command, Cert add and select, set websocket API command, Put focus on innerframe, innerloaded wrapper api command to add hashtag, Allow more file error on big sites, Keep worker running after stuked on done task, New more stable openSSL layer that works on OSX, Noparallel parameter bugfix, RateLimit allowed again interval bugfix, Updater skips non-writeable files, Try to close openssl dll before update
2015-05-25 01:26:33 +02:00
|
|
|
|
2015-03-19 21:19:14 +01:00
|
|
|
|
|
|
|
class SiteStorage:
|
2015-07-12 20:36:46 +02:00
|
|
|
|
|
|
|
def __init__(self, site, allow_create=True):
|
|
|
|
self.site = site
|
|
|
|
self.directory = "%s/%s" % (config.data_dir, self.site.address) # Site data diretory
|
2015-11-15 11:13:57 +01:00
|
|
|
self.allowed_dir = os.path.abspath(self.directory.decode(sys.getfilesystemencoding())) # Only serve/modify file within this dir
|
2015-07-12 20:36:46 +02:00
|
|
|
self.log = site.log
|
|
|
|
self.db = None # Db class
|
|
|
|
self.db_checked = False # Checked db tables since startup
|
|
|
|
self.event_db_busy = None # Gevent AsyncResult if db is working on rebuild
|
|
|
|
self.has_db = self.isFile("dbschema.json") # The site has schema
|
|
|
|
|
|
|
|
if not os.path.isdir(self.directory):
|
|
|
|
if allow_create:
|
|
|
|
os.mkdir(self.directory) # Create directory if not found
|
|
|
|
else:
|
|
|
|
raise Exception("Directory not exists: %s" % self.directory)
|
|
|
|
|
|
|
|
# Load db from dbschema.json
|
|
|
|
def openDb(self, check=True):
|
2016-03-04 00:40:26 +01:00
|
|
|
try:
|
|
|
|
schema = self.loadJson("dbschema.json")
|
|
|
|
db_path = self.getPath(schema["db_file"])
|
|
|
|
except Exception, err:
|
2016-03-04 00:47:26 +01:00
|
|
|
raise Exception("dbschema.json is not a valid JSON: %s" % err)
|
2016-03-04 00:40:26 +01:00
|
|
|
|
2015-07-12 20:36:46 +02:00
|
|
|
if check:
|
|
|
|
if not os.path.isfile(db_path) or os.path.getsize(db_path) == 0: # Not exist or null
|
|
|
|
self.rebuildDb()
|
2015-09-24 22:08:08 +02:00
|
|
|
|
|
|
|
if not self.db:
|
|
|
|
self.db = Db(schema, db_path)
|
|
|
|
|
2015-07-12 20:36:46 +02:00
|
|
|
if check and not self.db_checked:
|
|
|
|
changed_tables = self.db.checkTables()
|
|
|
|
if changed_tables:
|
2015-11-05 23:19:36 +01:00
|
|
|
self.rebuildDb(delete_db=False) # TODO: only update the changed table datas
|
2015-07-12 20:36:46 +02:00
|
|
|
|
|
|
|
def closeDb(self):
|
|
|
|
if self.db:
|
|
|
|
self.db.close()
|
|
|
|
self.event_db_busy = None
|
|
|
|
self.db = None
|
|
|
|
|
|
|
|
# Return db class
|
|
|
|
def getDb(self):
|
|
|
|
if not self.db:
|
|
|
|
self.log.debug("No database, waiting for dbschema.json...")
|
2015-09-27 02:08:53 +02:00
|
|
|
self.site.needFile("dbschema.json", priority=3)
|
2015-07-12 20:36:46 +02:00
|
|
|
self.has_db = self.isFile("dbschema.json") # Recheck if dbschema exist
|
|
|
|
if self.has_db:
|
|
|
|
self.openDb()
|
|
|
|
return self.db
|
|
|
|
|
|
|
|
# Rebuild sql cache
|
|
|
|
def rebuildDb(self, delete_db=True):
|
|
|
|
self.has_db = self.isFile("dbschema.json")
|
|
|
|
if not self.has_db:
|
|
|
|
return False
|
|
|
|
self.event_db_busy = gevent.event.AsyncResult()
|
|
|
|
schema = self.loadJson("dbschema.json")
|
|
|
|
db_path = self.getPath(schema["db_file"])
|
|
|
|
if os.path.isfile(db_path) and delete_db:
|
|
|
|
if self.db:
|
|
|
|
self.db.close() # Close db if open
|
|
|
|
self.log.info("Deleting %s" % db_path)
|
|
|
|
try:
|
|
|
|
os.unlink(db_path)
|
|
|
|
except Exception, err:
|
|
|
|
self.log.error("Delete error: %s" % err)
|
|
|
|
self.openDb(check=False)
|
|
|
|
self.log.info("Creating tables...")
|
|
|
|
self.db.checkTables()
|
|
|
|
self.log.info("Importing data...")
|
|
|
|
cur = self.db.getCursor()
|
|
|
|
cur.execute("BEGIN")
|
|
|
|
cur.logging = False
|
|
|
|
found = 0
|
|
|
|
s = time.time()
|
|
|
|
for content_inner_path, content in self.site.content_manager.contents.items():
|
|
|
|
content_path = self.getPath(content_inner_path)
|
|
|
|
if os.path.isfile(content_path): # Missing content.json file
|
|
|
|
if self.db.loadJson(content_path, cur=cur):
|
|
|
|
found += 1
|
|
|
|
else:
|
|
|
|
self.log.error("[MISSING] %s" % content_inner_path)
|
|
|
|
for file_relative_path in content["files"].keys():
|
|
|
|
if not file_relative_path.endswith(".json"):
|
|
|
|
continue # We only interesed in json files
|
Rev467, requirements.txt accept newer dependecies, Boost dbschema.json, Move getDirname getFilename to helper, Verify optional files, Includes not allowed in user files, Optional files rules, Peer hashfield functions, Test optional files signing, Test file info, Test verify file, Test helpers
2015-10-01 01:35:13 +02:00
|
|
|
content_inner_path_dir = helper.getDirname(content_inner_path) # Content.json dir relative to site
|
2015-07-12 20:36:46 +02:00
|
|
|
file_inner_path = content_inner_path_dir + file_relative_path # File Relative to site dir
|
|
|
|
file_inner_path = file_inner_path.strip("/") # Strip leading /
|
|
|
|
file_path = self.getPath(file_inner_path)
|
|
|
|
if os.path.isfile(file_path):
|
|
|
|
if self.db.loadJson(file_path, cur=cur):
|
|
|
|
found += 1
|
|
|
|
else:
|
|
|
|
self.log.error("[MISSING] %s" % file_inner_path)
|
|
|
|
cur.execute("END")
|
|
|
|
self.log.info("Imported %s data file in %ss" % (found, time.time() - s))
|
|
|
|
self.event_db_busy.set(True) # Event done, notify waiters
|
|
|
|
self.event_db_busy = None # Clear event
|
|
|
|
|
|
|
|
# Execute sql query or rebuild on dberror
|
|
|
|
def query(self, query, params=None):
|
|
|
|
if self.event_db_busy: # Db not ready for queries
|
|
|
|
self.log.debug("Wating for db...")
|
|
|
|
self.event_db_busy.get() # Wait for event
|
|
|
|
try:
|
|
|
|
res = self.getDb().execute(query, params)
|
|
|
|
except sqlite3.DatabaseError, err:
|
|
|
|
if err.__class__.__name__ == "DatabaseError":
|
|
|
|
self.log.error("Database error: %s, query: %s, try to rebuilding it..." % (err, query))
|
|
|
|
self.rebuildDb()
|
|
|
|
res = self.db.cur.execute(query, params)
|
|
|
|
else:
|
|
|
|
raise err
|
|
|
|
return res
|
|
|
|
|
|
|
|
# Open file object
|
|
|
|
def open(self, inner_path, mode="rb"):
|
|
|
|
return open(self.getPath(inner_path), mode)
|
|
|
|
|
|
|
|
# Open file object
|
|
|
|
def read(self, inner_path, mode="r"):
|
|
|
|
return open(self.getPath(inner_path), mode).read()
|
|
|
|
|
|
|
|
# Write content to file
|
|
|
|
def write(self, inner_path, content):
|
|
|
|
file_path = self.getPath(inner_path)
|
|
|
|
# Create dir if not exist
|
|
|
|
file_dir = os.path.dirname(file_path)
|
|
|
|
if not os.path.isdir(file_dir):
|
|
|
|
os.makedirs(file_dir)
|
|
|
|
# Write file
|
|
|
|
if hasattr(content, 'read'): # File-like object
|
|
|
|
with open(file_path, "wb") as file:
|
|
|
|
shutil.copyfileobj(content, file) # Write buff to disk
|
|
|
|
else: # Simple string
|
|
|
|
with open(file_path, "wb") as file:
|
|
|
|
file.write(content)
|
|
|
|
del content
|
|
|
|
self.onUpdated(inner_path)
|
|
|
|
|
2015-09-16 00:01:23 +02:00
|
|
|
# Remove file from filesystem
|
|
|
|
def delete(self, inner_path):
|
|
|
|
file_path = self.getPath(inner_path)
|
|
|
|
os.unlink(file_path)
|
|
|
|
|
2016-03-26 00:22:27 +01:00
|
|
|
def deleteDir(self, inner_path):
|
|
|
|
dir_path = self.getPath(inner_path)
|
|
|
|
os.rmdir(dir_path)
|
|
|
|
|
2015-09-28 22:07:26 +02:00
|
|
|
# List files from a directory
|
|
|
|
def list(self, dir_inner_path):
|
|
|
|
directory = self.getPath(dir_inner_path)
|
|
|
|
for root, dirs, files in os.walk(directory):
|
|
|
|
root = root.replace("\\", "/")
|
|
|
|
root_relative_path = re.sub("^%s" % re.escape(directory), "", root).lstrip("/")
|
|
|
|
for file_name in files:
|
|
|
|
if root_relative_path: # Not root dir
|
|
|
|
yield root_relative_path + "/" + file_name
|
|
|
|
else:
|
|
|
|
yield file_name
|
|
|
|
|
2015-07-12 20:36:46 +02:00
|
|
|
# Site content updated
|
|
|
|
def onUpdated(self, inner_path):
|
|
|
|
file_path = self.getPath(inner_path)
|
|
|
|
# Update Sql cache
|
|
|
|
if inner_path == "dbschema.json":
|
|
|
|
self.has_db = self.isFile("dbschema.json")
|
2015-11-05 23:19:36 +01:00
|
|
|
# Reopen DB to check changes
|
|
|
|
self.closeDb()
|
|
|
|
self.openDb()
|
2016-03-19 18:05:49 +01:00
|
|
|
elif not config.disable_db and inner_path.endswith(".json") and self.has_db: # Load json file to db
|
2016-04-06 13:39:17 +02:00
|
|
|
if config.verbose:
|
|
|
|
self.log.debug("Loading json file to db: %s" % inner_path)
|
2015-07-12 20:36:46 +02:00
|
|
|
try:
|
|
|
|
self.getDb().loadJson(file_path)
|
|
|
|
except Exception, err:
|
|
|
|
self.log.error("Json %s load error: %s" % (inner_path, Debug.formatException(err)))
|
|
|
|
self.closeDb()
|
|
|
|
|
|
|
|
# Load and parse json file
|
|
|
|
def loadJson(self, inner_path):
|
|
|
|
with self.open(inner_path) as file:
|
|
|
|
return json.load(file)
|
|
|
|
|
|
|
|
# Write formatted json file
|
|
|
|
def writeJson(self, inner_path, data):
|
2015-09-20 00:27:54 +02:00
|
|
|
content = json.dumps(data, indent=1, sort_keys=True)
|
2015-07-12 20:36:46 +02:00
|
|
|
# Make it a little more compact by removing unnecessary white space
|
|
|
|
|
|
|
|
def compact_list(match):
|
|
|
|
return "[ " + match.group(1).strip() + " ]"
|
|
|
|
|
|
|
|
def compact_dict(match):
|
|
|
|
return "{ " + match.group(1).strip() + " }"
|
|
|
|
|
|
|
|
content = re.sub("\[([^,\{\[]{10,100}?)\]", compact_list, content, flags=re.DOTALL)
|
|
|
|
content = re.sub("\{([^,\[\{]{10,100}?)\}", compact_dict, content, flags=re.DOTALL)
|
Rev900, Sidebar filestats bar width round fix, Sidebar WebGL not supported error, Sidebar optimalizations, Trayicon gray shadow, Trim end of line whitespace from json files, Fix testweb testcase, Implement experimental postMessage nonce security, Return None when testing external ip, Window opener security check and message, Increase timeout for large files
2016-02-10 02:30:04 +01:00
|
|
|
|
|
|
|
# Remove end of line whitespace
|
|
|
|
content = re.sub("(?m)[ ]+$", "", content)
|
|
|
|
|
2015-07-12 20:36:46 +02:00
|
|
|
# Write to disk
|
|
|
|
self.write(inner_path, content)
|
|
|
|
|
|
|
|
# Get file size
|
|
|
|
def getSize(self, inner_path):
|
|
|
|
path = self.getPath(inner_path)
|
|
|
|
if os.path.isfile(path):
|
|
|
|
return os.path.getsize(path)
|
|
|
|
else:
|
|
|
|
return 0
|
|
|
|
|
|
|
|
# File exist
|
|
|
|
def isFile(self, inner_path):
|
|
|
|
return os.path.isfile(self.getPath(inner_path))
|
|
|
|
|
2015-10-11 02:22:53 +02:00
|
|
|
# File or directory exist
|
|
|
|
def isExists(self, inner_path):
|
|
|
|
return os.path.exists(self.getPath(inner_path))
|
|
|
|
|
2015-07-12 20:36:46 +02:00
|
|
|
# Dir exist
|
|
|
|
def isDir(self, inner_path):
|
|
|
|
return os.path.isdir(self.getPath(inner_path))
|
|
|
|
|
|
|
|
# Security check and return path of site's file
|
|
|
|
def getPath(self, inner_path):
|
|
|
|
inner_path = inner_path.replace("\\", "/") # Windows separator fix
|
|
|
|
inner_path = re.sub("^%s/" % re.escape(self.directory), "", inner_path) # Remove site directory if begins with it
|
2015-09-08 03:07:44 +02:00
|
|
|
file_path = u"%s/%s" % (self.directory, inner_path)
|
2015-09-28 22:07:26 +02:00
|
|
|
if not inner_path:
|
|
|
|
return self.directory
|
|
|
|
|
2015-09-08 03:07:44 +02:00
|
|
|
file_abspath = os.path.dirname(os.path.abspath(file_path))
|
|
|
|
if ".." in file_path or not file_abspath.startswith(self.allowed_dir):
|
Rev957, Sidebar displays onion peers in graph, Sidebar display bad file retry number, Sidebar site Update/Pause/Delete, Ratelimit sidebar update, Encoded typo, Fix onion findHashId, More retry for bad files, Log file path errors, Testcase for self findhashIds, Testcase for Tor findHashId, Better Tor version parse, UiWebsocket callback on update/pause/resume/delete, Skip invalid postMessage messages
2016-03-09 00:48:57 +01:00
|
|
|
self.site.log.error(u"File %s not in allowed dir: %s" % (file_path, self.allowed_dir))
|
2015-09-08 03:07:44 +02:00
|
|
|
raise Exception(u"File not allowed: %s" % file_path)
|
2015-07-12 20:36:46 +02:00
|
|
|
return file_path
|
|
|
|
|
2015-08-16 11:51:00 +02:00
|
|
|
# Get site dir relative path
|
|
|
|
def getInnerPath(self, path):
|
2015-09-28 22:07:26 +02:00
|
|
|
if path == self.directory:
|
|
|
|
inner_path = ""
|
|
|
|
else:
|
|
|
|
inner_path = re.sub("^%s/" % re.escape(self.directory), "", path)
|
2015-08-16 11:51:00 +02:00
|
|
|
return inner_path
|
|
|
|
|
2015-07-12 20:36:46 +02:00
|
|
|
# Verify all files sha512sum using content.json
|
Rev571, Optional file sizes to sidebar, Download all optional files option in sidebar, Optional file number in peer stats, Delete removed or changed optional files, Auto download optional files if autodownloadoptional checked, SiteReload command, Peer use global file server if no site defined, Allow browser cache video files, Allow more keepalive connections, Gevent 1.1 ranged request bugfix, Dont sent optional files details on websocket, Remove files from workermanager tasks if no longer in bad_files, Notify local client about changes on external siteSign
2015-11-09 00:44:03 +01:00
|
|
|
def verifyFiles(self, quick_check=False, add_optional=False, add_changed=True):
|
2015-07-12 20:36:46 +02:00
|
|
|
bad_files = []
|
2016-03-12 23:07:06 +01:00
|
|
|
i = 0
|
2015-10-11 02:22:53 +02:00
|
|
|
|
2015-07-12 20:36:46 +02:00
|
|
|
if not self.site.content_manager.contents.get("content.json"): # No content.json, download it first
|
|
|
|
self.site.needFile("content.json", update=True) # Force update to fix corrupt file
|
|
|
|
self.site.content_manager.loadContent() # Reload content.json
|
|
|
|
for content_inner_path, content in self.site.content_manager.contents.items():
|
2016-03-12 23:07:06 +01:00
|
|
|
i += 1
|
2016-03-16 00:40:19 +01:00
|
|
|
if i % 50 == 0:
|
2016-03-12 23:07:06 +01:00
|
|
|
time.sleep(0.0001) # Context switch to avoid gevent hangs
|
2015-07-12 20:36:46 +02:00
|
|
|
if not os.path.isfile(self.getPath(content_inner_path)): # Missing content.json file
|
|
|
|
self.log.debug("[MISSING] %s" % content_inner_path)
|
|
|
|
bad_files.append(content_inner_path)
|
2015-10-11 02:22:53 +02:00
|
|
|
|
|
|
|
for file_relative_path in content.get("files", {}).keys():
|
Rev467, requirements.txt accept newer dependecies, Boost dbschema.json, Move getDirname getFilename to helper, Verify optional files, Includes not allowed in user files, Optional files rules, Peer hashfield functions, Test optional files signing, Test file info, Test verify file, Test helpers
2015-10-01 01:35:13 +02:00
|
|
|
file_inner_path = helper.getDirname(content_inner_path) + file_relative_path # Relative to site dir
|
2015-07-12 20:36:46 +02:00
|
|
|
file_inner_path = file_inner_path.strip("/") # Strip leading /
|
|
|
|
file_path = self.getPath(file_inner_path)
|
|
|
|
if not os.path.isfile(file_path):
|
|
|
|
self.log.debug("[MISSING] %s" % file_inner_path)
|
|
|
|
bad_files.append(file_inner_path)
|
|
|
|
continue
|
|
|
|
|
|
|
|
if quick_check:
|
|
|
|
ok = os.path.getsize(file_path) == content["files"][file_relative_path]["size"]
|
|
|
|
else:
|
|
|
|
ok = self.site.content_manager.verifyFile(file_inner_path, open(file_path, "rb"))
|
|
|
|
|
|
|
|
if not ok:
|
|
|
|
self.log.debug("[CHANGED] %s" % file_inner_path)
|
2016-03-11 12:39:39 +01:00
|
|
|
if add_changed or content.get("cert_sign"): # If updating own site only add changed user files
|
Rev571, Optional file sizes to sidebar, Download all optional files option in sidebar, Optional file number in peer stats, Delete removed or changed optional files, Auto download optional files if autodownloadoptional checked, SiteReload command, Peer use global file server if no site defined, Allow browser cache video files, Allow more keepalive connections, Gevent 1.1 ranged request bugfix, Dont sent optional files details on websocket, Remove files from workermanager tasks if no longer in bad_files, Notify local client about changes on external siteSign
2015-11-09 00:44:03 +01:00
|
|
|
bad_files.append(file_inner_path)
|
2015-10-11 02:22:53 +02:00
|
|
|
|
|
|
|
# Optional files
|
|
|
|
optional_added = 0
|
|
|
|
optional_removed = 0
|
|
|
|
for file_relative_path in content.get("files_optional", {}).keys():
|
|
|
|
file_inner_path = helper.getDirname(content_inner_path) + file_relative_path # Relative to site dir
|
|
|
|
file_inner_path = file_inner_path.strip("/") # Strip leading /
|
|
|
|
file_path = self.getPath(file_inner_path)
|
|
|
|
if not os.path.isfile(file_path):
|
|
|
|
self.site.content_manager.hashfield.removeHash(content["files_optional"][file_relative_path]["sha512"])
|
Rev571, Optional file sizes to sidebar, Download all optional files option in sidebar, Optional file number in peer stats, Delete removed or changed optional files, Auto download optional files if autodownloadoptional checked, SiteReload command, Peer use global file server if no site defined, Allow browser cache video files, Allow more keepalive connections, Gevent 1.1 ranged request bugfix, Dont sent optional files details on websocket, Remove files from workermanager tasks if no longer in bad_files, Notify local client about changes on external siteSign
2015-11-09 00:44:03 +01:00
|
|
|
if add_optional:
|
|
|
|
bad_files.append(file_inner_path)
|
2015-10-11 02:22:53 +02:00
|
|
|
continue
|
|
|
|
|
|
|
|
if quick_check:
|
|
|
|
ok = os.path.getsize(file_path) == content["files_optional"][file_relative_path]["size"]
|
|
|
|
else:
|
|
|
|
ok = self.site.content_manager.verifyFile(file_inner_path, open(file_path, "rb"))
|
|
|
|
|
|
|
|
if ok:
|
|
|
|
self.site.content_manager.hashfield.appendHash(content["files_optional"][file_relative_path]["sha512"])
|
|
|
|
optional_added += 1
|
|
|
|
else:
|
|
|
|
self.site.content_manager.hashfield.removeHash(content["files_optional"][file_relative_path]["sha512"])
|
|
|
|
optional_removed += 1
|
Rev571, Optional file sizes to sidebar, Download all optional files option in sidebar, Optional file number in peer stats, Delete removed or changed optional files, Auto download optional files if autodownloadoptional checked, SiteReload command, Peer use global file server if no site defined, Allow browser cache video files, Allow more keepalive connections, Gevent 1.1 ranged request bugfix, Dont sent optional files details on websocket, Remove files from workermanager tasks if no longer in bad_files, Notify local client about changes on external siteSign
2015-11-09 00:44:03 +01:00
|
|
|
if add_optional:
|
|
|
|
bad_files.append(file_inner_path)
|
2015-10-11 02:22:53 +02:00
|
|
|
self.log.debug("[OPTIONAL CHANGED] %s" % file_inner_path)
|
|
|
|
|
2016-03-06 00:55:50 +01:00
|
|
|
if config.verbose:
|
|
|
|
self.log.debug(
|
|
|
|
"%s verified: %s, quick: %s, bad: %s, optionals: +%s -%s" %
|
|
|
|
(content_inner_path, len(content["files"]), quick_check, bad_files, optional_added, optional_removed)
|
|
|
|
)
|
2015-07-12 20:36:46 +02:00
|
|
|
|
2016-03-12 23:07:06 +01:00
|
|
|
time.sleep(0.0001) # Context switch to avoid gevent hangs
|
2015-07-12 20:36:46 +02:00
|
|
|
return bad_files
|
|
|
|
|
|
|
|
# Check and try to fix site files integrity
|
|
|
|
def checkFiles(self, quick_check=True):
|
|
|
|
s = time.time()
|
Rev571, Optional file sizes to sidebar, Download all optional files option in sidebar, Optional file number in peer stats, Delete removed or changed optional files, Auto download optional files if autodownloadoptional checked, SiteReload command, Peer use global file server if no site defined, Allow browser cache video files, Allow more keepalive connections, Gevent 1.1 ranged request bugfix, Dont sent optional files details on websocket, Remove files from workermanager tasks if no longer in bad_files, Notify local client about changes on external siteSign
2015-11-09 00:44:03 +01:00
|
|
|
bad_files = self.verifyFiles(
|
|
|
|
quick_check,
|
|
|
|
add_optional=self.site.settings.get("autodownloadoptional"),
|
|
|
|
add_changed=not self.site.settings.get("own") # Don't overwrite changed files if site owned
|
|
|
|
)
|
|
|
|
self.site.bad_files = {}
|
2015-07-12 20:36:46 +02:00
|
|
|
if bad_files:
|
|
|
|
for bad_file in bad_files:
|
Rev571, Optional file sizes to sidebar, Download all optional files option in sidebar, Optional file number in peer stats, Delete removed or changed optional files, Auto download optional files if autodownloadoptional checked, SiteReload command, Peer use global file server if no site defined, Allow browser cache video files, Allow more keepalive connections, Gevent 1.1 ranged request bugfix, Dont sent optional files details on websocket, Remove files from workermanager tasks if no longer in bad_files, Notify local client about changes on external siteSign
2015-11-09 00:44:03 +01:00
|
|
|
self.site.bad_files[bad_file] = 1
|
2016-03-11 12:39:39 +01:00
|
|
|
self.log.debug("Checked files in %.2fs... Found bad files: %s, Quick:%s" % (time.time() - s, len(bad_files), quick_check))
|
2015-07-12 20:36:46 +02:00
|
|
|
|
|
|
|
# Delete site's all file
|
|
|
|
def deleteFiles(self):
|
|
|
|
if self.has_db:
|
|
|
|
self.log.debug("Deleting db file...")
|
|
|
|
self.closeDb()
|
|
|
|
try:
|
|
|
|
schema = self.loadJson("dbschema.json")
|
|
|
|
db_path = self.getPath(schema["db_file"])
|
|
|
|
if os.path.isfile(db_path):
|
|
|
|
os.unlink(db_path)
|
|
|
|
except Exception, err:
|
|
|
|
self.log.error("Db file delete error: %s" % err)
|
|
|
|
|
|
|
|
self.log.debug("Deleting files from content.json...")
|
|
|
|
files = [] # Get filenames
|
|
|
|
for content_inner_path, content in self.site.content_manager.contents.items():
|
|
|
|
files.append(content_inner_path)
|
Rev467, requirements.txt accept newer dependecies, Boost dbschema.json, Move getDirname getFilename to helper, Verify optional files, Includes not allowed in user files, Optional files rules, Peer hashfield functions, Test optional files signing, Test file info, Test verify file, Test helpers
2015-10-01 01:35:13 +02:00
|
|
|
# Add normal files
|
|
|
|
for file_relative_path in content.get("files", {}).keys():
|
|
|
|
file_inner_path = helper.getDirname(content_inner_path) + file_relative_path # Relative to site dir
|
|
|
|
files.append(file_inner_path)
|
|
|
|
# Add optional files
|
|
|
|
for file_relative_path in content.get("files_optional", {}).keys():
|
|
|
|
file_inner_path = helper.getDirname(content_inner_path) + file_relative_path # Relative to site dir
|
2015-07-12 20:36:46 +02:00
|
|
|
files.append(file_inner_path)
|
|
|
|
|
|
|
|
for inner_path in files:
|
|
|
|
path = self.getPath(inner_path)
|
|
|
|
if os.path.isfile(path):
|
|
|
|
os.unlink(path)
|
|
|
|
|
|
|
|
self.log.debug("Deleting empty dirs...")
|
|
|
|
for root, dirs, files in os.walk(self.directory, topdown=False):
|
|
|
|
for dir in dirs:
|
|
|
|
path = os.path.join(root, dir)
|
|
|
|
if os.path.isdir(path) and os.listdir(path) == []:
|
|
|
|
os.removedirs(path)
|
|
|
|
self.log.debug("Removing %s" % path)
|
|
|
|
if os.path.isdir(self.directory) and os.listdir(self.directory) == []:
|
|
|
|
os.removedirs(self.directory) # Remove sites directory if empty
|
|
|
|
|
|
|
|
if os.path.isdir(self.directory):
|
|
|
|
self.log.debug("Some unknown file remained in site data dir: %s..." % self.directory)
|
|
|
|
return False # Some files not deleted
|
|
|
|
else:
|
|
|
|
self.log.debug("Site data directory deleted: %s..." % self.directory)
|
2016-03-06 00:55:50 +01:00
|
|
|
return True # All clean
|