511 lines
14 KiB
Python
511 lines
14 KiB
Python
# Copyright (c) 2015, Frappe Technologies Pvt. Ltd. and Contributors
|
|
# MIT License. See license.txt
|
|
import json
|
|
import mimetypes
|
|
import os
|
|
import re
|
|
from functools import wraps
|
|
|
|
import yaml
|
|
from six import iteritems
|
|
from werkzeug.wrappers import Response
|
|
|
|
import frappe
|
|
from frappe import _
|
|
from frappe.model.document import Document
|
|
from frappe.utils import md_to_html
|
|
|
|
|
|
def delete_page_cache(path):
|
|
cache = frappe.cache()
|
|
cache.delete_value('full_index')
|
|
groups = ("website_page", "page_context")
|
|
if path:
|
|
for name in groups:
|
|
cache.hdel(name, path)
|
|
else:
|
|
for name in groups:
|
|
cache.delete_key(name)
|
|
|
|
def find_first_image(html):
|
|
m = re.finditer(r"""<img[^>]*src\s?=\s?['"]([^'"]*)['"]""", html)
|
|
try:
|
|
return next(m).groups()[0]
|
|
except StopIteration:
|
|
return None
|
|
|
|
def can_cache(no_cache=False):
|
|
if frappe.flags.force_website_cache:
|
|
return True
|
|
if frappe.conf.disable_website_cache or frappe.conf.developer_mode:
|
|
return False
|
|
if getattr(frappe.local, "no_cache", False):
|
|
return False
|
|
return not no_cache
|
|
|
|
|
|
def get_comment_list(doctype, name):
|
|
comments = frappe.get_all('Comment',
|
|
fields=['name', 'creation', 'owner',
|
|
'comment_email', 'comment_by', 'content'],
|
|
filters=dict(
|
|
reference_doctype=doctype,
|
|
reference_name=name,
|
|
comment_type='Comment',
|
|
),
|
|
or_filters=[
|
|
['owner', '=', frappe.session.user],
|
|
['published', '=', 1]])
|
|
|
|
communications = frappe.get_all("Communication",
|
|
fields=['name', 'creation', 'owner', 'owner as comment_email',
|
|
'sender_full_name as comment_by', 'content', 'recipients'],
|
|
filters=dict(
|
|
reference_doctype=doctype,
|
|
reference_name=name,
|
|
),
|
|
or_filters=[
|
|
['recipients', 'like', '%{0}%'.format(frappe.session.user)],
|
|
['cc', 'like', '%{0}%'.format(frappe.session.user)],
|
|
['bcc', 'like', '%{0}%'.format(frappe.session.user)]])
|
|
|
|
return sorted((comments + communications), key=lambda comment: comment['creation'], reverse=True)
|
|
|
|
|
|
def get_home_page():
|
|
if frappe.local.flags.home_page and not frappe.flags.in_test:
|
|
return frappe.local.flags.home_page
|
|
|
|
def _get_home_page():
|
|
home_page = None
|
|
|
|
# for user
|
|
if frappe.session.user != 'Guest':
|
|
# by role
|
|
for role in frappe.get_roles():
|
|
home_page = frappe.db.get_value('Role', role, 'home_page')
|
|
if home_page: break
|
|
|
|
# portal default
|
|
if not home_page:
|
|
home_page = frappe.db.get_value("Portal Settings", None, "default_portal_home")
|
|
|
|
# by hooks
|
|
if not home_page:
|
|
home_page = get_home_page_via_hooks()
|
|
|
|
# global
|
|
if not home_page:
|
|
home_page = frappe.db.get_value("Website Settings", None, "home_page")
|
|
|
|
if not home_page:
|
|
home_page = "login" if frappe.session.user == 'Guest' else "me"
|
|
|
|
home_page = home_page.strip('/')
|
|
|
|
return home_page
|
|
|
|
if frappe.local.dev_server:
|
|
# dont return cached homepage in development
|
|
return _get_home_page()
|
|
|
|
return frappe.cache().hget("home_page", frappe.session.user, _get_home_page)
|
|
|
|
def get_home_page_via_hooks():
|
|
home_page = None
|
|
|
|
home_page_method = frappe.get_hooks('get_website_user_home_page')
|
|
if home_page_method:
|
|
home_page = frappe.get_attr(home_page_method[-1])(frappe.session.user)
|
|
elif frappe.get_hooks('website_user_home_page'):
|
|
home_page = frappe.get_hooks('website_user_home_page')[-1]
|
|
|
|
if not home_page:
|
|
role_home_page = frappe.get_hooks("role_home_page")
|
|
if role_home_page:
|
|
for role in frappe.get_roles():
|
|
if role in role_home_page:
|
|
home_page = role_home_page[role][-1]
|
|
break
|
|
|
|
if not home_page:
|
|
home_page = frappe.get_hooks("home_page")
|
|
if home_page:
|
|
home_page = home_page[-1]
|
|
|
|
if home_page:
|
|
home_page = home_page.strip('/')
|
|
|
|
return home_page
|
|
|
|
|
|
def is_signup_disabled():
|
|
return frappe.db.get_single_value('Website Settings', 'disable_signup', True)
|
|
|
|
def cleanup_page_name(title):
|
|
"""make page name from title"""
|
|
if not title:
|
|
return ''
|
|
|
|
name = title.lower()
|
|
name = re.sub(r'[~!@#$%^&*+()<>,."\'\?]', '', name)
|
|
name = re.sub('[:/]', '-', name)
|
|
name = '-'.join(name.split())
|
|
# replace repeating hyphens
|
|
name = re.sub(r"(-)\1+", r"\1", name)
|
|
return name[:140]
|
|
|
|
|
|
def get_shade(color, percent=None):
|
|
frappe.msgprint(_('get_shade method has been deprecated.'))
|
|
return color
|
|
|
|
def abs_url(path):
|
|
"""Deconstructs and Reconstructs a URL into an absolute URL or a URL relative from root '/'"""
|
|
if not path:
|
|
return
|
|
if path.startswith('http://') or path.startswith('https://'):
|
|
return path
|
|
if path.startswith('data:'):
|
|
return path
|
|
if not path.startswith("/"):
|
|
path = "/" + path
|
|
return path
|
|
|
|
def get_toc(route, url_prefix=None, app=None):
|
|
'''Insert full index (table of contents) for {index} tag'''
|
|
|
|
full_index = get_full_index(app=app)
|
|
|
|
return frappe.get_template("templates/includes/full_index.html").render({
|
|
"full_index": full_index,
|
|
"url_prefix": url_prefix or "/",
|
|
"route": route.rstrip('/')
|
|
})
|
|
|
|
def get_next_link(route, url_prefix=None, app=None):
|
|
# insert next link
|
|
next_item = None
|
|
route = route.rstrip('/')
|
|
children_map = get_full_index(app=app)
|
|
parent_route = os.path.dirname(route)
|
|
children = children_map.get(parent_route, None)
|
|
|
|
if parent_route and children:
|
|
for i, c in enumerate(children):
|
|
if c.route == route and i < (len(children) - 1):
|
|
next_item = children[i+1]
|
|
next_item.url_prefix = url_prefix or "/"
|
|
|
|
if next_item:
|
|
if next_item.route and next_item.title:
|
|
html = ('<p class="btn-next-wrapper">' + frappe._("Next")\
|
|
+': <a class="btn-next" href="{url_prefix}{route}">{title}</a></p>').format(**next_item)
|
|
|
|
return html
|
|
|
|
return ''
|
|
|
|
def get_full_index(route=None, app=None):
|
|
"""Returns full index of the website for www upto the n-th level"""
|
|
from frappe.website.router import get_pages
|
|
|
|
if not frappe.local.flags.children_map:
|
|
def _build():
|
|
children_map = {}
|
|
added = []
|
|
pages = get_pages(app=app)
|
|
|
|
# make children map
|
|
for route, page_info in pages.items():
|
|
parent_route = os.path.dirname(route)
|
|
if parent_route not in added:
|
|
children_map.setdefault(parent_route, []).append(page_info)
|
|
|
|
# order as per index if present
|
|
for route, children in children_map.items():
|
|
if not route in pages:
|
|
# no parent (?)
|
|
continue
|
|
|
|
page_info = pages[route]
|
|
if page_info.index or ('index' in page_info.template):
|
|
new_children = []
|
|
page_info.extn = ''
|
|
for name in (page_info.index or []):
|
|
child_route = page_info.route + '/' + name
|
|
if child_route in pages:
|
|
if child_route not in added:
|
|
new_children.append(pages[child_route])
|
|
added.append(child_route)
|
|
|
|
# add remaining pages not in index.txt
|
|
_children = sorted(children, key=lambda x: os.path.basename(x.route))
|
|
|
|
for child_route in _children:
|
|
if child_route not in new_children:
|
|
if child_route not in added:
|
|
new_children.append(child_route)
|
|
added.append(child_route)
|
|
|
|
children_map[route] = new_children
|
|
|
|
return children_map
|
|
|
|
children_map = frappe.cache().get_value('website_full_index', _build)
|
|
|
|
frappe.local.flags.children_map = children_map
|
|
|
|
return frappe.local.flags.children_map
|
|
|
|
def extract_title(source, path):
|
|
'''Returns title from `<!-- title -->` or <h1> or path'''
|
|
title = extract_comment_tag(source, 'title')
|
|
|
|
if not title and "<h1>" in source:
|
|
# extract title from h1
|
|
match = re.findall('<h1>([^<]*)', source)
|
|
title_content = match[0].strip()[:300]
|
|
if '{{' not in title_content:
|
|
title = title_content
|
|
|
|
if not title:
|
|
# make title from name
|
|
title = os.path.basename(path.rsplit('.', )[0].rstrip('/')).replace('_', ' ').replace('-', ' ').title()
|
|
|
|
return title
|
|
|
|
def extract_comment_tag(source, tag):
|
|
'''Extract custom tags in comments from source.
|
|
|
|
:param source: raw template source in HTML
|
|
:param title: tag to search, example "title"
|
|
'''
|
|
|
|
if "<!-- {0}:".format(tag) in source:
|
|
return re.findall('<!-- {0}:([^>]*) -->'.format(tag), source)[0].strip()
|
|
else:
|
|
return None
|
|
|
|
|
|
def get_html_content_based_on_type(doc, fieldname, content_type):
|
|
'''
|
|
Set content based on content_type
|
|
'''
|
|
content = doc.get(fieldname)
|
|
|
|
if content_type == 'Markdown':
|
|
content = md_to_html(doc.get(fieldname + '_md'))
|
|
elif content_type == 'HTML':
|
|
content = doc.get(fieldname + '_html')
|
|
|
|
if content == None:
|
|
content = ''
|
|
|
|
return content
|
|
|
|
|
|
def clear_cache(path=None):
|
|
'''Clear website caches
|
|
:param path: (optional) for the given path'''
|
|
for key in ('website_generator_routes', 'website_pages',
|
|
'website_full_index', 'sitemap_routes'):
|
|
frappe.cache().delete_value(key)
|
|
|
|
frappe.cache().delete_value("website_404")
|
|
if path:
|
|
frappe.cache().hdel('website_redirects', path)
|
|
delete_page_cache(path)
|
|
else:
|
|
clear_sitemap()
|
|
frappe.clear_cache("Guest")
|
|
for key in ('portal_menu_items', 'home_page', 'website_route_rules',
|
|
'doctypes_with_web_view', 'website_redirects', 'page_context',
|
|
'website_page'):
|
|
frappe.cache().delete_value(key)
|
|
|
|
for method in frappe.get_hooks("website_clear_cache"):
|
|
frappe.get_attr(method)(path)
|
|
|
|
def clear_website_cache(path=None):
|
|
clear_cache(path)
|
|
|
|
def clear_sitemap():
|
|
delete_page_cache("*")
|
|
|
|
def get_frontmatter(string):
|
|
"Reference: https://github.com/jonbeebe/frontmatter"
|
|
frontmatter = ""
|
|
body = ""
|
|
result = re.compile(r'^\s*(?:---|\+\+\+)(.*?)(?:---|\+\+\+)\s*(.+)$', re.S | re.M).search(string)
|
|
if result:
|
|
frontmatter = result.group(1)
|
|
body = result.group(2)
|
|
|
|
return {
|
|
"attributes": yaml.safe_load(frontmatter),
|
|
"body": body,
|
|
}
|
|
|
|
def get_sidebar_items(parent_sidebar, basepath):
|
|
import frappe.www.list
|
|
sidebar_items = []
|
|
|
|
hooks = frappe.get_hooks('look_for_sidebar_json')
|
|
look_for_sidebar_json = hooks[0] if hooks else frappe.flags.look_for_sidebar
|
|
|
|
if basepath and look_for_sidebar_json:
|
|
sidebar_items = get_sidebar_items_from_sidebar_file(basepath, look_for_sidebar_json)
|
|
|
|
if not sidebar_items and parent_sidebar:
|
|
sidebar_items = frappe.get_all('Website Sidebar Item',
|
|
filters=dict(parent=parent_sidebar), fields=['title', 'route', '`group`'],
|
|
order_by='idx asc')
|
|
|
|
if not sidebar_items:
|
|
sidebar_items = get_portal_sidebar_items()
|
|
|
|
return sidebar_items
|
|
|
|
|
|
def get_portal_sidebar_items():
|
|
sidebar_items = frappe.cache().hget('portal_menu_items', frappe.session.user)
|
|
if sidebar_items is None:
|
|
sidebar_items = []
|
|
roles = frappe.get_roles()
|
|
portal_settings = frappe.get_doc('Portal Settings', 'Portal Settings')
|
|
|
|
def add_items(sidebar_items, items):
|
|
for d in items:
|
|
if d.get('enabled') and ((not d.get('role')) or d.get('role') in roles):
|
|
sidebar_items.append(d.as_dict() if isinstance(d, Document) else d)
|
|
|
|
if not portal_settings.hide_standard_menu:
|
|
add_items(sidebar_items, portal_settings.get('menu'))
|
|
|
|
if portal_settings.custom_menu:
|
|
add_items(sidebar_items, portal_settings.get('custom_menu'))
|
|
|
|
items_via_hooks = frappe.get_hooks('portal_menu_items')
|
|
if items_via_hooks:
|
|
for i in items_via_hooks:
|
|
i['enabled'] = 1
|
|
add_items(sidebar_items, items_via_hooks)
|
|
|
|
frappe.cache().hset('portal_menu_items', frappe.session.user, sidebar_items)
|
|
|
|
return sidebar_items
|
|
|
|
def get_sidebar_items_from_sidebar_file(basepath, look_for_sidebar_json):
|
|
sidebar_items = []
|
|
sidebar_json_path = get_sidebar_json_path(basepath, look_for_sidebar_json)
|
|
if not sidebar_json_path:
|
|
return sidebar_items
|
|
|
|
with open(sidebar_json_path, 'r') as sidebarfile:
|
|
try:
|
|
sidebar_json = sidebarfile.read()
|
|
sidebar_items = json.loads(sidebar_json)
|
|
except json.decoder.JSONDecodeError:
|
|
frappe.throw('Invalid Sidebar JSON at ' + sidebar_json_path)
|
|
|
|
return sidebar_items
|
|
|
|
def get_sidebar_json_path(path, look_for=False):
|
|
'''Get _sidebar.json path from directory path
|
|
:param path: path of the current diretory
|
|
:param look_for: if True, look for _sidebar.json going upwards from given path
|
|
:return: _sidebar.json path
|
|
'''
|
|
if os.path.split(path)[1] == 'www' or path == '/' or not path:
|
|
return ''
|
|
|
|
sidebar_json_path = os.path.join(path, '_sidebar.json')
|
|
if os.path.exists(sidebar_json_path):
|
|
return sidebar_json_path
|
|
else:
|
|
if look_for:
|
|
return get_sidebar_json_path(os.path.split(path)[0], look_for)
|
|
else:
|
|
return ''
|
|
|
|
def cache_html(func):
|
|
@wraps(func)
|
|
def cache_html_decorator(*args, **kwargs):
|
|
if can_cache():
|
|
html = None
|
|
page_cache = frappe.cache().hget("website_page", args[0].path)
|
|
if page_cache and frappe.local.lang in page_cache:
|
|
html = page_cache[frappe.local.lang]
|
|
if html:
|
|
frappe.local.response.from_cache = True
|
|
return html
|
|
html = func(*args, **kwargs)
|
|
context = args[0].context
|
|
if can_cache(context.no_cache):
|
|
page_cache = frappe.cache().hget("website_page", args[0].path) or {}
|
|
page_cache[frappe.local.lang] = html
|
|
frappe.cache().hset("website_page", args[0].path, page_cache)
|
|
|
|
return html
|
|
|
|
return cache_html_decorator
|
|
|
|
def build_response(path, data, http_status_code, headers=None):
|
|
# build response
|
|
response = Response()
|
|
response.data = set_content_type(response, data, path)
|
|
response.status_code = http_status_code
|
|
response.headers["X-Page-Name"] = path.encode("ascii", errors="xmlcharrefreplace")
|
|
response.headers["X-From-Cache"] = frappe.local.response.from_cache or False
|
|
|
|
add_preload_headers(response)
|
|
if headers:
|
|
for key, val in iteritems(headers):
|
|
response.headers[key] = val.encode("ascii", errors="xmlcharrefreplace")
|
|
|
|
return response
|
|
|
|
def set_content_type(response, data, path):
|
|
if isinstance(data, dict):
|
|
response.mimetype = 'application/json'
|
|
response.charset = 'utf-8'
|
|
data = json.dumps(data)
|
|
return data
|
|
|
|
response.mimetype = 'text/html'
|
|
response.charset = 'utf-8'
|
|
|
|
# ignore paths ending with .com to avoid unnecessary download
|
|
# https://bugs.python.org/issue22347
|
|
if "." in path and not path.endswith('.com'):
|
|
content_type, encoding = mimetypes.guess_type(path)
|
|
if content_type:
|
|
response.mimetype = content_type
|
|
if encoding:
|
|
response.charset = encoding
|
|
|
|
return data
|
|
|
|
def add_preload_headers(response):
|
|
from bs4 import BeautifulSoup, SoupStrainer
|
|
|
|
try:
|
|
preload = []
|
|
strainer = SoupStrainer(re.compile("script|link"))
|
|
soup = BeautifulSoup(response.data, "lxml", parse_only=strainer)
|
|
for elem in soup.find_all('script', src=re.compile(".*")):
|
|
preload.append(("script", elem.get("src")))
|
|
|
|
for elem in soup.find_all('link', rel="stylesheet"):
|
|
preload.append(("style", elem.get("href")))
|
|
|
|
links = []
|
|
for _type, link in preload:
|
|
links.append("<{}>; rel=preload; as={}".format(link, _type))
|
|
|
|
if links:
|
|
response.headers["Link"] = ",".join(links)
|
|
except Exception:
|
|
import traceback
|
|
traceback.print_exc()
|