""" This file contains multiple primitive tests for avoiding performance regressions. - Time bound tests: Benchmarks are done on GHA before adding numbers - Query count tests: More than expected # of queries for any action is frequent source of performance issues. This guards against such problems. E.g. We know get_controller is supposed to be cached and hence shouldn't make query post first query. This test can be written like this. >>> def test_controller_caching(self): >>> >>> get_controller("User") # <- "warm up code" >>> with self.assertQueryCount(0): >>> get_controller("User") """ import gc import itertools import sys import time from unittest.mock import patch from tenacity import retry, retry_if_exception_type, stop_after_attempt, wait_fixed import frappe from frappe.frappeclient import FrappeClient from frappe.model.base_document import get_controller from frappe.query_builder.utils import db_type_is from frappe.tests import IntegrationTestCase from frappe.tests.test_api import FrappeAPITestCase from frappe.tests.test_query_builder import run_only_if from frappe.utils import cint from frappe.utils.caching import redis_cache from frappe.website.path_resolver import PathResolver TEST_USER = "test@example.com" @run_only_if(db_type_is.MARIADB) class TestPerformance(IntegrationTestCase): def reset_request_specific_caches(self): # To simulate close to request level of handling frappe.destroy() # releases everything on frappe.local frappe.init(self.TEST_SITE) frappe.connect() frappe.clear_cache() def setUp(self) -> None: self.HOST = frappe.utils.get_site_url(frappe.local.site) self.reset_request_specific_caches() def test_meta_caching(self): frappe.clear_cache() frappe.get_meta("User") frappe.clear_cache(doctype="ToDo") with self.assertQueryCount(0): frappe.get_meta("User") def test_permitted_fieldnames(self): frappe.clear_cache() doc = frappe.new_doc("Prepared Report") # load permitted fieldnames once doc.permitted_fieldnames with patch("frappe.model.base_document.get_permitted_fields") as mocked: doc.as_dict() # get_permitted_fields should not be called again mocked.assert_not_called() def test_set_value_query_count(self): frappe.db.set_value("User", "Administrator", "interest", "Nothing") with self.assertQueryCount(1): frappe.db.set_value("User", "Administrator", "interest", "Nothing") with self.assertQueryCount(1): frappe.db.set_value("User", {"user_type": "System User"}, "interest", "Nothing") with self.assertQueryCount(1): frappe.db.set_value( "User", {"user_type": "System User"}, {"interest": "Nothing", "bio": "boring person"} ) def test_controller_caching(self): get_controller("User") with self.assertQueryCount(0): get_controller("User") def test_get_value_limits(self): # check both dict and list style filters filters = [{"enabled": 1}, [["enabled", "=", 1]]] # Warm up code frappe.db.get_values("User", filters=filters[0], limit=1) for filter in filters: with self.assertRowsRead(1): self.assertEqual(1, len(frappe.db.get_values("User", filters=filter, limit=1))) with self.assertRowsRead(2): self.assertEqual(2, len(frappe.db.get_values("User", filters=filter, limit=2))) self.assertEqual( len(frappe.db.get_values("User", filters=filter)), frappe.db.count("User", filter) ) with self.assertRowsRead(1): frappe.db.get_value("User", filters=filter) with self.assertRowsRead(1): frappe.db.exists("User", filter) def test_db_value_cache(self): """Link validation if repeated should just use db.value_cache, hence no extra queries""" doc = frappe.get_last_doc("User") doc.get_invalid_links() with self.assertQueryCount(0): doc.get_invalid_links() @retry( retry=retry_if_exception_type(AssertionError), stop=stop_after_attempt(3), wait=wait_fixed(0.5), reraise=True, ) def test_req_per_seconds_basic(self): """Ideally should be ran against gunicorn worker, though I have not seen any difference when using werkzeug's run_simple for synchronous requests.""" EXPECTED_RPS = 120 # measured on GHA FAILURE_THREASHOLD = 0.1 req_count = 1000 client = FrappeClient(self.HOST, "Administrator", self.ADMIN_PASSWORD) start = time.perf_counter() for _ in range(req_count): client.get_list("ToDo", limit_page_length=1) end = time.perf_counter() rps = req_count / (end - start) print(f"Completed {req_count} in {end - start} @ {rps} requests per seconds") self.assertGreaterEqual( rps, EXPECTED_RPS * (1 - FAILURE_THREASHOLD), "Possible performance regression in basic /api/Resource list requests", ) def test_homepage_resolver(self): paths = ["/", "/app"] for path in paths: PathResolver(path).resolve() with self.assertQueryCount(1): PathResolver(path).resolve() def test_consistent_build_version(self): from frappe.utils import get_build_version self.assertEqual(get_build_version(), get_build_version()) def test_get_list_single_query(self): """get_list should only perform single query.""" user = frappe.get_doc("User", TEST_USER) frappe.set_user(TEST_USER) # Give full read access, no share/user perm check should be done. user.add_roles("System Manager") frappe.get_list("User") with self.assertQueryCount(1): frappe.get_list("User") def test_no_ifnull_checks(self): query = frappe.get_all("DocType", {"autoname": ("is", "set")}, run=0).lower() self.assertNotIn("coalesce", query) self.assertNotIn("ifnull", query) def test_no_stale_ref_sql(self): """frappe.db.sql should not hold any internal references to result set. pymysql stores results internally. If your code reads a lot and doesn't make another query, for that entire duration there's copy of result consuming memory in internal attributes of pymysql. We clear it manually, this test ensures that it actually works. """ query = "select * from tabUser" for kwargs in ({}, {"as_dict": True}, {"as_list": True}): result = frappe.db.sql(query, **kwargs) self.assertEqual(sys.getrefcount(result), 2) # Note: This always returns +1 self.assertFalse(gc.get_referrers(result)) def test_no_cyclic_references(self): doc = frappe.get_doc("User", "Administrator") self.assertEqual(sys.getrefcount(doc), 2) # Note: This always returns +1 def test_get_doc_cache_calls(self): frappe.get_doc("User", "Administrator") with self.assertRedisCallCounts(0): frappe.get_doc("User", "Administrator") def test_local_caching(self): frappe.get_cached_doc("User", "Administrator") with self.assertRedisCallCounts(0): frappe.get_cached_doc("User", "Administrator") def test_redis_cache_calls(self): redis_cached_func() # warmup # Repeat call should use locally cached value with self.assertRedisCallCounts(0): redis_cached_func() frappe.local.cache.clear() # Without local cache - only one call required with self.assertRedisCallCounts(1): redis_cached_func() def test_one_time_setup(self): site = frappe.local.site frappe.init(site, force=True) run = frappe.qb._BuilderClasss.run frappe.init(site, force=True) patched_run = frappe.qb._BuilderClasss.run self.assertIs(run, patched_run, "frappe.init should run one-time patching code just once") def test_cpu_allocation(self): from frappe._optimizations import assign_core # Already allocated self.assertEqual(assign_core(0, 4, 8, [0], []), 0) # All physical, pid same as core for 0-7 siblings = [(i,) for i in range(8)] cores = list(range(8)) for pid in cores: self.assertEqual(assign_core(pid, len(cores), len(cores), cores, siblings), pid) # All physical, pid wraps for core for 8-15 for pid in range(8, 16): self.assertEqual(assign_core(pid, len(cores), len(cores), cores, siblings), pid % len(cores)) default_affinity_16 = list(range(16)) # "linear" siblings = (0,1) (2,3) ... linear_siblings_16 = list(itertools.batched(range(16), 2)) logical_cores = list(range(16)) expected_assignments = [*(l[0] for l in linear_siblings_16), *(l[1] for l in linear_siblings_16)] for pid, expected_core in zip(logical_cores, expected_assignments, strict=True): core = assign_core( pid, len(logical_cores) // 2, len(logical_cores), default_affinity_16, linear_siblings_16 ) self.assertEqual(core, expected_core) # "Block" siblings = (0,4) (1,5) ... block_siblings_16 = list(zip(range(8), range(8, 16), strict=True)) for pid in logical_cores: core = assign_core( pid, len(logical_cores) // 2, len(logical_cores), logical_cores, block_siblings_16 ) self.assertEqual(core, pid) # Few cores disabled enabled_cores = [0, 2, 4, 6] affinity = [(i,) for i in enabled_cores] core = assign_core(0, 4, 4, enabled_cores, affinity) self.assertEqual(core, 0) core = assign_core(1, 4, 4, enabled_cores, affinity) self.assertEqual(core, 2) @run_only_if(db_type_is.MARIADB) class TestOverheadCalls(FrappeAPITestCase): """Test that typical redis and db calls remain same overtime. If this tests fail on your PR, make sure you're not introducing something in hot-path of these endpoints. Only update values if you're really sure that's the right call. Every call increase here is an actual increase in cost! """ BASE_SQL_CALLS = 2 # rollback + begin def test_ping_overheads(self): self.get(self.method("ping"), {"sid": "Guest"}) with self.assertRedisCallCounts(2), self.assertQueryCount(self.BASE_SQL_CALLS): self.get(self.method("ping"), {"sid": "Guest"}) def test_ping_overheads_authenticated(self): sid = self.sid self.get(self.method("ping"), {"sid": sid}) with self.assertRedisCallCounts(3), self.assertQueryCount(self.BASE_SQL_CALLS): self.get(self.method("ping"), {"sid": sid}) def test_list_view_overheads(self): sid = self.sid self.get(self.resource("ToDo"), {"sid": sid}) self.get(self.resource("ToDo"), {"sid": sid}) with self.assertRedisCallCounts(6), self.assertQueryCount(self.BASE_SQL_CALLS + 1): self.get(self.resource("ToDo"), {"sid": sid}) def test_get_doc_overheads(self): sid = self.sid tables = len(frappe.get_meta("User").get_table_fields()) self.get(self.resource("User", "Administrator"), {"sid": sid}) self.get(self.resource("User", "Administrator"), {"sid": sid}) with self.assertRedisCallCounts(3), self.assertQueryCount(self.BASE_SQL_CALLS + 1 + tables): self.get(self.resource("User", "Administrator"), {"sid": sid}) @redis_cache def redis_cached_func(): return 42