timescaledb/scripts/test_memory_spikes.py
Alexander Kuzmenkov 21a3f8206c Run python linter and formatter in CI
Helps find some errors and cosmetic problems.
2023-01-30 13:49:26 +04:00

162 lines
5.1 KiB
Python

#!/usr/bin/env python
# This file and its contents are licensed under the Apache License 2.0.
# Please see the included NOTICE for copyright information and
# LICENSE-APACHE for a copy of the license.
# Python script to check if there are memory spikes when running queries
import psutil
import time
import sys
from datetime import datetime
DEFAULT_MEMCAP = 300 # in MB
THRESHOLD_RATIO = 1.5 # ratio above which considered memory spike
WAIT_TO_STABILIZE = 15 # wait in seconds before considering memory stable
CHECK_INTERVAL = 15
DEBUG = False
# finds processes with name as argument
def find_procs_by_name(name):
# Return a list of processes matching 'name'
ls = []
for p in psutil.process_iter():
if p.name() == name:
ls.append(p)
return ls
# return human readable form of number of bytes n
def bytes2human(n):
# http://code.activestate.com/recipes/578019
# >>> bytes2human(10000)
# '9.8K'
# >>> bytes2human(100001221)
# '95.4M'
symbols = ("K", "M", "G", "T", "P", "E", "Z", "Y")
prefix = {}
for i, s in enumerate(symbols):
prefix[s] = 1 << (i + 1) * 10
for s in reversed(symbols):
if n >= prefix[s]:
value = float(n) / prefix[s]
return f"{value:.1f}{s}"
return f"{n}B"
# prints pid of processes
def process_details(process):
return f"{process.pid} {''.join(process.cmdline()).strip()}"
def process_stats():
processes = find_procs_by_name("postgres")
for p in processes:
print(p, p.num_ctx_switches(), p.cpu_times(), p.memory_info(), flush=True)
# return process id of new postgres process created when running SQL file
def find_new_process():
# get postgres processes that are running before insertion starts
base_process = find_procs_by_name("postgres")
print("Processes running before inserts run: ")
for p in base_process:
print(process_details(p))
process_count = len(base_process)
print(
f"Waiting {WAIT_TO_STABILIZE} seconds for process running inserts to start",
flush=True,
)
time.sleep(
WAIT_TO_STABILIZE
) # wait WAIT_TO_STABILIZE seconds to get process that runs the inserts
# continuously check for creation of new postgres process
timeout = time.time() + 60
while True:
# prevent infinite loop
if time.time() > timeout:
print(
"Timed out on finding new process, should force quit SQL inserts",
flush=True,
)
sys.exit(4)
process = find_procs_by_name("postgres")
cnt = len(process)
print("process count ", cnt)
if cnt > process_count:
process = find_procs_by_name("postgres")
difference_set = set(process) - set(base_process)
new_process = None
# We assume that the backend is the first 'new' process to start, so it will have
# the lower PID
for p in difference_set:
print(f"found process: {process_details(p)}")
if new_process is None or p.pid < new_process.pid:
new_process = p
print(f"new_process: {process_details(new_process)}")
return new_process.pid
time.sleep(1)
def main():
# get process created from running insert test sql file
pid = find_new_process()
p = psutil.Process(pid)
print('*** Check this pid is the same as "pg_backend_pid" from SQL command ***')
print("New process backend process:", pid)
print(
f"Waiting {WAIT_TO_STABILIZE} seconds for memory consumption to stabilize",
flush=True,
)
time.sleep(WAIT_TO_STABILIZE)
# Calculate average memory consumption from 5 values over 15 seconds
total = 0
for _ in range(0, 5):
total += p.memory_info().rss
time.sleep(3)
avg = total / 5
print("Average memory consumption: ", bytes2human(avg), flush=True)
cap = int(sys.argv[1] if len(sys.argv) > 1 else DEFAULT_MEMCAP) * 1024 * 1024
upper_threshold = min(cap, avg * THRESHOLD_RATIO)
# check if memory consumption goes above threshold until process terminates every 30 seconds
timeout = time.time() + 45 * 60
while True:
# insert finished
if not psutil.pid_exists(pid):
break
# prevent infinite loop
if time.time() > timeout:
print("Timed out on running inserts (over 45 mins)")
print("Killing postgres process")
p.kill()
sys.exit(4)
rss = p.memory_info().rss
stamp = datetime.now().strftime("%H:%M:%S")
print(f"{stamp} Memory used by process {p.pid}: {bytes2human(rss)}", flush=True)
if DEBUG:
process_stats()
# exit with error if memory above threshold
if rss > upper_threshold:
print("Memory consumption exceeded upper threshold")
print("Killing postgres process", flush=True)
p.kill()
sys.exit(4)
time.sleep(CHECK_INTERVAL)
print("No memory leaks detected", flush=True)
sys.exit(0) # success
if __name__ == "__main__":
main()