mirror of
https://github.com/neondatabase/neon.git
synced 2025-12-26 07:39:58 +00:00
* Add yapf run to CircleCI * Pin yapf version * Enable `SPLIT_ALL_TOP_LEVEL_COMMA_SEPARATED_VALUES` setting * Reformat all existing code with slight manual adjustments * test_runner/README: note that yapf is forced
61 lines
2.3 KiB
Python
61 lines
2.3 KiB
Python
import os
|
|
from contextlib import closing
|
|
from fixtures.zenith_fixtures import PostgresFactory, ZenithPageserver
|
|
from fixtures.log_helper import log
|
|
|
|
pytest_plugins = ("fixtures.zenith_fixtures", "fixtures.benchmark_fixture")
|
|
|
|
|
|
#
|
|
# Run bulk INSERT test.
|
|
#
|
|
# Collects metrics:
|
|
#
|
|
# 1. Time to INSERT 5 million rows
|
|
# 2. Disk writes
|
|
# 3. Disk space used
|
|
# 4. Peak memory usage
|
|
#
|
|
def test_bulk_insert(postgres: PostgresFactory,
|
|
pageserver: ZenithPageserver,
|
|
pg_bin,
|
|
zenith_cli,
|
|
zenbenchmark,
|
|
repo_dir: str):
|
|
# Create a branch for us
|
|
zenith_cli.run(["branch", "test_bulk_insert", "empty"])
|
|
|
|
pg = postgres.create_start('test_bulk_insert')
|
|
log.info("postgres is running on 'test_bulk_insert' branch")
|
|
|
|
# Open a connection directly to the page server that we'll use to force
|
|
# flushing the layers to disk
|
|
psconn = pageserver.connect()
|
|
pscur = psconn.cursor()
|
|
|
|
# Get the timeline ID of our branch. We need it for the 'do_gc' command
|
|
with closing(pg.connect()) as conn:
|
|
with conn.cursor() as cur:
|
|
cur.execute("SHOW zenith.zenith_timeline")
|
|
timeline = cur.fetchone()[0]
|
|
|
|
cur.execute("create table huge (i int, j int);")
|
|
|
|
# Run INSERT, recording the time and I/O it takes
|
|
with zenbenchmark.record_pageserver_writes(pageserver, 'pageserver_writes'):
|
|
with zenbenchmark.record_duration('insert'):
|
|
cur.execute("insert into huge values (generate_series(1, 5000000), 0);")
|
|
|
|
# Flush the layers from memory to disk. This is included in the reported
|
|
# time and I/O
|
|
pscur.execute(f"do_gc {pageserver.initial_tenant} {timeline} 0")
|
|
|
|
# Record peak memory usage
|
|
zenbenchmark.record("peak_mem", zenbenchmark.get_peak_mem(pageserver) / 1024, 'MB')
|
|
|
|
# Report disk space used by the repository
|
|
timeline_size = zenbenchmark.get_timeline_size(repo_dir,
|
|
pageserver.initial_tenant,
|
|
timeline)
|
|
zenbenchmark.record('size', timeline_size / (1024 * 1024), 'MB')
|