mirror of
https://github.com/neondatabase/neon.git
synced 2025-12-26 15:49:58 +00:00
## Problem
In Neon DBaaS we adjust the shared_buffers to the size of the compute,
or better described we adjust the max number of connections to the
compute size and we adjust the shared_buffers size to the number of max
connections according to about the following sizes
`2 CU: 225mb; 4 CU: 450mb; 8 CU: 900mb`
[see](877e33b428/goapp/controlplane/internal/pkg/compute/computespec/pg_settings.go (L405))
## Summary of changes
We should run perf unit tests with settings that is realistic for a
paying customer and select 8 CU as the reference for those tests.
62 lines
1.9 KiB
Python
62 lines
1.9 KiB
Python
from __future__ import annotations
|
|
|
|
import asyncio
|
|
from io import BytesIO
|
|
from typing import TYPE_CHECKING
|
|
|
|
if TYPE_CHECKING:
|
|
from fixtures.neon_fixtures import Endpoint, NeonEnv
|
|
|
|
from fixtures.utils import shared_buffers_for_max_cu
|
|
|
|
|
|
async def repeat_bytes(buf, repetitions: int):
|
|
for _ in range(repetitions):
|
|
yield buf
|
|
|
|
|
|
async def copy_test_data_to_table(endpoint: Endpoint, worker_id: int, table_name: str):
|
|
buf = BytesIO()
|
|
for i in range(1000):
|
|
buf.write(
|
|
f"{i}\tLoaded by worker {worker_id}. Long string to consume some space.\n".encode()
|
|
)
|
|
buf.seek(0)
|
|
|
|
copy_input = repeat_bytes(buf.read(), 5000)
|
|
|
|
pg_conn = await endpoint.connect_async()
|
|
|
|
# PgProtocol.connect_async sets statement_timeout to 2 minutes.
|
|
# That's not enough for this test, on a slow system in debug mode.
|
|
await pg_conn.execute("SET statement_timeout='300s'")
|
|
|
|
await pg_conn.copy_to_table(table_name, source=copy_input)
|
|
|
|
|
|
async def parallel_load_same_table(endpoint: Endpoint, n_parallel: int):
|
|
workers = []
|
|
for worker_id in range(n_parallel):
|
|
worker = copy_test_data_to_table(endpoint, worker_id, "copytest")
|
|
workers.append(asyncio.create_task(worker))
|
|
|
|
# await all workers
|
|
await asyncio.gather(*workers)
|
|
|
|
|
|
# Load data into one table with COPY TO from 5 parallel connections
|
|
def test_parallel_copy(neon_simple_env: NeonEnv, n_parallel=5):
|
|
env = neon_simple_env
|
|
# use shared_buffers size like in production for 8 CU compute
|
|
endpoint = env.endpoints.create_start(
|
|
"main", config_lines=[f"shared_buffers={shared_buffers_for_max_cu(8.0)}"]
|
|
)
|
|
|
|
# Create test table
|
|
conn = endpoint.connect()
|
|
cur = conn.cursor()
|
|
cur.execute("CREATE TABLE copytest (i int, t text)")
|
|
|
|
# Run COPY TO to load the table with parallel connections.
|
|
asyncio.run(parallel_load_same_table(endpoint, n_parallel))
|