/*-------------------------------------------------------------------------- * * setup.c * Code to set up a dynamic shared memory segments and a specified * number of background workers for shared memory message queue * testing. * * Copyright (c) 2013-2021, PostgreSQL Global Development Group * * IDENTIFICATION * src/test/modules/test_shm_mq/setup.c * * ------------------------------------------------------------------------- */ #include "postgres.h" #include "miscadmin.h" #include "pgstat.h" #include "postmaster/bgworker.h" #include "storage/procsignal.h" #include "storage/shm_toc.h" #include "test_shm_mq.h" #include "utils/memutils.h" typedef struct { int nworkers; BackgroundWorkerHandle *handle[FLEXIBLE_ARRAY_MEMBER]; } worker_state; static void setup_dynamic_shared_memory(int64 queue_size, int nworkers, dsm_segment **segp, test_shm_mq_header **hdrp, shm_mq **outp, shm_mq **inp); static worker_state *setup_background_workers(int nworkers, dsm_segment *seg); static void cleanup_background_workers(dsm_segment *seg, Datum arg); static void wait_for_workers_to_become_ready(worker_state *wstate, volatile test_shm_mq_header *hdr); static bool check_worker_status(worker_state *wstate); /* * Set up a dynamic shared memory segment and zero or more background workers * for a test run. */ void test_shm_mq_setup(int64 queue_size, int32 nworkers, dsm_segment **segp, shm_mq_handle **output, shm_mq_handle **input) { dsm_segment *seg; test_shm_mq_header *hdr; shm_mq *outq = NULL; /* placate compiler */ shm_mq *inq = NULL; /* placate compiler */ worker_state *wstate; /* Set up a dynamic shared memory segment. */ setup_dynamic_shared_memory(queue_size, nworkers, &seg, &hdr, &outq, &inq); *segp = seg; /* Register background workers. */ wstate = setup_background_workers(nworkers, seg); /* Attach the queues. */ *output = shm_mq_attach(outq, seg, wstate->handle[0]); *input = shm_mq_attach(inq, seg, wstate->handle[nworkers - 1]); /* Wait for workers to become ready. */ wait_for_workers_to_become_ready(wstate, hdr); /* * Once we reach this point, all workers are ready. We no longer need to * kill them if we die; they'll die on their own as the message queues * shut down. */ cancel_on_dsm_detach(seg, cleanup_background_workers, PointerGetDatum(wstate)); pfree(wstate); } /* * Set up a dynamic shared memory segment. * * We set up a small control region that contains only a test_shm_mq_header, * plus one region per message queue. There are as many message queues as * the number of workers, plus one. */ static void setup_dynamic_shared_memory(int64 queue_size, int nworkers, dsm_segment **segp, test_shm_mq_header **hdrp, shm_mq **outp, shm_mq **inp) { shm_toc_estimator e; int i; Size segsize; dsm_segment *seg; shm_toc *toc; test_shm_mq_header *hdr; /* Ensure a valid queue size. */ if (queue_size < 0 || ((uint64) queue_size) < shm_mq_minimum_size) ereport(ERROR, (errcode(ERRCODE_INVALID_PARAMETER_VALUE), errmsg("queue size must be at least %zu bytes", shm_mq_minimum_size))); if (queue_size != ((Size) queue_size)) ereport(ERROR, (errcode(ERRCODE_INVALID_PARAMETER_VALUE), errmsg("queue size overflows size_t"))); /* * Estimate how much shared memory we need. * * Because the TOC machinery may choose to insert padding of oddly-sized * requests, we must estimate each chunk separately. * * We need one key to register the location of the header, and we need * nworkers + 1 keys to track the locations of the message queues. */ shm_toc_initialize_estimator(&e); shm_toc_estimate_chunk(&e, sizeof(test_shm_mq_header)); for (i = 0; i <= nworkers; ++i) shm_toc_estimate_chunk(&e, (Size) queue_size); shm_toc_estimate_keys(&e, 2 + nworkers); segsize = shm_toc_estimate(&e); /* Create the shared memory segment and establish a table of contents. */ seg = dsm_create(shm_toc_estimate(&e), 0); toc = shm_toc_create(PG_TEST_SHM_MQ_MAGIC, dsm_segment_address(seg), segsize); /* Set up the header region. */ hdr = shm_toc_allocate(toc, sizeof(test_shm_mq_header)); SpinLockInit(&hdr->mutex); hdr->workers_total = nworkers; hdr->workers_attached = 0; hdr->workers_ready = 0; shm_toc_insert(toc, 0, hdr); /* Set up one message queue per worker, plus one. */ for (i = 0; i <= nworkers; ++i) { shm_mq *mq; mq = shm_mq_create(shm_toc_allocate(toc, (Size) queue_size), (Size) queue_size); shm_toc_insert(toc, i + 1, mq); if (i == 0) { /* We send messages to the first queue. */ shm_mq_set_sender(mq, MyProc); *outp = mq; } if (i == nworkers) { /* We receive messages from the last queue. */ shm_mq_set_receiver(mq, MyProc); *inp = mq; } } /* Return results to caller. */ *segp = seg; *hdrp = hdr; } /* * Register background workers. */ static worker_state * setup_background_workers(int nworkers, dsm_segment *seg) { MemoryContext oldcontext; BackgroundWorker worker; worker_state *wstate; int i; /* * We need the worker_state object and the background worker handles to * which it points to be allocated in CurTransactionContext rather than * ExprContext; otherwise, they'll be destroyed before the on_dsm_detach * hooks run. */ oldcontext = MemoryContextSwitchTo(CurTransactionContext); /* Create worker state object. */ wstate = MemoryContextAlloc(TopTransactionContext, offsetof(worker_state, handle) + sizeof(BackgroundWorkerHandle *) * nworkers); wstate->nworkers = 0; /* * Arrange to kill all the workers if we abort before all workers are * finished hooking themselves up to the dynamic shared memory segment. * * If we die after all the workers have finished hooking themselves up to * the dynamic shared memory segment, we'll mark the two queues to which * we're directly connected as detached, and the worker(s) connected to * those queues will exit, marking any other queues to which they are * connected as detached. This will cause any as-yet-unaware workers * connected to those queues to exit in their turn, and so on, until * everybody exits. * * But suppose the workers which are supposed to connect to the queues to * which we're directly attached exit due to some error before they * actually attach the queues. The remaining workers will have no way of * knowing this. From their perspective, they're still waiting for those * workers to start, when in fact they've already died. */ on_dsm_detach(seg, cleanup_background_workers, PointerGetDatum(wstate)); /* Configure a worker. */ memset(&worker, 0, sizeof(worker)); worker.bgw_flags = BGWORKER_SHMEM_ACCESS; worker.bgw_start_time = BgWorkerStart_ConsistentState; worker.bgw_restart_time = BGW_NEVER_RESTART; sprintf(worker.bgw_library_name, "test_shm_mq"); sprintf(worker.bgw_function_name, "test_shm_mq_main"); snprintf(worker.bgw_type, BGW_MAXLEN, "test_shm_mq"); worker.bgw_main_arg = UInt32GetDatum(dsm_segment_handle(seg)); /* set bgw_notify_pid, so we can detect if the worker stops */ worker.bgw_notify_pid = MyProcPid; /* Register the workers. */ for (i = 0; i < nworkers; ++i) { if (!RegisterDynamicBackgroundWorker(&worker, &wstate->handle[i])) ereport(ERROR, (errcode(ERRCODE_INSUFFICIENT_RESOURCES), errmsg("could not register background process"), errhint("You may need to increase max_worker_processes."))); ++wstate->nworkers; } /* All done. */ MemoryContextSwitchTo(oldcontext); return wstate; } static void cleanup_background_workers(dsm_segment *seg, Datum arg) { worker_state *wstate = (worker_state *) DatumGetPointer(arg); while (wstate->nworkers > 0) { --wstate->nworkers; TerminateBackgroundWorker(wstate->handle[wstate->nworkers]); } } static void wait_for_workers_to_become_ready(worker_state *wstate, volatile test_shm_mq_header *hdr) { bool result = false; for (;;) { int workers_ready; /* If all the workers are ready, we have succeeded. */ SpinLockAcquire(&hdr->mutex); workers_ready = hdr->workers_ready; SpinLockRelease(&hdr->mutex); if (workers_ready >= wstate->nworkers) { result = true; break; } /* If any workers (or the postmaster) have died, we have failed. */ if (!check_worker_status(wstate)) { result = false; break; } /* Wait to be signaled. */ (void) WaitLatch(MyLatch, WL_LATCH_SET | WL_EXIT_ON_PM_DEATH, 0, PG_WAIT_EXTENSION); /* Reset the latch so we don't spin. */ ResetLatch(MyLatch); /* An interrupt may have occurred while we were waiting. */ CHECK_FOR_INTERRUPTS(); } if (!result) ereport(ERROR, (errcode(ERRCODE_INSUFFICIENT_RESOURCES), errmsg("one or more background workers failed to start"))); } static bool check_worker_status(worker_state *wstate) { int n; /* If any workers (or the postmaster) have died, we have failed. */ for (n = 0; n < wstate->nworkers; ++n) { BgwHandleStatus status; pid_t pid; status = GetBackgroundWorkerPid(wstate->handle[n], &pid); if (status == BGWH_STOPPED || status == BGWH_POSTMASTER_DIED) return false; } /* Otherwise, things still look OK. */ return true; }